var/home/core/zuul-output/0000755000175000017500000000000015071524743014535 5ustar corecorevar/home/core/zuul-output/logs/0000755000175000017500000000000015071535707015503 5ustar corecorevar/home/core/zuul-output/logs/kubelet.log0000644000000000000000004706757415071535701017725 0ustar rootrootOct 08 18:17:08 crc systemd[1]: Starting Kubernetes Kubelet... Oct 08 18:17:08 crc restorecon[4687]: Relabeled /var/lib/kubelet/config.json from system_u:object_r:unlabeled_t:s0 to system_u:object_r:container_var_lib_t:s0 Oct 08 18:17:08 crc restorecon[4687]: /var/lib/kubelet/device-plugins not reset as customized by admin to system_u:object_r:container_file_t:s0 Oct 08 18:17:08 crc restorecon[4687]: /var/lib/kubelet/device-plugins/kubelet.sock not reset as customized by admin to system_u:object_r:container_file_t:s0 Oct 08 18:17:08 crc restorecon[4687]: /var/lib/kubelet/pods/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8/volumes/kubernetes.io~configmap/nginx-conf/..2025_02_23_05_40_35.4114275528/nginx.conf not reset as customized by admin to system_u:object_r:container_file_t:s0:c15,c25 Oct 08 18:17:08 crc restorecon[4687]: /var/lib/kubelet/pods/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c15,c25 Oct 08 18:17:08 crc restorecon[4687]: /var/lib/kubelet/pods/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8/containers/networking-console-plugin/22e96971 not reset as customized by admin to system_u:object_r:container_file_t:s0:c15,c25 Oct 08 18:17:08 crc restorecon[4687]: /var/lib/kubelet/pods/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8/containers/networking-console-plugin/21c98286 not reset as customized by admin to system_u:object_r:container_file_t:s0:c15,c25 Oct 08 18:17:08 crc restorecon[4687]: /var/lib/kubelet/pods/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8/containers/networking-console-plugin/0f1869e1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c15,c25 Oct 08 18:17:08 crc restorecon[4687]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c215,c682 Oct 08 18:17:08 crc restorecon[4687]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/containers/setup/46889d52 not reset as customized by admin to system_u:object_r:container_file_t:s0:c225,c458 Oct 08 18:17:08 crc restorecon[4687]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/containers/setup/5b6a5969 not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c963 Oct 08 18:17:08 crc restorecon[4687]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/containers/setup/6c7921f5 not reset as customized by admin to system_u:object_r:container_file_t:s0:c215,c682 Oct 08 18:17:08 crc restorecon[4687]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/containers/kube-rbac-proxy-crio/4804f443 not reset as customized by admin to system_u:object_r:container_file_t:s0:c225,c458 Oct 08 18:17:08 crc restorecon[4687]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/containers/kube-rbac-proxy-crio/2a46b283 not reset as customized by admin to system_u:object_r:container_file_t:s0:c225,c458 Oct 08 18:17:08 crc restorecon[4687]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/containers/kube-rbac-proxy-crio/a6b5573e not reset as customized by admin to system_u:object_r:container_file_t:s0:c225,c458 Oct 08 18:17:08 crc restorecon[4687]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/containers/kube-rbac-proxy-crio/4f88ee5b not reset as customized by admin to system_u:object_r:container_file_t:s0:c225,c458 Oct 08 18:17:08 crc restorecon[4687]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/containers/kube-rbac-proxy-crio/5a4eee4b not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c963 Oct 08 18:17:08 crc restorecon[4687]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/containers/kube-rbac-proxy-crio/cd87c521 not reset as customized by admin to system_u:object_r:container_file_t:s0:c215,c682 Oct 08 18:17:08 crc restorecon[4687]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/volumes/kubernetes.io~configmap/service-ca-bundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Oct 08 18:17:08 crc restorecon[4687]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/volumes/kubernetes.io~configmap/service-ca-bundle/..2025_02_23_05_33_42.2574241751 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Oct 08 18:17:08 crc restorecon[4687]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/volumes/kubernetes.io~configmap/service-ca-bundle/..2025_02_23_05_33_42.2574241751/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Oct 08 18:17:08 crc restorecon[4687]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/volumes/kubernetes.io~configmap/service-ca-bundle/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Oct 08 18:17:08 crc restorecon[4687]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/volumes/kubernetes.io~configmap/service-ca-bundle/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Oct 08 18:17:08 crc restorecon[4687]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Oct 08 18:17:08 crc restorecon[4687]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/containers/router/38602af4 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Oct 08 18:17:08 crc restorecon[4687]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/containers/router/1483b002 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Oct 08 18:17:08 crc restorecon[4687]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/containers/router/0346718b not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Oct 08 18:17:08 crc restorecon[4687]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/containers/router/d3ed4ada not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Oct 08 18:17:08 crc restorecon[4687]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/containers/router/3bb473a5 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Oct 08 18:17:08 crc restorecon[4687]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/containers/router/8cd075a9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Oct 08 18:17:08 crc restorecon[4687]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/containers/router/00ab4760 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Oct 08 18:17:08 crc restorecon[4687]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/containers/router/54a21c09 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Oct 08 18:17:08 crc restorecon[4687]: /var/lib/kubelet/pods/37a5e44f-9a88-4405-be8a-b645485e7312/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c589,c726 Oct 08 18:17:08 crc restorecon[4687]: /var/lib/kubelet/pods/37a5e44f-9a88-4405-be8a-b645485e7312/containers/network-operator/70478888 not reset as customized by admin to system_u:object_r:container_file_t:s0:c176,c499 Oct 08 18:17:08 crc restorecon[4687]: /var/lib/kubelet/pods/37a5e44f-9a88-4405-be8a-b645485e7312/containers/network-operator/43802770 not reset as customized by admin to system_u:object_r:container_file_t:s0:c176,c499 Oct 08 18:17:08 crc restorecon[4687]: /var/lib/kubelet/pods/37a5e44f-9a88-4405-be8a-b645485e7312/containers/network-operator/955a0edc not reset as customized by admin to system_u:object_r:container_file_t:s0:c176,c499 Oct 08 18:17:08 crc restorecon[4687]: /var/lib/kubelet/pods/37a5e44f-9a88-4405-be8a-b645485e7312/containers/network-operator/bca2d009 not reset as customized by admin to system_u:object_r:container_file_t:s0:c140,c1009 Oct 08 18:17:08 crc restorecon[4687]: /var/lib/kubelet/pods/37a5e44f-9a88-4405-be8a-b645485e7312/containers/network-operator/b295f9bd not reset as customized by admin to system_u:object_r:container_file_t:s0:c589,c726 Oct 08 18:17:08 crc restorecon[4687]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-binary-copy not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Oct 08 18:17:08 crc restorecon[4687]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-binary-copy/..2025_02_23_05_21_22.3617465230 not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Oct 08 18:17:08 crc restorecon[4687]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-binary-copy/..2025_02_23_05_21_22.3617465230/cnibincopy.sh not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Oct 08 18:17:08 crc restorecon[4687]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-binary-copy/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Oct 08 18:17:08 crc restorecon[4687]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-binary-copy/cnibincopy.sh not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Oct 08 18:17:08 crc restorecon[4687]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-sysctl-allowlist not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Oct 08 18:17:08 crc restorecon[4687]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-sysctl-allowlist/..2025_02_23_05_21_22.2050650026 not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Oct 08 18:17:08 crc restorecon[4687]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-sysctl-allowlist/..2025_02_23_05_21_22.2050650026/allowlist.conf not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Oct 08 18:17:08 crc restorecon[4687]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-sysctl-allowlist/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Oct 08 18:17:08 crc restorecon[4687]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-sysctl-allowlist/allowlist.conf not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Oct 08 18:17:08 crc restorecon[4687]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Oct 08 18:17:08 crc restorecon[4687]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/egress-router-binary-copy/bc46ea27 not reset as customized by admin to system_u:object_r:container_file_t:s0:c203,c924 Oct 08 18:17:08 crc restorecon[4687]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/egress-router-binary-copy/5731fc1b not reset as customized by admin to system_u:object_r:container_file_t:s0:c138,c778 Oct 08 18:17:08 crc restorecon[4687]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/egress-router-binary-copy/5e1b2a3c not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Oct 08 18:17:08 crc restorecon[4687]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/cni-plugins/943f0936 not reset as customized by admin to system_u:object_r:container_file_t:s0:c203,c924 Oct 08 18:17:08 crc restorecon[4687]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/cni-plugins/3f764ee4 not reset as customized by admin to system_u:object_r:container_file_t:s0:c138,c778 Oct 08 18:17:08 crc restorecon[4687]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/cni-plugins/8695e3f9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Oct 08 18:17:08 crc restorecon[4687]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/bond-cni-plugin/aed7aa86 not reset as customized by admin to system_u:object_r:container_file_t:s0:c203,c924 Oct 08 18:17:08 crc restorecon[4687]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/bond-cni-plugin/c64d7448 not reset as customized by admin to system_u:object_r:container_file_t:s0:c138,c778 Oct 08 18:17:08 crc restorecon[4687]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/bond-cni-plugin/0ba16bd2 not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Oct 08 18:17:08 crc restorecon[4687]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/routeoverride-cni/207a939f not reset as customized by admin to system_u:object_r:container_file_t:s0:c203,c924 Oct 08 18:17:08 crc restorecon[4687]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/routeoverride-cni/54aa8cdb not reset as customized by admin to system_u:object_r:container_file_t:s0:c138,c778 Oct 08 18:17:08 crc restorecon[4687]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/routeoverride-cni/1f5fa595 not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Oct 08 18:17:08 crc restorecon[4687]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/whereabouts-cni-bincopy/bf9c8153 not reset as customized by admin to system_u:object_r:container_file_t:s0:c203,c924 Oct 08 18:17:08 crc restorecon[4687]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/whereabouts-cni-bincopy/47fba4ea not reset as customized by admin to system_u:object_r:container_file_t:s0:c138,c778 Oct 08 18:17:08 crc restorecon[4687]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/whereabouts-cni-bincopy/7ae55ce9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Oct 08 18:17:08 crc restorecon[4687]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/whereabouts-cni/7906a268 not reset as customized by admin to system_u:object_r:container_file_t:s0:c203,c924 Oct 08 18:17:08 crc restorecon[4687]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/whereabouts-cni/ce43fa69 not reset as customized by admin to system_u:object_r:container_file_t:s0:c138,c778 Oct 08 18:17:08 crc restorecon[4687]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/whereabouts-cni/7fc7ea3a not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Oct 08 18:17:08 crc restorecon[4687]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/kube-multus-additional-cni-plugins/d8c38b7d not reset as customized by admin to system_u:object_r:container_file_t:s0:c203,c924 Oct 08 18:17:08 crc restorecon[4687]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/kube-multus-additional-cni-plugins/9ef015fb not reset as customized by admin to system_u:object_r:container_file_t:s0:c138,c778 Oct 08 18:17:08 crc restorecon[4687]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/kube-multus-additional-cni-plugins/b9db6a41 not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Oct 08 18:17:08 crc restorecon[4687]: /var/lib/kubelet/pods/5b88f790-22fa-440e-b583-365168c0b23d/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c432,c991 Oct 08 18:17:08 crc restorecon[4687]: /var/lib/kubelet/pods/5b88f790-22fa-440e-b583-365168c0b23d/containers/network-metrics-daemon/b1733d79 not reset as customized by admin to system_u:object_r:container_file_t:s0:c476,c820 Oct 08 18:17:08 crc restorecon[4687]: /var/lib/kubelet/pods/5b88f790-22fa-440e-b583-365168c0b23d/containers/network-metrics-daemon/afccd338 not reset as customized by admin to system_u:object_r:container_file_t:s0:c272,c818 Oct 08 18:17:08 crc restorecon[4687]: /var/lib/kubelet/pods/5b88f790-22fa-440e-b583-365168c0b23d/containers/network-metrics-daemon/9df0a185 not reset as customized by admin to system_u:object_r:container_file_t:s0:c432,c991 Oct 08 18:17:08 crc restorecon[4687]: /var/lib/kubelet/pods/5b88f790-22fa-440e-b583-365168c0b23d/containers/kube-rbac-proxy/18938cf8 not reset as customized by admin to system_u:object_r:container_file_t:s0:c476,c820 Oct 08 18:17:08 crc restorecon[4687]: /var/lib/kubelet/pods/5b88f790-22fa-440e-b583-365168c0b23d/containers/kube-rbac-proxy/7ab4eb23 not reset as customized by admin to system_u:object_r:container_file_t:s0:c272,c818 Oct 08 18:17:08 crc restorecon[4687]: /var/lib/kubelet/pods/5b88f790-22fa-440e-b583-365168c0b23d/containers/kube-rbac-proxy/56930be6 not reset as customized by admin to system_u:object_r:container_file_t:s0:c432,c991 Oct 08 18:17:08 crc restorecon[4687]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/volumes/kubernetes.io~configmap/env-overrides not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Oct 08 18:17:08 crc restorecon[4687]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/volumes/kubernetes.io~configmap/env-overrides/..2025_02_23_05_21_35.630010865 not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Oct 08 18:17:08 crc restorecon[4687]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/volumes/kubernetes.io~configmap/env-overrides/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Oct 08 18:17:08 crc restorecon[4687]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/volumes/kubernetes.io~configmap/ovnkube-config not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Oct 08 18:17:08 crc restorecon[4687]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/volumes/kubernetes.io~configmap/ovnkube-config/..2025_02_23_05_21_35.1088506337 not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Oct 08 18:17:08 crc restorecon[4687]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/volumes/kubernetes.io~configmap/ovnkube-config/..2025_02_23_05_21_35.1088506337/ovnkube.conf not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Oct 08 18:17:08 crc restorecon[4687]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/volumes/kubernetes.io~configmap/ovnkube-config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Oct 08 18:17:08 crc restorecon[4687]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/volumes/kubernetes.io~configmap/ovnkube-config/ovnkube.conf not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Oct 08 18:17:08 crc restorecon[4687]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Oct 08 18:17:08 crc restorecon[4687]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/containers/kube-rbac-proxy/0d8e3722 not reset as customized by admin to system_u:object_r:container_file_t:s0:c89,c211 Oct 08 18:17:08 crc restorecon[4687]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/containers/kube-rbac-proxy/d22b2e76 not reset as customized by admin to system_u:object_r:container_file_t:s0:c382,c850 Oct 08 18:17:08 crc restorecon[4687]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/containers/kube-rbac-proxy/e036759f not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Oct 08 18:17:08 crc restorecon[4687]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/containers/ovnkube-cluster-manager/2734c483 not reset as customized by admin to system_u:object_r:container_file_t:s0:c89,c211 Oct 08 18:17:08 crc restorecon[4687]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/containers/ovnkube-cluster-manager/57878fe7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c89,c211 Oct 08 18:17:08 crc restorecon[4687]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/containers/ovnkube-cluster-manager/3f3c2e58 not reset as customized by admin to system_u:object_r:container_file_t:s0:c89,c211 Oct 08 18:17:08 crc restorecon[4687]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/containers/ovnkube-cluster-manager/375bec3e not reset as customized by admin to system_u:object_r:container_file_t:s0:c382,c850 Oct 08 18:17:08 crc restorecon[4687]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/containers/ovnkube-cluster-manager/7bc41e08 not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Oct 08 18:17:08 crc restorecon[4687]: /var/lib/kubelet/pods/cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Oct 08 18:17:08 crc restorecon[4687]: /var/lib/kubelet/pods/cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d/containers/download-server/48c7a72d not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Oct 08 18:17:08 crc restorecon[4687]: /var/lib/kubelet/pods/cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d/containers/download-server/4b66701f not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Oct 08 18:17:08 crc restorecon[4687]: /var/lib/kubelet/pods/cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d/containers/download-server/a5a1c202 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Oct 08 18:17:08 crc restorecon[4687]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/ovnkube-identity-cm not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Oct 08 18:17:08 crc restorecon[4687]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/ovnkube-identity-cm/..2025_02_23_05_21_40.3350632666 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Oct 08 18:17:08 crc restorecon[4687]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/ovnkube-identity-cm/..2025_02_23_05_21_40.3350632666/additional-cert-acceptance-cond.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Oct 08 18:17:08 crc restorecon[4687]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/ovnkube-identity-cm/..2025_02_23_05_21_40.3350632666/additional-pod-admission-cond.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Oct 08 18:17:08 crc restorecon[4687]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/ovnkube-identity-cm/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Oct 08 18:17:08 crc restorecon[4687]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/ovnkube-identity-cm/additional-cert-acceptance-cond.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Oct 08 18:17:08 crc restorecon[4687]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/ovnkube-identity-cm/additional-pod-admission-cond.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Oct 08 18:17:08 crc restorecon[4687]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/env-overrides not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Oct 08 18:17:08 crc restorecon[4687]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/env-overrides/..2025_02_23_05_21_40.1388695756 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Oct 08 18:17:08 crc restorecon[4687]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/env-overrides/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Oct 08 18:17:08 crc restorecon[4687]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Oct 08 18:17:08 crc restorecon[4687]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/containers/webhook/26f3df5b not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Oct 08 18:17:08 crc restorecon[4687]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/containers/webhook/6d8fb21d not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Oct 08 18:17:08 crc restorecon[4687]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/containers/webhook/50e94777 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Oct 08 18:17:08 crc restorecon[4687]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/containers/approver/208473b3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Oct 08 18:17:08 crc restorecon[4687]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/containers/approver/ec9e08ba not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Oct 08 18:17:08 crc restorecon[4687]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/containers/approver/3b787c39 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Oct 08 18:17:08 crc restorecon[4687]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/containers/approver/208eaed5 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Oct 08 18:17:08 crc restorecon[4687]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/containers/approver/93aa3a2b not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Oct 08 18:17:08 crc restorecon[4687]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/containers/approver/3c697968 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Oct 08 18:17:08 crc restorecon[4687]: /var/lib/kubelet/pods/3b6479f0-333b-4a96-9adf-2099afdc2447/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c20,c21 Oct 08 18:17:08 crc restorecon[4687]: /var/lib/kubelet/pods/3b6479f0-333b-4a96-9adf-2099afdc2447/containers/network-check-target-container/ba950ec9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c20,c21 Oct 08 18:17:08 crc restorecon[4687]: /var/lib/kubelet/pods/3b6479f0-333b-4a96-9adf-2099afdc2447/containers/network-check-target-container/cb5cdb37 not reset as customized by admin to system_u:object_r:container_file_t:s0:c20,c21 Oct 08 18:17:08 crc restorecon[4687]: /var/lib/kubelet/pods/3b6479f0-333b-4a96-9adf-2099afdc2447/containers/network-check-target-container/f2df9827 not reset as customized by admin to system_u:object_r:container_file_t:s0:c20,c21 Oct 08 18:17:08 crc restorecon[4687]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/images not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Oct 08 18:17:08 crc restorecon[4687]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/images/..2025_02_23_05_22_30.473230615 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Oct 08 18:17:08 crc restorecon[4687]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/images/..2025_02_23_05_22_30.473230615/images.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Oct 08 18:17:08 crc restorecon[4687]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/images/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Oct 08 18:17:08 crc restorecon[4687]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/images/images.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Oct 08 18:17:08 crc restorecon[4687]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/auth-proxy-config not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Oct 08 18:17:08 crc restorecon[4687]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/auth-proxy-config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Oct 08 18:17:08 crc restorecon[4687]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/auth-proxy-config/config-file.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Oct 08 18:17:08 crc restorecon[4687]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/auth-proxy-config/..2025_02_24_06_22_02.1904938450 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Oct 08 18:17:08 crc restorecon[4687]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/auth-proxy-config/..2025_02_24_06_22_02.1904938450/config-file.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Oct 08 18:17:08 crc restorecon[4687]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Oct 08 18:17:08 crc restorecon[4687]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/containers/machine-config-operator/fedaa673 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Oct 08 18:17:08 crc restorecon[4687]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/containers/machine-config-operator/9ca2df95 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Oct 08 18:17:08 crc restorecon[4687]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/containers/machine-config-operator/b2d7460e not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Oct 08 18:17:08 crc restorecon[4687]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/containers/kube-rbac-proxy/2207853c not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Oct 08 18:17:08 crc restorecon[4687]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/containers/kube-rbac-proxy/241c1c29 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Oct 08 18:17:08 crc restorecon[4687]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/containers/kube-rbac-proxy/2d910eaf not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Oct 08 18:17:08 crc restorecon[4687]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Oct 08 18:17:08 crc restorecon[4687]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Oct 08 18:17:08 crc restorecon[4687]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-ca/..2025_02_23_05_23_49.3726007728 not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Oct 08 18:17:08 crc restorecon[4687]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-ca/..2025_02_23_05_23_49.3726007728/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Oct 08 18:17:08 crc restorecon[4687]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-ca/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Oct 08 18:17:08 crc restorecon[4687]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-service-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Oct 08 18:17:08 crc restorecon[4687]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-service-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Oct 08 18:17:08 crc restorecon[4687]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-service-ca/..2025_02_23_05_23_49.841175008 not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Oct 08 18:17:08 crc restorecon[4687]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-service-ca/..2025_02_23_05_23_49.841175008/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Oct 08 18:17:08 crc restorecon[4687]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-service-ca/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Oct 08 18:17:08 crc restorecon[4687]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Oct 08 18:17:08 crc restorecon[4687]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.843437178 not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Oct 08 18:17:08 crc restorecon[4687]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.843437178/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Oct 08 18:17:08 crc restorecon[4687]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Oct 08 18:17:08 crc restorecon[4687]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Oct 08 18:17:08 crc restorecon[4687]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Oct 08 18:17:08 crc restorecon[4687]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/containers/etcd-operator/c6c0f2e7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c263,c871 Oct 08 18:17:08 crc restorecon[4687]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/containers/etcd-operator/399edc97 not reset as customized by admin to system_u:object_r:container_file_t:s0:c263,c871 Oct 08 18:17:08 crc restorecon[4687]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/containers/etcd-operator/8049f7cc not reset as customized by admin to system_u:object_r:container_file_t:s0:c263,c871 Oct 08 18:17:08 crc restorecon[4687]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/containers/etcd-operator/0cec5484 not reset as customized by admin to system_u:object_r:container_file_t:s0:c263,c871 Oct 08 18:17:08 crc restorecon[4687]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/containers/etcd-operator/312446d0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c406,c828 Oct 08 18:17:08 crc restorecon[4687]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/containers/etcd-operator/8e56a35d not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Oct 08 18:17:08 crc restorecon[4687]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c108,c511 Oct 08 18:17:08 crc restorecon[4687]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.133159589 not reset as customized by admin to system_u:object_r:container_file_t:s0:c108,c511 Oct 08 18:17:08 crc restorecon[4687]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.133159589/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c108,c511 Oct 08 18:17:08 crc restorecon[4687]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c108,c511 Oct 08 18:17:08 crc restorecon[4687]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c108,c511 Oct 08 18:17:08 crc restorecon[4687]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c108,c511 Oct 08 18:17:08 crc restorecon[4687]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/containers/kube-controller-manager-operator/2d30ddb9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c380,c909 Oct 08 18:17:08 crc restorecon[4687]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/containers/kube-controller-manager-operator/eca8053d not reset as customized by admin to system_u:object_r:container_file_t:s0:c380,c909 Oct 08 18:17:08 crc restorecon[4687]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/containers/kube-controller-manager-operator/c3a25c9a not reset as customized by admin to system_u:object_r:container_file_t:s0:c168,c522 Oct 08 18:17:08 crc restorecon[4687]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/containers/kube-controller-manager-operator/b9609c22 not reset as customized by admin to system_u:object_r:container_file_t:s0:c108,c511 Oct 08 18:17:08 crc restorecon[4687]: /var/lib/kubelet/pods/96b93a3a-6083-4aea-8eab-fe1aa8245ad9/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c968,c969 Oct 08 18:17:08 crc restorecon[4687]: /var/lib/kubelet/pods/96b93a3a-6083-4aea-8eab-fe1aa8245ad9/containers/dns-operator/e8b0eca9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c106,c418 Oct 08 18:17:08 crc restorecon[4687]: /var/lib/kubelet/pods/96b93a3a-6083-4aea-8eab-fe1aa8245ad9/containers/dns-operator/b36a9c3f not reset as customized by admin to system_u:object_r:container_file_t:s0:c529,c711 Oct 08 18:17:08 crc restorecon[4687]: /var/lib/kubelet/pods/96b93a3a-6083-4aea-8eab-fe1aa8245ad9/containers/dns-operator/38af7b07 not reset as customized by admin to system_u:object_r:container_file_t:s0:c968,c969 Oct 08 18:17:08 crc restorecon[4687]: /var/lib/kubelet/pods/96b93a3a-6083-4aea-8eab-fe1aa8245ad9/containers/kube-rbac-proxy/ae821620 not reset as customized by admin to system_u:object_r:container_file_t:s0:c106,c418 Oct 08 18:17:08 crc restorecon[4687]: /var/lib/kubelet/pods/96b93a3a-6083-4aea-8eab-fe1aa8245ad9/containers/kube-rbac-proxy/baa23338 not reset as customized by admin to system_u:object_r:container_file_t:s0:c529,c711 Oct 08 18:17:08 crc restorecon[4687]: /var/lib/kubelet/pods/96b93a3a-6083-4aea-8eab-fe1aa8245ad9/containers/kube-rbac-proxy/2c534809 not reset as customized by admin to system_u:object_r:container_file_t:s0:c968,c969 Oct 08 18:17:08 crc restorecon[4687]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c661,c999 Oct 08 18:17:08 crc restorecon[4687]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.3532625537 not reset as customized by admin to system_u:object_r:container_file_t:s0:c661,c999 Oct 08 18:17:08 crc restorecon[4687]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.3532625537/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c661,c999 Oct 08 18:17:08 crc restorecon[4687]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c661,c999 Oct 08 18:17:08 crc restorecon[4687]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c661,c999 Oct 08 18:17:08 crc restorecon[4687]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c661,c999 Oct 08 18:17:08 crc restorecon[4687]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/containers/kube-scheduler-operator-container/59b29eae not reset as customized by admin to system_u:object_r:container_file_t:s0:c338,c381 Oct 08 18:17:08 crc restorecon[4687]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/containers/kube-scheduler-operator-container/c91a8e4f not reset as customized by admin to system_u:object_r:container_file_t:s0:c338,c381 Oct 08 18:17:08 crc restorecon[4687]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/containers/kube-scheduler-operator-container/4d87494a not reset as customized by admin to system_u:object_r:container_file_t:s0:c442,c857 Oct 08 18:17:08 crc restorecon[4687]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/containers/kube-scheduler-operator-container/1e33ca63 not reset as customized by admin to system_u:object_r:container_file_t:s0:c661,c999 Oct 08 18:17:08 crc restorecon[4687]: /var/lib/kubelet/pods/3ab1a177-2de0-46d9-b765-d0d0649bb42e/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Oct 08 18:17:08 crc restorecon[4687]: /var/lib/kubelet/pods/3ab1a177-2de0-46d9-b765-d0d0649bb42e/containers/kube-rbac-proxy/8dea7be2 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Oct 08 18:17:08 crc restorecon[4687]: /var/lib/kubelet/pods/3ab1a177-2de0-46d9-b765-d0d0649bb42e/containers/kube-rbac-proxy/d0b04a99 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Oct 08 18:17:08 crc restorecon[4687]: /var/lib/kubelet/pods/3ab1a177-2de0-46d9-b765-d0d0649bb42e/containers/kube-rbac-proxy/d84f01e7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Oct 08 18:17:08 crc restorecon[4687]: /var/lib/kubelet/pods/3ab1a177-2de0-46d9-b765-d0d0649bb42e/containers/package-server-manager/4109059b not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Oct 08 18:17:08 crc restorecon[4687]: /var/lib/kubelet/pods/3ab1a177-2de0-46d9-b765-d0d0649bb42e/containers/package-server-manager/a7258a3e not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Oct 08 18:17:08 crc restorecon[4687]: /var/lib/kubelet/pods/3ab1a177-2de0-46d9-b765-d0d0649bb42e/containers/package-server-manager/05bdf2b6 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Oct 08 18:17:08 crc restorecon[4687]: /var/lib/kubelet/pods/6731426b-95fe-49ff-bb5f-40441049fde2/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Oct 08 18:17:08 crc restorecon[4687]: /var/lib/kubelet/pods/6731426b-95fe-49ff-bb5f-40441049fde2/containers/control-plane-machine-set-operator/f3261b51 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Oct 08 18:17:08 crc restorecon[4687]: /var/lib/kubelet/pods/6731426b-95fe-49ff-bb5f-40441049fde2/containers/control-plane-machine-set-operator/315d045e not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Oct 08 18:17:08 crc restorecon[4687]: /var/lib/kubelet/pods/6731426b-95fe-49ff-bb5f-40441049fde2/containers/control-plane-machine-set-operator/5fdcf278 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Oct 08 18:17:08 crc restorecon[4687]: /var/lib/kubelet/pods/6731426b-95fe-49ff-bb5f-40441049fde2/containers/control-plane-machine-set-operator/d053f757 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Oct 08 18:17:08 crc restorecon[4687]: /var/lib/kubelet/pods/6731426b-95fe-49ff-bb5f-40441049fde2/containers/control-plane-machine-set-operator/c2850dc7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Oct 08 18:17:08 crc restorecon[4687]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/volumes/kubernetes.io~configmap/marketplace-trusted-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:17:08 crc restorecon[4687]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/volumes/kubernetes.io~configmap/marketplace-trusted-ca/..2025_02_23_05_22_30.2390596521 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:17:08 crc restorecon[4687]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/volumes/kubernetes.io~configmap/marketplace-trusted-ca/..2025_02_23_05_22_30.2390596521/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:17:08 crc restorecon[4687]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/volumes/kubernetes.io~configmap/marketplace-trusted-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:17:08 crc restorecon[4687]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/volumes/kubernetes.io~configmap/marketplace-trusted-ca/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:17:08 crc restorecon[4687]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:17:08 crc restorecon[4687]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/containers/marketplace-operator/fcfb0b2b not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:17:08 crc restorecon[4687]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/containers/marketplace-operator/c7ac9b7d not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:17:08 crc restorecon[4687]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/containers/marketplace-operator/fa0c0d52 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:17:08 crc restorecon[4687]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/containers/marketplace-operator/c609b6ba not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:17:08 crc restorecon[4687]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/containers/marketplace-operator/2be6c296 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:17:08 crc restorecon[4687]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/containers/marketplace-operator/89a32653 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:17:08 crc restorecon[4687]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/containers/marketplace-operator/4eb9afeb not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:17:08 crc restorecon[4687]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/containers/marketplace-operator/13af6efa not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:17:08 crc restorecon[4687]: /var/lib/kubelet/pods/b6312bbd-5731-4ea0-a20f-81d5a57df44a/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Oct 08 18:17:08 crc restorecon[4687]: /var/lib/kubelet/pods/b6312bbd-5731-4ea0-a20f-81d5a57df44a/containers/olm-operator/b03f9724 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Oct 08 18:17:08 crc restorecon[4687]: /var/lib/kubelet/pods/b6312bbd-5731-4ea0-a20f-81d5a57df44a/containers/olm-operator/e3d105cc not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Oct 08 18:17:08 crc restorecon[4687]: /var/lib/kubelet/pods/b6312bbd-5731-4ea0-a20f-81d5a57df44a/containers/olm-operator/3aed4d83 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Oct 08 18:17:08 crc restorecon[4687]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Oct 08 18:17:08 crc restorecon[4687]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.1906041176 not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Oct 08 18:17:08 crc restorecon[4687]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.1906041176/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Oct 08 18:17:08 crc restorecon[4687]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Oct 08 18:17:08 crc restorecon[4687]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Oct 08 18:17:08 crc restorecon[4687]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Oct 08 18:17:08 crc restorecon[4687]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/containers/kube-storage-version-migrator-operator/0765fa6e not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Oct 08 18:17:08 crc restorecon[4687]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/containers/kube-storage-version-migrator-operator/2cefc627 not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Oct 08 18:17:08 crc restorecon[4687]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/containers/kube-storage-version-migrator-operator/3dcc6345 not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Oct 08 18:17:08 crc restorecon[4687]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/containers/kube-storage-version-migrator-operator/365af391 not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Oct 08 18:17:08 crc restorecon[4687]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/volumes/kubernetes.io~empty-dir/available-featuregates not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Oct 08 18:17:08 crc restorecon[4687]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/volumes/kubernetes.io~empty-dir/available-featuregates/featureGate-SelfManagedHA-Default.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Oct 08 18:17:08 crc restorecon[4687]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/volumes/kubernetes.io~empty-dir/available-featuregates/featureGate-SelfManagedHA-TechPreviewNoUpgrade.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Oct 08 18:17:08 crc restorecon[4687]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/volumes/kubernetes.io~empty-dir/available-featuregates/featureGate-SelfManagedHA-DevPreviewNoUpgrade.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Oct 08 18:17:08 crc restorecon[4687]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/volumes/kubernetes.io~empty-dir/available-featuregates/featureGate-Hypershift-TechPreviewNoUpgrade.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Oct 08 18:17:08 crc restorecon[4687]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/volumes/kubernetes.io~empty-dir/available-featuregates/featureGate-Hypershift-DevPreviewNoUpgrade.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Oct 08 18:17:08 crc restorecon[4687]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/volumes/kubernetes.io~empty-dir/available-featuregates/featureGate-Hypershift-Default.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Oct 08 18:17:08 crc restorecon[4687]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Oct 08 18:17:08 crc restorecon[4687]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/containers/openshift-api/b1130c0f not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Oct 08 18:17:08 crc restorecon[4687]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/containers/openshift-api/236a5913 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Oct 08 18:17:08 crc restorecon[4687]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/containers/openshift-api/b9432e26 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Oct 08 18:17:08 crc restorecon[4687]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/containers/openshift-config-operator/5ddb0e3f not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Oct 08 18:17:08 crc restorecon[4687]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/containers/openshift-config-operator/986dc4fd not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Oct 08 18:17:08 crc restorecon[4687]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/containers/openshift-config-operator/8a23ff9a not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Oct 08 18:17:08 crc restorecon[4687]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/containers/openshift-config-operator/9728ae68 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Oct 08 18:17:08 crc restorecon[4687]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/containers/openshift-config-operator/665f31d0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Oct 08 18:17:08 crc restorecon[4687]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Oct 08 18:17:08 crc restorecon[4687]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.1255385357 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Oct 08 18:17:08 crc restorecon[4687]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.1255385357/operator-config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Oct 08 18:17:08 crc restorecon[4687]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Oct 08 18:17:08 crc restorecon[4687]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/config/operator-config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Oct 08 18:17:08 crc restorecon[4687]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/service-ca-bundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Oct 08 18:17:08 crc restorecon[4687]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/service-ca-bundle/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Oct 08 18:17:08 crc restorecon[4687]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/service-ca-bundle/..2025_02_23_05_23_57.573792656 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Oct 08 18:17:08 crc restorecon[4687]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/service-ca-bundle/..2025_02_23_05_23_57.573792656/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Oct 08 18:17:08 crc restorecon[4687]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/service-ca-bundle/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Oct 08 18:17:08 crc restorecon[4687]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/trusted-ca-bundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Oct 08 18:17:08 crc restorecon[4687]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/trusted-ca-bundle/..2025_02_23_05_22_30.3254245399 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Oct 08 18:17:08 crc restorecon[4687]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/trusted-ca-bundle/..2025_02_23_05_22_30.3254245399/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Oct 08 18:17:08 crc restorecon[4687]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/trusted-ca-bundle/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Oct 08 18:17:08 crc restorecon[4687]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/trusted-ca-bundle/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Oct 08 18:17:08 crc restorecon[4687]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Oct 08 18:17:08 crc restorecon[4687]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/containers/authentication-operator/136c9b42 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Oct 08 18:17:08 crc restorecon[4687]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/containers/authentication-operator/98a1575b not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Oct 08 18:17:08 crc restorecon[4687]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/containers/authentication-operator/cac69136 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Oct 08 18:17:08 crc restorecon[4687]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/containers/authentication-operator/5deb77a7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Oct 08 18:17:08 crc restorecon[4687]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/containers/authentication-operator/2ae53400 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Oct 08 18:17:08 crc restorecon[4687]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Oct 08 18:17:08 crc restorecon[4687]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.3608339744 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Oct 08 18:17:08 crc restorecon[4687]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.3608339744/operator-config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Oct 08 18:17:08 crc restorecon[4687]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Oct 08 18:17:08 crc restorecon[4687]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/volumes/kubernetes.io~configmap/config/operator-config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Oct 08 18:17:08 crc restorecon[4687]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Oct 08 18:17:08 crc restorecon[4687]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/containers/service-ca-operator/e46f2326 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Oct 08 18:17:08 crc restorecon[4687]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/containers/service-ca-operator/dc688d3c not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Oct 08 18:17:08 crc restorecon[4687]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/containers/service-ca-operator/3497c3cd not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Oct 08 18:17:08 crc restorecon[4687]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/containers/service-ca-operator/177eb008 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Oct 08 18:17:08 crc restorecon[4687]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Oct 08 18:17:08 crc restorecon[4687]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.3819292994 not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Oct 08 18:17:08 crc restorecon[4687]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.3819292994/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Oct 08 18:17:08 crc restorecon[4687]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Oct 08 18:17:08 crc restorecon[4687]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Oct 08 18:17:08 crc restorecon[4687]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Oct 08 18:17:08 crc restorecon[4687]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/containers/openshift-apiserver-operator/af5a2afa not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Oct 08 18:17:08 crc restorecon[4687]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/containers/openshift-apiserver-operator/d780cb1f not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Oct 08 18:17:08 crc restorecon[4687]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/containers/openshift-apiserver-operator/49b0f374 not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Oct 08 18:17:08 crc restorecon[4687]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/containers/openshift-apiserver-operator/26fbb125 not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Oct 08 18:17:08 crc restorecon[4687]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/volumes/kubernetes.io~configmap/trusted-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Oct 08 18:17:08 crc restorecon[4687]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/volumes/kubernetes.io~configmap/trusted-ca/..2025_02_23_05_22_30.3244779536 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Oct 08 18:17:08 crc restorecon[4687]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/volumes/kubernetes.io~configmap/trusted-ca/..2025_02_23_05_22_30.3244779536/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Oct 08 18:17:08 crc restorecon[4687]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/volumes/kubernetes.io~configmap/trusted-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Oct 08 18:17:08 crc restorecon[4687]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/volumes/kubernetes.io~configmap/trusted-ca/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Oct 08 18:17:08 crc restorecon[4687]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Oct 08 18:17:08 crc restorecon[4687]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/cf14125a not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Oct 08 18:17:08 crc restorecon[4687]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/b7f86972 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Oct 08 18:17:08 crc restorecon[4687]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/e51d739c not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Oct 08 18:17:08 crc restorecon[4687]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/88ba6a69 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Oct 08 18:17:08 crc restorecon[4687]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/669a9acf not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Oct 08 18:17:08 crc restorecon[4687]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/5cd51231 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Oct 08 18:17:08 crc restorecon[4687]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/75349ec7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Oct 08 18:17:08 crc restorecon[4687]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/15c26839 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Oct 08 18:17:08 crc restorecon[4687]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/45023dcd not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Oct 08 18:17:08 crc restorecon[4687]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/2bb66a50 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Oct 08 18:17:08 crc restorecon[4687]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/kube-rbac-proxy/64d03bdd not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Oct 08 18:17:08 crc restorecon[4687]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/kube-rbac-proxy/ab8e7ca0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Oct 08 18:17:08 crc restorecon[4687]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/kube-rbac-proxy/bb9be25f not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Oct 08 18:17:08 crc restorecon[4687]: /var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/volumes/kubernetes.io~configmap/trusted-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 18:17:08 crc restorecon[4687]: /var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/volumes/kubernetes.io~configmap/trusted-ca/..2025_02_23_05_22_30.2034221258 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 18:17:08 crc restorecon[4687]: /var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/volumes/kubernetes.io~configmap/trusted-ca/..2025_02_23_05_22_30.2034221258/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 18:17:08 crc restorecon[4687]: /var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/volumes/kubernetes.io~configmap/trusted-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 18:17:08 crc restorecon[4687]: /var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/volumes/kubernetes.io~configmap/trusted-ca/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 18:17:08 crc restorecon[4687]: /var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 18:17:08 crc restorecon[4687]: /var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/containers/cluster-image-registry-operator/9a0b61d3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 18:17:08 crc restorecon[4687]: /var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/containers/cluster-image-registry-operator/d471b9d2 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 18:17:08 crc restorecon[4687]: /var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/containers/cluster-image-registry-operator/8cb76b8e not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 18:17:08 crc restorecon[4687]: /var/lib/kubelet/pods/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Oct 08 18:17:08 crc restorecon[4687]: /var/lib/kubelet/pods/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9/containers/catalog-operator/11a00840 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Oct 08 18:17:08 crc restorecon[4687]: /var/lib/kubelet/pods/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9/containers/catalog-operator/ec355a92 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Oct 08 18:17:08 crc restorecon[4687]: /var/lib/kubelet/pods/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9/containers/catalog-operator/992f735e not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Oct 08 18:17:08 crc restorecon[4687]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Oct 08 18:17:08 crc restorecon[4687]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.1782968797 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Oct 08 18:17:08 crc restorecon[4687]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.1782968797/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Oct 08 18:17:08 crc restorecon[4687]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Oct 08 18:17:08 crc restorecon[4687]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Oct 08 18:17:08 crc restorecon[4687]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Oct 08 18:17:08 crc restorecon[4687]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/containers/openshift-controller-manager-operator/d59cdbbc not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Oct 08 18:17:08 crc restorecon[4687]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/containers/openshift-controller-manager-operator/72133ff0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Oct 08 18:17:08 crc restorecon[4687]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/containers/openshift-controller-manager-operator/c56c834c not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Oct 08 18:17:08 crc restorecon[4687]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/containers/openshift-controller-manager-operator/d13724c7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Oct 08 18:17:08 crc restorecon[4687]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/containers/openshift-controller-manager-operator/0a498258 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Oct 08 18:17:08 crc restorecon[4687]: /var/lib/kubelet/pods/5fe579f8-e8a6-4643-bce5-a661393c4dde/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Oct 08 18:17:08 crc restorecon[4687]: /var/lib/kubelet/pods/5fe579f8-e8a6-4643-bce5-a661393c4dde/containers/machine-config-server/fa471982 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Oct 08 18:17:08 crc restorecon[4687]: /var/lib/kubelet/pods/5fe579f8-e8a6-4643-bce5-a661393c4dde/containers/machine-config-server/fc900d92 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Oct 08 18:17:08 crc restorecon[4687]: /var/lib/kubelet/pods/5fe579f8-e8a6-4643-bce5-a661393c4dde/containers/machine-config-server/fa7d68da not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Oct 08 18:17:08 crc restorecon[4687]: /var/lib/kubelet/pods/49ef4625-1d3a-4a9f-b595-c2433d32326d/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c22 Oct 08 18:17:08 crc restorecon[4687]: /var/lib/kubelet/pods/49ef4625-1d3a-4a9f-b595-c2433d32326d/containers/migrator/4bacf9b4 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c22 Oct 08 18:17:08 crc restorecon[4687]: /var/lib/kubelet/pods/49ef4625-1d3a-4a9f-b595-c2433d32326d/containers/migrator/424021b1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c22 Oct 08 18:17:08 crc restorecon[4687]: /var/lib/kubelet/pods/49ef4625-1d3a-4a9f-b595-c2433d32326d/containers/migrator/fc2e31a3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c22 Oct 08 18:17:08 crc restorecon[4687]: /var/lib/kubelet/pods/49ef4625-1d3a-4a9f-b595-c2433d32326d/containers/graceful-termination/f51eefac not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c22 Oct 08 18:17:08 crc restorecon[4687]: /var/lib/kubelet/pods/49ef4625-1d3a-4a9f-b595-c2433d32326d/containers/graceful-termination/c8997f2f not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c22 Oct 08 18:17:08 crc restorecon[4687]: /var/lib/kubelet/pods/49ef4625-1d3a-4a9f-b595-c2433d32326d/containers/graceful-termination/7481f599 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c22 Oct 08 18:17:08 crc restorecon[4687]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/volumes/kubernetes.io~configmap/signing-cabundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Oct 08 18:17:08 crc restorecon[4687]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/volumes/kubernetes.io~configmap/signing-cabundle/..2025_02_23_05_22_49.2255460704 not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Oct 08 18:17:08 crc restorecon[4687]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/volumes/kubernetes.io~configmap/signing-cabundle/..2025_02_23_05_22_49.2255460704/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Oct 08 18:17:08 crc restorecon[4687]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/volumes/kubernetes.io~configmap/signing-cabundle/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Oct 08 18:17:08 crc restorecon[4687]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/volumes/kubernetes.io~configmap/signing-cabundle/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Oct 08 18:17:08 crc restorecon[4687]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Oct 08 18:17:08 crc restorecon[4687]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/containers/service-ca-controller/fdafea19 not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Oct 08 18:17:08 crc restorecon[4687]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/containers/service-ca-controller/d0e1c571 not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Oct 08 18:17:08 crc restorecon[4687]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/containers/service-ca-controller/ee398915 not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Oct 08 18:17:08 crc restorecon[4687]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/containers/service-ca-controller/682bb6b8 not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Oct 08 18:17:08 crc restorecon[4687]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c920 Oct 08 18:17:08 crc restorecon[4687]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/setup/a3e67855 not reset as customized by admin to system_u:object_r:container_file_t:s0:c294,c884 Oct 08 18:17:08 crc restorecon[4687]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/setup/a989f289 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c1016 Oct 08 18:17:08 crc restorecon[4687]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/setup/915431bd not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c920 Oct 08 18:17:08 crc restorecon[4687]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-ensure-env-vars/7796fdab not reset as customized by admin to system_u:object_r:container_file_t:s0:c294,c884 Oct 08 18:17:08 crc restorecon[4687]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-ensure-env-vars/dcdb5f19 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c1016 Oct 08 18:17:08 crc restorecon[4687]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-ensure-env-vars/a3aaa88c not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c920 Oct 08 18:17:08 crc restorecon[4687]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-resources-copy/5508e3e6 not reset as customized by admin to system_u:object_r:container_file_t:s0:c294,c884 Oct 08 18:17:08 crc restorecon[4687]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-resources-copy/160585de not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c1016 Oct 08 18:17:08 crc restorecon[4687]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-resources-copy/e99f8da3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c920 Oct 08 18:17:08 crc restorecon[4687]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcdctl/8bc85570 not reset as customized by admin to system_u:object_r:container_file_t:s0:c294,c884 Oct 08 18:17:08 crc restorecon[4687]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcdctl/a5861c91 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c1016 Oct 08 18:17:08 crc restorecon[4687]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcdctl/84db1135 not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c920 Oct 08 18:17:08 crc restorecon[4687]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd/9e1a6043 not reset as customized by admin to system_u:object_r:container_file_t:s0:c294,c884 Oct 08 18:17:08 crc restorecon[4687]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd/c1aba1c2 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c1016 Oct 08 18:17:08 crc restorecon[4687]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd/d55ccd6d not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c920 Oct 08 18:17:08 crc restorecon[4687]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-metrics/971cc9f6 not reset as customized by admin to system_u:object_r:container_file_t:s0:c294,c884 Oct 08 18:17:08 crc restorecon[4687]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-metrics/8f2e3dcf not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c1016 Oct 08 18:17:08 crc restorecon[4687]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-metrics/ceb35e9c not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c920 Oct 08 18:17:08 crc restorecon[4687]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-readyz/1c192745 not reset as customized by admin to system_u:object_r:container_file_t:s0:c294,c884 Oct 08 18:17:08 crc restorecon[4687]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-readyz/5209e501 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c1016 Oct 08 18:17:08 crc restorecon[4687]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-readyz/f83de4df not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c920 Oct 08 18:17:08 crc restorecon[4687]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-rev/e7b978ac not reset as customized by admin to system_u:object_r:container_file_t:s0:c294,c884 Oct 08 18:17:08 crc restorecon[4687]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-rev/c64304a1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c1016 Oct 08 18:17:08 crc restorecon[4687]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-rev/5384386b not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c920 Oct 08 18:17:08 crc restorecon[4687]: /var/lib/kubelet/pods/efdd0498-1daa-4136-9a4a-3b948c2293fc/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c268,c620 Oct 08 18:17:08 crc restorecon[4687]: /var/lib/kubelet/pods/efdd0498-1daa-4136-9a4a-3b948c2293fc/containers/multus-admission-controller/cce3e3ff not reset as customized by admin to system_u:object_r:container_file_t:s0:c435,c756 Oct 08 18:17:08 crc restorecon[4687]: /var/lib/kubelet/pods/efdd0498-1daa-4136-9a4a-3b948c2293fc/containers/multus-admission-controller/8fb75465 not reset as customized by admin to system_u:object_r:container_file_t:s0:c268,c620 Oct 08 18:17:08 crc restorecon[4687]: /var/lib/kubelet/pods/efdd0498-1daa-4136-9a4a-3b948c2293fc/containers/kube-rbac-proxy/740f573e not reset as customized by admin to system_u:object_r:container_file_t:s0:c435,c756 Oct 08 18:17:08 crc restorecon[4687]: /var/lib/kubelet/pods/efdd0498-1daa-4136-9a4a-3b948c2293fc/containers/kube-rbac-proxy/32fd1134 not reset as customized by admin to system_u:object_r:container_file_t:s0:c268,c620 Oct 08 18:17:08 crc restorecon[4687]: /var/lib/kubelet/pods/20b0d48f-5fd6-431c-a545-e3c800c7b866/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c24 Oct 08 18:17:08 crc restorecon[4687]: /var/lib/kubelet/pods/20b0d48f-5fd6-431c-a545-e3c800c7b866/containers/serve-healthcheck-canary/0a861bd3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c24 Oct 08 18:17:08 crc restorecon[4687]: /var/lib/kubelet/pods/20b0d48f-5fd6-431c-a545-e3c800c7b866/containers/serve-healthcheck-canary/80363026 not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c24 Oct 08 18:17:08 crc restorecon[4687]: /var/lib/kubelet/pods/20b0d48f-5fd6-431c-a545-e3c800c7b866/containers/serve-healthcheck-canary/bfa952a8 not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c24 Oct 08 18:17:08 crc restorecon[4687]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/volumes/kubernetes.io~configmap/auth-proxy-config not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Oct 08 18:17:08 crc restorecon[4687]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/volumes/kubernetes.io~configmap/auth-proxy-config/..2025_02_23_05_33_31.2122464563 not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Oct 08 18:17:08 crc restorecon[4687]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/volumes/kubernetes.io~configmap/auth-proxy-config/..2025_02_23_05_33_31.2122464563/config-file.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Oct 08 18:17:08 crc restorecon[4687]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/volumes/kubernetes.io~configmap/auth-proxy-config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Oct 08 18:17:08 crc restorecon[4687]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/volumes/kubernetes.io~configmap/auth-proxy-config/config-file.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Oct 08 18:17:08 crc restorecon[4687]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Oct 08 18:17:08 crc restorecon[4687]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/volumes/kubernetes.io~configmap/config/..2025_02_23_05_33_31.333075221 not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Oct 08 18:17:08 crc restorecon[4687]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Oct 08 18:17:08 crc restorecon[4687]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Oct 08 18:17:08 crc restorecon[4687]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/containers/kube-rbac-proxy/793bf43d not reset as customized by admin to system_u:object_r:container_file_t:s0:c381,c387 Oct 08 18:17:08 crc restorecon[4687]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/containers/kube-rbac-proxy/7db1bb6e not reset as customized by admin to system_u:object_r:container_file_t:s0:c142,c438 Oct 08 18:17:08 crc restorecon[4687]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/containers/kube-rbac-proxy/4f6a0368 not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Oct 08 18:17:08 crc restorecon[4687]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/containers/machine-approver-controller/c12c7d86 not reset as customized by admin to system_u:object_r:container_file_t:s0:c381,c387 Oct 08 18:17:08 crc restorecon[4687]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/containers/machine-approver-controller/36c4a773 not reset as customized by admin to system_u:object_r:container_file_t:s0:c142,c438 Oct 08 18:17:08 crc restorecon[4687]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/containers/machine-approver-controller/4c1e98ae not reset as customized by admin to system_u:object_r:container_file_t:s0:c142,c438 Oct 08 18:17:08 crc restorecon[4687]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/containers/machine-approver-controller/a4c8115c not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Oct 08 18:17:08 crc restorecon[4687]: /var/lib/kubelet/pods/f4b27818a5e8e43d0dc095d08835c792/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c97,c980 Oct 08 18:17:08 crc restorecon[4687]: /var/lib/kubelet/pods/f4b27818a5e8e43d0dc095d08835c792/containers/setup/7db1802e not reset as customized by admin to system_u:object_r:container_file_t:s0:c97,c980 Oct 08 18:17:08 crc restorecon[4687]: /var/lib/kubelet/pods/f4b27818a5e8e43d0dc095d08835c792/containers/kube-apiserver/a008a7ab not reset as customized by admin to system_u:object_r:container_file_t:s0:c97,c980 Oct 08 18:17:08 crc restorecon[4687]: /var/lib/kubelet/pods/f4b27818a5e8e43d0dc095d08835c792/containers/kube-apiserver-cert-syncer/2c836bac not reset as customized by admin to system_u:object_r:container_file_t:s0:c97,c980 Oct 08 18:17:08 crc restorecon[4687]: /var/lib/kubelet/pods/f4b27818a5e8e43d0dc095d08835c792/containers/kube-apiserver-cert-regeneration-controller/0ce62299 not reset as customized by admin to system_u:object_r:container_file_t:s0:c97,c980 Oct 08 18:17:08 crc restorecon[4687]: /var/lib/kubelet/pods/f4b27818a5e8e43d0dc095d08835c792/containers/kube-apiserver-insecure-readyz/945d2457 not reset as customized by admin to system_u:object_r:container_file_t:s0:c97,c980 Oct 08 18:17:08 crc restorecon[4687]: /var/lib/kubelet/pods/f4b27818a5e8e43d0dc095d08835c792/containers/kube-apiserver-check-endpoints/7d5c1dd8 not reset as customized by admin to system_u:object_r:container_file_t:s0:c97,c980 Oct 08 18:17:08 crc restorecon[4687]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/utilities not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:17:08 crc restorecon[4687]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/utilities/copy-content not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:17:08 crc restorecon[4687]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:17:08 crc restorecon[4687]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:17:08 crc restorecon[4687]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/3scale-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:17:08 crc restorecon[4687]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/3scale-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:17:08 crc restorecon[4687]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/advanced-cluster-management not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:17:08 crc restorecon[4687]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/advanced-cluster-management/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:17:09 crc restorecon[4687]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq-broker-rhel8 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:17:09 crc restorecon[4687]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq-broker-rhel8/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:17:09 crc restorecon[4687]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq-online not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:17:09 crc restorecon[4687]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq-online/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:17:09 crc restorecon[4687]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq-streams not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:17:09 crc restorecon[4687]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq-streams/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:17:09 crc restorecon[4687]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq-streams-console not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:17:09 crc restorecon[4687]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq-streams-console/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:17:09 crc restorecon[4687]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq7-interconnect-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:17:09 crc restorecon[4687]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq7-interconnect-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:17:09 crc restorecon[4687]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ansible-automation-platform-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:17:09 crc restorecon[4687]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ansible-automation-platform-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:17:09 crc restorecon[4687]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ansible-cloud-addons-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:17:09 crc restorecon[4687]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ansible-cloud-addons-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:17:09 crc restorecon[4687]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicast-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:17:09 crc restorecon[4687]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicast-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:17:09 crc restorecon[4687]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicurio-registry-3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:17:09 crc restorecon[4687]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicurio-registry-3/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:17:09 crc restorecon[4687]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/authorino-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:17:09 crc restorecon[4687]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/authorino-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:17:09 crc restorecon[4687]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aws-load-balancer-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:17:09 crc restorecon[4687]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aws-load-balancer-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:17:09 crc restorecon[4687]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/bamoe-businessautomation-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:17:09 crc restorecon[4687]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/bamoe-businessautomation-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:17:09 crc restorecon[4687]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/bamoe-kogito-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:17:09 crc restorecon[4687]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/bamoe-kogito-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:17:09 crc restorecon[4687]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/bpfman-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:17:09 crc restorecon[4687]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/bpfman-operator/index.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:17:09 crc restorecon[4687]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/businessautomation-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:17:09 crc restorecon[4687]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/businessautomation-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:17:09 crc restorecon[4687]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cephcsi-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:17:09 crc restorecon[4687]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cephcsi-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:17:09 crc restorecon[4687]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cincinnati-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:17:09 crc restorecon[4687]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cincinnati-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:17:09 crc restorecon[4687]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-kube-descheduler-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:17:09 crc restorecon[4687]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-kube-descheduler-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:17:09 crc restorecon[4687]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-logging not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:17:09 crc restorecon[4687]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-logging/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:17:09 crc restorecon[4687]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-observability-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:17:09 crc restorecon[4687]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-observability-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:17:09 crc restorecon[4687]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/compliance-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:17:09 crc restorecon[4687]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/compliance-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:17:09 crc restorecon[4687]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/container-security-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:17:09 crc restorecon[4687]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/container-security-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:17:09 crc restorecon[4687]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/costmanagement-metrics-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:17:09 crc restorecon[4687]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/costmanagement-metrics-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:17:09 crc restorecon[4687]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cryostat-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:17:09 crc restorecon[4687]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cryostat-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:17:09 crc restorecon[4687]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datagrid not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:17:09 crc restorecon[4687]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datagrid/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:17:09 crc restorecon[4687]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/devspaces not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:17:09 crc restorecon[4687]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/devspaces/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:17:09 crc restorecon[4687]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/devworkspace-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:17:09 crc restorecon[4687]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/devworkspace-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:17:09 crc restorecon[4687]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dpu-network-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:17:09 crc restorecon[4687]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dpu-network-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:17:09 crc restorecon[4687]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eap not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:17:09 crc restorecon[4687]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eap/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:17:09 crc restorecon[4687]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/elasticsearch-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:17:09 crc restorecon[4687]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/elasticsearch-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:17:09 crc restorecon[4687]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/external-dns-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:17:09 crc restorecon[4687]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/external-dns-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:17:09 crc restorecon[4687]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fence-agents-remediation not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:17:09 crc restorecon[4687]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fence-agents-remediation/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:17:09 crc restorecon[4687]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/file-integrity-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:17:09 crc restorecon[4687]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/file-integrity-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:17:09 crc restorecon[4687]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fuse-apicurito not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:17:09 crc restorecon[4687]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fuse-apicurito/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:17:09 crc restorecon[4687]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fuse-console not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:17:09 crc restorecon[4687]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fuse-console/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:17:09 crc restorecon[4687]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fuse-online not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:17:09 crc restorecon[4687]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fuse-online/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:17:09 crc restorecon[4687]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/gatekeeper-operator-product not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:17:09 crc restorecon[4687]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/gatekeeper-operator-product/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:17:09 crc restorecon[4687]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/jaeger-product not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:17:09 crc restorecon[4687]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/jaeger-product/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:17:09 crc restorecon[4687]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/jws-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:17:09 crc restorecon[4687]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/jws-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:17:09 crc restorecon[4687]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kernel-module-management not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:17:09 crc restorecon[4687]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kernel-module-management/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:17:09 crc restorecon[4687]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kernel-module-management-hub not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:17:09 crc restorecon[4687]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kernel-module-management-hub/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:17:09 crc restorecon[4687]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kiali-ossm not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:17:09 crc restorecon[4687]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kiali-ossm/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:17:09 crc restorecon[4687]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubevirt-hyperconverged not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:17:09 crc restorecon[4687]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubevirt-hyperconverged/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:17:09 crc restorecon[4687]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/logic-operator-rhel8 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:17:09 crc restorecon[4687]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/logic-operator-rhel8/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:17:09 crc restorecon[4687]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/loki-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:17:09 crc restorecon[4687]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/loki-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:17:09 crc restorecon[4687]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/lvms-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:17:09 crc restorecon[4687]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/lvms-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:17:09 crc restorecon[4687]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/machine-deletion-remediation not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:17:09 crc restorecon[4687]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/machine-deletion-remediation/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:17:09 crc restorecon[4687]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mcg-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:17:09 crc restorecon[4687]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mcg-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:17:09 crc restorecon[4687]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mta-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:17:09 crc restorecon[4687]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mta-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:17:09 crc restorecon[4687]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mtc-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:17:09 crc restorecon[4687]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mtc-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:17:09 crc restorecon[4687]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mtr-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:17:09 crc restorecon[4687]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mtr-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:17:09 crc restorecon[4687]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mtv-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:17:09 crc restorecon[4687]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mtv-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:17:09 crc restorecon[4687]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/multicluster-engine not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:17:09 crc restorecon[4687]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/multicluster-engine/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:17:09 crc restorecon[4687]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netobserv-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:17:09 crc restorecon[4687]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netobserv-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:17:09 crc restorecon[4687]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-healthcheck-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:17:09 crc restorecon[4687]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-healthcheck-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:17:09 crc restorecon[4687]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-maintenance-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:17:09 crc restorecon[4687]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-maintenance-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:17:09 crc restorecon[4687]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-observability-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:17:09 crc restorecon[4687]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-observability-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:17:09 crc restorecon[4687]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ocs-client-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:17:09 crc restorecon[4687]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ocs-client-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:17:09 crc restorecon[4687]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ocs-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:17:09 crc restorecon[4687]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ocs-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:17:09 crc restorecon[4687]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odf-csi-addons-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:17:09 crc restorecon[4687]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odf-csi-addons-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:17:09 crc restorecon[4687]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odf-multicluster-orchestrator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:17:09 crc restorecon[4687]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odf-multicluster-orchestrator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:17:09 crc restorecon[4687]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odf-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:17:09 crc restorecon[4687]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odf-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:17:09 crc restorecon[4687]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odf-prometheus-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:17:09 crc restorecon[4687]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odf-prometheus-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:17:09 crc restorecon[4687]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odr-cluster-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:17:09 crc restorecon[4687]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odr-cluster-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:17:09 crc restorecon[4687]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odr-hub-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:17:09 crc restorecon[4687]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odr-hub-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:17:09 crc restorecon[4687]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-cert-manager-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:17:09 crc restorecon[4687]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-cert-manager-operator/bundle-v1.15.0.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:17:09 crc restorecon[4687]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-cert-manager-operator/channel.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:17:09 crc restorecon[4687]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-cert-manager-operator/package.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:17:09 crc restorecon[4687]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-custom-metrics-autoscaler-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:17:09 crc restorecon[4687]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-custom-metrics-autoscaler-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:17:09 crc restorecon[4687]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-gitops-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:17:09 crc restorecon[4687]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-gitops-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:17:09 crc restorecon[4687]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-pipelines-operator-rh not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:17:09 crc restorecon[4687]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-pipelines-operator-rh/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:17:09 crc restorecon[4687]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-secondary-scheduler-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:17:09 crc restorecon[4687]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-secondary-scheduler-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:17:09 crc restorecon[4687]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/opentelemetry-product not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:17:09 crc restorecon[4687]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/opentelemetry-product/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:17:09 crc restorecon[4687]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/quay-bridge-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:17:09 crc restorecon[4687]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/quay-bridge-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:17:09 crc restorecon[4687]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/quay-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:17:09 crc restorecon[4687]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/quay-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:17:09 crc restorecon[4687]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/recipe not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:17:09 crc restorecon[4687]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/recipe/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:17:09 crc restorecon[4687]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/red-hat-camel-k not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:17:09 crc restorecon[4687]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/red-hat-camel-k/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:17:09 crc restorecon[4687]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/red-hat-hawtio-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:17:09 crc restorecon[4687]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/red-hat-hawtio-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:17:09 crc restorecon[4687]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/redhat-oadp-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:17:09 crc restorecon[4687]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/redhat-oadp-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:17:09 crc restorecon[4687]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rh-service-binding-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:17:09 crc restorecon[4687]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rh-service-binding-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:17:09 crc restorecon[4687]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhacs-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:17:09 crc restorecon[4687]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhacs-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:17:09 crc restorecon[4687]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhbk-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:17:09 crc restorecon[4687]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhbk-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:17:09 crc restorecon[4687]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhdh not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:17:09 crc restorecon[4687]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhdh/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:17:09 crc restorecon[4687]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhods-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:17:09 crc restorecon[4687]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhods-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:17:09 crc restorecon[4687]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhods-prometheus-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:17:09 crc restorecon[4687]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhods-prometheus-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:17:09 crc restorecon[4687]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhpam-kogito-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:17:09 crc restorecon[4687]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhpam-kogito-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:17:09 crc restorecon[4687]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhsso-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:17:09 crc restorecon[4687]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhsso-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:17:09 crc restorecon[4687]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rook-ceph-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:17:09 crc restorecon[4687]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rook-ceph-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:17:09 crc restorecon[4687]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/run-once-duration-override-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:17:09 crc restorecon[4687]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/run-once-duration-override-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:17:09 crc restorecon[4687]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sandboxed-containers-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:17:09 crc restorecon[4687]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sandboxed-containers-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:17:09 crc restorecon[4687]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/security-profiles-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:17:09 crc restorecon[4687]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/security-profiles-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:17:09 crc restorecon[4687]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/self-node-remediation not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:17:09 crc restorecon[4687]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/self-node-remediation/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:17:09 crc restorecon[4687]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/serverless-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:17:09 crc restorecon[4687]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/serverless-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:17:09 crc restorecon[4687]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/service-registry-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:17:09 crc restorecon[4687]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/service-registry-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:17:09 crc restorecon[4687]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/servicemeshoperator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:17:09 crc restorecon[4687]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/servicemeshoperator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:17:09 crc restorecon[4687]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/servicemeshoperator3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:17:09 crc restorecon[4687]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/servicemeshoperator3/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:17:09 crc restorecon[4687]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/skupper-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:17:09 crc restorecon[4687]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/skupper-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:17:09 crc restorecon[4687]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/submariner not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:17:09 crc restorecon[4687]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/submariner/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:17:09 crc restorecon[4687]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tang-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:17:09 crc restorecon[4687]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tang-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:17:09 crc restorecon[4687]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tempo-product not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:17:09 crc restorecon[4687]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tempo-product/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:17:09 crc restorecon[4687]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/trustee-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:17:09 crc restorecon[4687]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/trustee-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:17:09 crc restorecon[4687]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/volsync-product not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:17:09 crc restorecon[4687]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/volsync-product/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:17:09 crc restorecon[4687]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/web-terminal not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:17:09 crc restorecon[4687]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/web-terminal/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:17:09 crc restorecon[4687]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:17:09 crc restorecon[4687]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:17:09 crc restorecon[4687]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:17:09 crc restorecon[4687]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/00000-1.psg not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:17:09 crc restorecon[4687]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/00000-1.psg.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:17:09 crc restorecon[4687]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/db.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:17:09 crc restorecon[4687]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/index.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:17:09 crc restorecon[4687]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/main.pix not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:17:09 crc restorecon[4687]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/overflow.pix not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:17:09 crc restorecon[4687]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/digest not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:17:09 crc restorecon[4687]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:17:09 crc restorecon[4687]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/containers/extract-utilities/bc8d0691 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:17:09 crc restorecon[4687]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/containers/extract-utilities/6b76097a not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:17:09 crc restorecon[4687]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/containers/extract-utilities/34d1af30 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:17:09 crc restorecon[4687]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/containers/extract-content/312ba61c not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:17:09 crc restorecon[4687]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/containers/extract-content/645d5dd1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:17:09 crc restorecon[4687]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/containers/extract-content/16e825f0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:17:09 crc restorecon[4687]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/containers/registry-server/4cf51fc9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:17:09 crc restorecon[4687]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/containers/registry-server/2a23d348 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:17:09 crc restorecon[4687]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/containers/registry-server/075dbd49 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:17:09 crc restorecon[4687]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes/kubernetes.io~configmap/serviceca not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Oct 08 18:17:09 crc restorecon[4687]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes/kubernetes.io~configmap/serviceca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Oct 08 18:17:09 crc restorecon[4687]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes/kubernetes.io~configmap/serviceca/..2025_02_24_06_09_13.3521195566 not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Oct 08 18:17:09 crc restorecon[4687]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes/kubernetes.io~configmap/serviceca/..2025_02_24_06_09_13.3521195566/image-registry.openshift-image-registry.svc..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Oct 08 18:17:09 crc restorecon[4687]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes/kubernetes.io~configmap/serviceca/..2025_02_24_06_09_13.3521195566/image-registry.openshift-image-registry.svc.cluster.local..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Oct 08 18:17:09 crc restorecon[4687]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes/kubernetes.io~configmap/serviceca/..2025_02_24_06_09_13.3521195566/default-route-openshift-image-registry.apps-crc.testing not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Oct 08 18:17:09 crc restorecon[4687]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes/kubernetes.io~configmap/serviceca/default-route-openshift-image-registry.apps-crc.testing not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Oct 08 18:17:09 crc restorecon[4687]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes/kubernetes.io~configmap/serviceca/image-registry.openshift-image-registry.svc..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Oct 08 18:17:09 crc restorecon[4687]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes/kubernetes.io~configmap/serviceca/image-registry.openshift-image-registry.svc.cluster.local..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Oct 08 18:17:09 crc restorecon[4687]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Oct 08 18:17:09 crc restorecon[4687]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/containers/node-ca/dd585ddd not reset as customized by admin to system_u:object_r:container_file_t:s0:c377,c642 Oct 08 18:17:09 crc restorecon[4687]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/containers/node-ca/17ebd0ab not reset as customized by admin to system_u:object_r:container_file_t:s0:c338,c343 Oct 08 18:17:09 crc restorecon[4687]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/containers/node-ca/005579f4 not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Oct 08 18:17:09 crc restorecon[4687]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/etcd-serving-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Oct 08 18:17:09 crc restorecon[4687]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/etcd-serving-ca/..2025_02_23_05_23_11.449897510 not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Oct 08 18:17:09 crc restorecon[4687]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/etcd-serving-ca/..2025_02_23_05_23_11.449897510/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Oct 08 18:17:09 crc restorecon[4687]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/etcd-serving-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Oct 08 18:17:09 crc restorecon[4687]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/etcd-serving-ca/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Oct 08 18:17:09 crc restorecon[4687]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/trusted-ca-bundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Oct 08 18:17:09 crc restorecon[4687]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/trusted-ca-bundle/..2025_02_23_05_23_11.1287037894 not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Oct 08 18:17:09 crc restorecon[4687]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/trusted-ca-bundle/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Oct 08 18:17:09 crc restorecon[4687]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/audit-policies not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Oct 08 18:17:09 crc restorecon[4687]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/audit-policies/..2025_02_23_05_23_11.1301053334 not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Oct 08 18:17:09 crc restorecon[4687]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/audit-policies/..2025_02_23_05_23_11.1301053334/policy.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Oct 08 18:17:09 crc restorecon[4687]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/audit-policies/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Oct 08 18:17:09 crc restorecon[4687]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/audit-policies/policy.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Oct 08 18:17:09 crc restorecon[4687]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Oct 08 18:17:09 crc restorecon[4687]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/containers/fix-audit-permissions/bf5f3b9c not reset as customized by admin to system_u:object_r:container_file_t:s0:c49,c263 Oct 08 18:17:09 crc restorecon[4687]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/containers/fix-audit-permissions/af276eb7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c701 Oct 08 18:17:09 crc restorecon[4687]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/containers/fix-audit-permissions/ea28e322 not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Oct 08 18:17:09 crc restorecon[4687]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/containers/oauth-apiserver/692e6683 not reset as customized by admin to system_u:object_r:container_file_t:s0:c49,c263 Oct 08 18:17:09 crc restorecon[4687]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/containers/oauth-apiserver/871746a7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c701 Oct 08 18:17:09 crc restorecon[4687]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/containers/oauth-apiserver/4eb2e958 not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Oct 08 18:17:09 crc restorecon[4687]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/console-config not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Oct 08 18:17:09 crc restorecon[4687]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/console-config/..2025_02_24_06_09_06.2875086261 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Oct 08 18:17:09 crc restorecon[4687]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/console-config/..2025_02_24_06_09_06.2875086261/console-config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Oct 08 18:17:09 crc restorecon[4687]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/console-config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Oct 08 18:17:09 crc restorecon[4687]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/console-config/console-config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Oct 08 18:17:09 crc restorecon[4687]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/trusted-ca-bundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Oct 08 18:17:09 crc restorecon[4687]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/trusted-ca-bundle/..2025_02_24_06_09_06.286118152 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Oct 08 18:17:09 crc restorecon[4687]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/trusted-ca-bundle/..2025_02_24_06_09_06.286118152/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Oct 08 18:17:09 crc restorecon[4687]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/trusted-ca-bundle/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Oct 08 18:17:09 crc restorecon[4687]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/trusted-ca-bundle/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Oct 08 18:17:09 crc restorecon[4687]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/oauth-serving-cert not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Oct 08 18:17:09 crc restorecon[4687]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/oauth-serving-cert/..2025_02_24_06_09_06.3865795478 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Oct 08 18:17:09 crc restorecon[4687]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/oauth-serving-cert/..2025_02_24_06_09_06.3865795478/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Oct 08 18:17:09 crc restorecon[4687]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/oauth-serving-cert/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Oct 08 18:17:09 crc restorecon[4687]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/oauth-serving-cert/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Oct 08 18:17:09 crc restorecon[4687]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/service-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Oct 08 18:17:09 crc restorecon[4687]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/service-ca/..2025_02_24_06_09_06.584414814 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Oct 08 18:17:09 crc restorecon[4687]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/service-ca/..2025_02_24_06_09_06.584414814/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Oct 08 18:17:09 crc restorecon[4687]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/service-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Oct 08 18:17:09 crc restorecon[4687]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/service-ca/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Oct 08 18:17:09 crc restorecon[4687]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Oct 08 18:17:09 crc restorecon[4687]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/containers/console/ca9b62da not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Oct 08 18:17:09 crc restorecon[4687]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/containers/console/0edd6fce not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Oct 08 18:17:09 crc restorecon[4687]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Oct 08 18:17:09 crc restorecon[4687]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/..2025_02_24_06_20_07.2406383837 not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Oct 08 18:17:09 crc restorecon[4687]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/..2025_02_24_06_20_07.2406383837/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Oct 08 18:17:09 crc restorecon[4687]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/..2025_02_24_06_20_07.2406383837/openshift-controller-manager.client-ca.configmap not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Oct 08 18:17:09 crc restorecon[4687]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/..2025_02_24_06_20_07.2406383837/openshift-controller-manager.openshift-global-ca.configmap not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Oct 08 18:17:09 crc restorecon[4687]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/..2025_02_24_06_20_07.2406383837/openshift-controller-manager.serving-cert.secret not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Oct 08 18:17:09 crc restorecon[4687]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Oct 08 18:17:09 crc restorecon[4687]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Oct 08 18:17:09 crc restorecon[4687]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/openshift-controller-manager.client-ca.configmap not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Oct 08 18:17:09 crc restorecon[4687]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/openshift-controller-manager.openshift-global-ca.configmap not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Oct 08 18:17:09 crc restorecon[4687]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/openshift-controller-manager.serving-cert.secret not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Oct 08 18:17:09 crc restorecon[4687]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/client-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Oct 08 18:17:09 crc restorecon[4687]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/client-ca/..2025_02_24_06_20_07.1071801880 not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Oct 08 18:17:09 crc restorecon[4687]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/client-ca/..2025_02_24_06_20_07.1071801880/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Oct 08 18:17:09 crc restorecon[4687]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/client-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Oct 08 18:17:09 crc restorecon[4687]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/client-ca/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Oct 08 18:17:09 crc restorecon[4687]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/proxy-ca-bundles not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Oct 08 18:17:09 crc restorecon[4687]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/proxy-ca-bundles/..2025_02_24_06_20_07.2494444877 not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Oct 08 18:17:09 crc restorecon[4687]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/proxy-ca-bundles/..2025_02_24_06_20_07.2494444877/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Oct 08 18:17:09 crc restorecon[4687]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/proxy-ca-bundles/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Oct 08 18:17:09 crc restorecon[4687]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/proxy-ca-bundles/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Oct 08 18:17:09 crc restorecon[4687]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Oct 08 18:17:09 crc restorecon[4687]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/containers/controller-manager/89b4555f not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Oct 08 18:17:09 crc restorecon[4687]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/volumes/kubernetes.io~configmap/config-volume not reset as customized by admin to system_u:object_r:container_file_t:s0:c466,c972 Oct 08 18:17:09 crc restorecon[4687]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/volumes/kubernetes.io~configmap/config-volume/..2025_02_23_05_23_22.4071100442 not reset as customized by admin to system_u:object_r:container_file_t:s0:c466,c972 Oct 08 18:17:09 crc restorecon[4687]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/volumes/kubernetes.io~configmap/config-volume/..2025_02_23_05_23_22.4071100442/Corefile not reset as customized by admin to system_u:object_r:container_file_t:s0:c466,c972 Oct 08 18:17:09 crc restorecon[4687]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/volumes/kubernetes.io~configmap/config-volume/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c466,c972 Oct 08 18:17:09 crc restorecon[4687]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/volumes/kubernetes.io~configmap/config-volume/Corefile not reset as customized by admin to system_u:object_r:container_file_t:s0:c466,c972 Oct 08 18:17:09 crc restorecon[4687]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c466,c972 Oct 08 18:17:09 crc restorecon[4687]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/containers/dns/655fcd71 not reset as customized by admin to system_u:object_r:container_file_t:s0:c457,c841 Oct 08 18:17:09 crc restorecon[4687]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/containers/dns/0d43c002 not reset as customized by admin to system_u:object_r:container_file_t:s0:c55,c1022 Oct 08 18:17:09 crc restorecon[4687]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/containers/dns/e68efd17 not reset as customized by admin to system_u:object_r:container_file_t:s0:c466,c972 Oct 08 18:17:09 crc restorecon[4687]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/containers/kube-rbac-proxy/9acf9b65 not reset as customized by admin to system_u:object_r:container_file_t:s0:c457,c841 Oct 08 18:17:09 crc restorecon[4687]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/containers/kube-rbac-proxy/5ae3ff11 not reset as customized by admin to system_u:object_r:container_file_t:s0:c55,c1022 Oct 08 18:17:09 crc restorecon[4687]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/containers/kube-rbac-proxy/1e59206a not reset as customized by admin to system_u:object_r:container_file_t:s0:c466,c972 Oct 08 18:17:09 crc restorecon[4687]: /var/lib/kubelet/pods/44663579-783b-4372-86d6-acf235a62d72/containers/dns-node-resolver/27af16d1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c304,c1017 Oct 08 18:17:09 crc restorecon[4687]: /var/lib/kubelet/pods/44663579-783b-4372-86d6-acf235a62d72/containers/dns-node-resolver/7918e729 not reset as customized by admin to system_u:object_r:container_file_t:s0:c853,c893 Oct 08 18:17:09 crc restorecon[4687]: /var/lib/kubelet/pods/44663579-783b-4372-86d6-acf235a62d72/containers/dns-node-resolver/5d976d0e not reset as customized by admin to system_u:object_r:container_file_t:s0:c585,c981 Oct 08 18:17:09 crc restorecon[4687]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Oct 08 18:17:09 crc restorecon[4687]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/config/..2025_02_23_05_38_56.1112187283 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Oct 08 18:17:09 crc restorecon[4687]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/config/..2025_02_23_05_38_56.1112187283/controller-config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Oct 08 18:17:09 crc restorecon[4687]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Oct 08 18:17:09 crc restorecon[4687]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/config/controller-config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Oct 08 18:17:09 crc restorecon[4687]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/trusted-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Oct 08 18:17:09 crc restorecon[4687]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/trusted-ca/..2025_02_23_05_38_56.2839772658 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Oct 08 18:17:09 crc restorecon[4687]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/trusted-ca/..2025_02_23_05_38_56.2839772658/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Oct 08 18:17:09 crc restorecon[4687]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/trusted-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Oct 08 18:17:09 crc restorecon[4687]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/trusted-ca/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Oct 08 18:17:09 crc restorecon[4687]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Oct 08 18:17:09 crc restorecon[4687]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/containers/console-operator/d7f55cbb not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Oct 08 18:17:09 crc restorecon[4687]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/containers/console-operator/f0812073 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Oct 08 18:17:09 crc restorecon[4687]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/containers/console-operator/1a56cbeb not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Oct 08 18:17:09 crc restorecon[4687]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/containers/console-operator/7fdd437e not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Oct 08 18:17:09 crc restorecon[4687]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/containers/console-operator/cdfb5652 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Oct 08 18:17:09 crc restorecon[4687]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/etcd-serving-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Oct 08 18:17:09 crc restorecon[4687]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/etcd-serving-ca/..2025_02_24_06_17_29.3844392896 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Oct 08 18:17:09 crc restorecon[4687]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/etcd-serving-ca/..2025_02_24_06_17_29.3844392896/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Oct 08 18:17:09 crc restorecon[4687]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/etcd-serving-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Oct 08 18:17:09 crc restorecon[4687]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/etcd-serving-ca/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Oct 08 18:17:09 crc restorecon[4687]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Oct 08 18:17:09 crc restorecon[4687]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/config/..2025_02_24_06_17_29.848549803 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Oct 08 18:17:09 crc restorecon[4687]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/config/..2025_02_24_06_17_29.848549803/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Oct 08 18:17:09 crc restorecon[4687]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Oct 08 18:17:09 crc restorecon[4687]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Oct 08 18:17:09 crc restorecon[4687]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/audit not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Oct 08 18:17:09 crc restorecon[4687]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/audit/..2025_02_24_06_17_29.780046231 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Oct 08 18:17:09 crc restorecon[4687]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/audit/..2025_02_24_06_17_29.780046231/policy.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Oct 08 18:17:09 crc restorecon[4687]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/audit/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Oct 08 18:17:09 crc restorecon[4687]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/audit/policy.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Oct 08 18:17:09 crc restorecon[4687]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/image-import-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Oct 08 18:17:09 crc restorecon[4687]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/image-import-ca/..2025_02_24_06_17_29.2926008347 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Oct 08 18:17:09 crc restorecon[4687]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/image-import-ca/..2025_02_24_06_17_29.2926008347/image-registry.openshift-image-registry.svc..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Oct 08 18:17:09 crc restorecon[4687]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/image-import-ca/..2025_02_24_06_17_29.2926008347/image-registry.openshift-image-registry.svc.cluster.local..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Oct 08 18:17:09 crc restorecon[4687]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/image-import-ca/..2025_02_24_06_17_29.2926008347/default-route-openshift-image-registry.apps-crc.testing not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Oct 08 18:17:09 crc restorecon[4687]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/image-import-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Oct 08 18:17:09 crc restorecon[4687]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/image-import-ca/image-registry.openshift-image-registry.svc..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Oct 08 18:17:09 crc restorecon[4687]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/image-import-ca/image-registry.openshift-image-registry.svc.cluster.local..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Oct 08 18:17:09 crc restorecon[4687]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/image-import-ca/default-route-openshift-image-registry.apps-crc.testing not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Oct 08 18:17:09 crc restorecon[4687]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/trusted-ca-bundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Oct 08 18:17:09 crc restorecon[4687]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/trusted-ca-bundle/..2025_02_24_06_17_29.2729721485 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Oct 08 18:17:09 crc restorecon[4687]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/trusted-ca-bundle/..2025_02_24_06_17_29.2729721485/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Oct 08 18:17:09 crc restorecon[4687]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/trusted-ca-bundle/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Oct 08 18:17:09 crc restorecon[4687]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/trusted-ca-bundle/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Oct 08 18:17:09 crc restorecon[4687]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Oct 08 18:17:09 crc restorecon[4687]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/containers/fix-audit-permissions/fb93119e not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Oct 08 18:17:09 crc restorecon[4687]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/containers/openshift-apiserver/f1e8fc0e not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Oct 08 18:17:09 crc restorecon[4687]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/containers/openshift-apiserver-check-endpoints/218511f3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Oct 08 18:17:09 crc restorecon[4687]: /var/lib/kubelet/pods/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b/volumes/kubernetes.io~empty-dir/tmpfs not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Oct 08 18:17:09 crc restorecon[4687]: /var/lib/kubelet/pods/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b/volumes/kubernetes.io~empty-dir/tmpfs/k8s-webhook-server not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Oct 08 18:17:09 crc restorecon[4687]: /var/lib/kubelet/pods/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b/volumes/kubernetes.io~empty-dir/tmpfs/k8s-webhook-server/serving-certs not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Oct 08 18:17:09 crc restorecon[4687]: /var/lib/kubelet/pods/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Oct 08 18:17:09 crc restorecon[4687]: /var/lib/kubelet/pods/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b/containers/packageserver/ca8af7b3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Oct 08 18:17:09 crc restorecon[4687]: /var/lib/kubelet/pods/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b/containers/packageserver/72cc8a75 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Oct 08 18:17:09 crc restorecon[4687]: /var/lib/kubelet/pods/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b/containers/packageserver/6e8a3760 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Oct 08 18:17:09 crc restorecon[4687]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/volumes/kubernetes.io~configmap/service-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Oct 08 18:17:09 crc restorecon[4687]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/volumes/kubernetes.io~configmap/service-ca/..2025_02_23_05_27_30.557428972 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Oct 08 18:17:09 crc restorecon[4687]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/volumes/kubernetes.io~configmap/service-ca/..2025_02_23_05_27_30.557428972/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Oct 08 18:17:09 crc restorecon[4687]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/volumes/kubernetes.io~configmap/service-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Oct 08 18:17:09 crc restorecon[4687]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/volumes/kubernetes.io~configmap/service-ca/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Oct 08 18:17:09 crc restorecon[4687]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Oct 08 18:17:09 crc restorecon[4687]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/containers/cluster-version-operator/4c3455c0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Oct 08 18:17:09 crc restorecon[4687]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/containers/cluster-version-operator/2278acb0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Oct 08 18:17:09 crc restorecon[4687]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/containers/cluster-version-operator/4b453e4f not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Oct 08 18:17:09 crc restorecon[4687]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/containers/cluster-version-operator/3ec09bda not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Oct 08 18:17:09 crc restorecon[4687]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/trusted-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 18:17:09 crc restorecon[4687]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/trusted-ca/..2025_02_24_06_25_03.422633132 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 18:17:09 crc restorecon[4687]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/trusted-ca/..2025_02_24_06_25_03.422633132/anchors not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 18:17:09 crc restorecon[4687]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/trusted-ca/..2025_02_24_06_25_03.422633132/anchors/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 18:17:09 crc restorecon[4687]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/trusted-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 18:17:09 crc restorecon[4687]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/trusted-ca/anchors not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 18:17:09 crc restorecon[4687]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/registry-certificates not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 18:17:09 crc restorecon[4687]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/registry-certificates/..2025_02_24_06_25_03.3594477318 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 18:17:09 crc restorecon[4687]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/registry-certificates/..2025_02_24_06_25_03.3594477318/image-registry.openshift-image-registry.svc..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 18:17:09 crc restorecon[4687]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/registry-certificates/..2025_02_24_06_25_03.3594477318/image-registry.openshift-image-registry.svc.cluster.local..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 18:17:09 crc restorecon[4687]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/registry-certificates/..2025_02_24_06_25_03.3594477318/default-route-openshift-image-registry.apps-crc.testing not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 18:17:09 crc restorecon[4687]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/registry-certificates/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 18:17:09 crc restorecon[4687]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/registry-certificates/image-registry.openshift-image-registry.svc..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 18:17:09 crc restorecon[4687]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/registry-certificates/image-registry.openshift-image-registry.svc.cluster.local..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 18:17:09 crc restorecon[4687]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/registry-certificates/default-route-openshift-image-registry.apps-crc.testing not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 18:17:09 crc restorecon[4687]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 18:17:09 crc restorecon[4687]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/edk2 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 18:17:09 crc restorecon[4687]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/edk2/cacerts.bin not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 18:17:09 crc restorecon[4687]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/java not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 18:17:09 crc restorecon[4687]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/java/cacerts not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 18:17:09 crc restorecon[4687]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/openssl not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 18:17:09 crc restorecon[4687]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/openssl/ca-bundle.trust.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 18:17:09 crc restorecon[4687]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 18:17:09 crc restorecon[4687]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 18:17:09 crc restorecon[4687]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/email-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 18:17:09 crc restorecon[4687]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/objsign-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 18:17:09 crc restorecon[4687]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 18:17:09 crc restorecon[4687]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/2ae6433e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 18:17:09 crc restorecon[4687]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/fde84897.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 18:17:09 crc restorecon[4687]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/75680d2e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 18:17:09 crc restorecon[4687]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/openshift-service-serving-signer_1740288168.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 18:17:09 crc restorecon[4687]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/facfc4fa.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 18:17:09 crc restorecon[4687]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8f5a969c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 18:17:09 crc restorecon[4687]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/CFCA_EV_ROOT.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 18:17:09 crc restorecon[4687]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9ef4a08a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 18:17:09 crc restorecon[4687]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ingress-operator_1740288202.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 18:17:09 crc restorecon[4687]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/2f332aed.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 18:17:09 crc restorecon[4687]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/248c8271.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 18:17:09 crc restorecon[4687]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8d10a21f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 18:17:09 crc restorecon[4687]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ACCVRAIZ1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 18:17:09 crc restorecon[4687]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/a94d09e5.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 18:17:09 crc restorecon[4687]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3c9a4d3b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 18:17:09 crc restorecon[4687]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/40193066.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 18:17:09 crc restorecon[4687]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/AC_RAIZ_FNMT-RCM.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 18:17:09 crc restorecon[4687]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/cd8c0d63.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 18:17:09 crc restorecon[4687]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b936d1c6.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 18:17:09 crc restorecon[4687]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/CA_Disig_Root_R2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 18:17:09 crc restorecon[4687]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/4fd49c6c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 18:17:09 crc restorecon[4687]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/AC_RAIZ_FNMT-RCM_SERVIDORES_SEGUROS.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 18:17:09 crc restorecon[4687]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b81b93f0.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 18:17:09 crc restorecon[4687]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5f9a69fa.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 18:17:09 crc restorecon[4687]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Certigna.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 18:17:09 crc restorecon[4687]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b30d5fda.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 18:17:09 crc restorecon[4687]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ANF_Secure_Server_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 18:17:09 crc restorecon[4687]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b433981b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 18:17:09 crc restorecon[4687]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/93851c9e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 18:17:09 crc restorecon[4687]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9282e51c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 18:17:09 crc restorecon[4687]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e7dd1bc4.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 18:17:09 crc restorecon[4687]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Actalis_Authentication_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 18:17:09 crc restorecon[4687]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/930ac5d2.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 18:17:09 crc restorecon[4687]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5f47b495.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 18:17:09 crc restorecon[4687]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e113c810.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 18:17:09 crc restorecon[4687]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5931b5bc.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 18:17:09 crc restorecon[4687]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/AffirmTrust_Commercial.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 18:17:09 crc restorecon[4687]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/2b349938.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 18:17:09 crc restorecon[4687]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e48193cf.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 18:17:09 crc restorecon[4687]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/302904dd.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 18:17:09 crc restorecon[4687]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/a716d4ed.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 18:17:09 crc restorecon[4687]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/AffirmTrust_Networking.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 18:17:09 crc restorecon[4687]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/93bc0acc.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 18:17:09 crc restorecon[4687]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/86212b19.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 18:17:09 crc restorecon[4687]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Certigna_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 18:17:09 crc restorecon[4687]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/AffirmTrust_Premium.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 18:17:09 crc restorecon[4687]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b727005e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 18:17:09 crc restorecon[4687]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/dbc54cab.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 18:17:09 crc restorecon[4687]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f51bb24c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 18:17:09 crc restorecon[4687]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/c28a8a30.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 18:17:09 crc restorecon[4687]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/AffirmTrust_Premium_ECC.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 18:17:09 crc restorecon[4687]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9c8dfbd4.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 18:17:09 crc restorecon[4687]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ccc52f49.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 18:17:09 crc restorecon[4687]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/cb1c3204.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 18:17:09 crc restorecon[4687]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Amazon_Root_CA_1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 18:17:09 crc restorecon[4687]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ce5e74ef.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 18:17:09 crc restorecon[4687]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/fd08c599.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 18:17:09 crc restorecon[4687]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Certum_Trusted_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 18:17:09 crc restorecon[4687]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Amazon_Root_CA_2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 18:17:09 crc restorecon[4687]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/6d41d539.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 18:17:09 crc restorecon[4687]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/fb5fa911.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 18:17:09 crc restorecon[4687]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e35234b1.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 18:17:09 crc restorecon[4687]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Amazon_Root_CA_3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 18:17:09 crc restorecon[4687]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8cb5ee0f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 18:17:09 crc restorecon[4687]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/7a7c655d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 18:17:09 crc restorecon[4687]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f8fc53da.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 18:17:09 crc restorecon[4687]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Amazon_Root_CA_4.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 18:17:09 crc restorecon[4687]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/de6d66f3.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 18:17:09 crc restorecon[4687]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d41b5e2a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 18:17:09 crc restorecon[4687]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/41a3f684.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 18:17:09 crc restorecon[4687]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1df5a75f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 18:17:09 crc restorecon[4687]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Atos_TrustedRoot_2011.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 18:17:09 crc restorecon[4687]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e36a6752.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 18:17:09 crc restorecon[4687]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b872f2b4.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 18:17:09 crc restorecon[4687]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9576d26b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 18:17:09 crc restorecon[4687]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/228f89db.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 18:17:09 crc restorecon[4687]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Atos_TrustedRoot_Root_CA_ECC_TLS_2021.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 18:17:09 crc restorecon[4687]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/fb717492.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 18:17:09 crc restorecon[4687]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/2d21b73c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 18:17:09 crc restorecon[4687]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/0b1b94ef.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 18:17:09 crc restorecon[4687]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/595e996b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 18:17:09 crc restorecon[4687]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Atos_TrustedRoot_Root_CA_RSA_TLS_2021.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 18:17:09 crc restorecon[4687]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9b46e03d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 18:17:09 crc restorecon[4687]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/128f4b91.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 18:17:09 crc restorecon[4687]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Buypass_Class_3_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 18:17:09 crc restorecon[4687]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/81f2d2b1.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 18:17:09 crc restorecon[4687]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Autoridad_de_Certificacion_Firmaprofesional_CIF_A62634068.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 18:17:09 crc restorecon[4687]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3bde41ac.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 18:17:09 crc restorecon[4687]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d16a5865.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 18:17:09 crc restorecon[4687]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Certum_EC-384_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 18:17:09 crc restorecon[4687]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/BJCA_Global_Root_CA1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 18:17:09 crc restorecon[4687]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/0179095f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 18:17:09 crc restorecon[4687]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ffa7f1eb.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 18:17:09 crc restorecon[4687]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9482e63a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 18:17:09 crc restorecon[4687]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d4dae3dd.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 18:17:09 crc restorecon[4687]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/BJCA_Global_Root_CA2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 18:17:09 crc restorecon[4687]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3e359ba6.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 18:17:09 crc restorecon[4687]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/7e067d03.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 18:17:09 crc restorecon[4687]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/95aff9e3.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 18:17:09 crc restorecon[4687]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d7746a63.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 18:17:09 crc restorecon[4687]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Baltimore_CyberTrust_Root.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 18:17:09 crc restorecon[4687]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/653b494a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 18:17:09 crc restorecon[4687]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3ad48a91.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 18:17:09 crc restorecon[4687]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Certum_Trusted_Network_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 18:17:09 crc restorecon[4687]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Buypass_Class_2_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 18:17:09 crc restorecon[4687]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/54657681.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 18:17:09 crc restorecon[4687]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/82223c44.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 18:17:09 crc restorecon[4687]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e8de2f56.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 18:17:09 crc restorecon[4687]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/2d9dafe4.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 18:17:09 crc restorecon[4687]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d96b65e2.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 18:17:09 crc restorecon[4687]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ee64a828.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 18:17:09 crc restorecon[4687]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/COMODO_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 18:17:09 crc restorecon[4687]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/40547a79.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 18:17:09 crc restorecon[4687]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5a3f0ff8.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 18:17:09 crc restorecon[4687]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/7a780d93.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 18:17:09 crc restorecon[4687]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/34d996fb.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 18:17:09 crc restorecon[4687]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/COMODO_ECC_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 18:17:09 crc restorecon[4687]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/eed8c118.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 18:17:09 crc restorecon[4687]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/89c02a45.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 18:17:09 crc restorecon[4687]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Certainly_Root_R1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 18:17:09 crc restorecon[4687]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b1159c4c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 18:17:09 crc restorecon[4687]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/COMODO_RSA_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 18:17:09 crc restorecon[4687]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d6325660.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 18:17:09 crc restorecon[4687]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d4c339cb.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 18:17:09 crc restorecon[4687]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8312c4c1.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 18:17:09 crc restorecon[4687]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Certainly_Root_E1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 18:17:09 crc restorecon[4687]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8508e720.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 18:17:09 crc restorecon[4687]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5fdd185d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 18:17:09 crc restorecon[4687]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/48bec511.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 18:17:09 crc restorecon[4687]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/69105f4f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 18:17:09 crc restorecon[4687]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GlobalSign.1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 18:17:09 crc restorecon[4687]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/0b9bc432.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 18:17:09 crc restorecon[4687]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Certum_Trusted_Network_CA_2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 18:17:09 crc restorecon[4687]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GTS_Root_R3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 18:17:09 crc restorecon[4687]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/32888f65.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 18:17:09 crc restorecon[4687]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/CommScope_Public_Trust_ECC_Root-01.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 18:17:09 crc restorecon[4687]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/6b03dec0.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 18:17:09 crc restorecon[4687]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/219d9499.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 18:17:09 crc restorecon[4687]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/CommScope_Public_Trust_ECC_Root-02.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 18:17:09 crc restorecon[4687]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5acf816d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 18:17:09 crc restorecon[4687]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/cbf06781.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 18:17:09 crc restorecon[4687]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/CommScope_Public_Trust_RSA_Root-01.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 18:17:09 crc restorecon[4687]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GTS_Root_R4.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 18:17:09 crc restorecon[4687]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/dc99f41e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 18:17:09 crc restorecon[4687]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/CommScope_Public_Trust_RSA_Root-02.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 18:17:09 crc restorecon[4687]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GlobalSign.3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 18:17:09 crc restorecon[4687]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/AAA_Certificate_Services.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 18:17:09 crc restorecon[4687]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/985c1f52.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 18:17:09 crc restorecon[4687]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8794b4e3.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 18:17:09 crc restorecon[4687]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/D-TRUST_BR_Root_CA_1_2020.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 18:17:09 crc restorecon[4687]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e7c037b4.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 18:17:09 crc restorecon[4687]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ef954a4e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 18:17:09 crc restorecon[4687]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/D-TRUST_EV_Root_CA_1_2020.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 18:17:09 crc restorecon[4687]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/2add47b6.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 18:17:09 crc restorecon[4687]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/90c5a3c8.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 18:17:09 crc restorecon[4687]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/D-TRUST_Root_Class_3_CA_2_2009.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 18:17:09 crc restorecon[4687]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b0f3e76e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 18:17:09 crc restorecon[4687]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/53a1b57a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 18:17:09 crc restorecon[4687]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/D-TRUST_Root_Class_3_CA_2_EV_2009.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 18:17:09 crc restorecon[4687]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GlobalSign_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 18:17:09 crc restorecon[4687]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_Assured_ID_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 18:17:09 crc restorecon[4687]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5ad8a5d6.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 18:17:09 crc restorecon[4687]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/68dd7389.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 18:17:09 crc restorecon[4687]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_Assured_ID_Root_G2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 18:17:09 crc restorecon[4687]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9d04f354.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 18:17:09 crc restorecon[4687]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8d6437c3.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 18:17:09 crc restorecon[4687]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/062cdee6.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 18:17:09 crc restorecon[4687]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/bd43e1dd.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 18:17:09 crc restorecon[4687]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_Assured_ID_Root_G3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 18:17:09 crc restorecon[4687]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/7f3d5d1d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 18:17:09 crc restorecon[4687]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/c491639e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 18:17:09 crc restorecon[4687]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GlobalSign_Root_E46.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 18:17:09 crc restorecon[4687]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_Global_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 18:17:09 crc restorecon[4687]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3513523f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 18:17:09 crc restorecon[4687]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/399e7759.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 18:17:09 crc restorecon[4687]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/feffd413.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 18:17:09 crc restorecon[4687]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d18e9066.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 18:17:09 crc restorecon[4687]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_Global_Root_G2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 18:17:09 crc restorecon[4687]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/607986c7.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 18:17:09 crc restorecon[4687]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/c90bc37d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 18:17:09 crc restorecon[4687]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1b0f7e5c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 18:17:09 crc restorecon[4687]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1e08bfd1.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 18:17:09 crc restorecon[4687]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_Global_Root_G3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 18:17:09 crc restorecon[4687]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/dd8e9d41.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 18:17:09 crc restorecon[4687]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ed39abd0.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 18:17:09 crc restorecon[4687]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/a3418fda.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 18:17:09 crc restorecon[4687]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/bc3f2570.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 18:17:09 crc restorecon[4687]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_High_Assurance_EV_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 18:17:09 crc restorecon[4687]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/244b5494.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 18:17:09 crc restorecon[4687]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/81b9768f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 18:17:09 crc restorecon[4687]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GlobalSign.2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 18:17:09 crc restorecon[4687]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/4be590e0.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 18:17:09 crc restorecon[4687]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_TLS_ECC_P384_Root_G5.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 18:17:09 crc restorecon[4687]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9846683b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 18:17:09 crc restorecon[4687]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/252252d2.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 18:17:09 crc restorecon[4687]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1e8e7201.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 18:17:09 crc restorecon[4687]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ISRG_Root_X1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 18:17:09 crc restorecon[4687]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_TLS_RSA4096_Root_G5.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 18:17:09 crc restorecon[4687]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d52c538d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 18:17:09 crc restorecon[4687]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/c44cc0c0.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 18:17:09 crc restorecon[4687]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GlobalSign_Root_R46.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 18:17:09 crc restorecon[4687]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_Trusted_Root_G4.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 18:17:09 crc restorecon[4687]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/75d1b2ed.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 18:17:09 crc restorecon[4687]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/a2c66da8.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 18:17:09 crc restorecon[4687]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GTS_Root_R2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 18:17:09 crc restorecon[4687]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ecccd8db.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 18:17:09 crc restorecon[4687]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Entrust.net_Certification_Authority__2048_.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 18:17:09 crc restorecon[4687]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/aee5f10d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 18:17:09 crc restorecon[4687]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3e7271e8.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 18:17:09 crc restorecon[4687]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b0e59380.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 18:17:09 crc restorecon[4687]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/4c3982f2.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 18:17:09 crc restorecon[4687]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Entrust_Root_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 18:17:09 crc restorecon[4687]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/6b99d060.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 18:17:09 crc restorecon[4687]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/bf64f35b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 18:17:09 crc restorecon[4687]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/0a775a30.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 18:17:09 crc restorecon[4687]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/002c0b4f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 18:17:09 crc restorecon[4687]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/cc450945.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 18:17:09 crc restorecon[4687]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Entrust_Root_Certification_Authority_-_EC1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 18:17:09 crc restorecon[4687]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/106f3e4d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 18:17:09 crc restorecon[4687]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b3fb433b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 18:17:09 crc restorecon[4687]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GlobalSign.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 18:17:09 crc restorecon[4687]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/4042bcee.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 18:17:09 crc restorecon[4687]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Entrust_Root_Certification_Authority_-_G2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 18:17:09 crc restorecon[4687]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/02265526.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 18:17:09 crc restorecon[4687]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/455f1b52.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 18:17:09 crc restorecon[4687]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/0d69c7e1.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 18:17:09 crc restorecon[4687]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9f727ac7.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 18:17:09 crc restorecon[4687]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Entrust_Root_Certification_Authority_-_G4.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 18:17:09 crc restorecon[4687]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5e98733a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 18:17:09 crc restorecon[4687]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f0cd152c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 18:17:09 crc restorecon[4687]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/dc4d6a89.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 18:17:09 crc restorecon[4687]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/6187b673.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 18:17:09 crc restorecon[4687]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/FIRMAPROFESIONAL_CA_ROOT-A_WEB.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 18:17:09 crc restorecon[4687]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ba8887ce.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 18:17:09 crc restorecon[4687]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/068570d1.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 18:17:09 crc restorecon[4687]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f081611a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 18:17:09 crc restorecon[4687]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/48a195d8.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 18:17:09 crc restorecon[4687]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GDCA_TrustAUTH_R5_ROOT.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 18:17:09 crc restorecon[4687]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/0f6fa695.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 18:17:09 crc restorecon[4687]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ab59055e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 18:17:09 crc restorecon[4687]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b92fd57f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 18:17:09 crc restorecon[4687]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GLOBALTRUST_2020.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 18:17:09 crc restorecon[4687]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/fa5da96b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 18:17:09 crc restorecon[4687]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1ec40989.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 18:17:09 crc restorecon[4687]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/7719f463.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 18:17:09 crc restorecon[4687]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GTS_Root_R1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 18:17:09 crc restorecon[4687]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1001acf7.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 18:17:09 crc restorecon[4687]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f013ecaf.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 18:17:09 crc restorecon[4687]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/626dceaf.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 18:17:09 crc restorecon[4687]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/c559d742.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 18:17:09 crc restorecon[4687]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1d3472b9.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 18:17:09 crc restorecon[4687]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9479c8c3.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 18:17:09 crc restorecon[4687]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/a81e292b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 18:17:09 crc restorecon[4687]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/4bfab552.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 18:17:09 crc restorecon[4687]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Go_Daddy_Class_2_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 18:17:09 crc restorecon[4687]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Sectigo_Public_Server_Authentication_Root_E46.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 18:17:09 crc restorecon[4687]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Go_Daddy_Root_Certificate_Authority_-_G2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 18:17:09 crc restorecon[4687]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e071171e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 18:17:09 crc restorecon[4687]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/57bcb2da.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 18:17:09 crc restorecon[4687]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/HARICA_TLS_ECC_Root_CA_2021.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 18:17:09 crc restorecon[4687]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ab5346f4.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 18:17:09 crc restorecon[4687]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5046c355.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 18:17:09 crc restorecon[4687]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/HARICA_TLS_RSA_Root_CA_2021.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 18:17:09 crc restorecon[4687]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/865fbdf9.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 18:17:09 crc restorecon[4687]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/da0cfd1d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 18:17:09 crc restorecon[4687]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/85cde254.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 18:17:09 crc restorecon[4687]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Hellenic_Academic_and_Research_Institutions_ECC_RootCA_2015.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 18:17:09 crc restorecon[4687]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/cbb3f32b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 18:17:09 crc restorecon[4687]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SecureSign_RootCA11.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 18:17:09 crc restorecon[4687]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Hellenic_Academic_and_Research_Institutions_RootCA_2015.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 18:17:09 crc restorecon[4687]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5860aaa6.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 18:17:09 crc restorecon[4687]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/31188b5e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 18:17:09 crc restorecon[4687]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/HiPKI_Root_CA_-_G1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 18:17:09 crc restorecon[4687]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/c7f1359b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 18:17:09 crc restorecon[4687]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5f15c80c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 18:17:09 crc restorecon[4687]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Hongkong_Post_Root_CA_3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 18:17:09 crc restorecon[4687]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/09789157.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 18:17:09 crc restorecon[4687]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ISRG_Root_X2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 18:17:09 crc restorecon[4687]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/18856ac4.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 18:17:09 crc restorecon[4687]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1e09d511.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 18:17:09 crc restorecon[4687]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/IdenTrust_Commercial_Root_CA_1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 18:17:09 crc restorecon[4687]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/cf701eeb.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 18:17:09 crc restorecon[4687]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d06393bb.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 18:17:09 crc restorecon[4687]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/IdenTrust_Public_Sector_Root_CA_1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 18:17:09 crc restorecon[4687]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/10531352.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 18:17:09 crc restorecon[4687]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Izenpe.com.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 18:17:09 crc restorecon[4687]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SecureTrust_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 18:17:09 crc restorecon[4687]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b0ed035a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 18:17:09 crc restorecon[4687]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Microsec_e-Szigno_Root_CA_2009.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 18:17:09 crc restorecon[4687]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8160b96c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 18:17:09 crc restorecon[4687]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e8651083.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 18:17:09 crc restorecon[4687]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/2c63f966.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 18:17:09 crc restorecon[4687]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Security_Communication_RootCA2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 18:17:09 crc restorecon[4687]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Microsoft_ECC_Root_Certificate_Authority_2017.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 18:17:09 crc restorecon[4687]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8d89cda1.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 18:17:09 crc restorecon[4687]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/01419da9.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 18:17:09 crc restorecon[4687]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SSL.com_TLS_RSA_Root_CA_2022.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 18:17:09 crc restorecon[4687]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b7a5b843.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 18:17:09 crc restorecon[4687]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Microsoft_RSA_Root_Certificate_Authority_2017.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 18:17:09 crc restorecon[4687]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/bf53fb88.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 18:17:09 crc restorecon[4687]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9591a472.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 18:17:09 crc restorecon[4687]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3afde786.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 18:17:09 crc restorecon[4687]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SwissSign_Gold_CA_-_G2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 18:17:09 crc restorecon[4687]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/NAVER_Global_Root_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 18:17:09 crc restorecon[4687]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3fb36b73.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 18:17:09 crc restorecon[4687]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d39b0a2c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 18:17:09 crc restorecon[4687]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/a89d74c2.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 18:17:09 crc restorecon[4687]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/cd58d51e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 18:17:09 crc restorecon[4687]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b7db1890.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 18:17:09 crc restorecon[4687]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/NetLock_Arany__Class_Gold__F__tan__s__tv__ny.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 18:17:09 crc restorecon[4687]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/988a38cb.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 18:17:09 crc restorecon[4687]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/60afe812.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 18:17:09 crc restorecon[4687]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f39fc864.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 18:17:09 crc restorecon[4687]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5443e9e3.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 18:17:09 crc restorecon[4687]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/OISTE_WISeKey_Global_Root_GB_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 18:17:09 crc restorecon[4687]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e73d606e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 18:17:09 crc restorecon[4687]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/dfc0fe80.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 18:17:09 crc restorecon[4687]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b66938e9.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 18:17:09 crc restorecon[4687]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1e1eab7c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 18:17:09 crc restorecon[4687]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/OISTE_WISeKey_Global_Root_GC_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 18:17:09 crc restorecon[4687]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/773e07ad.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 18:17:09 crc restorecon[4687]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3c899c73.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 18:17:09 crc restorecon[4687]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d59297b8.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 18:17:09 crc restorecon[4687]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ddcda989.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 18:17:09 crc restorecon[4687]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/QuoVadis_Root_CA_1_G3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 18:17:09 crc restorecon[4687]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/749e9e03.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 18:17:09 crc restorecon[4687]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/52b525c7.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 18:17:09 crc restorecon[4687]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Security_Communication_RootCA3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 18:17:09 crc restorecon[4687]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/QuoVadis_Root_CA_2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 18:17:09 crc restorecon[4687]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d7e8dc79.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 18:17:09 crc restorecon[4687]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/7a819ef2.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 18:17:09 crc restorecon[4687]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/08063a00.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 18:17:09 crc restorecon[4687]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/6b483515.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 18:17:09 crc restorecon[4687]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/QuoVadis_Root_CA_2_G3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 18:17:09 crc restorecon[4687]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/064e0aa9.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 18:17:09 crc restorecon[4687]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1f58a078.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 18:17:09 crc restorecon[4687]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/6f7454b3.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 18:17:09 crc restorecon[4687]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/7fa05551.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 18:17:09 crc restorecon[4687]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/QuoVadis_Root_CA_3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 18:17:09 crc restorecon[4687]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/76faf6c0.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 18:17:09 crc restorecon[4687]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9339512a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 18:17:09 crc restorecon[4687]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f387163d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 18:17:09 crc restorecon[4687]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ee37c333.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 18:17:09 crc restorecon[4687]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/QuoVadis_Root_CA_3_G3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 18:17:09 crc restorecon[4687]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e18bfb83.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 18:17:09 crc restorecon[4687]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e442e424.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 18:17:09 crc restorecon[4687]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/fe8a2cd8.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 18:17:09 crc restorecon[4687]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/23f4c490.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 18:17:09 crc restorecon[4687]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5cd81ad7.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 18:17:09 crc restorecon[4687]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SSL.com_EV_Root_Certification_Authority_ECC.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 18:17:09 crc restorecon[4687]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f0c70a8d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 18:17:09 crc restorecon[4687]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/7892ad52.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 18:17:09 crc restorecon[4687]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SZAFIR_ROOT_CA2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 18:17:09 crc restorecon[4687]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/4f316efb.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 18:17:09 crc restorecon[4687]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SSL.com_EV_Root_Certification_Authority_RSA_R2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 18:17:09 crc restorecon[4687]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/06dc52d5.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 18:17:09 crc restorecon[4687]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/583d0756.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 18:17:09 crc restorecon[4687]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Sectigo_Public_Server_Authentication_Root_R46.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 18:17:09 crc restorecon[4687]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SSL.com_Root_Certification_Authority_ECC.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 18:17:09 crc restorecon[4687]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/0bf05006.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 18:17:09 crc restorecon[4687]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/88950faa.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 18:17:09 crc restorecon[4687]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9046744a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 18:17:09 crc restorecon[4687]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3c860d51.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 18:17:09 crc restorecon[4687]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SSL.com_Root_Certification_Authority_RSA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 18:17:09 crc restorecon[4687]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/6fa5da56.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 18:17:09 crc restorecon[4687]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/33ee480d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 18:17:09 crc restorecon[4687]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Secure_Global_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 18:17:09 crc restorecon[4687]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/63a2c897.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 18:17:09 crc restorecon[4687]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SSL.com_TLS_ECC_Root_CA_2022.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 18:17:09 crc restorecon[4687]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/bdacca6f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 18:17:09 crc restorecon[4687]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ff34af3f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 18:17:09 crc restorecon[4687]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/dbff3a01.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 18:17:09 crc restorecon[4687]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Security_Communication_ECC_RootCA1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 18:17:09 crc restorecon[4687]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/emSign_Root_CA_-_C1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 18:17:09 crc restorecon[4687]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Starfield_Class_2_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 18:17:09 crc restorecon[4687]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/406c9bb1.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 18:17:09 crc restorecon[4687]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Starfield_Root_Certificate_Authority_-_G2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 18:17:09 crc restorecon[4687]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/emSign_ECC_Root_CA_-_C3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 18:17:09 crc restorecon[4687]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Starfield_Services_Root_Certificate_Authority_-_G2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 18:17:09 crc restorecon[4687]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SwissSign_Silver_CA_-_G2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 18:17:09 crc restorecon[4687]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/99e1b953.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 18:17:09 crc restorecon[4687]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/T-TeleSec_GlobalRoot_Class_2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 18:17:09 crc restorecon[4687]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/vTrus_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 18:17:09 crc restorecon[4687]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/T-TeleSec_GlobalRoot_Class_3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 18:17:09 crc restorecon[4687]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/14bc7599.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 18:17:09 crc restorecon[4687]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/TUBITAK_Kamu_SM_SSL_Kok_Sertifikasi_-_Surum_1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 18:17:09 crc restorecon[4687]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/TWCA_Global_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 18:17:09 crc restorecon[4687]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/7a3adc42.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 18:17:09 crc restorecon[4687]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/TWCA_Root_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 18:17:09 crc restorecon[4687]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f459871d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 18:17:09 crc restorecon[4687]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Telekom_Security_TLS_ECC_Root_2020.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 18:17:09 crc restorecon[4687]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/emSign_Root_CA_-_G1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 18:17:09 crc restorecon[4687]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Telekom_Security_TLS_RSA_Root_2023.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 18:17:09 crc restorecon[4687]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/TeliaSonera_Root_CA_v1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 18:17:09 crc restorecon[4687]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Telia_Root_CA_v2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 18:17:09 crc restorecon[4687]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8f103249.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 18:17:09 crc restorecon[4687]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f058632f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 18:17:09 crc restorecon[4687]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ca-certificates.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 18:17:09 crc restorecon[4687]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/TrustAsia_Global_Root_CA_G3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 18:17:09 crc restorecon[4687]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9bf03295.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 18:17:09 crc restorecon[4687]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/98aaf404.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 18:17:09 crc restorecon[4687]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 18:17:09 crc restorecon[4687]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/TrustAsia_Global_Root_CA_G4.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 18:17:09 crc restorecon[4687]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1cef98f5.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 18:17:09 crc restorecon[4687]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/073bfcc5.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 18:17:09 crc restorecon[4687]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/2923b3f9.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 18:17:09 crc restorecon[4687]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Trustwave_Global_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 18:17:09 crc restorecon[4687]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f249de83.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 18:17:09 crc restorecon[4687]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/edcbddb5.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 18:17:09 crc restorecon[4687]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/emSign_ECC_Root_CA_-_G3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 18:17:09 crc restorecon[4687]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Trustwave_Global_ECC_P256_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 18:17:09 crc restorecon[4687]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9b5697b0.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 18:17:09 crc restorecon[4687]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1ae85e5e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 18:17:09 crc restorecon[4687]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b74d2bd5.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 18:17:09 crc restorecon[4687]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Trustwave_Global_ECC_P384_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 18:17:09 crc restorecon[4687]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d887a5bb.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 18:17:09 crc restorecon[4687]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9aef356c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 18:17:09 crc restorecon[4687]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/TunTrust_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 18:17:09 crc restorecon[4687]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/fd64f3fc.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 18:17:09 crc restorecon[4687]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e13665f9.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 18:17:09 crc restorecon[4687]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/UCA_Extended_Validation_Root.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 18:17:09 crc restorecon[4687]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/0f5dc4f3.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 18:17:09 crc restorecon[4687]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/da7377f6.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 18:17:09 crc restorecon[4687]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/UCA_Global_G2_Root.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 18:17:09 crc restorecon[4687]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/c01eb047.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 18:17:09 crc restorecon[4687]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/304d27c3.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 18:17:09 crc restorecon[4687]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ed858448.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 18:17:09 crc restorecon[4687]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/USERTrust_ECC_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 18:17:09 crc restorecon[4687]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f30dd6ad.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 18:17:09 crc restorecon[4687]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/04f60c28.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 18:17:09 crc restorecon[4687]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/vTrus_ECC_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 18:17:09 crc restorecon[4687]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/USERTrust_RSA_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 18:17:09 crc restorecon[4687]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/fc5a8f99.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 18:17:09 crc restorecon[4687]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/35105088.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 18:17:09 crc restorecon[4687]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ee532fd5.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 18:17:09 crc restorecon[4687]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/XRamp_Global_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 18:17:09 crc restorecon[4687]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/706f604c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 18:17:09 crc restorecon[4687]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/76579174.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 18:17:09 crc restorecon[4687]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/certSIGN_ROOT_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 18:17:09 crc restorecon[4687]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8d86cdd1.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 18:17:09 crc restorecon[4687]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/882de061.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 18:17:09 crc restorecon[4687]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/certSIGN_ROOT_CA_G2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 18:17:09 crc restorecon[4687]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5f618aec.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 18:17:09 crc restorecon[4687]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/a9d40e02.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 18:17:09 crc restorecon[4687]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e-Szigno_Root_CA_2017.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 18:17:09 crc restorecon[4687]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e868b802.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 18:17:09 crc restorecon[4687]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/83e9984f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 18:17:09 crc restorecon[4687]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ePKI_Root_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 18:17:09 crc restorecon[4687]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ca6e4ad9.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 18:17:09 crc restorecon[4687]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9d6523ce.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 18:17:09 crc restorecon[4687]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/4b718d9b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 18:17:09 crc restorecon[4687]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/869fbf79.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 18:17:09 crc restorecon[4687]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 18:17:09 crc restorecon[4687]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/containers/registry/f8d22bdb not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 08 18:17:09 crc restorecon[4687]: /var/lib/kubelet/pods/a0128f3a-b052-44ed-a84e-c4c8aaf17c13/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c17 Oct 08 18:17:09 crc restorecon[4687]: /var/lib/kubelet/pods/a0128f3a-b052-44ed-a84e-c4c8aaf17c13/containers/cluster-samples-operator/6e8bbfac not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c17 Oct 08 18:17:09 crc restorecon[4687]: /var/lib/kubelet/pods/a0128f3a-b052-44ed-a84e-c4c8aaf17c13/containers/cluster-samples-operator/54dd7996 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c17 Oct 08 18:17:09 crc restorecon[4687]: /var/lib/kubelet/pods/a0128f3a-b052-44ed-a84e-c4c8aaf17c13/containers/cluster-samples-operator/a4f1bb05 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c17 Oct 08 18:17:09 crc restorecon[4687]: /var/lib/kubelet/pods/a0128f3a-b052-44ed-a84e-c4c8aaf17c13/containers/cluster-samples-operator-watch/207129da not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c17 Oct 08 18:17:09 crc restorecon[4687]: /var/lib/kubelet/pods/a0128f3a-b052-44ed-a84e-c4c8aaf17c13/containers/cluster-samples-operator-watch/c1df39e1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c17 Oct 08 18:17:09 crc restorecon[4687]: /var/lib/kubelet/pods/a0128f3a-b052-44ed-a84e-c4c8aaf17c13/containers/cluster-samples-operator-watch/15b8f1cd not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c17 Oct 08 18:17:09 crc restorecon[4687]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Oct 08 18:17:09 crc restorecon[4687]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/config/..2025_02_23_05_27_49.3523263858 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Oct 08 18:17:09 crc restorecon[4687]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/config/..2025_02_23_05_27_49.3523263858/config-file.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Oct 08 18:17:09 crc restorecon[4687]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Oct 08 18:17:09 crc restorecon[4687]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/config/config-file.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Oct 08 18:17:09 crc restorecon[4687]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/images not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Oct 08 18:17:09 crc restorecon[4687]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/images/..2025_02_23_05_27_49.3256605594 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Oct 08 18:17:09 crc restorecon[4687]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/images/..2025_02_23_05_27_49.3256605594/images.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Oct 08 18:17:09 crc restorecon[4687]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/images/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Oct 08 18:17:09 crc restorecon[4687]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/images/images.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Oct 08 18:17:09 crc restorecon[4687]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Oct 08 18:17:09 crc restorecon[4687]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/containers/kube-rbac-proxy/77bd6913 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Oct 08 18:17:09 crc restorecon[4687]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/containers/kube-rbac-proxy/2382c1b1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Oct 08 18:17:09 crc restorecon[4687]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/containers/kube-rbac-proxy/704ce128 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Oct 08 18:17:09 crc restorecon[4687]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/containers/machine-api-operator/70d16fe0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Oct 08 18:17:09 crc restorecon[4687]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/containers/machine-api-operator/bfb95535 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Oct 08 18:17:09 crc restorecon[4687]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/containers/machine-api-operator/57a8e8e2 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Oct 08 18:17:09 crc restorecon[4687]: /var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c219,c404 Oct 08 18:17:09 crc restorecon[4687]: /var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/volumes/kubernetes.io~configmap/config/..2025_02_23_05_27_49.3413793711 not reset as customized by admin to system_u:object_r:container_file_t:s0:c219,c404 Oct 08 18:17:09 crc restorecon[4687]: /var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/volumes/kubernetes.io~configmap/config/..2025_02_23_05_27_49.3413793711/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c219,c404 Oct 08 18:17:09 crc restorecon[4687]: /var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c219,c404 Oct 08 18:17:09 crc restorecon[4687]: /var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c219,c404 Oct 08 18:17:09 crc restorecon[4687]: /var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c219,c404 Oct 08 18:17:09 crc restorecon[4687]: /var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/containers/kube-apiserver-operator/1b9d3e5e not reset as customized by admin to system_u:object_r:container_file_t:s0:c107,c917 Oct 08 18:17:09 crc restorecon[4687]: /var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/containers/kube-apiserver-operator/fddb173c not reset as customized by admin to system_u:object_r:container_file_t:s0:c202,c983 Oct 08 18:17:09 crc restorecon[4687]: /var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/containers/kube-apiserver-operator/95d3c6c4 not reset as customized by admin to system_u:object_r:container_file_t:s0:c219,c404 Oct 08 18:17:09 crc restorecon[4687]: /var/lib/kubelet/pods/9d751cbb-f2e2-430d-9754-c882a5e924a5/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c20,c21 Oct 08 18:17:09 crc restorecon[4687]: /var/lib/kubelet/pods/9d751cbb-f2e2-430d-9754-c882a5e924a5/containers/check-endpoints/bfb5fff5 not reset as customized by admin to system_u:object_r:container_file_t:s0:c20,c21 Oct 08 18:17:09 crc restorecon[4687]: /var/lib/kubelet/pods/9d751cbb-f2e2-430d-9754-c882a5e924a5/containers/check-endpoints/2aef40aa not reset as customized by admin to system_u:object_r:container_file_t:s0:c20,c21 Oct 08 18:17:09 crc restorecon[4687]: /var/lib/kubelet/pods/9d751cbb-f2e2-430d-9754-c882a5e924a5/containers/check-endpoints/c0391cad not reset as customized by admin to system_u:object_r:container_file_t:s0:c20,c21 Oct 08 18:17:09 crc restorecon[4687]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c214,c928 Oct 08 18:17:09 crc restorecon[4687]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/containers/kube-controller-manager/1119e69d not reset as customized by admin to system_u:object_r:container_file_t:s0:c776,c1007 Oct 08 18:17:09 crc restorecon[4687]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/containers/kube-controller-manager/660608b4 not reset as customized by admin to system_u:object_r:container_file_t:s0:c214,c928 Oct 08 18:17:09 crc restorecon[4687]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/containers/kube-controller-manager/8220bd53 not reset as customized by admin to system_u:object_r:container_file_t:s0:c214,c928 Oct 08 18:17:09 crc restorecon[4687]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/containers/cluster-policy-controller/85f99d5c not reset as customized by admin to system_u:object_r:container_file_t:s0:c776,c1007 Oct 08 18:17:09 crc restorecon[4687]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/containers/cluster-policy-controller/4b0225f6 not reset as customized by admin to system_u:object_r:container_file_t:s0:c214,c928 Oct 08 18:17:09 crc restorecon[4687]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/containers/kube-controller-manager-cert-syncer/9c2a3394 not reset as customized by admin to system_u:object_r:container_file_t:s0:c776,c1007 Oct 08 18:17:09 crc restorecon[4687]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/containers/kube-controller-manager-cert-syncer/e820b243 not reset as customized by admin to system_u:object_r:container_file_t:s0:c214,c928 Oct 08 18:17:09 crc restorecon[4687]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/containers/kube-controller-manager-recovery-controller/1ca52ea0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c776,c1007 Oct 08 18:17:09 crc restorecon[4687]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/containers/kube-controller-manager-recovery-controller/e6988e45 not reset as customized by admin to system_u:object_r:container_file_t:s0:c214,c928 Oct 08 18:17:09 crc restorecon[4687]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/volumes/kubernetes.io~configmap/mcc-auth-proxy-config not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Oct 08 18:17:09 crc restorecon[4687]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/volumes/kubernetes.io~configmap/mcc-auth-proxy-config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Oct 08 18:17:09 crc restorecon[4687]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/volumes/kubernetes.io~configmap/mcc-auth-proxy-config/config-file.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Oct 08 18:17:09 crc restorecon[4687]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/volumes/kubernetes.io~configmap/mcc-auth-proxy-config/..2025_02_24_06_09_21.2517297950 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Oct 08 18:17:09 crc restorecon[4687]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/volumes/kubernetes.io~configmap/mcc-auth-proxy-config/..2025_02_24_06_09_21.2517297950/config-file.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Oct 08 18:17:09 crc restorecon[4687]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Oct 08 18:17:09 crc restorecon[4687]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/containers/machine-config-controller/6655f00b not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Oct 08 18:17:09 crc restorecon[4687]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/containers/machine-config-controller/98bc3986 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Oct 08 18:17:09 crc restorecon[4687]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/containers/machine-config-controller/08e3458a not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Oct 08 18:17:09 crc restorecon[4687]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/containers/kube-rbac-proxy/2a191cb0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Oct 08 18:17:09 crc restorecon[4687]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/containers/kube-rbac-proxy/6c4eeefb not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Oct 08 18:17:09 crc restorecon[4687]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/containers/kube-rbac-proxy/f61a549c not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Oct 08 18:17:09 crc restorecon[4687]: /var/lib/kubelet/pods/bd23aa5c-e532-4e53-bccf-e79f130c5ae8/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c318,c553 Oct 08 18:17:09 crc restorecon[4687]: /var/lib/kubelet/pods/bd23aa5c-e532-4e53-bccf-e79f130c5ae8/containers/hostpath-provisioner/24891863 not reset as customized by admin to system_u:object_r:container_file_t:s0:c37,c572 Oct 08 18:17:09 crc restorecon[4687]: /var/lib/kubelet/pods/bd23aa5c-e532-4e53-bccf-e79f130c5ae8/containers/hostpath-provisioner/fbdfd89c not reset as customized by admin to system_u:object_r:container_file_t:s0:c318,c553 Oct 08 18:17:09 crc restorecon[4687]: /var/lib/kubelet/pods/bd23aa5c-e532-4e53-bccf-e79f130c5ae8/containers/liveness-probe/9b63b3bc not reset as customized by admin to system_u:object_r:container_file_t:s0:c37,c572 Oct 08 18:17:09 crc restorecon[4687]: /var/lib/kubelet/pods/bd23aa5c-e532-4e53-bccf-e79f130c5ae8/containers/liveness-probe/8acde6d6 not reset as customized by admin to system_u:object_r:container_file_t:s0:c318,c553 Oct 08 18:17:09 crc restorecon[4687]: /var/lib/kubelet/pods/bd23aa5c-e532-4e53-bccf-e79f130c5ae8/containers/node-driver-registrar/59ecbba3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c318,c553 Oct 08 18:17:09 crc restorecon[4687]: /var/lib/kubelet/pods/bd23aa5c-e532-4e53-bccf-e79f130c5ae8/containers/csi-provisioner/685d4be3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c318,c553 Oct 08 18:17:09 crc restorecon[4687]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Oct 08 18:17:09 crc restorecon[4687]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/config/..2025_02_24_06_20_07.341639300 not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Oct 08 18:17:09 crc restorecon[4687]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/config/..2025_02_24_06_20_07.341639300/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Oct 08 18:17:09 crc restorecon[4687]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/config/..2025_02_24_06_20_07.341639300/openshift-route-controller-manager.client-ca.configmap not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Oct 08 18:17:09 crc restorecon[4687]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/config/..2025_02_24_06_20_07.341639300/openshift-route-controller-manager.serving-cert.secret not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Oct 08 18:17:09 crc restorecon[4687]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Oct 08 18:17:09 crc restorecon[4687]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Oct 08 18:17:09 crc restorecon[4687]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/config/openshift-route-controller-manager.client-ca.configmap not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Oct 08 18:17:09 crc restorecon[4687]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/config/openshift-route-controller-manager.serving-cert.secret not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Oct 08 18:17:09 crc restorecon[4687]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/client-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Oct 08 18:17:09 crc restorecon[4687]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/client-ca/..2025_02_24_06_20_07.2950937851 not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Oct 08 18:17:09 crc restorecon[4687]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/client-ca/..2025_02_24_06_20_07.2950937851/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Oct 08 18:17:09 crc restorecon[4687]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/client-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Oct 08 18:17:09 crc restorecon[4687]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/client-ca/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Oct 08 18:17:09 crc restorecon[4687]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Oct 08 18:17:09 crc restorecon[4687]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/containers/route-controller-manager/feaea55e not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Oct 08 18:17:09 crc restorecon[4687]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:17:09 crc restorecon[4687]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:17:09 crc restorecon[4687]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/abinitio-runtime-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:17:09 crc restorecon[4687]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/abinitio-runtime-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:17:09 crc restorecon[4687]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/accuknox-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:17:09 crc restorecon[4687]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/accuknox-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:17:09 crc restorecon[4687]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aci-containers-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:17:09 crc restorecon[4687]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aci-containers-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:17:09 crc restorecon[4687]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aikit-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:17:09 crc restorecon[4687]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aikit-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:17:09 crc restorecon[4687]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/airlock-microgateway not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:17:09 crc restorecon[4687]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/airlock-microgateway/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:17:09 crc restorecon[4687]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ako-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:17:09 crc restorecon[4687]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ako-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:17:09 crc restorecon[4687]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/alloy not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:17:09 crc restorecon[4687]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/alloy/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:17:09 crc restorecon[4687]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anchore-engine not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:17:09 crc restorecon[4687]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anchore-engine/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:17:09 crc restorecon[4687]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzo-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:17:09 crc restorecon[4687]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzo-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:17:09 crc restorecon[4687]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzograph-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:17:09 crc restorecon[4687]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzograph-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:17:09 crc restorecon[4687]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzounstructured-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:17:09 crc restorecon[4687]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzounstructured-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:17:09 crc restorecon[4687]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/appdynamics-cloud-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:17:09 crc restorecon[4687]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/appdynamics-cloud-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:17:09 crc restorecon[4687]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/appdynamics-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:17:09 crc restorecon[4687]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/appdynamics-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:17:09 crc restorecon[4687]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aqua-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:17:09 crc restorecon[4687]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aqua-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:17:09 crc restorecon[4687]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cass-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:17:09 crc restorecon[4687]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cass-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:17:09 crc restorecon[4687]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ccm-node-agent-dcap-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:17:09 crc restorecon[4687]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ccm-node-agent-dcap-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:17:09 crc restorecon[4687]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ccm-node-agent-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:17:09 crc restorecon[4687]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ccm-node-agent-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:17:09 crc restorecon[4687]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cfm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:17:09 crc restorecon[4687]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cfm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:17:09 crc restorecon[4687]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cilium not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:17:09 crc restorecon[4687]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cilium/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:17:09 crc restorecon[4687]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cilium-enterprise not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:17:09 crc restorecon[4687]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cilium-enterprise/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:17:09 crc restorecon[4687]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloud-native-postgresql not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:17:09 crc restorecon[4687]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloud-native-postgresql/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:17:09 crc restorecon[4687]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloudbees-ci not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:17:09 crc restorecon[4687]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloudbees-ci/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:17:09 crc restorecon[4687]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloudera-streams-messaging-kubernetes-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:17:09 crc restorecon[4687]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloudera-streams-messaging-kubernetes-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:17:09 crc restorecon[4687]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloudnative-pg not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:17:09 crc restorecon[4687]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloudnative-pg/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:17:09 crc restorecon[4687]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cnfv-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:17:09 crc restorecon[4687]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cnfv-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:17:09 crc restorecon[4687]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cockroachdb-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:17:09 crc restorecon[4687]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cockroachdb-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:17:09 crc restorecon[4687]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/conjur-follower-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:17:09 crc restorecon[4687]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/conjur-follower-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:17:09 crc restorecon[4687]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/coroot-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:17:09 crc restorecon[4687]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/coroot-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:17:09 crc restorecon[4687]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/crunchy-postgres-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:17:09 crc restorecon[4687]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/crunchy-postgres-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:17:09 crc restorecon[4687]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cte-k8s-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:17:09 crc restorecon[4687]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cte-k8s-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:17:09 crc restorecon[4687]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datadog-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:17:09 crc restorecon[4687]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datadog-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:17:09 crc restorecon[4687]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dell-csm-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:17:09 crc restorecon[4687]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dell-csm-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:17:09 crc restorecon[4687]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/digitalai-deploy-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:17:09 crc restorecon[4687]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/digitalai-deploy-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:17:09 crc restorecon[4687]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/digitalai-release-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:17:09 crc restorecon[4687]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/digitalai-release-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:17:09 crc restorecon[4687]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dynatrace-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:17:09 crc restorecon[4687]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dynatrace-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:17:09 crc restorecon[4687]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/edb-hcp-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:17:09 crc restorecon[4687]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/edb-hcp-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:17:09 crc restorecon[4687]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eginnovations-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:17:09 crc restorecon[4687]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eginnovations-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:17:09 crc restorecon[4687]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/elasticsearch-eck-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:17:09 crc restorecon[4687]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/elasticsearch-eck-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:17:09 crc restorecon[4687]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/falcon-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:17:09 crc restorecon[4687]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/falcon-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:17:09 crc restorecon[4687]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/federatorai-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:17:09 crc restorecon[4687]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/federatorai-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:17:09 crc restorecon[4687]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fujitsu-enterprise-postgres-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:17:09 crc restorecon[4687]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fujitsu-enterprise-postgres-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:17:09 crc restorecon[4687]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/function-mesh not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:17:09 crc restorecon[4687]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/function-mesh/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:17:09 crc restorecon[4687]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/harness-gitops-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:17:09 crc restorecon[4687]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/harness-gitops-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:17:09 crc restorecon[4687]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hazelcast-platform-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:17:09 crc restorecon[4687]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hazelcast-platform-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:17:09 crc restorecon[4687]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hcp-terraform-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:17:09 crc restorecon[4687]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hcp-terraform-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:17:09 crc restorecon[4687]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hpe-ezmeral-csi-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:17:09 crc restorecon[4687]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hpe-ezmeral-csi-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:17:09 crc restorecon[4687]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-application-gateway-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:17:09 crc restorecon[4687]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-application-gateway-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:17:09 crc restorecon[4687]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-block-csi-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:17:09 crc restorecon[4687]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-block-csi-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:17:09 crc restorecon[4687]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-security-verify-access-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:17:09 crc restorecon[4687]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-security-verify-access-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:17:09 crc restorecon[4687]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-security-verify-directory-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:17:09 crc restorecon[4687]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-security-verify-directory-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:17:09 crc restorecon[4687]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-security-verify-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:17:09 crc restorecon[4687]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-security-verify-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:17:09 crc restorecon[4687]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infoscale-dr-manager not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:17:09 crc restorecon[4687]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infoscale-dr-manager/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:17:09 crc restorecon[4687]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infoscale-licensing-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:17:09 crc restorecon[4687]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infoscale-licensing-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:17:09 crc restorecon[4687]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infoscale-sds-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:17:09 crc restorecon[4687]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infoscale-sds-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:17:09 crc restorecon[4687]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infrastructure-asset-orchestrator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:17:09 crc restorecon[4687]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infrastructure-asset-orchestrator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:17:09 crc restorecon[4687]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/instana-agent-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:17:09 crc restorecon[4687]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/instana-agent-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:17:09 crc restorecon[4687]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/intel-device-plugins-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:17:09 crc restorecon[4687]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/intel-device-plugins-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:17:09 crc restorecon[4687]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/intel-kubernetes-power-manager not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:17:09 crc restorecon[4687]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/intel-kubernetes-power-manager/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:17:09 crc restorecon[4687]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/iomesh-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:17:09 crc restorecon[4687]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/iomesh-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:17:09 crc restorecon[4687]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:17:09 crc restorecon[4687]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:17:09 crc restorecon[4687]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx8-openshift-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:17:09 crc restorecon[4687]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx8-openshift-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:17:09 crc restorecon[4687]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx8-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:17:09 crc restorecon[4687]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx8-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:17:09 crc restorecon[4687]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k8s-triliovault not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:17:09 crc restorecon[4687]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k8s-triliovault/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:17:09 crc restorecon[4687]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-ati-updates not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:17:09 crc restorecon[4687]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-ati-updates/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:17:09 crc restorecon[4687]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-kcos-framework not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:17:09 crc restorecon[4687]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-kcos-framework/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:17:09 crc restorecon[4687]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-kcos-ingress not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:17:09 crc restorecon[4687]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-kcos-ingress/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:17:09 crc restorecon[4687]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-kcos-licensing not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:17:09 crc restorecon[4687]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-kcos-licensing/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:17:09 crc restorecon[4687]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-kcos-sso not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:17:09 crc restorecon[4687]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-kcos-sso/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:17:09 crc restorecon[4687]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-keycloak-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:17:09 crc restorecon[4687]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-keycloak-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:17:09 crc restorecon[4687]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-load-core not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:17:09 crc restorecon[4687]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-load-core/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:17:09 crc restorecon[4687]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-loadcore-agents not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:17:09 crc restorecon[4687]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-loadcore-agents/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:17:09 crc restorecon[4687]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-nats-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:17:09 crc restorecon[4687]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-nats-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:17:09 crc restorecon[4687]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-nimbusmosaic-dusim not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:17:09 crc restorecon[4687]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-nimbusmosaic-dusim/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:17:09 crc restorecon[4687]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-rest-api-browser-v1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:17:09 crc restorecon[4687]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-rest-api-browser-v1/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:17:09 crc restorecon[4687]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-appsec not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:17:09 crc restorecon[4687]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-appsec/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:17:09 crc restorecon[4687]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-core not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:17:09 crc restorecon[4687]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-core/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:17:09 crc restorecon[4687]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-db not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:17:09 crc restorecon[4687]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-db/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:17:09 crc restorecon[4687]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-diagnostics not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:17:09 crc restorecon[4687]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-diagnostics/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:17:09 crc restorecon[4687]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-logging not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:17:09 crc restorecon[4687]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-logging/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:17:09 crc restorecon[4687]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-migration not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:17:09 crc restorecon[4687]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-migration/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:17:09 crc restorecon[4687]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-msg-broker not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:17:09 crc restorecon[4687]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-msg-broker/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:17:09 crc restorecon[4687]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-notifications not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:17:09 crc restorecon[4687]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-notifications/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:17:09 crc restorecon[4687]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-stats-dashboards not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:17:09 crc restorecon[4687]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-stats-dashboards/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:17:09 crc restorecon[4687]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-storage not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:17:09 crc restorecon[4687]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-storage/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:17:09 crc restorecon[4687]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-test-core not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:17:09 crc restorecon[4687]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-test-core/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:17:09 crc restorecon[4687]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-ui not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:17:09 crc restorecon[4687]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-ui/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:17:09 crc restorecon[4687]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-websocket-service not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:17:09 crc restorecon[4687]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-websocket-service/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:17:09 crc restorecon[4687]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kong-gateway-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:17:09 crc restorecon[4687]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kong-gateway-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:17:09 crc restorecon[4687]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubearmor-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:17:09 crc restorecon[4687]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubearmor-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:17:09 crc restorecon[4687]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubecost-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:17:09 crc restorecon[4687]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubecost-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:17:09 crc restorecon[4687]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubemq-operator-marketplace not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:17:09 crc restorecon[4687]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubemq-operator-marketplace/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:17:09 crc restorecon[4687]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubeturbo-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:17:09 crc restorecon[4687]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubeturbo-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:17:09 crc restorecon[4687]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/lenovo-locd-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:17:09 crc restorecon[4687]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/lenovo-locd-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:17:09 crc restorecon[4687]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/marketplace-games-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:17:09 crc restorecon[4687]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/marketplace-games-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:17:09 crc restorecon[4687]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/memcached-operator-ogaye not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:17:09 crc restorecon[4687]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/memcached-operator-ogaye/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:17:09 crc restorecon[4687]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/memory-machine-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:17:09 crc restorecon[4687]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/memory-machine-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:17:09 crc restorecon[4687]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/model-builder-for-vision-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:17:09 crc restorecon[4687]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/model-builder-for-vision-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:17:09 crc restorecon[4687]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mongodb-atlas-kubernetes not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:17:09 crc restorecon[4687]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mongodb-atlas-kubernetes/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:17:09 crc restorecon[4687]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mongodb-enterprise not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:17:09 crc restorecon[4687]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mongodb-enterprise/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:17:09 crc restorecon[4687]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netapp-spark-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:17:09 crc restorecon[4687]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netapp-spark-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:17:09 crc restorecon[4687]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netscaler-adm-agent-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:17:09 crc restorecon[4687]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netscaler-adm-agent-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:17:09 crc restorecon[4687]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netscaler-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:17:09 crc restorecon[4687]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netscaler-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:17:09 crc restorecon[4687]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/neuvector-certified-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:17:09 crc restorecon[4687]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/neuvector-certified-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:17:09 crc restorecon[4687]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nexus-repository-ha-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:17:09 crc restorecon[4687]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nexus-repository-ha-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:17:09 crc restorecon[4687]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nginx-ingress-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:17:09 crc restorecon[4687]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nginx-ingress-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:17:09 crc restorecon[4687]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pcc-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:17:09 crc restorecon[4687]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pcc-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:17:09 crc restorecon[4687]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nim-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:17:09 crc restorecon[4687]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nim-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:17:09 crc restorecon[4687]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nxiq-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:17:09 crc restorecon[4687]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nxiq-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:17:09 crc restorecon[4687]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nxrm-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:17:09 crc restorecon[4687]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nxrm-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:17:09 crc restorecon[4687]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odigos-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:17:09 crc restorecon[4687]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odigos-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:17:09 crc restorecon[4687]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/open-liberty-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:17:09 crc restorecon[4687]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/open-liberty-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:17:09 crc restorecon[4687]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshiftartifactoryha-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:17:09 crc restorecon[4687]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshiftartifactoryha-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:17:09 crc restorecon[4687]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshiftxray-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:17:09 crc restorecon[4687]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshiftxray-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:17:09 crc restorecon[4687]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/operator-certification-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:17:09 crc restorecon[4687]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/operator-certification-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:17:09 crc restorecon[4687]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ovms-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:17:09 crc restorecon[4687]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ovms-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:17:09 crc restorecon[4687]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pachyderm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:17:09 crc restorecon[4687]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pachyderm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:17:09 crc restorecon[4687]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pmem-csi-operator-os not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:17:09 crc restorecon[4687]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pmem-csi-operator-os/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:17:09 crc restorecon[4687]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/portworx-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:17:09 crc restorecon[4687]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/portworx-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:17:09 crc restorecon[4687]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/prometurbo-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:17:09 crc restorecon[4687]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/prometurbo-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:17:09 crc restorecon[4687]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pubsubplus-eventbroker-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:17:09 crc restorecon[4687]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pubsubplus-eventbroker-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:17:09 crc restorecon[4687]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/redis-enterprise-operator-cert not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:17:09 crc restorecon[4687]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/redis-enterprise-operator-cert/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:17:09 crc restorecon[4687]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/runtime-component-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:17:09 crc restorecon[4687]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/runtime-component-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:17:09 crc restorecon[4687]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/runtime-fabric-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:17:09 crc restorecon[4687]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/runtime-fabric-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:17:09 crc restorecon[4687]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sanstoragecsi-operator-bundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:17:09 crc restorecon[4687]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sanstoragecsi-operator-bundle/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:17:09 crc restorecon[4687]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/silicom-sts-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:17:09 crc restorecon[4687]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/silicom-sts-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:17:09 crc restorecon[4687]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/smilecdr-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:17:09 crc restorecon[4687]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/smilecdr-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:17:09 crc restorecon[4687]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sriov-fec not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:17:09 crc restorecon[4687]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sriov-fec/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:17:09 crc restorecon[4687]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/stackable-commons-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:17:09 crc restorecon[4687]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/stackable-commons-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:17:09 crc restorecon[4687]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/stackable-zookeeper-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:17:09 crc restorecon[4687]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/stackable-zookeeper-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:17:09 crc restorecon[4687]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/t8c-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:17:09 crc restorecon[4687]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/t8c-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:17:09 crc restorecon[4687]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/t8c-tsc-client-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:17:09 crc restorecon[4687]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/t8c-tsc-client-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:17:09 crc restorecon[4687]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tawon-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:17:09 crc restorecon[4687]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tawon-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:17:09 crc restorecon[4687]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tigera-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:17:09 crc restorecon[4687]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tigera-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:17:09 crc restorecon[4687]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/timemachine-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:17:09 crc restorecon[4687]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/timemachine-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:17:09 crc restorecon[4687]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/vault-secrets-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:17:09 crc restorecon[4687]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/vault-secrets-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:17:09 crc restorecon[4687]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/vcp-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:17:09 crc restorecon[4687]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/vcp-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:17:09 crc restorecon[4687]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/webotx-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:17:09 crc restorecon[4687]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/webotx-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:17:09 crc restorecon[4687]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/xcrypt-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:17:09 crc restorecon[4687]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/xcrypt-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:17:09 crc restorecon[4687]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/zabbix-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:17:09 crc restorecon[4687]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/zabbix-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:17:09 crc restorecon[4687]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:17:09 crc restorecon[4687]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:17:09 crc restorecon[4687]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:17:09 crc restorecon[4687]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/00000-1.psg not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:17:09 crc restorecon[4687]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/00000-1.psg.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:17:09 crc restorecon[4687]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/db.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:17:09 crc restorecon[4687]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/index.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:17:09 crc restorecon[4687]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/main.pix not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:17:09 crc restorecon[4687]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/overflow.pix not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:17:09 crc restorecon[4687]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/digest not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:17:09 crc restorecon[4687]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/utilities not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:17:09 crc restorecon[4687]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/utilities/copy-content not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:17:09 crc restorecon[4687]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:17:09 crc restorecon[4687]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/containers/extract-utilities/63709497 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:17:09 crc restorecon[4687]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/containers/extract-utilities/d966b7fd not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:17:09 crc restorecon[4687]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/containers/extract-utilities/f5773757 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:17:09 crc restorecon[4687]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/containers/extract-content/81c9edb9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:17:09 crc restorecon[4687]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/containers/extract-content/57bf57ee not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:17:09 crc restorecon[4687]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/containers/extract-content/86f5e6aa not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:17:09 crc restorecon[4687]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/containers/registry-server/0aabe31d not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:17:09 crc restorecon[4687]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/containers/registry-server/d2af85c2 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:17:09 crc restorecon[4687]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/containers/registry-server/09d157d9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:17:09 crc restorecon[4687]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:17:09 crc restorecon[4687]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:17:09 crc restorecon[4687]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:17:09 crc restorecon[4687]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:17:09 crc restorecon[4687]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/00000-1.psg not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:17:09 crc restorecon[4687]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/00000-1.psg.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:17:09 crc restorecon[4687]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/db.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:17:09 crc restorecon[4687]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/index.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:17:09 crc restorecon[4687]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/main.pix not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:17:09 crc restorecon[4687]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/overflow.pix not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:17:09 crc restorecon[4687]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/digest not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:17:09 crc restorecon[4687]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:17:09 crc restorecon[4687]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/3scale-community-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:17:09 crc restorecon[4687]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/3scale-community-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:17:09 crc restorecon[4687]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-acm-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:17:09 crc restorecon[4687]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-acm-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:17:09 crc restorecon[4687]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-acmpca-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:17:09 crc restorecon[4687]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-acmpca-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:17:09 crc restorecon[4687]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-apigateway-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:17:09 crc restorecon[4687]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-apigateway-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:17:09 crc restorecon[4687]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-apigatewayv2-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:17:09 crc restorecon[4687]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-apigatewayv2-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:17:09 crc restorecon[4687]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-applicationautoscaling-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:17:09 crc restorecon[4687]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-applicationautoscaling-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:17:09 crc restorecon[4687]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-athena-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:17:09 crc restorecon[4687]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-athena-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:17:09 crc restorecon[4687]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-cloudfront-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:17:09 crc restorecon[4687]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-cloudfront-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:17:09 crc restorecon[4687]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-cloudtrail-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:17:09 crc restorecon[4687]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-cloudtrail-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:17:09 crc restorecon[4687]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-cloudwatch-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:17:09 crc restorecon[4687]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-cloudwatch-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:17:09 crc restorecon[4687]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-cloudwatchlogs-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:17:09 crc restorecon[4687]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-cloudwatchlogs-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:17:09 crc restorecon[4687]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-documentdb-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:17:09 crc restorecon[4687]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-documentdb-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:17:09 crc restorecon[4687]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-dynamodb-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:17:09 crc restorecon[4687]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-dynamodb-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:17:09 crc restorecon[4687]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ec2-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:17:09 crc restorecon[4687]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ec2-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:17:09 crc restorecon[4687]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ecr-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:17:09 crc restorecon[4687]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ecr-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:17:09 crc restorecon[4687]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ecs-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:17:09 crc restorecon[4687]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ecs-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:17:09 crc restorecon[4687]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-efs-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:17:09 crc restorecon[4687]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-efs-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:17:09 crc restorecon[4687]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-eks-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:17:09 crc restorecon[4687]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-eks-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:17:09 crc restorecon[4687]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-elasticache-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:17:09 crc restorecon[4687]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-elasticache-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:17:09 crc restorecon[4687]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-elbv2-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:17:09 crc restorecon[4687]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-elbv2-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:17:09 crc restorecon[4687]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-emrcontainers-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:17:09 crc restorecon[4687]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-emrcontainers-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:17:09 crc restorecon[4687]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-eventbridge-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:17:09 crc restorecon[4687]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-eventbridge-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:17:09 crc restorecon[4687]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-iam-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:17:09 crc restorecon[4687]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-iam-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:17:09 crc restorecon[4687]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-kafka-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:17:09 crc restorecon[4687]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-kafka-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:17:09 crc restorecon[4687]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-keyspaces-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:17:09 crc restorecon[4687]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-keyspaces-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:17:09 crc restorecon[4687]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-kinesis-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:17:09 crc restorecon[4687]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-kinesis-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:17:09 crc restorecon[4687]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-kms-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:17:09 crc restorecon[4687]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-kms-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:17:09 crc restorecon[4687]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-lambda-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:17:09 crc restorecon[4687]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-lambda-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:17:09 crc restorecon[4687]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-memorydb-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:17:09 crc restorecon[4687]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-memorydb-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:17:09 crc restorecon[4687]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-mq-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:17:09 crc restorecon[4687]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-mq-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:17:09 crc restorecon[4687]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-networkfirewall-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:17:09 crc restorecon[4687]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-networkfirewall-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:17:09 crc restorecon[4687]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-opensearchservice-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:17:09 crc restorecon[4687]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-opensearchservice-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:17:09 crc restorecon[4687]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-organizations-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:17:09 crc restorecon[4687]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-organizations-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:17:09 crc restorecon[4687]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-pipes-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:17:09 crc restorecon[4687]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-pipes-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:17:09 crc restorecon[4687]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-prometheusservice-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:17:09 crc restorecon[4687]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-prometheusservice-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:17:09 crc restorecon[4687]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-rds-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:17:09 crc restorecon[4687]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-rds-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:17:09 crc restorecon[4687]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-recyclebin-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:17:09 crc restorecon[4687]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-recyclebin-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:17:09 crc restorecon[4687]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-route53-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:17:09 crc restorecon[4687]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-route53-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:17:09 crc restorecon[4687]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-route53resolver-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:17:09 crc restorecon[4687]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-route53resolver-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:17:09 crc restorecon[4687]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-s3-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:17:09 crc restorecon[4687]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-s3-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:17:09 crc restorecon[4687]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-sagemaker-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:17:09 crc restorecon[4687]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-sagemaker-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:17:09 crc restorecon[4687]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-secretsmanager-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:17:09 crc restorecon[4687]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-secretsmanager-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:17:09 crc restorecon[4687]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ses-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:17:09 crc restorecon[4687]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ses-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:17:09 crc restorecon[4687]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-sfn-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:17:09 crc restorecon[4687]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-sfn-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:17:09 crc restorecon[4687]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-sns-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:17:09 crc restorecon[4687]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-sns-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:17:09 crc restorecon[4687]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-sqs-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:17:09 crc restorecon[4687]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-sqs-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:17:09 crc restorecon[4687]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ssm-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:17:09 crc restorecon[4687]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ssm-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:17:09 crc restorecon[4687]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-wafv2-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:17:09 crc restorecon[4687]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-wafv2-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:17:09 crc restorecon[4687]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aerospike-kubernetes-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:17:09 crc restorecon[4687]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aerospike-kubernetes-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:17:09 crc restorecon[4687]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/airflow-helm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:17:09 crc restorecon[4687]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/airflow-helm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:17:09 crc restorecon[4687]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/alloydb-omni-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:17:09 crc restorecon[4687]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/alloydb-omni-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:17:09 crc restorecon[4687]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/alvearie-imaging-ingestion not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:17:09 crc restorecon[4687]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/alvearie-imaging-ingestion/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:17:09 crc restorecon[4687]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amd-gpu-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:17:09 crc restorecon[4687]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amd-gpu-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:17:09 crc restorecon[4687]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/analytics-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:17:09 crc restorecon[4687]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/analytics-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:17:09 crc restorecon[4687]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/annotationlab not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:17:09 crc restorecon[4687]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/annotationlab/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:17:09 crc restorecon[4687]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicast-community-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:17:09 crc restorecon[4687]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicast-community-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:17:09 crc restorecon[4687]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicurio-api-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:17:09 crc restorecon[4687]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicurio-api-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:17:09 crc restorecon[4687]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicurio-registry not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:17:09 crc restorecon[4687]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicurio-registry/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:17:09 crc restorecon[4687]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicurito not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:17:09 crc restorecon[4687]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicurito/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:17:09 crc restorecon[4687]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apimatic-kubernetes-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:17:09 crc restorecon[4687]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apimatic-kubernetes-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:17:09 crc restorecon[4687]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/application-services-metering-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:17:09 crc restorecon[4687]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/application-services-metering-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:17:09 crc restorecon[4687]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aqua not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:17:09 crc restorecon[4687]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aqua/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:17:09 crc restorecon[4687]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/argocd-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:17:09 crc restorecon[4687]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/argocd-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:17:09 crc restorecon[4687]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/assisted-service-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:17:09 crc restorecon[4687]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/assisted-service-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:17:09 crc restorecon[4687]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/authorino-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:17:09 crc restorecon[4687]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/authorino-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:17:09 crc restorecon[4687]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/automotive-infra not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:17:09 crc restorecon[4687]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/automotive-infra/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:17:09 crc restorecon[4687]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aws-efs-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:17:09 crc restorecon[4687]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aws-efs-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:17:09 crc restorecon[4687]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/awss3-operator-registry not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:17:09 crc restorecon[4687]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/awss3-operator-registry/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:17:09 crc restorecon[4687]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/azure-service-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:17:09 crc restorecon[4687]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/azure-service-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:17:09 crc restorecon[4687]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/beegfs-csi-driver-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:17:09 crc restorecon[4687]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/beegfs-csi-driver-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:17:09 crc restorecon[4687]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/bpfman-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:17:09 crc restorecon[4687]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/bpfman-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:17:09 crc restorecon[4687]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/camel-k not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:17:09 crc restorecon[4687]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/camel-k/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:17:09 crc restorecon[4687]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/camel-karavan-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:17:09 crc restorecon[4687]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/camel-karavan-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:17:09 crc restorecon[4687]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cass-operator-community not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:17:09 crc restorecon[4687]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cass-operator-community/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:17:09 crc restorecon[4687]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cert-manager not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:17:09 crc restorecon[4687]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cert-manager/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:17:09 crc restorecon[4687]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cert-utils-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:17:09 crc restorecon[4687]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cert-utils-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:17:09 crc restorecon[4687]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-aas-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:17:09 crc restorecon[4687]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-aas-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:17:09 crc restorecon[4687]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-impairment-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:17:09 crc restorecon[4687]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-impairment-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:17:09 crc restorecon[4687]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-manager not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:17:09 crc restorecon[4687]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-manager/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:17:09 crc restorecon[4687]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cockroachdb not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:17:09 crc restorecon[4687]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cockroachdb/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:17:09 crc restorecon[4687]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/codeflare-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:17:09 crc restorecon[4687]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/codeflare-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:17:09 crc restorecon[4687]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/community-kubevirt-hyperconverged not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:17:09 crc restorecon[4687]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/community-kubevirt-hyperconverged/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:17:09 crc restorecon[4687]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/community-trivy-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:17:09 crc restorecon[4687]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/community-trivy-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:17:09 crc restorecon[4687]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/community-windows-machine-config-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:17:09 crc restorecon[4687]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/community-windows-machine-config-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:17:09 crc restorecon[4687]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/customized-user-remediation not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:17:09 crc restorecon[4687]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/customized-user-remediation/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:17:09 crc restorecon[4687]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cxl-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:17:09 crc restorecon[4687]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cxl-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:17:09 crc restorecon[4687]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dapr-kubernetes-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:17:09 crc restorecon[4687]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dapr-kubernetes-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:17:09 crc restorecon[4687]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datadog-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:17:09 crc restorecon[4687]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datadog-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:17:09 crc restorecon[4687]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datatrucker-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:17:09 crc restorecon[4687]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datatrucker-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:17:09 crc restorecon[4687]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dbaas-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:17:09 crc restorecon[4687]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dbaas-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:17:09 crc restorecon[4687]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/debezium-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:17:09 crc restorecon[4687]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/debezium-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:17:09 crc restorecon[4687]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dell-csm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:17:09 crc restorecon[4687]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dell-csm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:17:09 crc restorecon[4687]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/deployment-validation-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:17:09 crc restorecon[4687]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/deployment-validation-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:17:09 crc restorecon[4687]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/devopsinabox not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:17:09 crc restorecon[4687]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/devopsinabox/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:17:09 crc restorecon[4687]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dns-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:17:09 crc restorecon[4687]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dns-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:17:09 crc restorecon[4687]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dynatrace-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:17:09 crc restorecon[4687]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dynatrace-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:17:09 crc restorecon[4687]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eclipse-amlen-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:17:09 crc restorecon[4687]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eclipse-amlen-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:17:09 crc restorecon[4687]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eclipse-che not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:17:09 crc restorecon[4687]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eclipse-che/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:17:09 crc restorecon[4687]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ecr-secret-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:17:09 crc restorecon[4687]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ecr-secret-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:17:09 crc restorecon[4687]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/edp-keycloak-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:17:09 crc restorecon[4687]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/edp-keycloak-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:17:09 crc restorecon[4687]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eginnovations-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:17:09 crc restorecon[4687]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eginnovations-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:17:09 crc restorecon[4687]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/egressip-ipam-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:17:09 crc restorecon[4687]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/egressip-ipam-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:17:09 crc restorecon[4687]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ember-csi-community-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:17:09 crc restorecon[4687]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ember-csi-community-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:17:09 crc restorecon[4687]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/etcd not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:17:09 crc restorecon[4687]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/etcd/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:17:09 crc restorecon[4687]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eventing-kogito not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:17:09 crc restorecon[4687]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eventing-kogito/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:17:09 crc restorecon[4687]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/external-secrets-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:17:09 crc restorecon[4687]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/external-secrets-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:17:09 crc restorecon[4687]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/falcon-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:17:09 crc restorecon[4687]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/falcon-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:17:09 crc restorecon[4687]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fence-agents-remediation not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:17:09 crc restorecon[4687]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fence-agents-remediation/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:17:09 crc restorecon[4687]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/flink-kubernetes-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:17:09 crc restorecon[4687]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/flink-kubernetes-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:17:09 crc restorecon[4687]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/flux not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:17:09 crc restorecon[4687]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/flux/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:17:09 crc restorecon[4687]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k8gb not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:17:09 crc restorecon[4687]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k8gb/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:17:09 crc restorecon[4687]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fossul-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:17:09 crc restorecon[4687]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fossul-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:17:09 crc restorecon[4687]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/github-arc-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:17:09 crc restorecon[4687]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/github-arc-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:17:09 crc restorecon[4687]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/gitops-primer not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:17:09 crc restorecon[4687]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/gitops-primer/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:17:09 crc restorecon[4687]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/gitwebhook-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:17:09 crc restorecon[4687]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/gitwebhook-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:17:09 crc restorecon[4687]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/global-load-balancer-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:17:09 crc restorecon[4687]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/global-load-balancer-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:17:09 crc restorecon[4687]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/grafana-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:17:09 crc restorecon[4687]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/grafana-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:17:09 crc restorecon[4687]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/group-sync-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:17:09 crc restorecon[4687]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/group-sync-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:17:09 crc restorecon[4687]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hawtio-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:17:09 crc restorecon[4687]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hawtio-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:17:09 crc restorecon[4687]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hazelcast-platform-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:17:09 crc restorecon[4687]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hazelcast-platform-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:17:09 crc restorecon[4687]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hedvig-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:17:09 crc restorecon[4687]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hedvig-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:17:09 crc restorecon[4687]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hive-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:17:09 crc restorecon[4687]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hive-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:17:09 crc restorecon[4687]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/horreum-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:17:09 crc restorecon[4687]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/horreum-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:17:09 crc restorecon[4687]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hyperfoil-bundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:17:09 crc restorecon[4687]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hyperfoil-bundle/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:17:09 crc restorecon[4687]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-block-csi-operator-community not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:17:09 crc restorecon[4687]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-block-csi-operator-community/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:17:09 crc restorecon[4687]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-security-verify-access-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:17:09 crc restorecon[4687]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-security-verify-access-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:17:09 crc restorecon[4687]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-spectrum-scale-csi-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:17:09 crc restorecon[4687]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-spectrum-scale-csi-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:17:09 crc restorecon[4687]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibmcloud-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:17:09 crc restorecon[4687]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibmcloud-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:17:09 crc restorecon[4687]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infinispan not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:17:09 crc restorecon[4687]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infinispan/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:17:09 crc restorecon[4687]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/integrity-shield-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:17:09 crc restorecon[4687]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/integrity-shield-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:17:09 crc restorecon[4687]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ipfs-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:17:09 crc restorecon[4687]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ipfs-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:17:09 crc restorecon[4687]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/istio-workspace-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:17:09 crc restorecon[4687]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/istio-workspace-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:17:09 crc restorecon[4687]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/jaeger not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:17:09 crc restorecon[4687]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/jaeger/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:17:09 crc restorecon[4687]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kaoto-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:17:09 crc restorecon[4687]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kaoto-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:17:09 crc restorecon[4687]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keda not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:17:09 crc restorecon[4687]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keda/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:17:09 crc restorecon[4687]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keepalived-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:17:09 crc restorecon[4687]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keepalived-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:17:09 crc restorecon[4687]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keycloak-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:17:09 crc restorecon[4687]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keycloak-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:17:09 crc restorecon[4687]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keycloak-permissions-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:17:09 crc restorecon[4687]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keycloak-permissions-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:17:09 crc restorecon[4687]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/klusterlet not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:17:09 crc restorecon[4687]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/klusterlet/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:17:09 crc restorecon[4687]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kogito-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:17:09 crc restorecon[4687]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kogito-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:17:09 crc restorecon[4687]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/koku-metrics-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:17:09 crc restorecon[4687]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/koku-metrics-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:17:09 crc restorecon[4687]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/konveyor-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:17:09 crc restorecon[4687]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/konveyor-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:17:09 crc restorecon[4687]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/korrel8r not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:17:09 crc restorecon[4687]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/korrel8r/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:17:09 crc restorecon[4687]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kuadrant-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:17:09 crc restorecon[4687]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kuadrant-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:17:09 crc restorecon[4687]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kube-green not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:17:09 crc restorecon[4687]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kube-green/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:17:09 crc restorecon[4687]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubecost not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:17:09 crc restorecon[4687]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubecost/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:17:09 crc restorecon[4687]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubernetes-imagepuller-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:17:09 crc restorecon[4687]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubernetes-imagepuller-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:17:09 crc restorecon[4687]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubeturbo not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:17:09 crc restorecon[4687]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubeturbo/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:17:09 crc restorecon[4687]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/l5-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:17:09 crc restorecon[4687]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/l5-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:17:09 crc restorecon[4687]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/layer7-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:17:09 crc restorecon[4687]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/layer7-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:17:09 crc restorecon[4687]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/lbconfig-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:17:09 crc restorecon[4687]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/lbconfig-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:17:09 crc restorecon[4687]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/lib-bucket-provisioner not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:17:09 crc restorecon[4687]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/lib-bucket-provisioner/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:17:09 crc restorecon[4687]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/limitador-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:17:09 crc restorecon[4687]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/limitador-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:17:09 crc restorecon[4687]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/logging-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:17:09 crc restorecon[4687]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/logging-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:17:09 crc restorecon[4687]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/loki-helm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:17:09 crc restorecon[4687]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/loki-helm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:17:09 crc restorecon[4687]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/loki-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:17:09 crc restorecon[4687]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/loki-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:17:09 crc restorecon[4687]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/machine-deletion-remediation not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:17:09 crc restorecon[4687]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/machine-deletion-remediation/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:17:09 crc restorecon[4687]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mariadb-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:17:09 crc restorecon[4687]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mariadb-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:17:09 crc restorecon[4687]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/marin3r not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:17:09 crc restorecon[4687]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/marin3r/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:17:09 crc restorecon[4687]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mercury-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:17:09 crc restorecon[4687]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mercury-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:17:09 crc restorecon[4687]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/microcks not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:17:09 crc restorecon[4687]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/microcks/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:17:09 crc restorecon[4687]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mongodb-atlas-kubernetes not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:17:09 crc restorecon[4687]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mongodb-atlas-kubernetes/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:17:09 crc restorecon[4687]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mongodb-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:17:09 crc restorecon[4687]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mongodb-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:17:09 crc restorecon[4687]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/move2kube-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:17:09 crc restorecon[4687]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/move2kube-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:17:09 crc restorecon[4687]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/multi-nic-cni-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:17:09 crc restorecon[4687]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/multi-nic-cni-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:17:09 crc restorecon[4687]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/multicluster-global-hub-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:17:09 crc restorecon[4687]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/multicluster-global-hub-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:17:09 crc restorecon[4687]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/multicluster-operators-subscription not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:17:09 crc restorecon[4687]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/multicluster-operators-subscription/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:17:09 crc restorecon[4687]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/must-gather-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:17:09 crc restorecon[4687]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/must-gather-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:17:09 crc restorecon[4687]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/namespace-configuration-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:17:09 crc restorecon[4687]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/namespace-configuration-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:17:09 crc restorecon[4687]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ncn-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:17:09 crc restorecon[4687]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ncn-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:17:09 crc restorecon[4687]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ndmspc-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:17:09 crc restorecon[4687]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ndmspc-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:17:09 crc restorecon[4687]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netobserv-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:17:09 crc restorecon[4687]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netobserv-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:17:09 crc restorecon[4687]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/neuvector-community-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:17:09 crc restorecon[4687]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/neuvector-community-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:17:09 crc restorecon[4687]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nexus-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:17:09 crc restorecon[4687]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nexus-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:17:09 crc restorecon[4687]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nexus-operator-m88i not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:17:09 crc restorecon[4687]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nexus-operator-m88i/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:17:09 crc restorecon[4687]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nfs-provisioner-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:17:09 crc restorecon[4687]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nfs-provisioner-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:17:09 crc restorecon[4687]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nlp-server not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:17:09 crc restorecon[4687]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nlp-server/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:17:09 crc restorecon[4687]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-discovery-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:17:09 crc restorecon[4687]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-discovery-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:17:09 crc restorecon[4687]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-healthcheck-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:17:09 crc restorecon[4687]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-healthcheck-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:17:09 crc restorecon[4687]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-maintenance-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:17:09 crc restorecon[4687]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-maintenance-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:17:09 crc restorecon[4687]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nsm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:17:09 crc restorecon[4687]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nsm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:17:09 crc restorecon[4687]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/oadp-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:17:09 crc restorecon[4687]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/oadp-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:17:09 crc restorecon[4687]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/observability-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:17:09 crc restorecon[4687]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/observability-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:17:09 crc restorecon[4687]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/oci-ccm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:17:09 crc restorecon[4687]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/oci-ccm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:17:09 crc restorecon[4687]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ocm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:17:09 crc restorecon[4687]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ocm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:17:09 crc restorecon[4687]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odoo-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:17:09 crc restorecon[4687]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odoo-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:17:09 crc restorecon[4687]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/opendatahub-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:17:09 crc restorecon[4687]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/opendatahub-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:17:09 crc restorecon[4687]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openebs not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:17:09 crc restorecon[4687]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openebs/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:17:09 crc restorecon[4687]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-nfd-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:17:09 crc restorecon[4687]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-nfd-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:17:09 crc restorecon[4687]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-node-upgrade-mutex-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:17:09 crc restorecon[4687]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-node-upgrade-mutex-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:17:09 crc restorecon[4687]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-qiskit-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:17:09 crc restorecon[4687]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-qiskit-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:17:09 crc restorecon[4687]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/opentelemetry-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:17:09 crc restorecon[4687]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/opentelemetry-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:17:09 crc restorecon[4687]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/patch-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:17:09 crc restorecon[4687]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/patch-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:17:09 crc restorecon[4687]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/patterns-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:17:09 crc restorecon[4687]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/patterns-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:17:09 crc restorecon[4687]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pcc-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:17:09 crc restorecon[4687]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pcc-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:17:09 crc restorecon[4687]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pelorus-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:17:09 crc restorecon[4687]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pelorus-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:17:09 crc restorecon[4687]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/percona-xtradb-cluster-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:17:09 crc restorecon[4687]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/percona-xtradb-cluster-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:17:09 crc restorecon[4687]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/portworx-essentials not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:17:09 crc restorecon[4687]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/portworx-essentials/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:17:09 crc restorecon[4687]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/postgresql not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:17:09 crc restorecon[4687]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/postgresql/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:17:09 crc restorecon[4687]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/proactive-node-scaling-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:17:09 crc restorecon[4687]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/proactive-node-scaling-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:17:09 crc restorecon[4687]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/project-quay not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:17:09 crc restorecon[4687]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/project-quay/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:17:09 crc restorecon[4687]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/prometheus not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:17:09 crc restorecon[4687]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/prometheus/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:17:09 crc restorecon[4687]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/prometheus-exporter-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:17:09 crc restorecon[4687]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/prometheus-exporter-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:17:09 crc restorecon[4687]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/prometurbo not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:17:09 crc restorecon[4687]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/prometurbo/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:17:09 crc restorecon[4687]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pubsubplus-eventbroker-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:17:09 crc restorecon[4687]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pubsubplus-eventbroker-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:17:09 crc restorecon[4687]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pulp-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:17:09 crc restorecon[4687]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pulp-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:17:09 crc restorecon[4687]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rabbitmq-cluster-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:17:09 crc restorecon[4687]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rabbitmq-cluster-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:17:09 crc restorecon[4687]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rabbitmq-messaging-topology-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:17:09 crc restorecon[4687]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rabbitmq-messaging-topology-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:17:09 crc restorecon[4687]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/redis-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:17:09 crc restorecon[4687]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/redis-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:17:09 crc restorecon[4687]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/reportportal-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:17:09 crc restorecon[4687]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/reportportal-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:17:09 crc restorecon[4687]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/resource-locker-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:17:09 crc restorecon[4687]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/resource-locker-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:17:09 crc restorecon[4687]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhoas-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:17:09 crc restorecon[4687]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhoas-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:17:09 crc restorecon[4687]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ripsaw not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:17:09 crc restorecon[4687]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ripsaw/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:17:09 crc restorecon[4687]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sailoperator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:17:09 crc restorecon[4687]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sailoperator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:17:09 crc restorecon[4687]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sap-commerce-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:17:09 crc restorecon[4687]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sap-commerce-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:17:09 crc restorecon[4687]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sap-data-intelligence-observer-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:17:09 crc restorecon[4687]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sap-data-intelligence-observer-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:17:09 crc restorecon[4687]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sap-hana-express-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:17:09 crc restorecon[4687]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sap-hana-express-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:17:09 crc restorecon[4687]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/seldon-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:17:09 crc restorecon[4687]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/seldon-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:17:09 crc restorecon[4687]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/self-node-remediation not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:17:09 crc restorecon[4687]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/self-node-remediation/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:17:09 crc restorecon[4687]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/service-binding-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:17:09 crc restorecon[4687]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/service-binding-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:17:09 crc restorecon[4687]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/shipwright-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:17:09 crc restorecon[4687]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/shipwright-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:17:09 crc restorecon[4687]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sigstore-helm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:17:09 crc restorecon[4687]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sigstore-helm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:17:09 crc restorecon[4687]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/silicom-sts-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:17:09 crc restorecon[4687]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/silicom-sts-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:17:09 crc restorecon[4687]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/skupper-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:17:09 crc restorecon[4687]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/skupper-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:17:09 crc restorecon[4687]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/snapscheduler not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:17:09 crc restorecon[4687]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/snapscheduler/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:17:09 crc restorecon[4687]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/snyk-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:17:09 crc restorecon[4687]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/snyk-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:17:09 crc restorecon[4687]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/socmmd not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:17:09 crc restorecon[4687]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/socmmd/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:17:09 crc restorecon[4687]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sonar-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:17:09 crc restorecon[4687]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sonar-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:17:09 crc restorecon[4687]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sosivio not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:17:09 crc restorecon[4687]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sosivio/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:17:09 crc restorecon[4687]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sonataflow-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:17:09 crc restorecon[4687]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sonataflow-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:17:09 crc restorecon[4687]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sosreport-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:17:09 crc restorecon[4687]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sosreport-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:17:09 crc restorecon[4687]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/spark-helm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:17:09 crc restorecon[4687]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/spark-helm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:17:09 crc restorecon[4687]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/special-resource-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:17:09 crc restorecon[4687]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/special-resource-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:17:09 crc restorecon[4687]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/stolostron not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:17:09 crc restorecon[4687]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/stolostron/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:17:09 crc restorecon[4687]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/stolostron-engine not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:17:09 crc restorecon[4687]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/stolostron-engine/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:17:09 crc restorecon[4687]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/strimzi-kafka-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:17:09 crc restorecon[4687]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/strimzi-kafka-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:17:09 crc restorecon[4687]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/syndesis not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:17:09 crc restorecon[4687]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/syndesis/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:17:09 crc restorecon[4687]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/t8c not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:17:09 crc restorecon[4687]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/t8c/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:17:09 crc restorecon[4687]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tagger not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:17:09 crc restorecon[4687]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tagger/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:17:09 crc restorecon[4687]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tempo-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:17:09 crc restorecon[4687]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tempo-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:17:09 crc restorecon[4687]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tf-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:17:09 crc restorecon[4687]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tf-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:17:09 crc restorecon[4687]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tidb-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:17:09 crc restorecon[4687]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tidb-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:17:09 crc restorecon[4687]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/trident-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:17:09 crc restorecon[4687]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/trident-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:17:09 crc restorecon[4687]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/trustify-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:17:09 crc restorecon[4687]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/trustify-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:17:09 crc restorecon[4687]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ucs-ci-solutions-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:17:09 crc restorecon[4687]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ucs-ci-solutions-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:17:09 crc restorecon[4687]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/universal-crossplane not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:17:09 crc restorecon[4687]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/universal-crossplane/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:17:09 crc restorecon[4687]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/varnish-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:17:09 crc restorecon[4687]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/varnish-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:17:09 crc restorecon[4687]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/vault-config-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:17:09 crc restorecon[4687]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/vault-config-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:17:09 crc restorecon[4687]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/verticadb-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:17:09 crc restorecon[4687]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/verticadb-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:17:09 crc restorecon[4687]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/volume-expander-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:17:09 crc restorecon[4687]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/volume-expander-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:17:09 crc restorecon[4687]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/wandb-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:17:09 crc restorecon[4687]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/wandb-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:17:09 crc restorecon[4687]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/windup-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:17:09 crc restorecon[4687]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/windup-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:17:09 crc restorecon[4687]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/yaks not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:17:09 crc restorecon[4687]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/yaks/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:17:09 crc restorecon[4687]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/utilities not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:17:09 crc restorecon[4687]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/utilities/copy-content not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:17:09 crc restorecon[4687]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:17:09 crc restorecon[4687]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/containers/extract-utilities/c0fe7256 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:17:09 crc restorecon[4687]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/containers/extract-utilities/c30319e4 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:17:09 crc restorecon[4687]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/containers/extract-utilities/e6b1dd45 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:17:09 crc restorecon[4687]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/containers/extract-content/2bb643f0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:17:09 crc restorecon[4687]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/containers/extract-content/920de426 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:17:09 crc restorecon[4687]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/containers/extract-content/70fa1e87 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:17:09 crc restorecon[4687]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/containers/registry-server/a1c12a2f not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:17:09 crc restorecon[4687]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/containers/registry-server/9442e6c7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:17:09 crc restorecon[4687]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/containers/registry-server/5b45ec72 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:17:09 crc restorecon[4687]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:17:09 crc restorecon[4687]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:17:09 crc restorecon[4687]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/abot-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:17:09 crc restorecon[4687]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/abot-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:17:09 crc restorecon[4687]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aerospike-kubernetes-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:17:09 crc restorecon[4687]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aerospike-kubernetes-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:17:09 crc restorecon[4687]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aikit-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:17:09 crc restorecon[4687]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aikit-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:17:09 crc restorecon[4687]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzo-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:17:09 crc restorecon[4687]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzo-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:17:09 crc restorecon[4687]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzograph-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:17:09 crc restorecon[4687]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzograph-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:17:09 crc restorecon[4687]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzounstructured-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:17:09 crc restorecon[4687]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzounstructured-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:17:09 crc restorecon[4687]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloudbees-ci-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:17:09 crc restorecon[4687]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloudbees-ci-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:17:09 crc restorecon[4687]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cockroachdb-certified-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:17:09 crc restorecon[4687]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cockroachdb-certified-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:17:09 crc restorecon[4687]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/crunchy-postgres-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:17:09 crc restorecon[4687]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/crunchy-postgres-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:17:09 crc restorecon[4687]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datadog-operator-certified-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:17:09 crc restorecon[4687]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datadog-operator-certified-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:17:09 crc restorecon[4687]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dynatrace-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:17:09 crc restorecon[4687]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dynatrace-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:17:09 crc restorecon[4687]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/entando-k8s-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:17:09 crc restorecon[4687]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/entando-k8s-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:17:09 crc restorecon[4687]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/flux not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:17:09 crc restorecon[4687]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/flux/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:17:09 crc restorecon[4687]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/instana-agent-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:17:09 crc restorecon[4687]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/instana-agent-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:17:09 crc restorecon[4687]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/iomesh-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:17:09 crc restorecon[4687]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/iomesh-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:17:09 crc restorecon[4687]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:17:09 crc restorecon[4687]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:17:09 crc restorecon[4687]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx8-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:17:09 crc restorecon[4687]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx8-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:17:09 crc restorecon[4687]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k10-kasten-operator-paygo-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:17:09 crc restorecon[4687]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k10-kasten-operator-paygo-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:17:09 crc restorecon[4687]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k10-kasten-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:17:09 crc restorecon[4687]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k10-kasten-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:17:09 crc restorecon[4687]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k10-kasten-operator-term-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:17:09 crc restorecon[4687]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k10-kasten-operator-term-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:17:09 crc restorecon[4687]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubemq-operator-marketplace-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:17:09 crc restorecon[4687]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubemq-operator-marketplace-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:17:09 crc restorecon[4687]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubeturbo-certified-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:17:09 crc restorecon[4687]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubeturbo-certified-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:17:09 crc restorecon[4687]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/linstor-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:17:09 crc restorecon[4687]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/linstor-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:17:09 crc restorecon[4687]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/marketplace-games-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:17:09 crc restorecon[4687]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/marketplace-games-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:17:09 crc restorecon[4687]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/model-builder-for-vision-certified-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:17:09 crc restorecon[4687]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/model-builder-for-vision-certified-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:17:09 crc restorecon[4687]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/neuvector-certified-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:17:09 crc restorecon[4687]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/neuvector-certified-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:17:09 crc restorecon[4687]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ovms-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:17:09 crc restorecon[4687]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ovms-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:17:09 crc restorecon[4687]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pachyderm-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:17:09 crc restorecon[4687]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pachyderm-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:17:09 crc restorecon[4687]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/redis-enterprise-operator-cert-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:17:09 crc restorecon[4687]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/redis-enterprise-operator-cert-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:17:09 crc restorecon[4687]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/seldon-deploy-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:17:09 crc restorecon[4687]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/seldon-deploy-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:17:09 crc restorecon[4687]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/starburst-enterprise-helm-operator-paygo-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:17:09 crc restorecon[4687]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/starburst-enterprise-helm-operator-paygo-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:17:09 crc restorecon[4687]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/starburst-enterprise-helm-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:17:09 crc restorecon[4687]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/starburst-enterprise-helm-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:17:09 crc restorecon[4687]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/t8c-certified-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:17:09 crc restorecon[4687]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/t8c-certified-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:17:09 crc restorecon[4687]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/timemachine-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:17:09 crc restorecon[4687]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/timemachine-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:17:09 crc restorecon[4687]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/vfunction-server-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:17:09 crc restorecon[4687]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/vfunction-server-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:17:09 crc restorecon[4687]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/xcrypt-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:17:09 crc restorecon[4687]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/xcrypt-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:17:09 crc restorecon[4687]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/yugabyte-platform-operator-bundle-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:17:09 crc restorecon[4687]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/yugabyte-platform-operator-bundle-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:17:09 crc restorecon[4687]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/zabbix-operator-certified-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:17:09 crc restorecon[4687]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/zabbix-operator-certified-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:17:09 crc restorecon[4687]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:17:09 crc restorecon[4687]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:17:09 crc restorecon[4687]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:17:09 crc restorecon[4687]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/00000-1.psg not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:17:09 crc restorecon[4687]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/00000-1.psg.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:17:09 crc restorecon[4687]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/db.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:17:09 crc restorecon[4687]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/index.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:17:09 crc restorecon[4687]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/main.pix not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:17:09 crc restorecon[4687]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/overflow.pix not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:17:09 crc restorecon[4687]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/digest not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:17:09 crc restorecon[4687]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/utilities not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:17:09 crc restorecon[4687]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/utilities/copy-content not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:17:09 crc restorecon[4687]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:17:09 crc restorecon[4687]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/containers/extract-utilities/3c9f3a59 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:17:09 crc restorecon[4687]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/containers/extract-utilities/1091c11b not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:17:09 crc restorecon[4687]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/containers/extract-utilities/9a6821c6 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:17:09 crc restorecon[4687]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/containers/extract-content/ec0c35e2 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:17:09 crc restorecon[4687]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/containers/extract-content/517f37e7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:17:09 crc restorecon[4687]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/containers/extract-content/6214fe78 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:17:09 crc restorecon[4687]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/containers/registry-server/ba189c8b not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:17:09 crc restorecon[4687]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/containers/registry-server/351e4f31 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:17:09 crc restorecon[4687]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/containers/registry-server/c0f219ff not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 08 18:17:09 crc restorecon[4687]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c247,c522 Oct 08 18:17:09 crc restorecon[4687]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/wait-for-host-port/8069f607 not reset as customized by admin to system_u:object_r:container_file_t:s0:c378,c723 Oct 08 18:17:09 crc restorecon[4687]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/wait-for-host-port/559c3d82 not reset as customized by admin to system_u:object_r:container_file_t:s0:c133,c223 Oct 08 18:17:09 crc restorecon[4687]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/wait-for-host-port/605ad488 not reset as customized by admin to system_u:object_r:container_file_t:s0:c247,c522 Oct 08 18:17:09 crc restorecon[4687]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/kube-scheduler/148df488 not reset as customized by admin to system_u:object_r:container_file_t:s0:c378,c723 Oct 08 18:17:09 crc restorecon[4687]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/kube-scheduler/3bf6dcb4 not reset as customized by admin to system_u:object_r:container_file_t:s0:c133,c223 Oct 08 18:17:09 crc restorecon[4687]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/kube-scheduler/022a2feb not reset as customized by admin to system_u:object_r:container_file_t:s0:c247,c522 Oct 08 18:17:09 crc restorecon[4687]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/kube-scheduler-cert-syncer/938c3924 not reset as customized by admin to system_u:object_r:container_file_t:s0:c378,c723 Oct 08 18:17:09 crc restorecon[4687]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/kube-scheduler-cert-syncer/729fe23e not reset as customized by admin to system_u:object_r:container_file_t:s0:c133,c223 Oct 08 18:17:09 crc restorecon[4687]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/kube-scheduler-cert-syncer/1fd5cbd4 not reset as customized by admin to system_u:object_r:container_file_t:s0:c247,c522 Oct 08 18:17:09 crc restorecon[4687]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/kube-scheduler-recovery-controller/a96697e1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c378,c723 Oct 08 18:17:09 crc restorecon[4687]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/kube-scheduler-recovery-controller/e155ddca not reset as customized by admin to system_u:object_r:container_file_t:s0:c133,c223 Oct 08 18:17:09 crc restorecon[4687]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/kube-scheduler-recovery-controller/10dd0e0f not reset as customized by admin to system_u:object_r:container_file_t:s0:c247,c522 Oct 08 18:17:09 crc restorecon[4687]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-trusted-ca-bundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Oct 08 18:17:09 crc restorecon[4687]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-trusted-ca-bundle/..2025_02_24_06_09_35.3018472960 not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Oct 08 18:17:09 crc restorecon[4687]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-trusted-ca-bundle/..2025_02_24_06_09_35.3018472960/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Oct 08 18:17:09 crc restorecon[4687]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-trusted-ca-bundle/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Oct 08 18:17:09 crc restorecon[4687]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-trusted-ca-bundle/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Oct 08 18:17:09 crc restorecon[4687]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/audit-policies not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Oct 08 18:17:09 crc restorecon[4687]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/audit-policies/..2025_02_24_06_09_35.4262376737 not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Oct 08 18:17:09 crc restorecon[4687]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/audit-policies/..2025_02_24_06_09_35.4262376737/audit.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Oct 08 18:17:09 crc restorecon[4687]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/audit-policies/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Oct 08 18:17:09 crc restorecon[4687]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/audit-policies/audit.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Oct 08 18:17:09 crc restorecon[4687]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-cliconfig not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Oct 08 18:17:09 crc restorecon[4687]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-cliconfig/..2025_02_24_06_09_35.2630275752 not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Oct 08 18:17:09 crc restorecon[4687]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-cliconfig/..2025_02_24_06_09_35.2630275752/v4-0-config-system-cliconfig not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Oct 08 18:17:09 crc restorecon[4687]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-cliconfig/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Oct 08 18:17:09 crc restorecon[4687]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-cliconfig/v4-0-config-system-cliconfig not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Oct 08 18:17:09 crc restorecon[4687]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-service-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Oct 08 18:17:09 crc restorecon[4687]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-service-ca/..2025_02_24_06_09_35.2376963788 not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Oct 08 18:17:09 crc restorecon[4687]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-service-ca/..2025_02_24_06_09_35.2376963788/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Oct 08 18:17:09 crc restorecon[4687]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-service-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Oct 08 18:17:09 crc restorecon[4687]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-service-ca/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Oct 08 18:17:09 crc restorecon[4687]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Oct 08 18:17:09 crc restorecon[4687]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/containers/oauth-openshift/6f2c8392 not reset as customized by admin to system_u:object_r:container_file_t:s0:c267,c588 Oct 08 18:17:09 crc restorecon[4687]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/containers/oauth-openshift/bd241ad9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Oct 08 18:17:09 crc restorecon[4687]: /var/lib/kubelet/plugins not reset as customized by admin to system_u:object_r:container_file_t:s0 Oct 08 18:17:09 crc restorecon[4687]: /var/lib/kubelet/plugins/csi-hostpath not reset as customized by admin to system_u:object_r:container_file_t:s0 Oct 08 18:17:09 crc restorecon[4687]: /var/lib/kubelet/plugins/csi-hostpath/csi.sock not reset as customized by admin to system_u:object_r:container_file_t:s0 Oct 08 18:17:09 crc restorecon[4687]: /var/lib/kubelet/plugins/kubernetes.io not reset as customized by admin to system_u:object_r:container_file_t:s0 Oct 08 18:17:09 crc restorecon[4687]: /var/lib/kubelet/plugins/kubernetes.io/csi not reset as customized by admin to system_u:object_r:container_file_t:s0 Oct 08 18:17:09 crc restorecon[4687]: /var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner not reset as customized by admin to system_u:object_r:container_file_t:s0 Oct 08 18:17:09 crc restorecon[4687]: /var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner/1f4776af88835e41c12b831b4c9fed40233456d14189815a54dbe7f892fc1983 not reset as customized by admin to system_u:object_r:container_file_t:s0 Oct 08 18:17:09 crc restorecon[4687]: /var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner/1f4776af88835e41c12b831b4c9fed40233456d14189815a54dbe7f892fc1983/globalmount not reset as customized by admin to system_u:object_r:container_file_t:s0 Oct 08 18:17:09 crc restorecon[4687]: /var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner/1f4776af88835e41c12b831b4c9fed40233456d14189815a54dbe7f892fc1983/vol_data.json not reset as customized by admin to system_u:object_r:container_file_t:s0 Oct 08 18:17:09 crc restorecon[4687]: /var/lib/kubelet/plugins_registry not reset as customized by admin to system_u:object_r:container_file_t:s0 Oct 08 18:17:09 crc restorecon[4687]: Relabeled /var/usrlocal/bin/kubenswrapper from system_u:object_r:bin_t:s0 to system_u:object_r:kubelet_exec_t:s0 Oct 08 18:17:10 crc kubenswrapper[4859]: Flag --container-runtime-endpoint has been deprecated, This parameter should be set via the config file specified by the Kubelet's --config flag. See https://kubernetes.io/docs/tasks/administer-cluster/kubelet-config-file/ for more information. Oct 08 18:17:10 crc kubenswrapper[4859]: Flag --minimum-container-ttl-duration has been deprecated, Use --eviction-hard or --eviction-soft instead. Will be removed in a future version. Oct 08 18:17:10 crc kubenswrapper[4859]: Flag --volume-plugin-dir has been deprecated, This parameter should be set via the config file specified by the Kubelet's --config flag. See https://kubernetes.io/docs/tasks/administer-cluster/kubelet-config-file/ for more information. Oct 08 18:17:10 crc kubenswrapper[4859]: Flag --register-with-taints has been deprecated, This parameter should be set via the config file specified by the Kubelet's --config flag. See https://kubernetes.io/docs/tasks/administer-cluster/kubelet-config-file/ for more information. Oct 08 18:17:10 crc kubenswrapper[4859]: Flag --pod-infra-container-image has been deprecated, will be removed in a future release. Image garbage collector will get sandbox image information from CRI. Oct 08 18:17:10 crc kubenswrapper[4859]: Flag --system-reserved has been deprecated, This parameter should be set via the config file specified by the Kubelet's --config flag. See https://kubernetes.io/docs/tasks/administer-cluster/kubelet-config-file/ for more information. Oct 08 18:17:10 crc kubenswrapper[4859]: I1008 18:17:10.184829 4859 server.go:211] "--pod-infra-container-image will not be pruned by the image garbage collector in kubelet and should also be set in the remote runtime" Oct 08 18:17:10 crc kubenswrapper[4859]: W1008 18:17:10.190254 4859 feature_gate.go:330] unrecognized feature gate: RouteAdvertisements Oct 08 18:17:10 crc kubenswrapper[4859]: W1008 18:17:10.190424 4859 feature_gate.go:330] unrecognized feature gate: GCPClusterHostedDNS Oct 08 18:17:10 crc kubenswrapper[4859]: W1008 18:17:10.190537 4859 feature_gate.go:330] unrecognized feature gate: ClusterAPIInstall Oct 08 18:17:10 crc kubenswrapper[4859]: W1008 18:17:10.190664 4859 feature_gate.go:330] unrecognized feature gate: MultiArchInstallGCP Oct 08 18:17:10 crc kubenswrapper[4859]: W1008 18:17:10.190833 4859 feature_gate.go:330] unrecognized feature gate: OVNObservability Oct 08 18:17:10 crc kubenswrapper[4859]: W1008 18:17:10.190936 4859 feature_gate.go:330] unrecognized feature gate: InsightsConfig Oct 08 18:17:10 crc kubenswrapper[4859]: W1008 18:17:10.191026 4859 feature_gate.go:330] unrecognized feature gate: ManagedBootImages Oct 08 18:17:10 crc kubenswrapper[4859]: W1008 18:17:10.191139 4859 feature_gate.go:330] unrecognized feature gate: BareMetalLoadBalancer Oct 08 18:17:10 crc kubenswrapper[4859]: W1008 18:17:10.191233 4859 feature_gate.go:330] unrecognized feature gate: VolumeGroupSnapshot Oct 08 18:17:10 crc kubenswrapper[4859]: W1008 18:17:10.191322 4859 feature_gate.go:330] unrecognized feature gate: GatewayAPI Oct 08 18:17:10 crc kubenswrapper[4859]: W1008 18:17:10.191411 4859 feature_gate.go:330] unrecognized feature gate: NetworkLiveMigration Oct 08 18:17:10 crc kubenswrapper[4859]: W1008 18:17:10.191500 4859 feature_gate.go:330] unrecognized feature gate: VSphereMultiVCenters Oct 08 18:17:10 crc kubenswrapper[4859]: W1008 18:17:10.191599 4859 feature_gate.go:330] unrecognized feature gate: VSphereStaticIPs Oct 08 18:17:10 crc kubenswrapper[4859]: W1008 18:17:10.191697 4859 feature_gate.go:330] unrecognized feature gate: NetworkDiagnosticsConfig Oct 08 18:17:10 crc kubenswrapper[4859]: W1008 18:17:10.191831 4859 feature_gate.go:330] unrecognized feature gate: NewOLM Oct 08 18:17:10 crc kubenswrapper[4859]: W1008 18:17:10.191926 4859 feature_gate.go:353] Setting GA feature gate ValidatingAdmissionPolicy=true. It will be removed in a future release. Oct 08 18:17:10 crc kubenswrapper[4859]: W1008 18:17:10.192020 4859 feature_gate.go:353] Setting GA feature gate DisableKubeletCloudCredentialProviders=true. It will be removed in a future release. Oct 08 18:17:10 crc kubenswrapper[4859]: W1008 18:17:10.192113 4859 feature_gate.go:330] unrecognized feature gate: HardwareSpeed Oct 08 18:17:10 crc kubenswrapper[4859]: W1008 18:17:10.192216 4859 feature_gate.go:330] unrecognized feature gate: NutanixMultiSubnets Oct 08 18:17:10 crc kubenswrapper[4859]: W1008 18:17:10.192333 4859 feature_gate.go:330] unrecognized feature gate: MultiArchInstallAWS Oct 08 18:17:10 crc kubenswrapper[4859]: W1008 18:17:10.192432 4859 feature_gate.go:353] Setting GA feature gate CloudDualStackNodeIPs=true. It will be removed in a future release. Oct 08 18:17:10 crc kubenswrapper[4859]: W1008 18:17:10.192524 4859 feature_gate.go:330] unrecognized feature gate: OnClusterBuild Oct 08 18:17:10 crc kubenswrapper[4859]: W1008 18:17:10.192613 4859 feature_gate.go:330] unrecognized feature gate: CSIDriverSharedResource Oct 08 18:17:10 crc kubenswrapper[4859]: W1008 18:17:10.192708 4859 feature_gate.go:330] unrecognized feature gate: VSphereMultiNetworks Oct 08 18:17:10 crc kubenswrapper[4859]: W1008 18:17:10.192829 4859 feature_gate.go:330] unrecognized feature gate: DNSNameResolver Oct 08 18:17:10 crc kubenswrapper[4859]: W1008 18:17:10.192918 4859 feature_gate.go:330] unrecognized feature gate: ManagedBootImagesAWS Oct 08 18:17:10 crc kubenswrapper[4859]: W1008 18:17:10.193023 4859 feature_gate.go:330] unrecognized feature gate: MetricsCollectionProfiles Oct 08 18:17:10 crc kubenswrapper[4859]: W1008 18:17:10.193115 4859 feature_gate.go:330] unrecognized feature gate: BuildCSIVolumes Oct 08 18:17:10 crc kubenswrapper[4859]: W1008 18:17:10.193205 4859 feature_gate.go:330] unrecognized feature gate: MachineAPIProviderOpenStack Oct 08 18:17:10 crc kubenswrapper[4859]: W1008 18:17:10.193293 4859 feature_gate.go:330] unrecognized feature gate: AdminNetworkPolicy Oct 08 18:17:10 crc kubenswrapper[4859]: W1008 18:17:10.193380 4859 feature_gate.go:330] unrecognized feature gate: SignatureStores Oct 08 18:17:10 crc kubenswrapper[4859]: W1008 18:17:10.193467 4859 feature_gate.go:330] unrecognized feature gate: IngressControllerDynamicConfigurationManager Oct 08 18:17:10 crc kubenswrapper[4859]: W1008 18:17:10.193553 4859 feature_gate.go:330] unrecognized feature gate: EtcdBackendQuota Oct 08 18:17:10 crc kubenswrapper[4859]: W1008 18:17:10.193653 4859 feature_gate.go:330] unrecognized feature gate: PrivateHostedZoneAWS Oct 08 18:17:10 crc kubenswrapper[4859]: W1008 18:17:10.193798 4859 feature_gate.go:330] unrecognized feature gate: MinimumKubeletVersion Oct 08 18:17:10 crc kubenswrapper[4859]: W1008 18:17:10.193907 4859 feature_gate.go:330] unrecognized feature gate: UpgradeStatus Oct 08 18:17:10 crc kubenswrapper[4859]: W1008 18:17:10.193999 4859 feature_gate.go:330] unrecognized feature gate: MultiArchInstallAzure Oct 08 18:17:10 crc kubenswrapper[4859]: W1008 18:17:10.194088 4859 feature_gate.go:330] unrecognized feature gate: MachineAPIMigration Oct 08 18:17:10 crc kubenswrapper[4859]: W1008 18:17:10.194178 4859 feature_gate.go:330] unrecognized feature gate: ChunkSizeMiB Oct 08 18:17:10 crc kubenswrapper[4859]: W1008 18:17:10.194283 4859 feature_gate.go:330] unrecognized feature gate: InsightsConfigAPI Oct 08 18:17:10 crc kubenswrapper[4859]: W1008 18:17:10.194387 4859 feature_gate.go:351] Setting deprecated feature gate KMSv1=true. It will be removed in a future release. Oct 08 18:17:10 crc kubenswrapper[4859]: W1008 18:17:10.194483 4859 feature_gate.go:330] unrecognized feature gate: NetworkSegmentation Oct 08 18:17:10 crc kubenswrapper[4859]: W1008 18:17:10.194575 4859 feature_gate.go:330] unrecognized feature gate: OpenShiftPodSecurityAdmission Oct 08 18:17:10 crc kubenswrapper[4859]: W1008 18:17:10.194664 4859 feature_gate.go:330] unrecognized feature gate: InsightsOnDemandDataGather Oct 08 18:17:10 crc kubenswrapper[4859]: W1008 18:17:10.194825 4859 feature_gate.go:330] unrecognized feature gate: ImageStreamImportMode Oct 08 18:17:10 crc kubenswrapper[4859]: W1008 18:17:10.194948 4859 feature_gate.go:330] unrecognized feature gate: ExternalOIDC Oct 08 18:17:10 crc kubenswrapper[4859]: W1008 18:17:10.195062 4859 feature_gate.go:330] unrecognized feature gate: InsightsRuntimeExtractor Oct 08 18:17:10 crc kubenswrapper[4859]: W1008 18:17:10.195171 4859 feature_gate.go:330] unrecognized feature gate: AutomatedEtcdBackup Oct 08 18:17:10 crc kubenswrapper[4859]: W1008 18:17:10.195265 4859 feature_gate.go:330] unrecognized feature gate: AWSClusterHostedDNS Oct 08 18:17:10 crc kubenswrapper[4859]: W1008 18:17:10.195353 4859 feature_gate.go:330] unrecognized feature gate: AdditionalRoutingCapabilities Oct 08 18:17:10 crc kubenswrapper[4859]: W1008 18:17:10.195442 4859 feature_gate.go:330] unrecognized feature gate: ClusterMonitoringConfig Oct 08 18:17:10 crc kubenswrapper[4859]: W1008 18:17:10.195540 4859 feature_gate.go:330] unrecognized feature gate: PinnedImages Oct 08 18:17:10 crc kubenswrapper[4859]: W1008 18:17:10.195631 4859 feature_gate.go:330] unrecognized feature gate: SigstoreImageVerification Oct 08 18:17:10 crc kubenswrapper[4859]: W1008 18:17:10.195763 4859 feature_gate.go:330] unrecognized feature gate: VSphereDriverConfiguration Oct 08 18:17:10 crc kubenswrapper[4859]: W1008 18:17:10.195882 4859 feature_gate.go:330] unrecognized feature gate: IngressControllerLBSubnetsAWS Oct 08 18:17:10 crc kubenswrapper[4859]: W1008 18:17:10.195984 4859 feature_gate.go:330] unrecognized feature gate: ConsolePluginContentSecurityPolicy Oct 08 18:17:10 crc kubenswrapper[4859]: W1008 18:17:10.196075 4859 feature_gate.go:330] unrecognized feature gate: VSphereControlPlaneMachineSet Oct 08 18:17:10 crc kubenswrapper[4859]: W1008 18:17:10.196173 4859 feature_gate.go:330] unrecognized feature gate: AlibabaPlatform Oct 08 18:17:10 crc kubenswrapper[4859]: W1008 18:17:10.196266 4859 feature_gate.go:330] unrecognized feature gate: NodeDisruptionPolicy Oct 08 18:17:10 crc kubenswrapper[4859]: W1008 18:17:10.196357 4859 feature_gate.go:330] unrecognized feature gate: MachineConfigNodes Oct 08 18:17:10 crc kubenswrapper[4859]: W1008 18:17:10.196445 4859 feature_gate.go:330] unrecognized feature gate: MixedCPUsAllocation Oct 08 18:17:10 crc kubenswrapper[4859]: W1008 18:17:10.196545 4859 feature_gate.go:330] unrecognized feature gate: MachineAPIOperatorDisableMachineHealthCheckController Oct 08 18:17:10 crc kubenswrapper[4859]: W1008 18:17:10.196638 4859 feature_gate.go:330] unrecognized feature gate: BootcNodeManagement Oct 08 18:17:10 crc kubenswrapper[4859]: W1008 18:17:10.196774 4859 feature_gate.go:330] unrecognized feature gate: GCPLabelsTags Oct 08 18:17:10 crc kubenswrapper[4859]: W1008 18:17:10.196876 4859 feature_gate.go:330] unrecognized feature gate: Example Oct 08 18:17:10 crc kubenswrapper[4859]: W1008 18:17:10.196968 4859 feature_gate.go:330] unrecognized feature gate: AzureWorkloadIdentity Oct 08 18:17:10 crc kubenswrapper[4859]: W1008 18:17:10.197058 4859 feature_gate.go:330] unrecognized feature gate: SetEIPForNLBIngressController Oct 08 18:17:10 crc kubenswrapper[4859]: W1008 18:17:10.197149 4859 feature_gate.go:330] unrecognized feature gate: AWSEFSDriverVolumeMetrics Oct 08 18:17:10 crc kubenswrapper[4859]: W1008 18:17:10.197257 4859 feature_gate.go:330] unrecognized feature gate: ClusterAPIInstallIBMCloud Oct 08 18:17:10 crc kubenswrapper[4859]: W1008 18:17:10.197352 4859 feature_gate.go:330] unrecognized feature gate: PlatformOperators Oct 08 18:17:10 crc kubenswrapper[4859]: W1008 18:17:10.197446 4859 feature_gate.go:330] unrecognized feature gate: PersistentIPsForVirtualization Oct 08 18:17:10 crc kubenswrapper[4859]: I1008 18:17:10.197753 4859 flags.go:64] FLAG: --address="0.0.0.0" Oct 08 18:17:10 crc kubenswrapper[4859]: I1008 18:17:10.197876 4859 flags.go:64] FLAG: --allowed-unsafe-sysctls="[]" Oct 08 18:17:10 crc kubenswrapper[4859]: I1008 18:17:10.197999 4859 flags.go:64] FLAG: --anonymous-auth="true" Oct 08 18:17:10 crc kubenswrapper[4859]: I1008 18:17:10.198099 4859 flags.go:64] FLAG: --application-metrics-count-limit="100" Oct 08 18:17:10 crc kubenswrapper[4859]: I1008 18:17:10.198207 4859 flags.go:64] FLAG: --authentication-token-webhook="false" Oct 08 18:17:10 crc kubenswrapper[4859]: I1008 18:17:10.198302 4859 flags.go:64] FLAG: --authentication-token-webhook-cache-ttl="2m0s" Oct 08 18:17:10 crc kubenswrapper[4859]: I1008 18:17:10.198398 4859 flags.go:64] FLAG: --authorization-mode="AlwaysAllow" Oct 08 18:17:10 crc kubenswrapper[4859]: I1008 18:17:10.198503 4859 flags.go:64] FLAG: --authorization-webhook-cache-authorized-ttl="5m0s" Oct 08 18:17:10 crc kubenswrapper[4859]: I1008 18:17:10.198606 4859 flags.go:64] FLAG: --authorization-webhook-cache-unauthorized-ttl="30s" Oct 08 18:17:10 crc kubenswrapper[4859]: I1008 18:17:10.198708 4859 flags.go:64] FLAG: --boot-id-file="/proc/sys/kernel/random/boot_id" Oct 08 18:17:10 crc kubenswrapper[4859]: I1008 18:17:10.198834 4859 flags.go:64] FLAG: --bootstrap-kubeconfig="/etc/kubernetes/kubeconfig" Oct 08 18:17:10 crc kubenswrapper[4859]: I1008 18:17:10.199011 4859 flags.go:64] FLAG: --cert-dir="/var/lib/kubelet/pki" Oct 08 18:17:10 crc kubenswrapper[4859]: I1008 18:17:10.199135 4859 flags.go:64] FLAG: --cgroup-driver="cgroupfs" Oct 08 18:17:10 crc kubenswrapper[4859]: I1008 18:17:10.199241 4859 flags.go:64] FLAG: --cgroup-root="" Oct 08 18:17:10 crc kubenswrapper[4859]: I1008 18:17:10.199335 4859 flags.go:64] FLAG: --cgroups-per-qos="true" Oct 08 18:17:10 crc kubenswrapper[4859]: I1008 18:17:10.199427 4859 flags.go:64] FLAG: --client-ca-file="" Oct 08 18:17:10 crc kubenswrapper[4859]: I1008 18:17:10.199533 4859 flags.go:64] FLAG: --cloud-config="" Oct 08 18:17:10 crc kubenswrapper[4859]: I1008 18:17:10.199634 4859 flags.go:64] FLAG: --cloud-provider="" Oct 08 18:17:10 crc kubenswrapper[4859]: I1008 18:17:10.199792 4859 flags.go:64] FLAG: --cluster-dns="[]" Oct 08 18:17:10 crc kubenswrapper[4859]: I1008 18:17:10.199899 4859 flags.go:64] FLAG: --cluster-domain="" Oct 08 18:17:10 crc kubenswrapper[4859]: I1008 18:17:10.200010 4859 flags.go:64] FLAG: --config="/etc/kubernetes/kubelet.conf" Oct 08 18:17:10 crc kubenswrapper[4859]: I1008 18:17:10.200107 4859 flags.go:64] FLAG: --config-dir="" Oct 08 18:17:10 crc kubenswrapper[4859]: I1008 18:17:10.200206 4859 flags.go:64] FLAG: --container-hints="/etc/cadvisor/container_hints.json" Oct 08 18:17:10 crc kubenswrapper[4859]: I1008 18:17:10.200300 4859 flags.go:64] FLAG: --container-log-max-files="5" Oct 08 18:17:10 crc kubenswrapper[4859]: I1008 18:17:10.200396 4859 flags.go:64] FLAG: --container-log-max-size="10Mi" Oct 08 18:17:10 crc kubenswrapper[4859]: I1008 18:17:10.200497 4859 flags.go:64] FLAG: --container-runtime-endpoint="/var/run/crio/crio.sock" Oct 08 18:17:10 crc kubenswrapper[4859]: I1008 18:17:10.200593 4859 flags.go:64] FLAG: --containerd="/run/containerd/containerd.sock" Oct 08 18:17:10 crc kubenswrapper[4859]: I1008 18:17:10.200691 4859 flags.go:64] FLAG: --containerd-namespace="k8s.io" Oct 08 18:17:10 crc kubenswrapper[4859]: I1008 18:17:10.200830 4859 flags.go:64] FLAG: --contention-profiling="false" Oct 08 18:17:10 crc kubenswrapper[4859]: I1008 18:17:10.200928 4859 flags.go:64] FLAG: --cpu-cfs-quota="true" Oct 08 18:17:10 crc kubenswrapper[4859]: I1008 18:17:10.201028 4859 flags.go:64] FLAG: --cpu-cfs-quota-period="100ms" Oct 08 18:17:10 crc kubenswrapper[4859]: I1008 18:17:10.201130 4859 flags.go:64] FLAG: --cpu-manager-policy="none" Oct 08 18:17:10 crc kubenswrapper[4859]: I1008 18:17:10.201230 4859 flags.go:64] FLAG: --cpu-manager-policy-options="" Oct 08 18:17:10 crc kubenswrapper[4859]: I1008 18:17:10.201326 4859 flags.go:64] FLAG: --cpu-manager-reconcile-period="10s" Oct 08 18:17:10 crc kubenswrapper[4859]: I1008 18:17:10.201417 4859 flags.go:64] FLAG: --enable-controller-attach-detach="true" Oct 08 18:17:10 crc kubenswrapper[4859]: I1008 18:17:10.201507 4859 flags.go:64] FLAG: --enable-debugging-handlers="true" Oct 08 18:17:10 crc kubenswrapper[4859]: I1008 18:17:10.201608 4859 flags.go:64] FLAG: --enable-load-reader="false" Oct 08 18:17:10 crc kubenswrapper[4859]: I1008 18:17:10.201708 4859 flags.go:64] FLAG: --enable-server="true" Oct 08 18:17:10 crc kubenswrapper[4859]: I1008 18:17:10.201846 4859 flags.go:64] FLAG: --enforce-node-allocatable="[pods]" Oct 08 18:17:10 crc kubenswrapper[4859]: I1008 18:17:10.201950 4859 flags.go:64] FLAG: --event-burst="100" Oct 08 18:17:10 crc kubenswrapper[4859]: I1008 18:17:10.202193 4859 flags.go:64] FLAG: --event-qps="50" Oct 08 18:17:10 crc kubenswrapper[4859]: I1008 18:17:10.202308 4859 flags.go:64] FLAG: --event-storage-age-limit="default=0" Oct 08 18:17:10 crc kubenswrapper[4859]: I1008 18:17:10.202404 4859 flags.go:64] FLAG: --event-storage-event-limit="default=0" Oct 08 18:17:10 crc kubenswrapper[4859]: I1008 18:17:10.202504 4859 flags.go:64] FLAG: --eviction-hard="" Oct 08 18:17:10 crc kubenswrapper[4859]: I1008 18:17:10.202603 4859 flags.go:64] FLAG: --eviction-max-pod-grace-period="0" Oct 08 18:17:10 crc kubenswrapper[4859]: I1008 18:17:10.202701 4859 flags.go:64] FLAG: --eviction-minimum-reclaim="" Oct 08 18:17:10 crc kubenswrapper[4859]: I1008 18:17:10.202836 4859 flags.go:64] FLAG: --eviction-pressure-transition-period="5m0s" Oct 08 18:17:10 crc kubenswrapper[4859]: I1008 18:17:10.202931 4859 flags.go:64] FLAG: --eviction-soft="" Oct 08 18:17:10 crc kubenswrapper[4859]: I1008 18:17:10.203095 4859 flags.go:64] FLAG: --eviction-soft-grace-period="" Oct 08 18:17:10 crc kubenswrapper[4859]: I1008 18:17:10.203203 4859 flags.go:64] FLAG: --exit-on-lock-contention="false" Oct 08 18:17:10 crc kubenswrapper[4859]: I1008 18:17:10.203296 4859 flags.go:64] FLAG: --experimental-allocatable-ignore-eviction="false" Oct 08 18:17:10 crc kubenswrapper[4859]: I1008 18:17:10.203387 4859 flags.go:64] FLAG: --experimental-mounter-path="" Oct 08 18:17:10 crc kubenswrapper[4859]: I1008 18:17:10.203479 4859 flags.go:64] FLAG: --fail-cgroupv1="false" Oct 08 18:17:10 crc kubenswrapper[4859]: I1008 18:17:10.203600 4859 flags.go:64] FLAG: --fail-swap-on="true" Oct 08 18:17:10 crc kubenswrapper[4859]: I1008 18:17:10.203704 4859 flags.go:64] FLAG: --feature-gates="" Oct 08 18:17:10 crc kubenswrapper[4859]: I1008 18:17:10.203880 4859 flags.go:64] FLAG: --file-check-frequency="20s" Oct 08 18:17:10 crc kubenswrapper[4859]: I1008 18:17:10.203992 4859 flags.go:64] FLAG: --global-housekeeping-interval="1m0s" Oct 08 18:17:10 crc kubenswrapper[4859]: I1008 18:17:10.204092 4859 flags.go:64] FLAG: --hairpin-mode="promiscuous-bridge" Oct 08 18:17:10 crc kubenswrapper[4859]: I1008 18:17:10.204181 4859 flags.go:64] FLAG: --healthz-bind-address="127.0.0.1" Oct 08 18:17:10 crc kubenswrapper[4859]: I1008 18:17:10.204273 4859 flags.go:64] FLAG: --healthz-port="10248" Oct 08 18:17:10 crc kubenswrapper[4859]: I1008 18:17:10.204364 4859 flags.go:64] FLAG: --help="false" Oct 08 18:17:10 crc kubenswrapper[4859]: I1008 18:17:10.204463 4859 flags.go:64] FLAG: --hostname-override="" Oct 08 18:17:10 crc kubenswrapper[4859]: I1008 18:17:10.204556 4859 flags.go:64] FLAG: --housekeeping-interval="10s" Oct 08 18:17:10 crc kubenswrapper[4859]: I1008 18:17:10.204656 4859 flags.go:64] FLAG: --http-check-frequency="20s" Oct 08 18:17:10 crc kubenswrapper[4859]: I1008 18:17:10.204788 4859 flags.go:64] FLAG: --image-credential-provider-bin-dir="" Oct 08 18:17:10 crc kubenswrapper[4859]: I1008 18:17:10.204901 4859 flags.go:64] FLAG: --image-credential-provider-config="" Oct 08 18:17:10 crc kubenswrapper[4859]: I1008 18:17:10.205003 4859 flags.go:64] FLAG: --image-gc-high-threshold="85" Oct 08 18:17:10 crc kubenswrapper[4859]: I1008 18:17:10.205098 4859 flags.go:64] FLAG: --image-gc-low-threshold="80" Oct 08 18:17:10 crc kubenswrapper[4859]: I1008 18:17:10.205188 4859 flags.go:64] FLAG: --image-service-endpoint="" Oct 08 18:17:10 crc kubenswrapper[4859]: I1008 18:17:10.205276 4859 flags.go:64] FLAG: --kernel-memcg-notification="false" Oct 08 18:17:10 crc kubenswrapper[4859]: I1008 18:17:10.205375 4859 flags.go:64] FLAG: --kube-api-burst="100" Oct 08 18:17:10 crc kubenswrapper[4859]: I1008 18:17:10.205468 4859 flags.go:64] FLAG: --kube-api-content-type="application/vnd.kubernetes.protobuf" Oct 08 18:17:10 crc kubenswrapper[4859]: I1008 18:17:10.205559 4859 flags.go:64] FLAG: --kube-api-qps="50" Oct 08 18:17:10 crc kubenswrapper[4859]: I1008 18:17:10.205661 4859 flags.go:64] FLAG: --kube-reserved="" Oct 08 18:17:10 crc kubenswrapper[4859]: I1008 18:17:10.205873 4859 flags.go:64] FLAG: --kube-reserved-cgroup="" Oct 08 18:17:10 crc kubenswrapper[4859]: I1008 18:17:10.205910 4859 flags.go:64] FLAG: --kubeconfig="/var/lib/kubelet/kubeconfig" Oct 08 18:17:10 crc kubenswrapper[4859]: I1008 18:17:10.205925 4859 flags.go:64] FLAG: --kubelet-cgroups="" Oct 08 18:17:10 crc kubenswrapper[4859]: I1008 18:17:10.205936 4859 flags.go:64] FLAG: --local-storage-capacity-isolation="true" Oct 08 18:17:10 crc kubenswrapper[4859]: I1008 18:17:10.205948 4859 flags.go:64] FLAG: --lock-file="" Oct 08 18:17:10 crc kubenswrapper[4859]: I1008 18:17:10.205959 4859 flags.go:64] FLAG: --log-cadvisor-usage="false" Oct 08 18:17:10 crc kubenswrapper[4859]: I1008 18:17:10.205972 4859 flags.go:64] FLAG: --log-flush-frequency="5s" Oct 08 18:17:10 crc kubenswrapper[4859]: I1008 18:17:10.205985 4859 flags.go:64] FLAG: --log-json-info-buffer-size="0" Oct 08 18:17:10 crc kubenswrapper[4859]: I1008 18:17:10.206003 4859 flags.go:64] FLAG: --log-json-split-stream="false" Oct 08 18:17:10 crc kubenswrapper[4859]: I1008 18:17:10.206015 4859 flags.go:64] FLAG: --log-text-info-buffer-size="0" Oct 08 18:17:10 crc kubenswrapper[4859]: I1008 18:17:10.206026 4859 flags.go:64] FLAG: --log-text-split-stream="false" Oct 08 18:17:10 crc kubenswrapper[4859]: I1008 18:17:10.206038 4859 flags.go:64] FLAG: --logging-format="text" Oct 08 18:17:10 crc kubenswrapper[4859]: I1008 18:17:10.206049 4859 flags.go:64] FLAG: --machine-id-file="/etc/machine-id,/var/lib/dbus/machine-id" Oct 08 18:17:10 crc kubenswrapper[4859]: I1008 18:17:10.206062 4859 flags.go:64] FLAG: --make-iptables-util-chains="true" Oct 08 18:17:10 crc kubenswrapper[4859]: I1008 18:17:10.206073 4859 flags.go:64] FLAG: --manifest-url="" Oct 08 18:17:10 crc kubenswrapper[4859]: I1008 18:17:10.206089 4859 flags.go:64] FLAG: --manifest-url-header="" Oct 08 18:17:10 crc kubenswrapper[4859]: I1008 18:17:10.206105 4859 flags.go:64] FLAG: --max-housekeeping-interval="15s" Oct 08 18:17:10 crc kubenswrapper[4859]: I1008 18:17:10.206118 4859 flags.go:64] FLAG: --max-open-files="1000000" Oct 08 18:17:10 crc kubenswrapper[4859]: I1008 18:17:10.206133 4859 flags.go:64] FLAG: --max-pods="110" Oct 08 18:17:10 crc kubenswrapper[4859]: I1008 18:17:10.206145 4859 flags.go:64] FLAG: --maximum-dead-containers="-1" Oct 08 18:17:10 crc kubenswrapper[4859]: I1008 18:17:10.206156 4859 flags.go:64] FLAG: --maximum-dead-containers-per-container="1" Oct 08 18:17:10 crc kubenswrapper[4859]: I1008 18:17:10.206168 4859 flags.go:64] FLAG: --memory-manager-policy="None" Oct 08 18:17:10 crc kubenswrapper[4859]: I1008 18:17:10.206180 4859 flags.go:64] FLAG: --minimum-container-ttl-duration="6m0s" Oct 08 18:17:10 crc kubenswrapper[4859]: I1008 18:17:10.206192 4859 flags.go:64] FLAG: --minimum-image-ttl-duration="2m0s" Oct 08 18:17:10 crc kubenswrapper[4859]: I1008 18:17:10.206204 4859 flags.go:64] FLAG: --node-ip="192.168.126.11" Oct 08 18:17:10 crc kubenswrapper[4859]: I1008 18:17:10.206215 4859 flags.go:64] FLAG: --node-labels="node-role.kubernetes.io/control-plane=,node-role.kubernetes.io/master=,node.openshift.io/os_id=rhcos" Oct 08 18:17:10 crc kubenswrapper[4859]: I1008 18:17:10.206244 4859 flags.go:64] FLAG: --node-status-max-images="50" Oct 08 18:17:10 crc kubenswrapper[4859]: I1008 18:17:10.206256 4859 flags.go:64] FLAG: --node-status-update-frequency="10s" Oct 08 18:17:10 crc kubenswrapper[4859]: I1008 18:17:10.206269 4859 flags.go:64] FLAG: --oom-score-adj="-999" Oct 08 18:17:10 crc kubenswrapper[4859]: I1008 18:17:10.206280 4859 flags.go:64] FLAG: --pod-cidr="" Oct 08 18:17:10 crc kubenswrapper[4859]: I1008 18:17:10.206292 4859 flags.go:64] FLAG: --pod-infra-container-image="quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:33549946e22a9ffa738fd94b1345f90921bc8f92fa6137784cb33c77ad806f9d" Oct 08 18:17:10 crc kubenswrapper[4859]: I1008 18:17:10.206309 4859 flags.go:64] FLAG: --pod-manifest-path="" Oct 08 18:17:10 crc kubenswrapper[4859]: I1008 18:17:10.206321 4859 flags.go:64] FLAG: --pod-max-pids="-1" Oct 08 18:17:10 crc kubenswrapper[4859]: I1008 18:17:10.206333 4859 flags.go:64] FLAG: --pods-per-core="0" Oct 08 18:17:10 crc kubenswrapper[4859]: I1008 18:17:10.206344 4859 flags.go:64] FLAG: --port="10250" Oct 08 18:17:10 crc kubenswrapper[4859]: I1008 18:17:10.206355 4859 flags.go:64] FLAG: --protect-kernel-defaults="false" Oct 08 18:17:10 crc kubenswrapper[4859]: I1008 18:17:10.206366 4859 flags.go:64] FLAG: --provider-id="" Oct 08 18:17:10 crc kubenswrapper[4859]: I1008 18:17:10.206376 4859 flags.go:64] FLAG: --qos-reserved="" Oct 08 18:17:10 crc kubenswrapper[4859]: I1008 18:17:10.206388 4859 flags.go:64] FLAG: --read-only-port="10255" Oct 08 18:17:10 crc kubenswrapper[4859]: I1008 18:17:10.206399 4859 flags.go:64] FLAG: --register-node="true" Oct 08 18:17:10 crc kubenswrapper[4859]: I1008 18:17:10.206410 4859 flags.go:64] FLAG: --register-schedulable="true" Oct 08 18:17:10 crc kubenswrapper[4859]: I1008 18:17:10.206421 4859 flags.go:64] FLAG: --register-with-taints="node-role.kubernetes.io/master=:NoSchedule" Oct 08 18:17:10 crc kubenswrapper[4859]: I1008 18:17:10.206442 4859 flags.go:64] FLAG: --registry-burst="10" Oct 08 18:17:10 crc kubenswrapper[4859]: I1008 18:17:10.206454 4859 flags.go:64] FLAG: --registry-qps="5" Oct 08 18:17:10 crc kubenswrapper[4859]: I1008 18:17:10.206465 4859 flags.go:64] FLAG: --reserved-cpus="" Oct 08 18:17:10 crc kubenswrapper[4859]: I1008 18:17:10.206475 4859 flags.go:64] FLAG: --reserved-memory="" Oct 08 18:17:10 crc kubenswrapper[4859]: I1008 18:17:10.206490 4859 flags.go:64] FLAG: --resolv-conf="/etc/resolv.conf" Oct 08 18:17:10 crc kubenswrapper[4859]: I1008 18:17:10.206501 4859 flags.go:64] FLAG: --root-dir="/var/lib/kubelet" Oct 08 18:17:10 crc kubenswrapper[4859]: I1008 18:17:10.206513 4859 flags.go:64] FLAG: --rotate-certificates="false" Oct 08 18:17:10 crc kubenswrapper[4859]: I1008 18:17:10.206525 4859 flags.go:64] FLAG: --rotate-server-certificates="false" Oct 08 18:17:10 crc kubenswrapper[4859]: I1008 18:17:10.206536 4859 flags.go:64] FLAG: --runonce="false" Oct 08 18:17:10 crc kubenswrapper[4859]: I1008 18:17:10.206547 4859 flags.go:64] FLAG: --runtime-cgroups="/system.slice/crio.service" Oct 08 18:17:10 crc kubenswrapper[4859]: I1008 18:17:10.206558 4859 flags.go:64] FLAG: --runtime-request-timeout="2m0s" Oct 08 18:17:10 crc kubenswrapper[4859]: I1008 18:17:10.206573 4859 flags.go:64] FLAG: --seccomp-default="false" Oct 08 18:17:10 crc kubenswrapper[4859]: I1008 18:17:10.206586 4859 flags.go:64] FLAG: --serialize-image-pulls="true" Oct 08 18:17:10 crc kubenswrapper[4859]: I1008 18:17:10.206597 4859 flags.go:64] FLAG: --storage-driver-buffer-duration="1m0s" Oct 08 18:17:10 crc kubenswrapper[4859]: I1008 18:17:10.206609 4859 flags.go:64] FLAG: --storage-driver-db="cadvisor" Oct 08 18:17:10 crc kubenswrapper[4859]: I1008 18:17:10.206620 4859 flags.go:64] FLAG: --storage-driver-host="localhost:8086" Oct 08 18:17:10 crc kubenswrapper[4859]: I1008 18:17:10.206632 4859 flags.go:64] FLAG: --storage-driver-password="root" Oct 08 18:17:10 crc kubenswrapper[4859]: I1008 18:17:10.206644 4859 flags.go:64] FLAG: --storage-driver-secure="false" Oct 08 18:17:10 crc kubenswrapper[4859]: I1008 18:17:10.206656 4859 flags.go:64] FLAG: --storage-driver-table="stats" Oct 08 18:17:10 crc kubenswrapper[4859]: I1008 18:17:10.206667 4859 flags.go:64] FLAG: --storage-driver-user="root" Oct 08 18:17:10 crc kubenswrapper[4859]: I1008 18:17:10.206678 4859 flags.go:64] FLAG: --streaming-connection-idle-timeout="4h0m0s" Oct 08 18:17:10 crc kubenswrapper[4859]: I1008 18:17:10.206700 4859 flags.go:64] FLAG: --sync-frequency="1m0s" Oct 08 18:17:10 crc kubenswrapper[4859]: I1008 18:17:10.206744 4859 flags.go:64] FLAG: --system-cgroups="" Oct 08 18:17:10 crc kubenswrapper[4859]: I1008 18:17:10.206757 4859 flags.go:64] FLAG: --system-reserved="cpu=200m,ephemeral-storage=350Mi,memory=350Mi" Oct 08 18:17:10 crc kubenswrapper[4859]: I1008 18:17:10.206776 4859 flags.go:64] FLAG: --system-reserved-cgroup="" Oct 08 18:17:10 crc kubenswrapper[4859]: I1008 18:17:10.206786 4859 flags.go:64] FLAG: --tls-cert-file="" Oct 08 18:17:10 crc kubenswrapper[4859]: I1008 18:17:10.206797 4859 flags.go:64] FLAG: --tls-cipher-suites="[]" Oct 08 18:17:10 crc kubenswrapper[4859]: I1008 18:17:10.206812 4859 flags.go:64] FLAG: --tls-min-version="" Oct 08 18:17:10 crc kubenswrapper[4859]: I1008 18:17:10.206823 4859 flags.go:64] FLAG: --tls-private-key-file="" Oct 08 18:17:10 crc kubenswrapper[4859]: I1008 18:17:10.206834 4859 flags.go:64] FLAG: --topology-manager-policy="none" Oct 08 18:17:10 crc kubenswrapper[4859]: I1008 18:17:10.206845 4859 flags.go:64] FLAG: --topology-manager-policy-options="" Oct 08 18:17:10 crc kubenswrapper[4859]: I1008 18:17:10.206856 4859 flags.go:64] FLAG: --topology-manager-scope="container" Oct 08 18:17:10 crc kubenswrapper[4859]: I1008 18:17:10.206868 4859 flags.go:64] FLAG: --v="2" Oct 08 18:17:10 crc kubenswrapper[4859]: I1008 18:17:10.206883 4859 flags.go:64] FLAG: --version="false" Oct 08 18:17:10 crc kubenswrapper[4859]: I1008 18:17:10.206898 4859 flags.go:64] FLAG: --vmodule="" Oct 08 18:17:10 crc kubenswrapper[4859]: I1008 18:17:10.206911 4859 flags.go:64] FLAG: --volume-plugin-dir="/etc/kubernetes/kubelet-plugins/volume/exec" Oct 08 18:17:10 crc kubenswrapper[4859]: I1008 18:17:10.206923 4859 flags.go:64] FLAG: --volume-stats-agg-period="1m0s" Oct 08 18:17:10 crc kubenswrapper[4859]: W1008 18:17:10.207244 4859 feature_gate.go:330] unrecognized feature gate: BareMetalLoadBalancer Oct 08 18:17:10 crc kubenswrapper[4859]: W1008 18:17:10.207262 4859 feature_gate.go:330] unrecognized feature gate: MachineConfigNodes Oct 08 18:17:10 crc kubenswrapper[4859]: W1008 18:17:10.207274 4859 feature_gate.go:330] unrecognized feature gate: SigstoreImageVerification Oct 08 18:17:10 crc kubenswrapper[4859]: W1008 18:17:10.207286 4859 feature_gate.go:330] unrecognized feature gate: InsightsOnDemandDataGather Oct 08 18:17:10 crc kubenswrapper[4859]: W1008 18:17:10.207296 4859 feature_gate.go:330] unrecognized feature gate: OVNObservability Oct 08 18:17:10 crc kubenswrapper[4859]: W1008 18:17:10.207305 4859 feature_gate.go:330] unrecognized feature gate: GCPClusterHostedDNS Oct 08 18:17:10 crc kubenswrapper[4859]: W1008 18:17:10.207318 4859 feature_gate.go:330] unrecognized feature gate: PersistentIPsForVirtualization Oct 08 18:17:10 crc kubenswrapper[4859]: W1008 18:17:10.207328 4859 feature_gate.go:330] unrecognized feature gate: InsightsRuntimeExtractor Oct 08 18:17:10 crc kubenswrapper[4859]: W1008 18:17:10.207338 4859 feature_gate.go:330] unrecognized feature gate: MinimumKubeletVersion Oct 08 18:17:10 crc kubenswrapper[4859]: W1008 18:17:10.207348 4859 feature_gate.go:330] unrecognized feature gate: AWSClusterHostedDNS Oct 08 18:17:10 crc kubenswrapper[4859]: W1008 18:17:10.207359 4859 feature_gate.go:330] unrecognized feature gate: VSphereStaticIPs Oct 08 18:17:10 crc kubenswrapper[4859]: W1008 18:17:10.207370 4859 feature_gate.go:330] unrecognized feature gate: OpenShiftPodSecurityAdmission Oct 08 18:17:10 crc kubenswrapper[4859]: W1008 18:17:10.207380 4859 feature_gate.go:330] unrecognized feature gate: MachineAPIOperatorDisableMachineHealthCheckController Oct 08 18:17:10 crc kubenswrapper[4859]: W1008 18:17:10.207390 4859 feature_gate.go:330] unrecognized feature gate: ChunkSizeMiB Oct 08 18:17:10 crc kubenswrapper[4859]: W1008 18:17:10.207401 4859 feature_gate.go:330] unrecognized feature gate: AWSEFSDriverVolumeMetrics Oct 08 18:17:10 crc kubenswrapper[4859]: W1008 18:17:10.207411 4859 feature_gate.go:330] unrecognized feature gate: VSphereMultiVCenters Oct 08 18:17:10 crc kubenswrapper[4859]: W1008 18:17:10.207426 4859 feature_gate.go:353] Setting GA feature gate ValidatingAdmissionPolicy=true. It will be removed in a future release. Oct 08 18:17:10 crc kubenswrapper[4859]: W1008 18:17:10.207438 4859 feature_gate.go:351] Setting deprecated feature gate KMSv1=true. It will be removed in a future release. Oct 08 18:17:10 crc kubenswrapper[4859]: W1008 18:17:10.207451 4859 feature_gate.go:353] Setting GA feature gate CloudDualStackNodeIPs=true. It will be removed in a future release. Oct 08 18:17:10 crc kubenswrapper[4859]: W1008 18:17:10.207466 4859 feature_gate.go:330] unrecognized feature gate: NewOLM Oct 08 18:17:10 crc kubenswrapper[4859]: W1008 18:17:10.207478 4859 feature_gate.go:330] unrecognized feature gate: GatewayAPI Oct 08 18:17:10 crc kubenswrapper[4859]: W1008 18:17:10.207489 4859 feature_gate.go:330] unrecognized feature gate: HardwareSpeed Oct 08 18:17:10 crc kubenswrapper[4859]: W1008 18:17:10.207500 4859 feature_gate.go:330] unrecognized feature gate: ManagedBootImagesAWS Oct 08 18:17:10 crc kubenswrapper[4859]: W1008 18:17:10.207510 4859 feature_gate.go:330] unrecognized feature gate: NodeDisruptionPolicy Oct 08 18:17:10 crc kubenswrapper[4859]: W1008 18:17:10.207521 4859 feature_gate.go:330] unrecognized feature gate: MultiArchInstallAWS Oct 08 18:17:10 crc kubenswrapper[4859]: W1008 18:17:10.207534 4859 feature_gate.go:330] unrecognized feature gate: InsightsConfig Oct 08 18:17:10 crc kubenswrapper[4859]: W1008 18:17:10.207545 4859 feature_gate.go:330] unrecognized feature gate: IngressControllerLBSubnetsAWS Oct 08 18:17:10 crc kubenswrapper[4859]: W1008 18:17:10.207555 4859 feature_gate.go:330] unrecognized feature gate: SignatureStores Oct 08 18:17:10 crc kubenswrapper[4859]: W1008 18:17:10.207565 4859 feature_gate.go:330] unrecognized feature gate: AlibabaPlatform Oct 08 18:17:10 crc kubenswrapper[4859]: W1008 18:17:10.207575 4859 feature_gate.go:330] unrecognized feature gate: Example Oct 08 18:17:10 crc kubenswrapper[4859]: W1008 18:17:10.207585 4859 feature_gate.go:330] unrecognized feature gate: MultiArchInstallGCP Oct 08 18:17:10 crc kubenswrapper[4859]: W1008 18:17:10.207595 4859 feature_gate.go:330] unrecognized feature gate: PrivateHostedZoneAWS Oct 08 18:17:10 crc kubenswrapper[4859]: W1008 18:17:10.207605 4859 feature_gate.go:330] unrecognized feature gate: VolumeGroupSnapshot Oct 08 18:17:10 crc kubenswrapper[4859]: W1008 18:17:10.207615 4859 feature_gate.go:330] unrecognized feature gate: OnClusterBuild Oct 08 18:17:10 crc kubenswrapper[4859]: W1008 18:17:10.207625 4859 feature_gate.go:330] unrecognized feature gate: BootcNodeManagement Oct 08 18:17:10 crc kubenswrapper[4859]: W1008 18:17:10.207634 4859 feature_gate.go:330] unrecognized feature gate: ClusterAPIInstall Oct 08 18:17:10 crc kubenswrapper[4859]: W1008 18:17:10.207645 4859 feature_gate.go:330] unrecognized feature gate: ConsolePluginContentSecurityPolicy Oct 08 18:17:10 crc kubenswrapper[4859]: W1008 18:17:10.207654 4859 feature_gate.go:330] unrecognized feature gate: PinnedImages Oct 08 18:17:10 crc kubenswrapper[4859]: W1008 18:17:10.207664 4859 feature_gate.go:330] unrecognized feature gate: EtcdBackendQuota Oct 08 18:17:10 crc kubenswrapper[4859]: W1008 18:17:10.207674 4859 feature_gate.go:330] unrecognized feature gate: AzureWorkloadIdentity Oct 08 18:17:10 crc kubenswrapper[4859]: W1008 18:17:10.207684 4859 feature_gate.go:330] unrecognized feature gate: NetworkDiagnosticsConfig Oct 08 18:17:10 crc kubenswrapper[4859]: W1008 18:17:10.207701 4859 feature_gate.go:330] unrecognized feature gate: NetworkLiveMigration Oct 08 18:17:10 crc kubenswrapper[4859]: W1008 18:17:10.207751 4859 feature_gate.go:330] unrecognized feature gate: AutomatedEtcdBackup Oct 08 18:17:10 crc kubenswrapper[4859]: W1008 18:17:10.207763 4859 feature_gate.go:330] unrecognized feature gate: AdditionalRoutingCapabilities Oct 08 18:17:10 crc kubenswrapper[4859]: W1008 18:17:10.207773 4859 feature_gate.go:330] unrecognized feature gate: MachineAPIMigration Oct 08 18:17:10 crc kubenswrapper[4859]: W1008 18:17:10.207783 4859 feature_gate.go:330] unrecognized feature gate: VSphereMultiNetworks Oct 08 18:17:10 crc kubenswrapper[4859]: W1008 18:17:10.207809 4859 feature_gate.go:353] Setting GA feature gate DisableKubeletCloudCredentialProviders=true. It will be removed in a future release. Oct 08 18:17:10 crc kubenswrapper[4859]: W1008 18:17:10.207826 4859 feature_gate.go:330] unrecognized feature gate: MixedCPUsAllocation Oct 08 18:17:10 crc kubenswrapper[4859]: W1008 18:17:10.207841 4859 feature_gate.go:330] unrecognized feature gate: SetEIPForNLBIngressController Oct 08 18:17:10 crc kubenswrapper[4859]: W1008 18:17:10.207851 4859 feature_gate.go:330] unrecognized feature gate: DNSNameResolver Oct 08 18:17:10 crc kubenswrapper[4859]: W1008 18:17:10.207861 4859 feature_gate.go:330] unrecognized feature gate: IngressControllerDynamicConfigurationManager Oct 08 18:17:10 crc kubenswrapper[4859]: W1008 18:17:10.207871 4859 feature_gate.go:330] unrecognized feature gate: GCPLabelsTags Oct 08 18:17:10 crc kubenswrapper[4859]: W1008 18:17:10.207882 4859 feature_gate.go:330] unrecognized feature gate: ExternalOIDC Oct 08 18:17:10 crc kubenswrapper[4859]: W1008 18:17:10.207892 4859 feature_gate.go:330] unrecognized feature gate: VSphereDriverConfiguration Oct 08 18:17:10 crc kubenswrapper[4859]: W1008 18:17:10.207902 4859 feature_gate.go:330] unrecognized feature gate: MachineAPIProviderOpenStack Oct 08 18:17:10 crc kubenswrapper[4859]: W1008 18:17:10.207912 4859 feature_gate.go:330] unrecognized feature gate: NetworkSegmentation Oct 08 18:17:10 crc kubenswrapper[4859]: W1008 18:17:10.207922 4859 feature_gate.go:330] unrecognized feature gate: UpgradeStatus Oct 08 18:17:10 crc kubenswrapper[4859]: W1008 18:17:10.207932 4859 feature_gate.go:330] unrecognized feature gate: NutanixMultiSubnets Oct 08 18:17:10 crc kubenswrapper[4859]: W1008 18:17:10.207943 4859 feature_gate.go:330] unrecognized feature gate: ClusterAPIInstallIBMCloud Oct 08 18:17:10 crc kubenswrapper[4859]: W1008 18:17:10.207952 4859 feature_gate.go:330] unrecognized feature gate: VSphereControlPlaneMachineSet Oct 08 18:17:10 crc kubenswrapper[4859]: W1008 18:17:10.207961 4859 feature_gate.go:330] unrecognized feature gate: PlatformOperators Oct 08 18:17:10 crc kubenswrapper[4859]: W1008 18:17:10.207971 4859 feature_gate.go:330] unrecognized feature gate: InsightsConfigAPI Oct 08 18:17:10 crc kubenswrapper[4859]: W1008 18:17:10.207981 4859 feature_gate.go:330] unrecognized feature gate: CSIDriverSharedResource Oct 08 18:17:10 crc kubenswrapper[4859]: W1008 18:17:10.207991 4859 feature_gate.go:330] unrecognized feature gate: ImageStreamImportMode Oct 08 18:17:10 crc kubenswrapper[4859]: W1008 18:17:10.208005 4859 feature_gate.go:330] unrecognized feature gate: MetricsCollectionProfiles Oct 08 18:17:10 crc kubenswrapper[4859]: W1008 18:17:10.208016 4859 feature_gate.go:330] unrecognized feature gate: BuildCSIVolumes Oct 08 18:17:10 crc kubenswrapper[4859]: W1008 18:17:10.208026 4859 feature_gate.go:330] unrecognized feature gate: ManagedBootImages Oct 08 18:17:10 crc kubenswrapper[4859]: W1008 18:17:10.208036 4859 feature_gate.go:330] unrecognized feature gate: MultiArchInstallAzure Oct 08 18:17:10 crc kubenswrapper[4859]: W1008 18:17:10.208046 4859 feature_gate.go:330] unrecognized feature gate: RouteAdvertisements Oct 08 18:17:10 crc kubenswrapper[4859]: W1008 18:17:10.208056 4859 feature_gate.go:330] unrecognized feature gate: AdminNetworkPolicy Oct 08 18:17:10 crc kubenswrapper[4859]: W1008 18:17:10.208066 4859 feature_gate.go:330] unrecognized feature gate: ClusterMonitoringConfig Oct 08 18:17:10 crc kubenswrapper[4859]: I1008 18:17:10.208096 4859 feature_gate.go:386] feature gates: {map[CloudDualStackNodeIPs:true DisableKubeletCloudCredentialProviders:true DynamicResourceAllocation:false EventedPLEG:false KMSv1:true MaxUnavailableStatefulSet:false NodeSwap:false ProcMountType:false RouteExternalCertificate:false ServiceAccountTokenNodeBinding:false TranslateStreamCloseWebsocketRequests:false UserNamespacesPodSecurityStandards:false UserNamespacesSupport:false ValidatingAdmissionPolicy:true VolumeAttributesClass:false]} Oct 08 18:17:10 crc kubenswrapper[4859]: I1008 18:17:10.228967 4859 server.go:491] "Kubelet version" kubeletVersion="v1.31.5" Oct 08 18:17:10 crc kubenswrapper[4859]: I1008 18:17:10.229042 4859 server.go:493] "Golang settings" GOGC="" GOMAXPROCS="" GOTRACEBACK="" Oct 08 18:17:10 crc kubenswrapper[4859]: W1008 18:17:10.229163 4859 feature_gate.go:330] unrecognized feature gate: CSIDriverSharedResource Oct 08 18:17:10 crc kubenswrapper[4859]: W1008 18:17:10.229178 4859 feature_gate.go:330] unrecognized feature gate: UpgradeStatus Oct 08 18:17:10 crc kubenswrapper[4859]: W1008 18:17:10.229188 4859 feature_gate.go:351] Setting deprecated feature gate KMSv1=true. It will be removed in a future release. Oct 08 18:17:10 crc kubenswrapper[4859]: W1008 18:17:10.229198 4859 feature_gate.go:330] unrecognized feature gate: Example Oct 08 18:17:10 crc kubenswrapper[4859]: W1008 18:17:10.229207 4859 feature_gate.go:330] unrecognized feature gate: MultiArchInstallGCP Oct 08 18:17:10 crc kubenswrapper[4859]: W1008 18:17:10.229214 4859 feature_gate.go:330] unrecognized feature gate: BareMetalLoadBalancer Oct 08 18:17:10 crc kubenswrapper[4859]: W1008 18:17:10.229224 4859 feature_gate.go:353] Setting GA feature gate DisableKubeletCloudCredentialProviders=true. It will be removed in a future release. Oct 08 18:17:10 crc kubenswrapper[4859]: W1008 18:17:10.229235 4859 feature_gate.go:353] Setting GA feature gate CloudDualStackNodeIPs=true. It will be removed in a future release. Oct 08 18:17:10 crc kubenswrapper[4859]: W1008 18:17:10.229243 4859 feature_gate.go:330] unrecognized feature gate: ClusterMonitoringConfig Oct 08 18:17:10 crc kubenswrapper[4859]: W1008 18:17:10.229250 4859 feature_gate.go:330] unrecognized feature gate: MachineConfigNodes Oct 08 18:17:10 crc kubenswrapper[4859]: W1008 18:17:10.229258 4859 feature_gate.go:330] unrecognized feature gate: MixedCPUsAllocation Oct 08 18:17:10 crc kubenswrapper[4859]: W1008 18:17:10.229265 4859 feature_gate.go:330] unrecognized feature gate: AzureWorkloadIdentity Oct 08 18:17:10 crc kubenswrapper[4859]: W1008 18:17:10.229273 4859 feature_gate.go:330] unrecognized feature gate: AWSEFSDriverVolumeMetrics Oct 08 18:17:10 crc kubenswrapper[4859]: W1008 18:17:10.229281 4859 feature_gate.go:330] unrecognized feature gate: PersistentIPsForVirtualization Oct 08 18:17:10 crc kubenswrapper[4859]: W1008 18:17:10.229289 4859 feature_gate.go:330] unrecognized feature gate: AdminNetworkPolicy Oct 08 18:17:10 crc kubenswrapper[4859]: W1008 18:17:10.229296 4859 feature_gate.go:330] unrecognized feature gate: AWSClusterHostedDNS Oct 08 18:17:10 crc kubenswrapper[4859]: W1008 18:17:10.229304 4859 feature_gate.go:330] unrecognized feature gate: SigstoreImageVerification Oct 08 18:17:10 crc kubenswrapper[4859]: W1008 18:17:10.229313 4859 feature_gate.go:330] unrecognized feature gate: MachineAPIOperatorDisableMachineHealthCheckController Oct 08 18:17:10 crc kubenswrapper[4859]: W1008 18:17:10.229320 4859 feature_gate.go:330] unrecognized feature gate: SetEIPForNLBIngressController Oct 08 18:17:10 crc kubenswrapper[4859]: W1008 18:17:10.229327 4859 feature_gate.go:330] unrecognized feature gate: GatewayAPI Oct 08 18:17:10 crc kubenswrapper[4859]: W1008 18:17:10.229337 4859 feature_gate.go:353] Setting GA feature gate ValidatingAdmissionPolicy=true. It will be removed in a future release. Oct 08 18:17:10 crc kubenswrapper[4859]: W1008 18:17:10.229346 4859 feature_gate.go:330] unrecognized feature gate: PinnedImages Oct 08 18:17:10 crc kubenswrapper[4859]: W1008 18:17:10.229353 4859 feature_gate.go:330] unrecognized feature gate: GCPLabelsTags Oct 08 18:17:10 crc kubenswrapper[4859]: W1008 18:17:10.229361 4859 feature_gate.go:330] unrecognized feature gate: InsightsConfig Oct 08 18:17:10 crc kubenswrapper[4859]: W1008 18:17:10.229368 4859 feature_gate.go:330] unrecognized feature gate: NewOLM Oct 08 18:17:10 crc kubenswrapper[4859]: W1008 18:17:10.229375 4859 feature_gate.go:330] unrecognized feature gate: MultiArchInstallAzure Oct 08 18:17:10 crc kubenswrapper[4859]: W1008 18:17:10.229382 4859 feature_gate.go:330] unrecognized feature gate: IngressControllerLBSubnetsAWS Oct 08 18:17:10 crc kubenswrapper[4859]: W1008 18:17:10.229389 4859 feature_gate.go:330] unrecognized feature gate: AlibabaPlatform Oct 08 18:17:10 crc kubenswrapper[4859]: W1008 18:17:10.229396 4859 feature_gate.go:330] unrecognized feature gate: RouteAdvertisements Oct 08 18:17:10 crc kubenswrapper[4859]: W1008 18:17:10.229402 4859 feature_gate.go:330] unrecognized feature gate: OVNObservability Oct 08 18:17:10 crc kubenswrapper[4859]: W1008 18:17:10.229423 4859 feature_gate.go:330] unrecognized feature gate: HardwareSpeed Oct 08 18:17:10 crc kubenswrapper[4859]: W1008 18:17:10.229430 4859 feature_gate.go:330] unrecognized feature gate: ChunkSizeMiB Oct 08 18:17:10 crc kubenswrapper[4859]: W1008 18:17:10.229439 4859 feature_gate.go:330] unrecognized feature gate: NodeDisruptionPolicy Oct 08 18:17:10 crc kubenswrapper[4859]: W1008 18:17:10.229446 4859 feature_gate.go:330] unrecognized feature gate: BuildCSIVolumes Oct 08 18:17:10 crc kubenswrapper[4859]: W1008 18:17:10.229453 4859 feature_gate.go:330] unrecognized feature gate: VSphereDriverConfiguration Oct 08 18:17:10 crc kubenswrapper[4859]: W1008 18:17:10.229461 4859 feature_gate.go:330] unrecognized feature gate: EtcdBackendQuota Oct 08 18:17:10 crc kubenswrapper[4859]: W1008 18:17:10.229468 4859 feature_gate.go:330] unrecognized feature gate: MinimumKubeletVersion Oct 08 18:17:10 crc kubenswrapper[4859]: W1008 18:17:10.229475 4859 feature_gate.go:330] unrecognized feature gate: AdditionalRoutingCapabilities Oct 08 18:17:10 crc kubenswrapper[4859]: W1008 18:17:10.229482 4859 feature_gate.go:330] unrecognized feature gate: VolumeGroupSnapshot Oct 08 18:17:10 crc kubenswrapper[4859]: W1008 18:17:10.229489 4859 feature_gate.go:330] unrecognized feature gate: NutanixMultiSubnets Oct 08 18:17:10 crc kubenswrapper[4859]: W1008 18:17:10.229495 4859 feature_gate.go:330] unrecognized feature gate: ExternalOIDC Oct 08 18:17:10 crc kubenswrapper[4859]: W1008 18:17:10.229501 4859 feature_gate.go:330] unrecognized feature gate: NetworkDiagnosticsConfig Oct 08 18:17:10 crc kubenswrapper[4859]: W1008 18:17:10.229508 4859 feature_gate.go:330] unrecognized feature gate: ManagedBootImagesAWS Oct 08 18:17:10 crc kubenswrapper[4859]: W1008 18:17:10.229514 4859 feature_gate.go:330] unrecognized feature gate: PrivateHostedZoneAWS Oct 08 18:17:10 crc kubenswrapper[4859]: W1008 18:17:10.229520 4859 feature_gate.go:330] unrecognized feature gate: NetworkSegmentation Oct 08 18:17:10 crc kubenswrapper[4859]: W1008 18:17:10.229527 4859 feature_gate.go:330] unrecognized feature gate: OpenShiftPodSecurityAdmission Oct 08 18:17:10 crc kubenswrapper[4859]: W1008 18:17:10.229534 4859 feature_gate.go:330] unrecognized feature gate: BootcNodeManagement Oct 08 18:17:10 crc kubenswrapper[4859]: W1008 18:17:10.229542 4859 feature_gate.go:330] unrecognized feature gate: VSphereMultiVCenters Oct 08 18:17:10 crc kubenswrapper[4859]: W1008 18:17:10.229550 4859 feature_gate.go:330] unrecognized feature gate: IngressControllerDynamicConfigurationManager Oct 08 18:17:10 crc kubenswrapper[4859]: W1008 18:17:10.229557 4859 feature_gate.go:330] unrecognized feature gate: InsightsConfigAPI Oct 08 18:17:10 crc kubenswrapper[4859]: W1008 18:17:10.229565 4859 feature_gate.go:330] unrecognized feature gate: InsightsOnDemandDataGather Oct 08 18:17:10 crc kubenswrapper[4859]: W1008 18:17:10.229571 4859 feature_gate.go:330] unrecognized feature gate: ConsolePluginContentSecurityPolicy Oct 08 18:17:10 crc kubenswrapper[4859]: W1008 18:17:10.229577 4859 feature_gate.go:330] unrecognized feature gate: ManagedBootImages Oct 08 18:17:10 crc kubenswrapper[4859]: W1008 18:17:10.229585 4859 feature_gate.go:330] unrecognized feature gate: AutomatedEtcdBackup Oct 08 18:17:10 crc kubenswrapper[4859]: W1008 18:17:10.229591 4859 feature_gate.go:330] unrecognized feature gate: ClusterAPIInstall Oct 08 18:17:10 crc kubenswrapper[4859]: W1008 18:17:10.229597 4859 feature_gate.go:330] unrecognized feature gate: MultiArchInstallAWS Oct 08 18:17:10 crc kubenswrapper[4859]: W1008 18:17:10.229602 4859 feature_gate.go:330] unrecognized feature gate: DNSNameResolver Oct 08 18:17:10 crc kubenswrapper[4859]: W1008 18:17:10.229608 4859 feature_gate.go:330] unrecognized feature gate: MetricsCollectionProfiles Oct 08 18:17:10 crc kubenswrapper[4859]: W1008 18:17:10.229614 4859 feature_gate.go:330] unrecognized feature gate: MachineAPIProviderOpenStack Oct 08 18:17:10 crc kubenswrapper[4859]: W1008 18:17:10.229621 4859 feature_gate.go:330] unrecognized feature gate: InsightsRuntimeExtractor Oct 08 18:17:10 crc kubenswrapper[4859]: W1008 18:17:10.229626 4859 feature_gate.go:330] unrecognized feature gate: ClusterAPIInstallIBMCloud Oct 08 18:17:10 crc kubenswrapper[4859]: W1008 18:17:10.229632 4859 feature_gate.go:330] unrecognized feature gate: PlatformOperators Oct 08 18:17:10 crc kubenswrapper[4859]: W1008 18:17:10.229638 4859 feature_gate.go:330] unrecognized feature gate: VSphereStaticIPs Oct 08 18:17:10 crc kubenswrapper[4859]: W1008 18:17:10.229644 4859 feature_gate.go:330] unrecognized feature gate: SignatureStores Oct 08 18:17:10 crc kubenswrapper[4859]: W1008 18:17:10.229650 4859 feature_gate.go:330] unrecognized feature gate: ImageStreamImportMode Oct 08 18:17:10 crc kubenswrapper[4859]: W1008 18:17:10.229656 4859 feature_gate.go:330] unrecognized feature gate: VSphereControlPlaneMachineSet Oct 08 18:17:10 crc kubenswrapper[4859]: W1008 18:17:10.229662 4859 feature_gate.go:330] unrecognized feature gate: GCPClusterHostedDNS Oct 08 18:17:10 crc kubenswrapper[4859]: W1008 18:17:10.229669 4859 feature_gate.go:330] unrecognized feature gate: NetworkLiveMigration Oct 08 18:17:10 crc kubenswrapper[4859]: W1008 18:17:10.229675 4859 feature_gate.go:330] unrecognized feature gate: OnClusterBuild Oct 08 18:17:10 crc kubenswrapper[4859]: W1008 18:17:10.229680 4859 feature_gate.go:330] unrecognized feature gate: VSphereMultiNetworks Oct 08 18:17:10 crc kubenswrapper[4859]: W1008 18:17:10.229691 4859 feature_gate.go:330] unrecognized feature gate: MachineAPIMigration Oct 08 18:17:10 crc kubenswrapper[4859]: I1008 18:17:10.229703 4859 feature_gate.go:386] feature gates: {map[CloudDualStackNodeIPs:true DisableKubeletCloudCredentialProviders:true DynamicResourceAllocation:false EventedPLEG:false KMSv1:true MaxUnavailableStatefulSet:false NodeSwap:false ProcMountType:false RouteExternalCertificate:false ServiceAccountTokenNodeBinding:false TranslateStreamCloseWebsocketRequests:false UserNamespacesPodSecurityStandards:false UserNamespacesSupport:false ValidatingAdmissionPolicy:true VolumeAttributesClass:false]} Oct 08 18:17:10 crc kubenswrapper[4859]: W1008 18:17:10.229970 4859 feature_gate.go:330] unrecognized feature gate: ClusterMonitoringConfig Oct 08 18:17:10 crc kubenswrapper[4859]: W1008 18:17:10.229985 4859 feature_gate.go:330] unrecognized feature gate: InsightsRuntimeExtractor Oct 08 18:17:10 crc kubenswrapper[4859]: W1008 18:17:10.229991 4859 feature_gate.go:330] unrecognized feature gate: ImageStreamImportMode Oct 08 18:17:10 crc kubenswrapper[4859]: W1008 18:17:10.229998 4859 feature_gate.go:330] unrecognized feature gate: SetEIPForNLBIngressController Oct 08 18:17:10 crc kubenswrapper[4859]: W1008 18:17:10.230007 4859 feature_gate.go:330] unrecognized feature gate: ClusterAPIInstallIBMCloud Oct 08 18:17:10 crc kubenswrapper[4859]: W1008 18:17:10.230013 4859 feature_gate.go:330] unrecognized feature gate: ManagedBootImages Oct 08 18:17:10 crc kubenswrapper[4859]: W1008 18:17:10.230020 4859 feature_gate.go:330] unrecognized feature gate: NewOLM Oct 08 18:17:10 crc kubenswrapper[4859]: W1008 18:17:10.230027 4859 feature_gate.go:330] unrecognized feature gate: AWSEFSDriverVolumeMetrics Oct 08 18:17:10 crc kubenswrapper[4859]: W1008 18:17:10.230033 4859 feature_gate.go:330] unrecognized feature gate: MultiArchInstallAzure Oct 08 18:17:10 crc kubenswrapper[4859]: W1008 18:17:10.230039 4859 feature_gate.go:330] unrecognized feature gate: AzureWorkloadIdentity Oct 08 18:17:10 crc kubenswrapper[4859]: W1008 18:17:10.230047 4859 feature_gate.go:330] unrecognized feature gate: NetworkLiveMigration Oct 08 18:17:10 crc kubenswrapper[4859]: W1008 18:17:10.230053 4859 feature_gate.go:330] unrecognized feature gate: AlibabaPlatform Oct 08 18:17:10 crc kubenswrapper[4859]: W1008 18:17:10.230061 4859 feature_gate.go:351] Setting deprecated feature gate KMSv1=true. It will be removed in a future release. Oct 08 18:17:10 crc kubenswrapper[4859]: W1008 18:17:10.230071 4859 feature_gate.go:330] unrecognized feature gate: MinimumKubeletVersion Oct 08 18:17:10 crc kubenswrapper[4859]: W1008 18:17:10.230080 4859 feature_gate.go:330] unrecognized feature gate: BootcNodeManagement Oct 08 18:17:10 crc kubenswrapper[4859]: W1008 18:17:10.230086 4859 feature_gate.go:330] unrecognized feature gate: MachineAPIOperatorDisableMachineHealthCheckController Oct 08 18:17:10 crc kubenswrapper[4859]: W1008 18:17:10.230093 4859 feature_gate.go:330] unrecognized feature gate: SigstoreImageVerification Oct 08 18:17:10 crc kubenswrapper[4859]: W1008 18:17:10.230100 4859 feature_gate.go:330] unrecognized feature gate: MixedCPUsAllocation Oct 08 18:17:10 crc kubenswrapper[4859]: W1008 18:17:10.230105 4859 feature_gate.go:330] unrecognized feature gate: OpenShiftPodSecurityAdmission Oct 08 18:17:10 crc kubenswrapper[4859]: W1008 18:17:10.230112 4859 feature_gate.go:330] unrecognized feature gate: GatewayAPI Oct 08 18:17:10 crc kubenswrapper[4859]: W1008 18:17:10.230118 4859 feature_gate.go:330] unrecognized feature gate: InsightsConfigAPI Oct 08 18:17:10 crc kubenswrapper[4859]: W1008 18:17:10.230124 4859 feature_gate.go:330] unrecognized feature gate: AutomatedEtcdBackup Oct 08 18:17:10 crc kubenswrapper[4859]: W1008 18:17:10.230130 4859 feature_gate.go:330] unrecognized feature gate: VSphereMultiVCenters Oct 08 18:17:10 crc kubenswrapper[4859]: W1008 18:17:10.230136 4859 feature_gate.go:330] unrecognized feature gate: EtcdBackendQuota Oct 08 18:17:10 crc kubenswrapper[4859]: W1008 18:17:10.230142 4859 feature_gate.go:330] unrecognized feature gate: MultiArchInstallAWS Oct 08 18:17:10 crc kubenswrapper[4859]: W1008 18:17:10.230150 4859 feature_gate.go:353] Setting GA feature gate ValidatingAdmissionPolicy=true. It will be removed in a future release. Oct 08 18:17:10 crc kubenswrapper[4859]: W1008 18:17:10.230156 4859 feature_gate.go:330] unrecognized feature gate: HardwareSpeed Oct 08 18:17:10 crc kubenswrapper[4859]: W1008 18:17:10.230163 4859 feature_gate.go:330] unrecognized feature gate: ChunkSizeMiB Oct 08 18:17:10 crc kubenswrapper[4859]: W1008 18:17:10.230171 4859 feature_gate.go:330] unrecognized feature gate: PinnedImages Oct 08 18:17:10 crc kubenswrapper[4859]: W1008 18:17:10.230177 4859 feature_gate.go:330] unrecognized feature gate: NutanixMultiSubnets Oct 08 18:17:10 crc kubenswrapper[4859]: W1008 18:17:10.230183 4859 feature_gate.go:330] unrecognized feature gate: PrivateHostedZoneAWS Oct 08 18:17:10 crc kubenswrapper[4859]: W1008 18:17:10.230189 4859 feature_gate.go:330] unrecognized feature gate: InsightsOnDemandDataGather Oct 08 18:17:10 crc kubenswrapper[4859]: W1008 18:17:10.230196 4859 feature_gate.go:330] unrecognized feature gate: BareMetalLoadBalancer Oct 08 18:17:10 crc kubenswrapper[4859]: W1008 18:17:10.230203 4859 feature_gate.go:330] unrecognized feature gate: PlatformOperators Oct 08 18:17:10 crc kubenswrapper[4859]: W1008 18:17:10.230209 4859 feature_gate.go:330] unrecognized feature gate: VSphereStaticIPs Oct 08 18:17:10 crc kubenswrapper[4859]: W1008 18:17:10.230216 4859 feature_gate.go:330] unrecognized feature gate: MultiArchInstallGCP Oct 08 18:17:10 crc kubenswrapper[4859]: W1008 18:17:10.230222 4859 feature_gate.go:330] unrecognized feature gate: AWSClusterHostedDNS Oct 08 18:17:10 crc kubenswrapper[4859]: W1008 18:17:10.230233 4859 feature_gate.go:353] Setting GA feature gate CloudDualStackNodeIPs=true. It will be removed in a future release. Oct 08 18:17:10 crc kubenswrapper[4859]: W1008 18:17:10.230241 4859 feature_gate.go:330] unrecognized feature gate: AdminNetworkPolicy Oct 08 18:17:10 crc kubenswrapper[4859]: W1008 18:17:10.230249 4859 feature_gate.go:330] unrecognized feature gate: RouteAdvertisements Oct 08 18:17:10 crc kubenswrapper[4859]: W1008 18:17:10.230257 4859 feature_gate.go:330] unrecognized feature gate: SignatureStores Oct 08 18:17:10 crc kubenswrapper[4859]: W1008 18:17:10.230265 4859 feature_gate.go:330] unrecognized feature gate: VSphereMultiNetworks Oct 08 18:17:10 crc kubenswrapper[4859]: W1008 18:17:10.230272 4859 feature_gate.go:330] unrecognized feature gate: Example Oct 08 18:17:10 crc kubenswrapper[4859]: W1008 18:17:10.230279 4859 feature_gate.go:330] unrecognized feature gate: IngressControllerDynamicConfigurationManager Oct 08 18:17:10 crc kubenswrapper[4859]: W1008 18:17:10.230287 4859 feature_gate.go:330] unrecognized feature gate: MetricsCollectionProfiles Oct 08 18:17:10 crc kubenswrapper[4859]: W1008 18:17:10.230294 4859 feature_gate.go:330] unrecognized feature gate: ManagedBootImagesAWS Oct 08 18:17:10 crc kubenswrapper[4859]: W1008 18:17:10.230302 4859 feature_gate.go:353] Setting GA feature gate DisableKubeletCloudCredentialProviders=true. It will be removed in a future release. Oct 08 18:17:10 crc kubenswrapper[4859]: W1008 18:17:10.230310 4859 feature_gate.go:330] unrecognized feature gate: CSIDriverSharedResource Oct 08 18:17:10 crc kubenswrapper[4859]: W1008 18:17:10.230317 4859 feature_gate.go:330] unrecognized feature gate: PersistentIPsForVirtualization Oct 08 18:17:10 crc kubenswrapper[4859]: W1008 18:17:10.230324 4859 feature_gate.go:330] unrecognized feature gate: NetworkDiagnosticsConfig Oct 08 18:17:10 crc kubenswrapper[4859]: W1008 18:17:10.230331 4859 feature_gate.go:330] unrecognized feature gate: GCPLabelsTags Oct 08 18:17:10 crc kubenswrapper[4859]: W1008 18:17:10.230339 4859 feature_gate.go:330] unrecognized feature gate: BuildCSIVolumes Oct 08 18:17:10 crc kubenswrapper[4859]: W1008 18:17:10.230345 4859 feature_gate.go:330] unrecognized feature gate: DNSNameResolver Oct 08 18:17:10 crc kubenswrapper[4859]: W1008 18:17:10.230350 4859 feature_gate.go:330] unrecognized feature gate: UpgradeStatus Oct 08 18:17:10 crc kubenswrapper[4859]: W1008 18:17:10.230356 4859 feature_gate.go:330] unrecognized feature gate: IngressControllerLBSubnetsAWS Oct 08 18:17:10 crc kubenswrapper[4859]: W1008 18:17:10.230363 4859 feature_gate.go:330] unrecognized feature gate: InsightsConfig Oct 08 18:17:10 crc kubenswrapper[4859]: W1008 18:17:10.230369 4859 feature_gate.go:330] unrecognized feature gate: OVNObservability Oct 08 18:17:10 crc kubenswrapper[4859]: W1008 18:17:10.230375 4859 feature_gate.go:330] unrecognized feature gate: VSphereDriverConfiguration Oct 08 18:17:10 crc kubenswrapper[4859]: W1008 18:17:10.230381 4859 feature_gate.go:330] unrecognized feature gate: AdditionalRoutingCapabilities Oct 08 18:17:10 crc kubenswrapper[4859]: W1008 18:17:10.230388 4859 feature_gate.go:330] unrecognized feature gate: MachineAPIProviderOpenStack Oct 08 18:17:10 crc kubenswrapper[4859]: W1008 18:17:10.230396 4859 feature_gate.go:330] unrecognized feature gate: MachineConfigNodes Oct 08 18:17:10 crc kubenswrapper[4859]: W1008 18:17:10.230403 4859 feature_gate.go:330] unrecognized feature gate: NetworkSegmentation Oct 08 18:17:10 crc kubenswrapper[4859]: W1008 18:17:10.230409 4859 feature_gate.go:330] unrecognized feature gate: VolumeGroupSnapshot Oct 08 18:17:10 crc kubenswrapper[4859]: W1008 18:17:10.230415 4859 feature_gate.go:330] unrecognized feature gate: ConsolePluginContentSecurityPolicy Oct 08 18:17:10 crc kubenswrapper[4859]: W1008 18:17:10.230422 4859 feature_gate.go:330] unrecognized feature gate: ExternalOIDC Oct 08 18:17:10 crc kubenswrapper[4859]: W1008 18:17:10.230428 4859 feature_gate.go:330] unrecognized feature gate: VSphereControlPlaneMachineSet Oct 08 18:17:10 crc kubenswrapper[4859]: W1008 18:17:10.230434 4859 feature_gate.go:330] unrecognized feature gate: GCPClusterHostedDNS Oct 08 18:17:10 crc kubenswrapper[4859]: W1008 18:17:10.230441 4859 feature_gate.go:330] unrecognized feature gate: MachineAPIMigration Oct 08 18:17:10 crc kubenswrapper[4859]: W1008 18:17:10.230447 4859 feature_gate.go:330] unrecognized feature gate: NodeDisruptionPolicy Oct 08 18:17:10 crc kubenswrapper[4859]: W1008 18:17:10.230453 4859 feature_gate.go:330] unrecognized feature gate: OnClusterBuild Oct 08 18:17:10 crc kubenswrapper[4859]: W1008 18:17:10.230459 4859 feature_gate.go:330] unrecognized feature gate: ClusterAPIInstall Oct 08 18:17:10 crc kubenswrapper[4859]: I1008 18:17:10.230470 4859 feature_gate.go:386] feature gates: {map[CloudDualStackNodeIPs:true DisableKubeletCloudCredentialProviders:true DynamicResourceAllocation:false EventedPLEG:false KMSv1:true MaxUnavailableStatefulSet:false NodeSwap:false ProcMountType:false RouteExternalCertificate:false ServiceAccountTokenNodeBinding:false TranslateStreamCloseWebsocketRequests:false UserNamespacesPodSecurityStandards:false UserNamespacesSupport:false ValidatingAdmissionPolicy:true VolumeAttributesClass:false]} Oct 08 18:17:10 crc kubenswrapper[4859]: I1008 18:17:10.231733 4859 server.go:940] "Client rotation is on, will bootstrap in background" Oct 08 18:17:10 crc kubenswrapper[4859]: I1008 18:17:10.236942 4859 bootstrap.go:85] "Current kubeconfig file contents are still valid, no bootstrap necessary" Oct 08 18:17:10 crc kubenswrapper[4859]: I1008 18:17:10.237068 4859 certificate_store.go:130] Loading cert/key pair from "/var/lib/kubelet/pki/kubelet-client-current.pem". Oct 08 18:17:10 crc kubenswrapper[4859]: I1008 18:17:10.239292 4859 server.go:997] "Starting client certificate rotation" Oct 08 18:17:10 crc kubenswrapper[4859]: I1008 18:17:10.239325 4859 certificate_manager.go:356] kubernetes.io/kube-apiserver-client-kubelet: Certificate rotation is enabled Oct 08 18:17:10 crc kubenswrapper[4859]: I1008 18:17:10.240361 4859 certificate_manager.go:356] kubernetes.io/kube-apiserver-client-kubelet: Certificate expiration is 2026-02-24 05:52:08 +0000 UTC, rotation deadline is 2026-01-10 12:54:14.981669614 +0000 UTC Oct 08 18:17:10 crc kubenswrapper[4859]: I1008 18:17:10.240503 4859 certificate_manager.go:356] kubernetes.io/kube-apiserver-client-kubelet: Waiting 2250h37m4.741171886s for next certificate rotation Oct 08 18:17:10 crc kubenswrapper[4859]: I1008 18:17:10.269581 4859 dynamic_cafile_content.go:123] "Loaded a new CA Bundle and Verifier" name="client-ca-bundle::/etc/kubernetes/kubelet-ca.crt" Oct 08 18:17:10 crc kubenswrapper[4859]: I1008 18:17:10.272673 4859 dynamic_cafile_content.go:161] "Starting controller" name="client-ca-bundle::/etc/kubernetes/kubelet-ca.crt" Oct 08 18:17:10 crc kubenswrapper[4859]: I1008 18:17:10.290254 4859 log.go:25] "Validated CRI v1 runtime API" Oct 08 18:17:10 crc kubenswrapper[4859]: I1008 18:17:10.329741 4859 log.go:25] "Validated CRI v1 image API" Oct 08 18:17:10 crc kubenswrapper[4859]: I1008 18:17:10.332925 4859 server.go:1437] "Using cgroup driver setting received from the CRI runtime" cgroupDriver="systemd" Oct 08 18:17:10 crc kubenswrapper[4859]: I1008 18:17:10.339046 4859 fs.go:133] Filesystem UUIDs: map[0b076daa-c26a-46d2-b3a6-72a8dbc6e257:/dev/vda4 2025-10-08-16-45-26-00:/dev/sr0 7B77-95E7:/dev/vda2 de0497b0-db1b-465a-b278-03db02455c71:/dev/vda3] Oct 08 18:17:10 crc kubenswrapper[4859]: I1008 18:17:10.339122 4859 fs.go:134] Filesystem partitions: map[/dev/shm:{mountpoint:/dev/shm major:0 minor:22 fsType:tmpfs blockSize:0} /dev/vda3:{mountpoint:/boot major:252 minor:3 fsType:ext4 blockSize:0} /dev/vda4:{mountpoint:/var major:252 minor:4 fsType:xfs blockSize:0} /run:{mountpoint:/run major:0 minor:24 fsType:tmpfs blockSize:0} /run/user/1000:{mountpoint:/run/user/1000 major:0 minor:42 fsType:tmpfs blockSize:0} /tmp:{mountpoint:/tmp major:0 minor:30 fsType:tmpfs blockSize:0} /var/lib/etcd:{mountpoint:/var/lib/etcd major:0 minor:43 fsType:tmpfs blockSize:0}] Oct 08 18:17:10 crc kubenswrapper[4859]: I1008 18:17:10.369700 4859 manager.go:217] Machine: {Timestamp:2025-10-08 18:17:10.367290523 +0000 UTC m=+0.614129932 CPUVendorID:AuthenticAMD NumCores:12 NumPhysicalCores:1 NumSockets:12 CpuFrequency:2800000 MemoryCapacity:33654120448 SwapCapacity:0 MemoryByType:map[] NVMInfo:{MemoryModeCapacity:0 AppDirectModeCapacity:0 AvgPowerBudget:0} HugePages:[{PageSize:1048576 NumPages:0} {PageSize:2048 NumPages:0}] MachineID:21801e6708c44f15b81395eb736a7cec SystemUUID:8f73c7ed-44aa-4d16-bae1-d4e684803cfc BootID:edb3ee6d-476b-407c-bb7d-4480786e8777 Filesystems:[{Device:/tmp DeviceMajor:0 DeviceMinor:30 Capacity:16827060224 Type:vfs Inodes:1048576 HasInodes:true} {Device:/dev/vda3 DeviceMajor:252 DeviceMinor:3 Capacity:366869504 Type:vfs Inodes:98304 HasInodes:true} {Device:/run/user/1000 DeviceMajor:0 DeviceMinor:42 Capacity:3365408768 Type:vfs Inodes:821633 HasInodes:true} {Device:/var/lib/etcd DeviceMajor:0 DeviceMinor:43 Capacity:1073741824 Type:vfs Inodes:4108169 HasInodes:true} {Device:/dev/shm DeviceMajor:0 DeviceMinor:22 Capacity:16827060224 Type:vfs Inodes:4108169 HasInodes:true} {Device:/run DeviceMajor:0 DeviceMinor:24 Capacity:6730825728 Type:vfs Inodes:819200 HasInodes:true} {Device:/dev/vda4 DeviceMajor:252 DeviceMinor:4 Capacity:85292941312 Type:vfs Inodes:41679680 HasInodes:true}] DiskMap:map[252:0:{Name:vda Major:252 Minor:0 Size:214748364800 Scheduler:none}] NetworkDevices:[{Name:br-ex MacAddress:fa:16:3e:c2:13:e5 Speed:0 Mtu:1500} {Name:br-int MacAddress:d6:39:55:2e:22:71 Speed:0 Mtu:1400} {Name:ens3 MacAddress:fa:16:3e:c2:13:e5 Speed:-1 Mtu:1500} {Name:ens7 MacAddress:fa:16:3e:fe:a2:b5 Speed:-1 Mtu:1500} {Name:ens7.20 MacAddress:52:54:00:d1:0f:95 Speed:-1 Mtu:1496} {Name:ens7.21 MacAddress:52:54:00:ae:7c:08 Speed:-1 Mtu:1496} {Name:ens7.22 MacAddress:52:54:00:7b:47:f6 Speed:-1 Mtu:1496} {Name:eth10 MacAddress:e2:fb:d0:8d:84:f4 Speed:0 Mtu:1500} {Name:ovn-k8s-mp0 MacAddress:0a:58:0a:d9:00:02 Speed:0 Mtu:1400} {Name:ovs-system MacAddress:9a:fd:ba:37:5a:2f Speed:0 Mtu:1500}] Topology:[{Id:0 Memory:33654120448 HugePages:[{PageSize:1048576 NumPages:0} {PageSize:2048 NumPages:0}] Cores:[{Id:0 Threads:[0] Caches:[{Id:0 Size:32768 Type:Data Level:1} {Id:0 Size:32768 Type:Instruction Level:1} {Id:0 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:0 Size:16777216 Type:Unified Level:3}] SocketID:0 BookID: DrawerID:} {Id:0 Threads:[1] Caches:[{Id:1 Size:32768 Type:Data Level:1} {Id:1 Size:32768 Type:Instruction Level:1} {Id:1 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:1 Size:16777216 Type:Unified Level:3}] SocketID:1 BookID: DrawerID:} {Id:0 Threads:[10] Caches:[{Id:10 Size:32768 Type:Data Level:1} {Id:10 Size:32768 Type:Instruction Level:1} {Id:10 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:10 Size:16777216 Type:Unified Level:3}] SocketID:10 BookID: DrawerID:} {Id:0 Threads:[11] Caches:[{Id:11 Size:32768 Type:Data Level:1} {Id:11 Size:32768 Type:Instruction Level:1} {Id:11 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:11 Size:16777216 Type:Unified Level:3}] SocketID:11 BookID: DrawerID:} {Id:0 Threads:[2] Caches:[{Id:2 Size:32768 Type:Data Level:1} {Id:2 Size:32768 Type:Instruction Level:1} {Id:2 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:2 Size:16777216 Type:Unified Level:3}] SocketID:2 BookID: DrawerID:} {Id:0 Threads:[3] Caches:[{Id:3 Size:32768 Type:Data Level:1} {Id:3 Size:32768 Type:Instruction Level:1} {Id:3 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:3 Size:16777216 Type:Unified Level:3}] SocketID:3 BookID: DrawerID:} {Id:0 Threads:[4] Caches:[{Id:4 Size:32768 Type:Data Level:1} {Id:4 Size:32768 Type:Instruction Level:1} {Id:4 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:4 Size:16777216 Type:Unified Level:3}] SocketID:4 BookID: DrawerID:} {Id:0 Threads:[5] Caches:[{Id:5 Size:32768 Type:Data Level:1} {Id:5 Size:32768 Type:Instruction Level:1} {Id:5 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:5 Size:16777216 Type:Unified Level:3}] SocketID:5 BookID: DrawerID:} {Id:0 Threads:[6] Caches:[{Id:6 Size:32768 Type:Data Level:1} {Id:6 Size:32768 Type:Instruction Level:1} {Id:6 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:6 Size:16777216 Type:Unified Level:3}] SocketID:6 BookID: DrawerID:} {Id:0 Threads:[7] Caches:[{Id:7 Size:32768 Type:Data Level:1} {Id:7 Size:32768 Type:Instruction Level:1} {Id:7 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:7 Size:16777216 Type:Unified Level:3}] SocketID:7 BookID: DrawerID:} {Id:0 Threads:[8] Caches:[{Id:8 Size:32768 Type:Data Level:1} {Id:8 Size:32768 Type:Instruction Level:1} {Id:8 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:8 Size:16777216 Type:Unified Level:3}] SocketID:8 BookID: DrawerID:} {Id:0 Threads:[9] Caches:[{Id:9 Size:32768 Type:Data Level:1} {Id:9 Size:32768 Type:Instruction Level:1} {Id:9 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:9 Size:16777216 Type:Unified Level:3}] SocketID:9 BookID: DrawerID:}] Caches:[] Distances:[10]}] CloudProvider:Unknown InstanceType:Unknown InstanceID:None} Oct 08 18:17:10 crc kubenswrapper[4859]: I1008 18:17:10.370040 4859 manager_no_libpfm.go:29] cAdvisor is build without cgo and/or libpfm support. Perf event counters are not available. Oct 08 18:17:10 crc kubenswrapper[4859]: I1008 18:17:10.370242 4859 manager.go:233] Version: {KernelVersion:5.14.0-427.50.2.el9_4.x86_64 ContainerOsVersion:Red Hat Enterprise Linux CoreOS 418.94.202502100215-0 DockerVersion: DockerAPIVersion: CadvisorVersion: CadvisorRevision:} Oct 08 18:17:10 crc kubenswrapper[4859]: I1008 18:17:10.370965 4859 swap_util.go:113] "Swap is on" /proc/swaps contents="Filename\t\t\t\tType\t\tSize\t\tUsed\t\tPriority" Oct 08 18:17:10 crc kubenswrapper[4859]: I1008 18:17:10.371344 4859 container_manager_linux.go:267] "Container manager verified user specified cgroup-root exists" cgroupRoot=[] Oct 08 18:17:10 crc kubenswrapper[4859]: I1008 18:17:10.371416 4859 container_manager_linux.go:272] "Creating Container Manager object based on Node Config" nodeConfig={"NodeName":"crc","RuntimeCgroupsName":"/system.slice/crio.service","SystemCgroupsName":"/system.slice","KubeletCgroupsName":"","KubeletOOMScoreAdj":-999,"ContainerRuntime":"","CgroupsPerQOS":true,"CgroupRoot":"/","CgroupDriver":"systemd","KubeletRootDir":"/var/lib/kubelet","ProtectKernelDefaults":true,"KubeReservedCgroupName":"","SystemReservedCgroupName":"","ReservedSystemCPUs":{},"EnforceNodeAllocatable":{"pods":{}},"KubeReserved":null,"SystemReserved":{"cpu":"200m","ephemeral-storage":"350Mi","memory":"350Mi"},"HardEvictionThresholds":[{"Signal":"memory.available","Operator":"LessThan","Value":{"Quantity":"100Mi","Percentage":0},"GracePeriod":0,"MinReclaim":null},{"Signal":"nodefs.available","Operator":"LessThan","Value":{"Quantity":null,"Percentage":0.1},"GracePeriod":0,"MinReclaim":null},{"Signal":"nodefs.inodesFree","Operator":"LessThan","Value":{"Quantity":null,"Percentage":0.05},"GracePeriod":0,"MinReclaim":null},{"Signal":"imagefs.available","Operator":"LessThan","Value":{"Quantity":null,"Percentage":0.15},"GracePeriod":0,"MinReclaim":null},{"Signal":"imagefs.inodesFree","Operator":"LessThan","Value":{"Quantity":null,"Percentage":0.05},"GracePeriod":0,"MinReclaim":null}],"QOSReserved":{},"CPUManagerPolicy":"none","CPUManagerPolicyOptions":null,"TopologyManagerScope":"container","CPUManagerReconcilePeriod":10000000000,"ExperimentalMemoryManagerPolicy":"None","ExperimentalMemoryManagerReservedMemory":null,"PodPidsLimit":4096,"EnforceCPULimits":true,"CPUCFSQuotaPeriod":100000000,"TopologyManagerPolicy":"none","TopologyManagerPolicyOptions":null,"CgroupVersion":2} Oct 08 18:17:10 crc kubenswrapper[4859]: I1008 18:17:10.371931 4859 topology_manager.go:138] "Creating topology manager with none policy" Oct 08 18:17:10 crc kubenswrapper[4859]: I1008 18:17:10.371954 4859 container_manager_linux.go:303] "Creating device plugin manager" Oct 08 18:17:10 crc kubenswrapper[4859]: I1008 18:17:10.372665 4859 manager.go:142] "Creating Device Plugin manager" path="/var/lib/kubelet/device-plugins/kubelet.sock" Oct 08 18:17:10 crc kubenswrapper[4859]: I1008 18:17:10.372750 4859 server.go:66] "Creating device plugin registration server" version="v1beta1" socket="/var/lib/kubelet/device-plugins/kubelet.sock" Oct 08 18:17:10 crc kubenswrapper[4859]: I1008 18:17:10.373120 4859 state_mem.go:36] "Initialized new in-memory state store" Oct 08 18:17:10 crc kubenswrapper[4859]: I1008 18:17:10.373887 4859 server.go:1245] "Using root directory" path="/var/lib/kubelet" Oct 08 18:17:10 crc kubenswrapper[4859]: I1008 18:17:10.378261 4859 kubelet.go:418] "Attempting to sync node with API server" Oct 08 18:17:10 crc kubenswrapper[4859]: I1008 18:17:10.378302 4859 kubelet.go:313] "Adding static pod path" path="/etc/kubernetes/manifests" Oct 08 18:17:10 crc kubenswrapper[4859]: I1008 18:17:10.378350 4859 file.go:69] "Watching path" path="/etc/kubernetes/manifests" Oct 08 18:17:10 crc kubenswrapper[4859]: I1008 18:17:10.378374 4859 kubelet.go:324] "Adding apiserver pod source" Oct 08 18:17:10 crc kubenswrapper[4859]: I1008 18:17:10.378393 4859 apiserver.go:42] "Waiting for node sync before watching apiserver pods" Oct 08 18:17:10 crc kubenswrapper[4859]: I1008 18:17:10.384062 4859 kuberuntime_manager.go:262] "Container runtime initialized" containerRuntime="cri-o" version="1.31.5-4.rhaos4.18.gitdad78d5.el9" apiVersion="v1" Oct 08 18:17:10 crc kubenswrapper[4859]: I1008 18:17:10.385588 4859 certificate_store.go:130] Loading cert/key pair from "/var/lib/kubelet/pki/kubelet-server-current.pem". Oct 08 18:17:10 crc kubenswrapper[4859]: W1008 18:17:10.386756 4859 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.Node: Get "https://api-int.crc.testing:6443/api/v1/nodes?fieldSelector=metadata.name%3Dcrc&limit=500&resourceVersion=0": dial tcp 38.102.83.13:6443: connect: connection refused Oct 08 18:17:10 crc kubenswrapper[4859]: E1008 18:17:10.386859 4859 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.Node: failed to list *v1.Node: Get \"https://api-int.crc.testing:6443/api/v1/nodes?fieldSelector=metadata.name%3Dcrc&limit=500&resourceVersion=0\": dial tcp 38.102.83.13:6443: connect: connection refused" logger="UnhandledError" Oct 08 18:17:10 crc kubenswrapper[4859]: W1008 18:17:10.386862 4859 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.Service: Get "https://api-int.crc.testing:6443/api/v1/services?fieldSelector=spec.clusterIP%21%3DNone&limit=500&resourceVersion=0": dial tcp 38.102.83.13:6443: connect: connection refused Oct 08 18:17:10 crc kubenswrapper[4859]: E1008 18:17:10.387002 4859 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.Service: failed to list *v1.Service: Get \"https://api-int.crc.testing:6443/api/v1/services?fieldSelector=spec.clusterIP%21%3DNone&limit=500&resourceVersion=0\": dial tcp 38.102.83.13:6443: connect: connection refused" logger="UnhandledError" Oct 08 18:17:10 crc kubenswrapper[4859]: I1008 18:17:10.388121 4859 kubelet.go:854] "Not starting ClusterTrustBundle informer because we are in static kubelet mode" Oct 08 18:17:10 crc kubenswrapper[4859]: I1008 18:17:10.390918 4859 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/portworx-volume" Oct 08 18:17:10 crc kubenswrapper[4859]: I1008 18:17:10.390945 4859 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/empty-dir" Oct 08 18:17:10 crc kubenswrapper[4859]: I1008 18:17:10.390954 4859 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/git-repo" Oct 08 18:17:10 crc kubenswrapper[4859]: I1008 18:17:10.390964 4859 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/host-path" Oct 08 18:17:10 crc kubenswrapper[4859]: I1008 18:17:10.390983 4859 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/nfs" Oct 08 18:17:10 crc kubenswrapper[4859]: I1008 18:17:10.390992 4859 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/secret" Oct 08 18:17:10 crc kubenswrapper[4859]: I1008 18:17:10.391001 4859 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/iscsi" Oct 08 18:17:10 crc kubenswrapper[4859]: I1008 18:17:10.391014 4859 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/downward-api" Oct 08 18:17:10 crc kubenswrapper[4859]: I1008 18:17:10.391026 4859 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/fc" Oct 08 18:17:10 crc kubenswrapper[4859]: I1008 18:17:10.391037 4859 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/configmap" Oct 08 18:17:10 crc kubenswrapper[4859]: I1008 18:17:10.391052 4859 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/projected" Oct 08 18:17:10 crc kubenswrapper[4859]: I1008 18:17:10.391063 4859 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/local-volume" Oct 08 18:17:10 crc kubenswrapper[4859]: I1008 18:17:10.391548 4859 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/csi" Oct 08 18:17:10 crc kubenswrapper[4859]: I1008 18:17:10.392181 4859 server.go:1280] "Started kubelet" Oct 08 18:17:10 crc kubenswrapper[4859]: I1008 18:17:10.392446 4859 ratelimit.go:55] "Setting rate limiting for endpoint" service="podresources" qps=100 burstTokens=10 Oct 08 18:17:10 crc kubenswrapper[4859]: I1008 18:17:10.392523 4859 server.go:163] "Starting to listen" address="0.0.0.0" port=10250 Oct 08 18:17:10 crc kubenswrapper[4859]: I1008 18:17:10.392626 4859 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csinodes/crc?resourceVersion=0": dial tcp 38.102.83.13:6443: connect: connection refused Oct 08 18:17:10 crc kubenswrapper[4859]: I1008 18:17:10.393243 4859 server.go:236] "Starting to serve the podresources API" endpoint="unix:/var/lib/kubelet/pod-resources/kubelet.sock" Oct 08 18:17:10 crc systemd[1]: Started Kubernetes Kubelet. Oct 08 18:17:10 crc kubenswrapper[4859]: I1008 18:17:10.395167 4859 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate rotation is enabled Oct 08 18:17:10 crc kubenswrapper[4859]: I1008 18:17:10.395279 4859 fs_resource_analyzer.go:67] "Starting FS ResourceAnalyzer" Oct 08 18:17:10 crc kubenswrapper[4859]: I1008 18:17:10.395772 4859 volume_manager.go:287] "The desired_state_of_world populator starts" Oct 08 18:17:10 crc kubenswrapper[4859]: I1008 18:17:10.395809 4859 volume_manager.go:289] "Starting Kubelet Volume Manager" Oct 08 18:17:10 crc kubenswrapper[4859]: I1008 18:17:10.395879 4859 desired_state_of_world_populator.go:146] "Desired state populator starts to run" Oct 08 18:17:10 crc kubenswrapper[4859]: I1008 18:17:10.396364 4859 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-12-06 14:11:54.710929961 +0000 UTC Oct 08 18:17:10 crc kubenswrapper[4859]: I1008 18:17:10.396461 4859 certificate_manager.go:356] kubernetes.io/kubelet-serving: Waiting 1411h54m44.314473685s for next certificate rotation Oct 08 18:17:10 crc kubenswrapper[4859]: E1008 18:17:10.396616 4859 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Oct 08 18:17:10 crc kubenswrapper[4859]: W1008 18:17:10.396630 4859 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.CSIDriver: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csidrivers?limit=500&resourceVersion=0": dial tcp 38.102.83.13:6443: connect: connection refused Oct 08 18:17:10 crc kubenswrapper[4859]: E1008 18:17:10.396761 4859 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.CSIDriver: failed to list *v1.CSIDriver: Get \"https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csidrivers?limit=500&resourceVersion=0\": dial tcp 38.102.83.13:6443: connect: connection refused" logger="UnhandledError" Oct 08 18:17:10 crc kubenswrapper[4859]: I1008 18:17:10.398963 4859 factory.go:55] Registering systemd factory Oct 08 18:17:10 crc kubenswrapper[4859]: I1008 18:17:10.399011 4859 factory.go:221] Registration of the systemd container factory successfully Oct 08 18:17:10 crc kubenswrapper[4859]: I1008 18:17:10.399583 4859 server.go:460] "Adding debug handlers to kubelet server" Oct 08 18:17:10 crc kubenswrapper[4859]: E1008 18:17:10.399634 4859 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.13:6443: connect: connection refused" interval="200ms" Oct 08 18:17:10 crc kubenswrapper[4859]: I1008 18:17:10.406392 4859 factory.go:153] Registering CRI-O factory Oct 08 18:17:10 crc kubenswrapper[4859]: I1008 18:17:10.406494 4859 factory.go:221] Registration of the crio container factory successfully Oct 08 18:17:10 crc kubenswrapper[4859]: I1008 18:17:10.406658 4859 factory.go:219] Registration of the containerd container factory failed: unable to create containerd client: containerd: cannot unix dial containerd api service: dial unix /run/containerd/containerd.sock: connect: no such file or directory Oct 08 18:17:10 crc kubenswrapper[4859]: I1008 18:17:10.406795 4859 factory.go:103] Registering Raw factory Oct 08 18:17:10 crc kubenswrapper[4859]: I1008 18:17:10.406925 4859 manager.go:1196] Started watching for new ooms in manager Oct 08 18:17:10 crc kubenswrapper[4859]: I1008 18:17:10.412353 4859 manager.go:319] Starting recovery of all containers Oct 08 18:17:10 crc kubenswrapper[4859]: E1008 18:17:10.413151 4859 event.go:368] "Unable to write event (may retry after sleeping)" err="Post \"https://api-int.crc.testing:6443/api/v1/namespaces/default/events\": dial tcp 38.102.83.13:6443: connect: connection refused" event="&Event{ObjectMeta:{crc.186c96dd03e21ec7 default 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Node,Namespace:,Name:crc,UID:crc,APIVersion:,ResourceVersion:,FieldPath:,},Reason:Starting,Message:Starting kubelet.,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2025-10-08 18:17:10.392110791 +0000 UTC m=+0.638950170,LastTimestamp:2025-10-08 18:17:10.392110791 +0000 UTC m=+0.638950170,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Oct 08 18:17:10 crc kubenswrapper[4859]: I1008 18:17:10.417142 4859 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="31d8b7a1-420e-4252-a5b7-eebe8a111292" volumeName="kubernetes.io/secret/31d8b7a1-420e-4252-a5b7-eebe8a111292-proxy-tls" seLinuxMountContext="" Oct 08 18:17:10 crc kubenswrapper[4859]: I1008 18:17:10.417191 4859 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="37a5e44f-9a88-4405-be8a-b645485e7312" volumeName="kubernetes.io/projected/37a5e44f-9a88-4405-be8a-b645485e7312-kube-api-access-rdwmf" seLinuxMountContext="" Oct 08 18:17:10 crc kubenswrapper[4859]: I1008 18:17:10.417207 4859 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="3cb93b32-e0ae-4377-b9c8-fdb9842c6d59" volumeName="kubernetes.io/configmap/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59-serviceca" seLinuxMountContext="" Oct 08 18:17:10 crc kubenswrapper[4859]: I1008 18:17:10.417225 4859 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="43509403-f426-496e-be36-56cef71462f5" volumeName="kubernetes.io/secret/43509403-f426-496e-be36-56cef71462f5-console-serving-cert" seLinuxMountContext="" Oct 08 18:17:10 crc kubenswrapper[4859]: I1008 18:17:10.417239 4859 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-error" seLinuxMountContext="" Oct 08 18:17:10 crc kubenswrapper[4859]: I1008 18:17:10.417253 4859 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5441d097-087c-4d9a-baa8-b210afa90fc9" volumeName="kubernetes.io/configmap/5441d097-087c-4d9a-baa8-b210afa90fc9-config" seLinuxMountContext="" Oct 08 18:17:10 crc kubenswrapper[4859]: I1008 18:17:10.417269 4859 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" volumeName="kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf" seLinuxMountContext="" Oct 08 18:17:10 crc kubenswrapper[4859]: I1008 18:17:10.417290 4859 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="0b574797-001e-440a-8f4e-c0be86edad0f" volumeName="kubernetes.io/secret/0b574797-001e-440a-8f4e-c0be86edad0f-proxy-tls" seLinuxMountContext="" Oct 08 18:17:10 crc kubenswrapper[4859]: I1008 18:17:10.417307 4859 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="9d4552c7-cd75-42dd-8880-30dd377c49a4" volumeName="kubernetes.io/configmap/9d4552c7-cd75-42dd-8880-30dd377c49a4-config" seLinuxMountContext="" Oct 08 18:17:10 crc kubenswrapper[4859]: I1008 18:17:10.417321 4859 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="e7e6199b-1264-4501-8953-767f51328d08" volumeName="kubernetes.io/secret/e7e6199b-1264-4501-8953-767f51328d08-serving-cert" seLinuxMountContext="" Oct 08 18:17:10 crc kubenswrapper[4859]: I1008 18:17:10.417339 4859 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="f88749ec-7931-4ee7-b3fc-1ec5e11f92e9" volumeName="kubernetes.io/secret/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-srv-cert" seLinuxMountContext="" Oct 08 18:17:10 crc kubenswrapper[4859]: I1008 18:17:10.417361 4859 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8f668bae-612b-4b75-9490-919e737c6a3b" volumeName="kubernetes.io/configmap/8f668bae-612b-4b75-9490-919e737c6a3b-trusted-ca" seLinuxMountContext="" Oct 08 18:17:10 crc kubenswrapper[4859]: I1008 18:17:10.417383 4859 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1bf7eb37-55a3-4c65-b768-a94c82151e69" volumeName="kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-etcd-client" seLinuxMountContext="" Oct 08 18:17:10 crc kubenswrapper[4859]: I1008 18:17:10.417406 4859 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="22c825df-677d-4ca6-82db-3454ed06e783" volumeName="kubernetes.io/secret/22c825df-677d-4ca6-82db-3454ed06e783-machine-approver-tls" seLinuxMountContext="" Oct 08 18:17:10 crc kubenswrapper[4859]: I1008 18:17:10.417455 4859 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="31d8b7a1-420e-4252-a5b7-eebe8a111292" volumeName="kubernetes.io/configmap/31d8b7a1-420e-4252-a5b7-eebe8a111292-auth-proxy-config" seLinuxMountContext="" Oct 08 18:17:10 crc kubenswrapper[4859]: I1008 18:17:10.417473 4859 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1386a44e-36a2-460c-96d0-0359d2b6f0f5" volumeName="kubernetes.io/projected/1386a44e-36a2-460c-96d0-0359d2b6f0f5-kube-api-access" seLinuxMountContext="" Oct 08 18:17:10 crc kubenswrapper[4859]: I1008 18:17:10.417494 4859 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1bf7eb37-55a3-4c65-b768-a94c82151e69" volumeName="kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-encryption-config" seLinuxMountContext="" Oct 08 18:17:10 crc kubenswrapper[4859]: I1008 18:17:10.417510 4859 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="25e176fe-21b4-4974-b1ed-c8b94f112a7f" volumeName="kubernetes.io/configmap/25e176fe-21b4-4974-b1ed-c8b94f112a7f-signing-cabundle" seLinuxMountContext="" Oct 08 18:17:10 crc kubenswrapper[4859]: I1008 18:17:10.417526 4859 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="308be0ea-9f5f-4b29-aeb1-5abd31a0b17b" volumeName="kubernetes.io/secret/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-webhook-cert" seLinuxMountContext="" Oct 08 18:17:10 crc kubenswrapper[4859]: I1008 18:17:10.417542 4859 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="43509403-f426-496e-be36-56cef71462f5" volumeName="kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-trusted-ca-bundle" seLinuxMountContext="" Oct 08 18:17:10 crc kubenswrapper[4859]: I1008 18:17:10.417560 4859 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="496e6271-fb68-4057-954e-a0d97a4afa3f" volumeName="kubernetes.io/secret/496e6271-fb68-4057-954e-a0d97a4afa3f-serving-cert" seLinuxMountContext="" Oct 08 18:17:10 crc kubenswrapper[4859]: I1008 18:17:10.417577 4859 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6509e943-70c6-444c-bc41-48a544e36fbd" volumeName="kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-trusted-ca-bundle" seLinuxMountContext="" Oct 08 18:17:10 crc kubenswrapper[4859]: I1008 18:17:10.417594 4859 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8f668bae-612b-4b75-9490-919e737c6a3b" volumeName="kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-kube-api-access-kfwg7" seLinuxMountContext="" Oct 08 18:17:10 crc kubenswrapper[4859]: I1008 18:17:10.417611 4859 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="01ab3dd5-8196-46d0-ad33-122e2ca51def" volumeName="kubernetes.io/projected/01ab3dd5-8196-46d0-ad33-122e2ca51def-kube-api-access-w7l8j" seLinuxMountContext="" Oct 08 18:17:10 crc kubenswrapper[4859]: I1008 18:17:10.417629 4859 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="a31745f5-9847-4afe-82a5-3161cc66ca93" volumeName="kubernetes.io/configmap/a31745f5-9847-4afe-82a5-3161cc66ca93-trusted-ca" seLinuxMountContext="" Oct 08 18:17:10 crc kubenswrapper[4859]: I1008 18:17:10.417647 4859 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6402fda4-df10-493c-b4e5-d0569419652d" volumeName="kubernetes.io/configmap/6402fda4-df10-493c-b4e5-d0569419652d-images" seLinuxMountContext="" Oct 08 18:17:10 crc kubenswrapper[4859]: I1008 18:17:10.417669 4859 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6731426b-95fe-49ff-bb5f-40441049fde2" volumeName="kubernetes.io/secret/6731426b-95fe-49ff-bb5f-40441049fde2-control-plane-machine-set-operator-tls" seLinuxMountContext="" Oct 08 18:17:10 crc kubenswrapper[4859]: I1008 18:17:10.417694 4859 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6ea678ab-3438-413e-bfe3-290ae7725660" volumeName="kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-ovnkube-script-lib" seLinuxMountContext="" Oct 08 18:17:10 crc kubenswrapper[4859]: I1008 18:17:10.417746 4859 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="b6312bbd-5731-4ea0-a20f-81d5a57df44a" volumeName="kubernetes.io/secret/b6312bbd-5731-4ea0-a20f-81d5a57df44a-srv-cert" seLinuxMountContext="" Oct 08 18:17:10 crc kubenswrapper[4859]: I1008 18:17:10.417770 4859 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="b6cd30de-2eeb-49a2-ab40-9167f4560ff5" volumeName="kubernetes.io/projected/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-kube-api-access-pj782" seLinuxMountContext="" Oct 08 18:17:10 crc kubenswrapper[4859]: I1008 18:17:10.417815 4859 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="b6cd30de-2eeb-49a2-ab40-9167f4560ff5" volumeName="kubernetes.io/secret/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-marketplace-operator-metrics" seLinuxMountContext="" Oct 08 18:17:10 crc kubenswrapper[4859]: I1008 18:17:10.417835 4859 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="bf126b07-da06-4140-9a57-dfd54fc6b486" volumeName="kubernetes.io/secret/bf126b07-da06-4140-9a57-dfd54fc6b486-image-registry-operator-tls" seLinuxMountContext="" Oct 08 18:17:10 crc kubenswrapper[4859]: I1008 18:17:10.417853 4859 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09ae3b1a-e8e7-4524-b54b-61eab6f9239a" volumeName="kubernetes.io/projected/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-kube-api-access-zkvpv" seLinuxMountContext="" Oct 08 18:17:10 crc kubenswrapper[4859]: I1008 18:17:10.417871 4859 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="ef543e1b-8068-4ea3-b32a-61027b32e95d" volumeName="kubernetes.io/secret/ef543e1b-8068-4ea3-b32a-61027b32e95d-webhook-cert" seLinuxMountContext="" Oct 08 18:17:10 crc kubenswrapper[4859]: I1008 18:17:10.417886 4859 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5225d0e4-402f-4861-b410-819f433b1803" volumeName="kubernetes.io/projected/5225d0e4-402f-4861-b410-819f433b1803-kube-api-access-9xfj7" seLinuxMountContext="" Oct 08 18:17:10 crc kubenswrapper[4859]: I1008 18:17:10.417904 4859 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5fe579f8-e8a6-4643-bce5-a661393c4dde" volumeName="kubernetes.io/projected/5fe579f8-e8a6-4643-bce5-a661393c4dde-kube-api-access-fcqwp" seLinuxMountContext="" Oct 08 18:17:10 crc kubenswrapper[4859]: I1008 18:17:10.417924 4859 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="bf126b07-da06-4140-9a57-dfd54fc6b486" volumeName="kubernetes.io/projected/bf126b07-da06-4140-9a57-dfd54fc6b486-bound-sa-token" seLinuxMountContext="" Oct 08 18:17:10 crc kubenswrapper[4859]: I1008 18:17:10.417942 4859 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="c03ee662-fb2f-4fc4-a2c1-af487c19d254" volumeName="kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-stats-auth" seLinuxMountContext="" Oct 08 18:17:10 crc kubenswrapper[4859]: I1008 18:17:10.417958 4859 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="01ab3dd5-8196-46d0-ad33-122e2ca51def" volumeName="kubernetes.io/configmap/01ab3dd5-8196-46d0-ad33-122e2ca51def-config" seLinuxMountContext="" Oct 08 18:17:10 crc kubenswrapper[4859]: I1008 18:17:10.417975 4859 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="0b78653f-4ff9-4508-8672-245ed9b561e3" volumeName="kubernetes.io/configmap/0b78653f-4ff9-4508-8672-245ed9b561e3-service-ca" seLinuxMountContext="" Oct 08 18:17:10 crc kubenswrapper[4859]: I1008 18:17:10.417994 4859 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="4bb40260-dbaa-4fb0-84df-5e680505d512" volumeName="kubernetes.io/configmap/4bb40260-dbaa-4fb0-84df-5e680505d512-cni-binary-copy" seLinuxMountContext="" Oct 08 18:17:10 crc kubenswrapper[4859]: I1008 18:17:10.418011 4859 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7583ce53-e0fe-4a16-9e4d-50516596a136" volumeName="kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-proxy-ca-bundles" seLinuxMountContext="" Oct 08 18:17:10 crc kubenswrapper[4859]: I1008 18:17:10.418028 4859 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="b11524ee-3fca-4b1b-9cdf-6da289fdbc7d" volumeName="kubernetes.io/projected/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-kube-api-access-x4zgh" seLinuxMountContext="" Oct 08 18:17:10 crc kubenswrapper[4859]: I1008 18:17:10.418046 4859 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="0b574797-001e-440a-8f4e-c0be86edad0f" volumeName="kubernetes.io/projected/0b574797-001e-440a-8f4e-c0be86edad0f-kube-api-access-lzf88" seLinuxMountContext="" Oct 08 18:17:10 crc kubenswrapper[4859]: I1008 18:17:10.418073 4859 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09efc573-dbb6-4249-bd59-9b87aba8dd28" volumeName="kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-service-ca" seLinuxMountContext="" Oct 08 18:17:10 crc kubenswrapper[4859]: I1008 18:17:10.418094 4859 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="0b574797-001e-440a-8f4e-c0be86edad0f" volumeName="kubernetes.io/configmap/0b574797-001e-440a-8f4e-c0be86edad0f-mcc-auth-proxy-config" seLinuxMountContext="" Oct 08 18:17:10 crc kubenswrapper[4859]: I1008 18:17:10.418113 4859 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-router-certs" seLinuxMountContext="" Oct 08 18:17:10 crc kubenswrapper[4859]: I1008 18:17:10.418132 4859 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5fe579f8-e8a6-4643-bce5-a661393c4dde" volumeName="kubernetes.io/secret/5fe579f8-e8a6-4643-bce5-a661393c4dde-node-bootstrap-token" seLinuxMountContext="" Oct 08 18:17:10 crc kubenswrapper[4859]: I1008 18:17:10.418150 4859 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6ea678ab-3438-413e-bfe3-290ae7725660" volumeName="kubernetes.io/secret/6ea678ab-3438-413e-bfe3-290ae7725660-ovn-node-metrics-cert" seLinuxMountContext="" Oct 08 18:17:10 crc kubenswrapper[4859]: I1008 18:17:10.418168 4859 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="bd23aa5c-e532-4e53-bccf-e79f130c5ae8" volumeName="kubernetes.io/projected/bd23aa5c-e532-4e53-bccf-e79f130c5ae8-kube-api-access-jhbk2" seLinuxMountContext="" Oct 08 18:17:10 crc kubenswrapper[4859]: I1008 18:17:10.418335 4859 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d" volumeName="kubernetes.io/projected/cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d-kube-api-access-x2m85" seLinuxMountContext="" Oct 08 18:17:10 crc kubenswrapper[4859]: I1008 18:17:10.418356 4859 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="01ab3dd5-8196-46d0-ad33-122e2ca51def" volumeName="kubernetes.io/secret/01ab3dd5-8196-46d0-ad33-122e2ca51def-serving-cert" seLinuxMountContext="" Oct 08 18:17:10 crc kubenswrapper[4859]: I1008 18:17:10.418382 4859 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-session" seLinuxMountContext="" Oct 08 18:17:10 crc kubenswrapper[4859]: I1008 18:17:10.418400 4859 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="57a731c4-ef35-47a8-b875-bfb08a7f8011" volumeName="kubernetes.io/projected/57a731c4-ef35-47a8-b875-bfb08a7f8011-kube-api-access-cfbct" seLinuxMountContext="" Oct 08 18:17:10 crc kubenswrapper[4859]: I1008 18:17:10.418418 4859 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6ea678ab-3438-413e-bfe3-290ae7725660" volumeName="kubernetes.io/projected/6ea678ab-3438-413e-bfe3-290ae7725660-kube-api-access-htfz6" seLinuxMountContext="" Oct 08 18:17:10 crc kubenswrapper[4859]: I1008 18:17:10.418441 4859 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7539238d-5fe0-46ed-884e-1c3b566537ec" volumeName="kubernetes.io/secret/7539238d-5fe0-46ed-884e-1c3b566537ec-serving-cert" seLinuxMountContext="" Oct 08 18:17:10 crc kubenswrapper[4859]: I1008 18:17:10.418462 4859 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="87cf06ed-a83f-41a7-828d-70653580a8cb" volumeName="kubernetes.io/projected/87cf06ed-a83f-41a7-828d-70653580a8cb-kube-api-access-d6qdx" seLinuxMountContext="" Oct 08 18:17:10 crc kubenswrapper[4859]: I1008 18:17:10.418480 4859 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="96b93a3a-6083-4aea-8eab-fe1aa8245ad9" volumeName="kubernetes.io/projected/96b93a3a-6083-4aea-8eab-fe1aa8245ad9-kube-api-access-nzwt7" seLinuxMountContext="" Oct 08 18:17:10 crc kubenswrapper[4859]: I1008 18:17:10.418497 4859 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="9d4552c7-cd75-42dd-8880-30dd377c49a4" volumeName="kubernetes.io/secret/9d4552c7-cd75-42dd-8880-30dd377c49a4-serving-cert" seLinuxMountContext="" Oct 08 18:17:10 crc kubenswrapper[4859]: I1008 18:17:10.418514 4859 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1386a44e-36a2-460c-96d0-0359d2b6f0f5" volumeName="kubernetes.io/configmap/1386a44e-36a2-460c-96d0-0359d2b6f0f5-config" seLinuxMountContext="" Oct 08 18:17:10 crc kubenswrapper[4859]: I1008 18:17:10.418533 4859 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="ef543e1b-8068-4ea3-b32a-61027b32e95d" volumeName="kubernetes.io/configmap/ef543e1b-8068-4ea3-b32a-61027b32e95d-ovnkube-identity-cm" seLinuxMountContext="" Oct 08 18:17:10 crc kubenswrapper[4859]: I1008 18:17:10.418551 4859 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-idp-0-file-data" seLinuxMountContext="" Oct 08 18:17:10 crc kubenswrapper[4859]: I1008 18:17:10.418592 4859 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5441d097-087c-4d9a-baa8-b210afa90fc9" volumeName="kubernetes.io/secret/5441d097-087c-4d9a-baa8-b210afa90fc9-serving-cert" seLinuxMountContext="" Oct 08 18:17:10 crc kubenswrapper[4859]: I1008 18:17:10.418610 4859 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5fe579f8-e8a6-4643-bce5-a661393c4dde" volumeName="kubernetes.io/secret/5fe579f8-e8a6-4643-bce5-a661393c4dde-certs" seLinuxMountContext="" Oct 08 18:17:10 crc kubenswrapper[4859]: I1008 18:17:10.418627 4859 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="a0128f3a-b052-44ed-a84e-c4c8aaf17c13" volumeName="kubernetes.io/projected/a0128f3a-b052-44ed-a84e-c4c8aaf17c13-kube-api-access-gf66m" seLinuxMountContext="" Oct 08 18:17:10 crc kubenswrapper[4859]: I1008 18:17:10.418646 4859 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="b6312bbd-5731-4ea0-a20f-81d5a57df44a" volumeName="kubernetes.io/projected/b6312bbd-5731-4ea0-a20f-81d5a57df44a-kube-api-access-249nr" seLinuxMountContext="" Oct 08 18:17:10 crc kubenswrapper[4859]: I1008 18:17:10.418665 4859 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="b6cd30de-2eeb-49a2-ab40-9167f4560ff5" volumeName="kubernetes.io/configmap/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-marketplace-trusted-ca" seLinuxMountContext="" Oct 08 18:17:10 crc kubenswrapper[4859]: I1008 18:17:10.418683 4859 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09efc573-dbb6-4249-bd59-9b87aba8dd28" volumeName="kubernetes.io/projected/09efc573-dbb6-4249-bd59-9b87aba8dd28-kube-api-access-8tdtz" seLinuxMountContext="" Oct 08 18:17:10 crc kubenswrapper[4859]: I1008 18:17:10.418706 4859 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="3cb93b32-e0ae-4377-b9c8-fdb9842c6d59" volumeName="kubernetes.io/projected/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59-kube-api-access-wxkg8" seLinuxMountContext="" Oct 08 18:17:10 crc kubenswrapper[4859]: I1008 18:17:10.418746 4859 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-login" seLinuxMountContext="" Oct 08 18:17:10 crc kubenswrapper[4859]: I1008 18:17:10.418763 4859 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7539238d-5fe0-46ed-884e-1c3b566537ec" volumeName="kubernetes.io/configmap/7539238d-5fe0-46ed-884e-1c3b566537ec-config" seLinuxMountContext="" Oct 08 18:17:10 crc kubenswrapper[4859]: I1008 18:17:10.418780 4859 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="c03ee662-fb2f-4fc4-a2c1-af487c19d254" volumeName="kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-default-certificate" seLinuxMountContext="" Oct 08 18:17:10 crc kubenswrapper[4859]: I1008 18:17:10.418797 4859 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="fda69060-fa79-4696-b1a6-7980f124bf7c" volumeName="kubernetes.io/secret/fda69060-fa79-4696-b1a6-7980f124bf7c-proxy-tls" seLinuxMountContext="" Oct 08 18:17:10 crc kubenswrapper[4859]: I1008 18:17:10.418813 4859 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="22c825df-677d-4ca6-82db-3454ed06e783" volumeName="kubernetes.io/projected/22c825df-677d-4ca6-82db-3454ed06e783-kube-api-access-7c4vf" seLinuxMountContext="" Oct 08 18:17:10 crc kubenswrapper[4859]: I1008 18:17:10.418830 4859 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5b88f790-22fa-440e-b583-365168c0b23d" volumeName="kubernetes.io/secret/5b88f790-22fa-440e-b583-365168c0b23d-metrics-certs" seLinuxMountContext="" Oct 08 18:17:10 crc kubenswrapper[4859]: I1008 18:17:10.418848 4859 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6402fda4-df10-493c-b4e5-d0569419652d" volumeName="kubernetes.io/secret/6402fda4-df10-493c-b4e5-d0569419652d-machine-api-operator-tls" seLinuxMountContext="" Oct 08 18:17:10 crc kubenswrapper[4859]: I1008 18:17:10.418865 4859 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6509e943-70c6-444c-bc41-48a544e36fbd" volumeName="kubernetes.io/projected/6509e943-70c6-444c-bc41-48a544e36fbd-kube-api-access-6g6sz" seLinuxMountContext="" Oct 08 18:17:10 crc kubenswrapper[4859]: I1008 18:17:10.418881 4859 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="b11524ee-3fca-4b1b-9cdf-6da289fdbc7d" volumeName="kubernetes.io/empty-dir/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-catalog-content" seLinuxMountContext="" Oct 08 18:17:10 crc kubenswrapper[4859]: I1008 18:17:10.418926 4859 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-service-ca" seLinuxMountContext="" Oct 08 18:17:10 crc kubenswrapper[4859]: I1008 18:17:10.418946 4859 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1d611f23-29be-4491-8495-bee1670e935f" volumeName="kubernetes.io/projected/1d611f23-29be-4491-8495-bee1670e935f-kube-api-access-bf2bz" seLinuxMountContext="" Oct 08 18:17:10 crc kubenswrapper[4859]: I1008 18:17:10.418964 4859 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="3b6479f0-333b-4a96-9adf-2099afdc2447" volumeName="kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr" seLinuxMountContext="" Oct 08 18:17:10 crc kubenswrapper[4859]: I1008 18:17:10.418981 4859 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="44663579-783b-4372-86d6-acf235a62d72" volumeName="kubernetes.io/projected/44663579-783b-4372-86d6-acf235a62d72-kube-api-access-vt5rc" seLinuxMountContext="" Oct 08 18:17:10 crc kubenswrapper[4859]: I1008 18:17:10.418998 4859 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8f668bae-612b-4b75-9490-919e737c6a3b" volumeName="kubernetes.io/configmap/8f668bae-612b-4b75-9490-919e737c6a3b-registry-certificates" seLinuxMountContext="" Oct 08 18:17:10 crc kubenswrapper[4859]: I1008 18:17:10.419016 4859 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="b11524ee-3fca-4b1b-9cdf-6da289fdbc7d" volumeName="kubernetes.io/empty-dir/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-utilities" seLinuxMountContext="" Oct 08 18:17:10 crc kubenswrapper[4859]: I1008 18:17:10.419032 4859 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="bc5039c0-ea34-426b-a2b7-fbbc87b49a6d" volumeName="kubernetes.io/secret/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-serving-cert" seLinuxMountContext="" Oct 08 18:17:10 crc kubenswrapper[4859]: I1008 18:17:10.419049 4859 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="0b78653f-4ff9-4508-8672-245ed9b561e3" volumeName="kubernetes.io/projected/0b78653f-4ff9-4508-8672-245ed9b561e3-kube-api-access" seLinuxMountContext="" Oct 08 18:17:10 crc kubenswrapper[4859]: I1008 18:17:10.419066 4859 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="efdd0498-1daa-4136-9a4a-3b948c2293fc" volumeName="kubernetes.io/secret/efdd0498-1daa-4136-9a4a-3b948c2293fc-webhook-certs" seLinuxMountContext="" Oct 08 18:17:10 crc kubenswrapper[4859]: I1008 18:17:10.419084 4859 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="57a731c4-ef35-47a8-b875-bfb08a7f8011" volumeName="kubernetes.io/empty-dir/57a731c4-ef35-47a8-b875-bfb08a7f8011-utilities" seLinuxMountContext="" Oct 08 18:17:10 crc kubenswrapper[4859]: I1008 18:17:10.419101 4859 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="43509403-f426-496e-be36-56cef71462f5" volumeName="kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-console-config" seLinuxMountContext="" Oct 08 18:17:10 crc kubenswrapper[4859]: I1008 18:17:10.419118 4859 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-cliconfig" seLinuxMountContext="" Oct 08 18:17:10 crc kubenswrapper[4859]: I1008 18:17:10.419136 4859 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="a31745f5-9847-4afe-82a5-3161cc66ca93" volumeName="kubernetes.io/projected/a31745f5-9847-4afe-82a5-3161cc66ca93-bound-sa-token" seLinuxMountContext="" Oct 08 18:17:10 crc kubenswrapper[4859]: I1008 18:17:10.419153 4859 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="d75a4c96-2883-4a0b-bab2-0fab2b6c0b49" volumeName="kubernetes.io/projected/d75a4c96-2883-4a0b-bab2-0fab2b6c0b49-kube-api-access-rczfb" seLinuxMountContext="" Oct 08 18:17:10 crc kubenswrapper[4859]: I1008 18:17:10.419172 4859 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="fda69060-fa79-4696-b1a6-7980f124bf7c" volumeName="kubernetes.io/projected/fda69060-fa79-4696-b1a6-7980f124bf7c-kube-api-access-xcgwh" seLinuxMountContext="" Oct 08 18:17:10 crc kubenswrapper[4859]: I1008 18:17:10.419190 4859 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09efc573-dbb6-4249-bd59-9b87aba8dd28" volumeName="kubernetes.io/secret/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-client" seLinuxMountContext="" Oct 08 18:17:10 crc kubenswrapper[4859]: I1008 18:17:10.419209 4859 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1d611f23-29be-4491-8495-bee1670e935f" volumeName="kubernetes.io/empty-dir/1d611f23-29be-4491-8495-bee1670e935f-utilities" seLinuxMountContext="" Oct 08 18:17:10 crc kubenswrapper[4859]: I1008 18:17:10.419227 4859 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7bb08738-c794-4ee8-9972-3a62ca171029" volumeName="kubernetes.io/configmap/7bb08738-c794-4ee8-9972-3a62ca171029-cni-sysctl-allowlist" seLinuxMountContext="" Oct 08 18:17:10 crc kubenswrapper[4859]: I1008 18:17:10.419242 4859 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8f668bae-612b-4b75-9490-919e737c6a3b" volumeName="kubernetes.io/secret/8f668bae-612b-4b75-9490-919e737c6a3b-installation-pull-secrets" seLinuxMountContext="" Oct 08 18:17:10 crc kubenswrapper[4859]: I1008 18:17:10.419259 4859 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="b6312bbd-5731-4ea0-a20f-81d5a57df44a" volumeName="kubernetes.io/secret/b6312bbd-5731-4ea0-a20f-81d5a57df44a-profile-collector-cert" seLinuxMountContext="" Oct 08 18:17:10 crc kubenswrapper[4859]: I1008 18:17:10.419276 4859 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1bf7eb37-55a3-4c65-b768-a94c82151e69" volumeName="kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-image-import-ca" seLinuxMountContext="" Oct 08 18:17:10 crc kubenswrapper[4859]: I1008 18:17:10.419294 4859 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c" volumeName="kubernetes.io/projected/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-kube-api-access-qs4fp" seLinuxMountContext="" Oct 08 18:17:10 crc kubenswrapper[4859]: I1008 18:17:10.419311 4859 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="43509403-f426-496e-be36-56cef71462f5" volumeName="kubernetes.io/projected/43509403-f426-496e-be36-56cef71462f5-kube-api-access-qg5z5" seLinuxMountContext="" Oct 08 18:17:10 crc kubenswrapper[4859]: I1008 18:17:10.419327 4859 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7583ce53-e0fe-4a16-9e4d-50516596a136" volumeName="kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-client-ca" seLinuxMountContext="" Oct 08 18:17:10 crc kubenswrapper[4859]: I1008 18:17:10.419343 4859 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="925f1c65-6136-48ba-85aa-3a3b50560753" volumeName="kubernetes.io/configmap/925f1c65-6136-48ba-85aa-3a3b50560753-env-overrides" seLinuxMountContext="" Oct 08 18:17:10 crc kubenswrapper[4859]: I1008 18:17:10.419358 4859 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="bc5039c0-ea34-426b-a2b7-fbbc87b49a6d" volumeName="kubernetes.io/empty-dir/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-available-featuregates" seLinuxMountContext="" Oct 08 18:17:10 crc kubenswrapper[4859]: I1008 18:17:10.419386 4859 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="0b78653f-4ff9-4508-8672-245ed9b561e3" volumeName="kubernetes.io/secret/0b78653f-4ff9-4508-8672-245ed9b561e3-serving-cert" seLinuxMountContext="" Oct 08 18:17:10 crc kubenswrapper[4859]: I1008 18:17:10.419406 4859 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-serving-cert" seLinuxMountContext="" Oct 08 18:17:10 crc kubenswrapper[4859]: I1008 18:17:10.419425 4859 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="925f1c65-6136-48ba-85aa-3a3b50560753" volumeName="kubernetes.io/configmap/925f1c65-6136-48ba-85aa-3a3b50560753-ovnkube-config" seLinuxMountContext="" Oct 08 18:17:10 crc kubenswrapper[4859]: I1008 18:17:10.419444 4859 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="bf126b07-da06-4140-9a57-dfd54fc6b486" volumeName="kubernetes.io/configmap/bf126b07-da06-4140-9a57-dfd54fc6b486-trusted-ca" seLinuxMountContext="" Oct 08 18:17:10 crc kubenswrapper[4859]: I1008 18:17:10.419461 4859 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="25e176fe-21b4-4974-b1ed-c8b94f112a7f" volumeName="kubernetes.io/projected/25e176fe-21b4-4974-b1ed-c8b94f112a7f-kube-api-access-d4lsv" seLinuxMountContext="" Oct 08 18:17:10 crc kubenswrapper[4859]: I1008 18:17:10.419478 4859 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1bf7eb37-55a3-4c65-b768-a94c82151e69" volumeName="kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-audit" seLinuxMountContext="" Oct 08 18:17:10 crc kubenswrapper[4859]: I1008 18:17:10.419498 4859 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8cea82b4-6893-4ddc-af9f-1bb5ae425c5b" volumeName="kubernetes.io/projected/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-kube-api-access-w4xd4" seLinuxMountContext="" Oct 08 18:17:10 crc kubenswrapper[4859]: I1008 18:17:10.419518 4859 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="c03ee662-fb2f-4fc4-a2c1-af487c19d254" volumeName="kubernetes.io/configmap/c03ee662-fb2f-4fc4-a2c1-af487c19d254-service-ca-bundle" seLinuxMountContext="" Oct 08 18:17:10 crc kubenswrapper[4859]: I1008 18:17:10.419535 4859 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09ae3b1a-e8e7-4524-b54b-61eab6f9239a" volumeName="kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-audit-policies" seLinuxMountContext="" Oct 08 18:17:10 crc kubenswrapper[4859]: I1008 18:17:10.419556 4859 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="20b0d48f-5fd6-431c-a545-e3c800c7b866" volumeName="kubernetes.io/secret/20b0d48f-5fd6-431c-a545-e3c800c7b866-cert" seLinuxMountContext="" Oct 08 18:17:10 crc kubenswrapper[4859]: I1008 18:17:10.419586 4859 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="496e6271-fb68-4057-954e-a0d97a4afa3f" volumeName="kubernetes.io/configmap/496e6271-fb68-4057-954e-a0d97a4afa3f-config" seLinuxMountContext="" Oct 08 18:17:10 crc kubenswrapper[4859]: I1008 18:17:10.419605 4859 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5441d097-087c-4d9a-baa8-b210afa90fc9" volumeName="kubernetes.io/projected/5441d097-087c-4d9a-baa8-b210afa90fc9-kube-api-access-2d4wz" seLinuxMountContext="" Oct 08 18:17:10 crc kubenswrapper[4859]: I1008 18:17:10.419621 4859 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6509e943-70c6-444c-bc41-48a544e36fbd" volumeName="kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-service-ca-bundle" seLinuxMountContext="" Oct 08 18:17:10 crc kubenswrapper[4859]: I1008 18:17:10.419637 4859 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7bb08738-c794-4ee8-9972-3a62ca171029" volumeName="kubernetes.io/configmap/7bb08738-c794-4ee8-9972-3a62ca171029-cni-binary-copy" seLinuxMountContext="" Oct 08 18:17:10 crc kubenswrapper[4859]: I1008 18:17:10.419755 4859 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="e7e6199b-1264-4501-8953-767f51328d08" volumeName="kubernetes.io/configmap/e7e6199b-1264-4501-8953-767f51328d08-config" seLinuxMountContext="" Oct 08 18:17:10 crc kubenswrapper[4859]: I1008 18:17:10.419776 4859 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="ef543e1b-8068-4ea3-b32a-61027b32e95d" volumeName="kubernetes.io/configmap/ef543e1b-8068-4ea3-b32a-61027b32e95d-env-overrides" seLinuxMountContext="" Oct 08 18:17:10 crc kubenswrapper[4859]: I1008 18:17:10.419794 4859 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1bf7eb37-55a3-4c65-b768-a94c82151e69" volumeName="kubernetes.io/projected/1bf7eb37-55a3-4c65-b768-a94c82151e69-kube-api-access-sb6h7" seLinuxMountContext="" Oct 08 18:17:10 crc kubenswrapper[4859]: I1008 18:17:10.419813 4859 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="fda69060-fa79-4696-b1a6-7980f124bf7c" volumeName="kubernetes.io/configmap/fda69060-fa79-4696-b1a6-7980f124bf7c-mcd-auth-proxy-config" seLinuxMountContext="" Oct 08 18:17:10 crc kubenswrapper[4859]: I1008 18:17:10.419832 4859 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="efdd0498-1daa-4136-9a4a-3b948c2293fc" volumeName="kubernetes.io/projected/efdd0498-1daa-4136-9a4a-3b948c2293fc-kube-api-access-fqsjt" seLinuxMountContext="" Oct 08 18:17:10 crc kubenswrapper[4859]: I1008 18:17:10.419852 4859 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8f668bae-612b-4b75-9490-919e737c6a3b" volumeName="kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-registry-tls" seLinuxMountContext="" Oct 08 18:17:10 crc kubenswrapper[4859]: I1008 18:17:10.419873 4859 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="bf126b07-da06-4140-9a57-dfd54fc6b486" volumeName="kubernetes.io/projected/bf126b07-da06-4140-9a57-dfd54fc6b486-kube-api-access-rnphk" seLinuxMountContext="" Oct 08 18:17:10 crc kubenswrapper[4859]: I1008 18:17:10.419892 4859 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="25e176fe-21b4-4974-b1ed-c8b94f112a7f" volumeName="kubernetes.io/secret/25e176fe-21b4-4974-b1ed-c8b94f112a7f-signing-key" seLinuxMountContext="" Oct 08 18:17:10 crc kubenswrapper[4859]: I1008 18:17:10.419912 4859 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="308be0ea-9f5f-4b29-aeb1-5abd31a0b17b" volumeName="kubernetes.io/empty-dir/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-tmpfs" seLinuxMountContext="" Oct 08 18:17:10 crc kubenswrapper[4859]: I1008 18:17:10.419929 4859 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5441d097-087c-4d9a-baa8-b210afa90fc9" volumeName="kubernetes.io/configmap/5441d097-087c-4d9a-baa8-b210afa90fc9-client-ca" seLinuxMountContext="" Oct 08 18:17:10 crc kubenswrapper[4859]: I1008 18:17:10.419947 4859 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7583ce53-e0fe-4a16-9e4d-50516596a136" volumeName="kubernetes.io/secret/7583ce53-e0fe-4a16-9e4d-50516596a136-serving-cert" seLinuxMountContext="" Oct 08 18:17:10 crc kubenswrapper[4859]: I1008 18:17:10.419966 4859 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="87cf06ed-a83f-41a7-828d-70653580a8cb" volumeName="kubernetes.io/secret/87cf06ed-a83f-41a7-828d-70653580a8cb-metrics-tls" seLinuxMountContext="" Oct 08 18:17:10 crc kubenswrapper[4859]: I1008 18:17:10.419983 4859 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="a31745f5-9847-4afe-82a5-3161cc66ca93" volumeName="kubernetes.io/projected/a31745f5-9847-4afe-82a5-3161cc66ca93-kube-api-access-lz9wn" seLinuxMountContext="" Oct 08 18:17:10 crc kubenswrapper[4859]: I1008 18:17:10.420004 4859 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="bc5039c0-ea34-426b-a2b7-fbbc87b49a6d" volumeName="kubernetes.io/projected/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-kube-api-access-mnrrd" seLinuxMountContext="" Oct 08 18:17:10 crc kubenswrapper[4859]: I1008 18:17:10.420026 4859 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09ae3b1a-e8e7-4524-b54b-61eab6f9239a" volumeName="kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-trusted-ca-bundle" seLinuxMountContext="" Oct 08 18:17:10 crc kubenswrapper[4859]: I1008 18:17:10.420044 4859 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="925f1c65-6136-48ba-85aa-3a3b50560753" volumeName="kubernetes.io/projected/925f1c65-6136-48ba-85aa-3a3b50560753-kube-api-access-s4n52" seLinuxMountContext="" Oct 08 18:17:10 crc kubenswrapper[4859]: I1008 18:17:10.420061 4859 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="3ab1a177-2de0-46d9-b765-d0d0649bb42e" volumeName="kubernetes.io/secret/3ab1a177-2de0-46d9-b765-d0d0649bb42e-package-server-manager-serving-cert" seLinuxMountContext="" Oct 08 18:17:10 crc kubenswrapper[4859]: I1008 18:17:10.420079 4859 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1bf7eb37-55a3-4c65-b768-a94c82151e69" volumeName="kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-etcd-serving-ca" seLinuxMountContext="" Oct 08 18:17:10 crc kubenswrapper[4859]: I1008 18:17:10.420099 4859 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1bf7eb37-55a3-4c65-b768-a94c82151e69" volumeName="kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-serving-cert" seLinuxMountContext="" Oct 08 18:17:10 crc kubenswrapper[4859]: I1008 18:17:10.420117 4859 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c" volumeName="kubernetes.io/configmap/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-config" seLinuxMountContext="" Oct 08 18:17:10 crc kubenswrapper[4859]: I1008 18:17:10.420134 4859 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="43509403-f426-496e-be36-56cef71462f5" volumeName="kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-oauth-serving-cert" seLinuxMountContext="" Oct 08 18:17:10 crc kubenswrapper[4859]: I1008 18:17:10.420153 4859 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-audit-policies" seLinuxMountContext="" Oct 08 18:17:10 crc kubenswrapper[4859]: I1008 18:17:10.420169 4859 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5225d0e4-402f-4861-b410-819f433b1803" volumeName="kubernetes.io/empty-dir/5225d0e4-402f-4861-b410-819f433b1803-catalog-content" seLinuxMountContext="" Oct 08 18:17:10 crc kubenswrapper[4859]: I1008 18:17:10.420186 4859 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6ea678ab-3438-413e-bfe3-290ae7725660" volumeName="kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-ovnkube-config" seLinuxMountContext="" Oct 08 18:17:10 crc kubenswrapper[4859]: I1008 18:17:10.420208 4859 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1386a44e-36a2-460c-96d0-0359d2b6f0f5" volumeName="kubernetes.io/secret/1386a44e-36a2-460c-96d0-0359d2b6f0f5-serving-cert" seLinuxMountContext="" Oct 08 18:17:10 crc kubenswrapper[4859]: I1008 18:17:10.420226 4859 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="f88749ec-7931-4ee7-b3fc-1ec5e11f92e9" volumeName="kubernetes.io/projected/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-kube-api-access-dbsvg" seLinuxMountContext="" Oct 08 18:17:10 crc kubenswrapper[4859]: I1008 18:17:10.420245 4859 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09efc573-dbb6-4249-bd59-9b87aba8dd28" volumeName="kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-config" seLinuxMountContext="" Oct 08 18:17:10 crc kubenswrapper[4859]: I1008 18:17:10.420263 4859 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c" volumeName="kubernetes.io/secret/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-serving-cert" seLinuxMountContext="" Oct 08 18:17:10 crc kubenswrapper[4859]: I1008 18:17:10.420281 4859 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="308be0ea-9f5f-4b29-aeb1-5abd31a0b17b" volumeName="kubernetes.io/projected/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-kube-api-access-6ccd8" seLinuxMountContext="" Oct 08 18:17:10 crc kubenswrapper[4859]: I1008 18:17:10.420298 4859 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="37a5e44f-9a88-4405-be8a-b645485e7312" volumeName="kubernetes.io/secret/37a5e44f-9a88-4405-be8a-b645485e7312-metrics-tls" seLinuxMountContext="" Oct 08 18:17:10 crc kubenswrapper[4859]: I1008 18:17:10.420314 4859 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5225d0e4-402f-4861-b410-819f433b1803" volumeName="kubernetes.io/empty-dir/5225d0e4-402f-4861-b410-819f433b1803-utilities" seLinuxMountContext="" Oct 08 18:17:10 crc kubenswrapper[4859]: I1008 18:17:10.420331 4859 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6731426b-95fe-49ff-bb5f-40441049fde2" volumeName="kubernetes.io/projected/6731426b-95fe-49ff-bb5f-40441049fde2-kube-api-access-x7zkh" seLinuxMountContext="" Oct 08 18:17:10 crc kubenswrapper[4859]: I1008 18:17:10.420348 4859 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7539238d-5fe0-46ed-884e-1c3b566537ec" volumeName="kubernetes.io/projected/7539238d-5fe0-46ed-884e-1c3b566537ec-kube-api-access-tk88c" seLinuxMountContext="" Oct 08 18:17:10 crc kubenswrapper[4859]: I1008 18:17:10.420367 4859 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09ae3b1a-e8e7-4524-b54b-61eab6f9239a" volumeName="kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-etcd-client" seLinuxMountContext="" Oct 08 18:17:10 crc kubenswrapper[4859]: I1008 18:17:10.420389 4859 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="925f1c65-6136-48ba-85aa-3a3b50560753" volumeName="kubernetes.io/secret/925f1c65-6136-48ba-85aa-3a3b50560753-ovn-control-plane-metrics-cert" seLinuxMountContext="" Oct 08 18:17:10 crc kubenswrapper[4859]: I1008 18:17:10.420406 4859 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="9d4552c7-cd75-42dd-8880-30dd377c49a4" volumeName="kubernetes.io/configmap/9d4552c7-cd75-42dd-8880-30dd377c49a4-trusted-ca" seLinuxMountContext="" Oct 08 18:17:10 crc kubenswrapper[4859]: I1008 18:17:10.420425 4859 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="f88749ec-7931-4ee7-b3fc-1ec5e11f92e9" volumeName="kubernetes.io/secret/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-profile-collector-cert" seLinuxMountContext="" Oct 08 18:17:10 crc kubenswrapper[4859]: I1008 18:17:10.420445 4859 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="87cf06ed-a83f-41a7-828d-70653580a8cb" volumeName="kubernetes.io/configmap/87cf06ed-a83f-41a7-828d-70653580a8cb-config-volume" seLinuxMountContext="" Oct 08 18:17:10 crc kubenswrapper[4859]: I1008 18:17:10.420461 4859 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1d611f23-29be-4491-8495-bee1670e935f" volumeName="kubernetes.io/empty-dir/1d611f23-29be-4491-8495-bee1670e935f-catalog-content" seLinuxMountContext="" Oct 08 18:17:10 crc kubenswrapper[4859]: I1008 18:17:10.420480 4859 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49ef4625-1d3a-4a9f-b595-c2433d32326d" volumeName="kubernetes.io/projected/49ef4625-1d3a-4a9f-b595-c2433d32326d-kube-api-access-pjr6v" seLinuxMountContext="" Oct 08 18:17:10 crc kubenswrapper[4859]: I1008 18:17:10.420500 4859 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5b88f790-22fa-440e-b583-365168c0b23d" volumeName="kubernetes.io/projected/5b88f790-22fa-440e-b583-365168c0b23d-kube-api-access-jkwtn" seLinuxMountContext="" Oct 08 18:17:10 crc kubenswrapper[4859]: I1008 18:17:10.420518 4859 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6402fda4-df10-493c-b4e5-d0569419652d" volumeName="kubernetes.io/configmap/6402fda4-df10-493c-b4e5-d0569419652d-config" seLinuxMountContext="" Oct 08 18:17:10 crc kubenswrapper[4859]: I1008 18:17:10.420537 4859 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7583ce53-e0fe-4a16-9e4d-50516596a136" volumeName="kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-config" seLinuxMountContext="" Oct 08 18:17:10 crc kubenswrapper[4859]: I1008 18:17:10.420555 4859 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8f668bae-612b-4b75-9490-919e737c6a3b" volumeName="kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" seLinuxMountContext="" Oct 08 18:17:10 crc kubenswrapper[4859]: I1008 18:17:10.423866 4859 reconstruct.go:144] "Volume is marked device as uncertain and added into the actual state" volumeName="kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" deviceMountPath="/var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner/1f4776af88835e41c12b831b4c9fed40233456d14189815a54dbe7f892fc1983/globalmount" Oct 08 18:17:10 crc kubenswrapper[4859]: I1008 18:17:10.423918 4859 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="96b93a3a-6083-4aea-8eab-fe1aa8245ad9" volumeName="kubernetes.io/secret/96b93a3a-6083-4aea-8eab-fe1aa8245ad9-metrics-tls" seLinuxMountContext="" Oct 08 18:17:10 crc kubenswrapper[4859]: I1008 18:17:10.423940 4859 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1bf7eb37-55a3-4c65-b768-a94c82151e69" volumeName="kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-trusted-ca-bundle" seLinuxMountContext="" Oct 08 18:17:10 crc kubenswrapper[4859]: I1008 18:17:10.423961 4859 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09ae3b1a-e8e7-4524-b54b-61eab6f9239a" volumeName="kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-encryption-config" seLinuxMountContext="" Oct 08 18:17:10 crc kubenswrapper[4859]: I1008 18:17:10.423980 4859 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="308be0ea-9f5f-4b29-aeb1-5abd31a0b17b" volumeName="kubernetes.io/secret/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-apiservice-cert" seLinuxMountContext="" Oct 08 18:17:10 crc kubenswrapper[4859]: I1008 18:17:10.424009 4859 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="31d8b7a1-420e-4252-a5b7-eebe8a111292" volumeName="kubernetes.io/configmap/31d8b7a1-420e-4252-a5b7-eebe8a111292-images" seLinuxMountContext="" Oct 08 18:17:10 crc kubenswrapper[4859]: I1008 18:17:10.424041 4859 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="4bb40260-dbaa-4fb0-84df-5e680505d512" volumeName="kubernetes.io/configmap/4bb40260-dbaa-4fb0-84df-5e680505d512-multus-daemon-config" seLinuxMountContext="" Oct 08 18:17:10 crc kubenswrapper[4859]: I1008 18:17:10.424073 4859 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09ae3b1a-e8e7-4524-b54b-61eab6f9239a" volumeName="kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-etcd-serving-ca" seLinuxMountContext="" Oct 08 18:17:10 crc kubenswrapper[4859]: I1008 18:17:10.424105 4859 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8cea82b4-6893-4ddc-af9f-1bb5ae425c5b" volumeName="kubernetes.io/secret/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-serving-cert" seLinuxMountContext="" Oct 08 18:17:10 crc kubenswrapper[4859]: I1008 18:17:10.424134 4859 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="c03ee662-fb2f-4fc4-a2c1-af487c19d254" volumeName="kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-metrics-certs" seLinuxMountContext="" Oct 08 18:17:10 crc kubenswrapper[4859]: I1008 18:17:10.424164 4859 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7583ce53-e0fe-4a16-9e4d-50516596a136" volumeName="kubernetes.io/projected/7583ce53-e0fe-4a16-9e4d-50516596a136-kube-api-access-xcphl" seLinuxMountContext="" Oct 08 18:17:10 crc kubenswrapper[4859]: I1008 18:17:10.424192 4859 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="3ab1a177-2de0-46d9-b765-d0d0649bb42e" volumeName="kubernetes.io/projected/3ab1a177-2de0-46d9-b765-d0d0649bb42e-kube-api-access-4d4hj" seLinuxMountContext="" Oct 08 18:17:10 crc kubenswrapper[4859]: I1008 18:17:10.424226 4859 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-trusted-ca-bundle" seLinuxMountContext="" Oct 08 18:17:10 crc kubenswrapper[4859]: I1008 18:17:10.424257 4859 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="9d751cbb-f2e2-430d-9754-c882a5e924a5" volumeName="kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl" seLinuxMountContext="" Oct 08 18:17:10 crc kubenswrapper[4859]: I1008 18:17:10.424287 4859 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="e7e6199b-1264-4501-8953-767f51328d08" volumeName="kubernetes.io/projected/e7e6199b-1264-4501-8953-767f51328d08-kube-api-access" seLinuxMountContext="" Oct 08 18:17:10 crc kubenswrapper[4859]: I1008 18:17:10.424319 4859 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09efc573-dbb6-4249-bd59-9b87aba8dd28" volumeName="kubernetes.io/secret/09efc573-dbb6-4249-bd59-9b87aba8dd28-serving-cert" seLinuxMountContext="" Oct 08 18:17:10 crc kubenswrapper[4859]: I1008 18:17:10.424350 4859 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7bb08738-c794-4ee8-9972-3a62ca171029" volumeName="kubernetes.io/projected/7bb08738-c794-4ee8-9972-3a62ca171029-kube-api-access-279lb" seLinuxMountContext="" Oct 08 18:17:10 crc kubenswrapper[4859]: I1008 18:17:10.424380 4859 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8f668bae-612b-4b75-9490-919e737c6a3b" volumeName="kubernetes.io/empty-dir/8f668bae-612b-4b75-9490-919e737c6a3b-ca-trust-extracted" seLinuxMountContext="" Oct 08 18:17:10 crc kubenswrapper[4859]: I1008 18:17:10.424409 4859 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="a0128f3a-b052-44ed-a84e-c4c8aaf17c13" volumeName="kubernetes.io/secret/a0128f3a-b052-44ed-a84e-c4c8aaf17c13-samples-operator-tls" seLinuxMountContext="" Oct 08 18:17:10 crc kubenswrapper[4859]: I1008 18:17:10.424443 4859 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="22c825df-677d-4ca6-82db-3454ed06e783" volumeName="kubernetes.io/configmap/22c825df-677d-4ca6-82db-3454ed06e783-auth-proxy-config" seLinuxMountContext="" Oct 08 18:17:10 crc kubenswrapper[4859]: I1008 18:17:10.424473 4859 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="43509403-f426-496e-be36-56cef71462f5" volumeName="kubernetes.io/secret/43509403-f426-496e-be36-56cef71462f5-console-oauth-config" seLinuxMountContext="" Oct 08 18:17:10 crc kubenswrapper[4859]: I1008 18:17:10.424503 4859 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="496e6271-fb68-4057-954e-a0d97a4afa3f" volumeName="kubernetes.io/projected/496e6271-fb68-4057-954e-a0d97a4afa3f-kube-api-access" seLinuxMountContext="" Oct 08 18:17:10 crc kubenswrapper[4859]: I1008 18:17:10.424558 4859 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-ocp-branding-template" seLinuxMountContext="" Oct 08 18:17:10 crc kubenswrapper[4859]: I1008 18:17:10.424588 4859 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="4bb40260-dbaa-4fb0-84df-5e680505d512" volumeName="kubernetes.io/projected/4bb40260-dbaa-4fb0-84df-5e680505d512-kube-api-access-2w9zh" seLinuxMountContext="" Oct 08 18:17:10 crc kubenswrapper[4859]: I1008 18:17:10.424616 4859 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="57a731c4-ef35-47a8-b875-bfb08a7f8011" volumeName="kubernetes.io/empty-dir/57a731c4-ef35-47a8-b875-bfb08a7f8011-catalog-content" seLinuxMountContext="" Oct 08 18:17:10 crc kubenswrapper[4859]: I1008 18:17:10.424643 4859 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6ea678ab-3438-413e-bfe3-290ae7725660" volumeName="kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-env-overrides" seLinuxMountContext="" Oct 08 18:17:10 crc kubenswrapper[4859]: I1008 18:17:10.424788 4859 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8f668bae-612b-4b75-9490-919e737c6a3b" volumeName="kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-bound-sa-token" seLinuxMountContext="" Oct 08 18:17:10 crc kubenswrapper[4859]: I1008 18:17:10.424823 4859 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="22c825df-677d-4ca6-82db-3454ed06e783" volumeName="kubernetes.io/configmap/22c825df-677d-4ca6-82db-3454ed06e783-config" seLinuxMountContext="" Oct 08 18:17:10 crc kubenswrapper[4859]: I1008 18:17:10.424854 4859 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/projected/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-kube-api-access-ngvvp" seLinuxMountContext="" Oct 08 18:17:10 crc kubenswrapper[4859]: I1008 18:17:10.424883 4859 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-provider-selection" seLinuxMountContext="" Oct 08 18:17:10 crc kubenswrapper[4859]: I1008 18:17:10.424911 4859 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6509e943-70c6-444c-bc41-48a544e36fbd" volumeName="kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-config" seLinuxMountContext="" Oct 08 18:17:10 crc kubenswrapper[4859]: I1008 18:17:10.424937 4859 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6509e943-70c6-444c-bc41-48a544e36fbd" volumeName="kubernetes.io/secret/6509e943-70c6-444c-bc41-48a544e36fbd-serving-cert" seLinuxMountContext="" Oct 08 18:17:10 crc kubenswrapper[4859]: I1008 18:17:10.424964 4859 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8cea82b4-6893-4ddc-af9f-1bb5ae425c5b" volumeName="kubernetes.io/configmap/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-config" seLinuxMountContext="" Oct 08 18:17:10 crc kubenswrapper[4859]: I1008 18:17:10.424989 4859 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="c03ee662-fb2f-4fc4-a2c1-af487c19d254" volumeName="kubernetes.io/projected/c03ee662-fb2f-4fc4-a2c1-af487c19d254-kube-api-access-v47cf" seLinuxMountContext="" Oct 08 18:17:10 crc kubenswrapper[4859]: I1008 18:17:10.425014 4859 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="ef543e1b-8068-4ea3-b32a-61027b32e95d" volumeName="kubernetes.io/projected/ef543e1b-8068-4ea3-b32a-61027b32e95d-kube-api-access-s2kz5" seLinuxMountContext="" Oct 08 18:17:10 crc kubenswrapper[4859]: I1008 18:17:10.425080 4859 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09ae3b1a-e8e7-4524-b54b-61eab6f9239a" volumeName="kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-serving-cert" seLinuxMountContext="" Oct 08 18:17:10 crc kubenswrapper[4859]: I1008 18:17:10.425109 4859 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1bf7eb37-55a3-4c65-b768-a94c82151e69" volumeName="kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-config" seLinuxMountContext="" Oct 08 18:17:10 crc kubenswrapper[4859]: I1008 18:17:10.425137 4859 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="20b0d48f-5fd6-431c-a545-e3c800c7b866" volumeName="kubernetes.io/projected/20b0d48f-5fd6-431c-a545-e3c800c7b866-kube-api-access-w9rds" seLinuxMountContext="" Oct 08 18:17:10 crc kubenswrapper[4859]: I1008 18:17:10.425166 4859 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="31d8b7a1-420e-4252-a5b7-eebe8a111292" volumeName="kubernetes.io/projected/31d8b7a1-420e-4252-a5b7-eebe8a111292-kube-api-access-zgdk5" seLinuxMountContext="" Oct 08 18:17:10 crc kubenswrapper[4859]: I1008 18:17:10.425195 4859 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="43509403-f426-496e-be36-56cef71462f5" volumeName="kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-service-ca" seLinuxMountContext="" Oct 08 18:17:10 crc kubenswrapper[4859]: I1008 18:17:10.425221 4859 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" volumeName="kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert" seLinuxMountContext="" Oct 08 18:17:10 crc kubenswrapper[4859]: I1008 18:17:10.425248 4859 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6402fda4-df10-493c-b4e5-d0569419652d" volumeName="kubernetes.io/projected/6402fda4-df10-493c-b4e5-d0569419652d-kube-api-access-mg5zb" seLinuxMountContext="" Oct 08 18:17:10 crc kubenswrapper[4859]: I1008 18:17:10.425274 4859 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="9d4552c7-cd75-42dd-8880-30dd377c49a4" volumeName="kubernetes.io/projected/9d4552c7-cd75-42dd-8880-30dd377c49a4-kube-api-access-pcxfs" seLinuxMountContext="" Oct 08 18:17:10 crc kubenswrapper[4859]: I1008 18:17:10.425299 4859 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09efc573-dbb6-4249-bd59-9b87aba8dd28" volumeName="kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-ca" seLinuxMountContext="" Oct 08 18:17:10 crc kubenswrapper[4859]: I1008 18:17:10.425327 4859 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="d75a4c96-2883-4a0b-bab2-0fab2b6c0b49" volumeName="kubernetes.io/configmap/d75a4c96-2883-4a0b-bab2-0fab2b6c0b49-iptables-alerter-script" seLinuxMountContext="" Oct 08 18:17:10 crc kubenswrapper[4859]: I1008 18:17:10.425354 4859 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="a31745f5-9847-4afe-82a5-3161cc66ca93" volumeName="kubernetes.io/secret/a31745f5-9847-4afe-82a5-3161cc66ca93-metrics-tls" seLinuxMountContext="" Oct 08 18:17:10 crc kubenswrapper[4859]: I1008 18:17:10.425378 4859 reconstruct.go:97] "Volume reconstruction finished" Oct 08 18:17:10 crc kubenswrapper[4859]: I1008 18:17:10.425394 4859 reconciler.go:26] "Reconciler: start to sync state" Oct 08 18:17:10 crc kubenswrapper[4859]: I1008 18:17:10.437620 4859 manager.go:324] Recovery completed Oct 08 18:17:10 crc kubenswrapper[4859]: I1008 18:17:10.449489 4859 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 08 18:17:10 crc kubenswrapper[4859]: I1008 18:17:10.451187 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 18:17:10 crc kubenswrapper[4859]: I1008 18:17:10.451253 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 18:17:10 crc kubenswrapper[4859]: I1008 18:17:10.451268 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 18:17:10 crc kubenswrapper[4859]: I1008 18:17:10.452207 4859 cpu_manager.go:225] "Starting CPU manager" policy="none" Oct 08 18:17:10 crc kubenswrapper[4859]: I1008 18:17:10.452228 4859 cpu_manager.go:226] "Reconciling" reconcilePeriod="10s" Oct 08 18:17:10 crc kubenswrapper[4859]: I1008 18:17:10.452270 4859 state_mem.go:36] "Initialized new in-memory state store" Oct 08 18:17:10 crc kubenswrapper[4859]: I1008 18:17:10.465450 4859 kubelet_network_linux.go:50] "Initialized iptables rules." protocol="IPv4" Oct 08 18:17:10 crc kubenswrapper[4859]: I1008 18:17:10.467682 4859 kubelet_network_linux.go:50] "Initialized iptables rules." protocol="IPv6" Oct 08 18:17:10 crc kubenswrapper[4859]: I1008 18:17:10.468572 4859 status_manager.go:217] "Starting to sync pod status with apiserver" Oct 08 18:17:10 crc kubenswrapper[4859]: I1008 18:17:10.468609 4859 kubelet.go:2335] "Starting kubelet main sync loop" Oct 08 18:17:10 crc kubenswrapper[4859]: E1008 18:17:10.468654 4859 kubelet.go:2359] "Skipping pod synchronization" err="[container runtime status check may not have completed yet, PLEG is not healthy: pleg has yet to be successful]" Oct 08 18:17:10 crc kubenswrapper[4859]: W1008 18:17:10.475189 4859 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.RuntimeClass: Get "https://api-int.crc.testing:6443/apis/node.k8s.io/v1/runtimeclasses?limit=500&resourceVersion=0": dial tcp 38.102.83.13:6443: connect: connection refused Oct 08 18:17:10 crc kubenswrapper[4859]: E1008 18:17:10.475320 4859 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.RuntimeClass: failed to list *v1.RuntimeClass: Get \"https://api-int.crc.testing:6443/apis/node.k8s.io/v1/runtimeclasses?limit=500&resourceVersion=0\": dial tcp 38.102.83.13:6443: connect: connection refused" logger="UnhandledError" Oct 08 18:17:10 crc kubenswrapper[4859]: I1008 18:17:10.477052 4859 policy_none.go:49] "None policy: Start" Oct 08 18:17:10 crc kubenswrapper[4859]: I1008 18:17:10.479202 4859 memory_manager.go:170] "Starting memorymanager" policy="None" Oct 08 18:17:10 crc kubenswrapper[4859]: I1008 18:17:10.479260 4859 state_mem.go:35] "Initializing new in-memory state store" Oct 08 18:17:10 crc kubenswrapper[4859]: E1008 18:17:10.496686 4859 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Oct 08 18:17:10 crc kubenswrapper[4859]: I1008 18:17:10.536081 4859 manager.go:334] "Starting Device Plugin manager" Oct 08 18:17:10 crc kubenswrapper[4859]: I1008 18:17:10.536148 4859 manager.go:513] "Failed to read data from checkpoint" checkpoint="kubelet_internal_checkpoint" err="checkpoint is not found" Oct 08 18:17:10 crc kubenswrapper[4859]: I1008 18:17:10.536162 4859 server.go:79] "Starting device plugin registration server" Oct 08 18:17:10 crc kubenswrapper[4859]: I1008 18:17:10.536588 4859 eviction_manager.go:189] "Eviction manager: starting control loop" Oct 08 18:17:10 crc kubenswrapper[4859]: I1008 18:17:10.536601 4859 container_log_manager.go:189] "Initializing container log rotate workers" workers=1 monitorPeriod="10s" Oct 08 18:17:10 crc kubenswrapper[4859]: I1008 18:17:10.537072 4859 plugin_watcher.go:51] "Plugin Watcher Start" path="/var/lib/kubelet/plugins_registry" Oct 08 18:17:10 crc kubenswrapper[4859]: I1008 18:17:10.537178 4859 plugin_manager.go:116] "The desired_state_of_world populator (plugin watcher) starts" Oct 08 18:17:10 crc kubenswrapper[4859]: I1008 18:17:10.537188 4859 plugin_manager.go:118] "Starting Kubelet Plugin Manager" Oct 08 18:17:10 crc kubenswrapper[4859]: E1008 18:17:10.551431 4859 eviction_manager.go:285] "Eviction manager: failed to get summary stats" err="failed to get node info: node \"crc\" not found" Oct 08 18:17:10 crc kubenswrapper[4859]: I1008 18:17:10.569676 4859 kubelet.go:2421] "SyncLoop ADD" source="file" pods=["openshift-machine-config-operator/kube-rbac-proxy-crio-crc","openshift-etcd/etcd-crc","openshift-kube-apiserver/kube-apiserver-crc","openshift-kube-controller-manager/kube-controller-manager-crc","openshift-kube-scheduler/openshift-kube-scheduler-crc"] Oct 08 18:17:10 crc kubenswrapper[4859]: I1008 18:17:10.569914 4859 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 08 18:17:10 crc kubenswrapper[4859]: I1008 18:17:10.571281 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 18:17:10 crc kubenswrapper[4859]: I1008 18:17:10.571337 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 18:17:10 crc kubenswrapper[4859]: I1008 18:17:10.571348 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 18:17:10 crc kubenswrapper[4859]: I1008 18:17:10.571533 4859 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 08 18:17:10 crc kubenswrapper[4859]: I1008 18:17:10.572178 4859 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" Oct 08 18:17:10 crc kubenswrapper[4859]: I1008 18:17:10.572250 4859 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 08 18:17:10 crc kubenswrapper[4859]: I1008 18:17:10.572489 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 18:17:10 crc kubenswrapper[4859]: I1008 18:17:10.572533 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 18:17:10 crc kubenswrapper[4859]: I1008 18:17:10.572549 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 18:17:10 crc kubenswrapper[4859]: I1008 18:17:10.572750 4859 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 08 18:17:10 crc kubenswrapper[4859]: I1008 18:17:10.573083 4859 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-etcd/etcd-crc" Oct 08 18:17:10 crc kubenswrapper[4859]: I1008 18:17:10.573132 4859 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 08 18:17:10 crc kubenswrapper[4859]: I1008 18:17:10.573542 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 18:17:10 crc kubenswrapper[4859]: I1008 18:17:10.573581 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 18:17:10 crc kubenswrapper[4859]: I1008 18:17:10.573598 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 18:17:10 crc kubenswrapper[4859]: I1008 18:17:10.573811 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 18:17:10 crc kubenswrapper[4859]: I1008 18:17:10.573831 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 18:17:10 crc kubenswrapper[4859]: I1008 18:17:10.573843 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 18:17:10 crc kubenswrapper[4859]: I1008 18:17:10.574009 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 18:17:10 crc kubenswrapper[4859]: I1008 18:17:10.574020 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 18:17:10 crc kubenswrapper[4859]: I1008 18:17:10.574029 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 18:17:10 crc kubenswrapper[4859]: I1008 18:17:10.574142 4859 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 08 18:17:10 crc kubenswrapper[4859]: I1008 18:17:10.574385 4859 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/kube-apiserver-crc" Oct 08 18:17:10 crc kubenswrapper[4859]: I1008 18:17:10.574434 4859 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 08 18:17:10 crc kubenswrapper[4859]: I1008 18:17:10.574956 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 18:17:10 crc kubenswrapper[4859]: I1008 18:17:10.574984 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 18:17:10 crc kubenswrapper[4859]: I1008 18:17:10.574993 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 18:17:10 crc kubenswrapper[4859]: I1008 18:17:10.575118 4859 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 08 18:17:10 crc kubenswrapper[4859]: I1008 18:17:10.575380 4859 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Oct 08 18:17:10 crc kubenswrapper[4859]: I1008 18:17:10.575416 4859 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 08 18:17:10 crc kubenswrapper[4859]: I1008 18:17:10.578541 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 18:17:10 crc kubenswrapper[4859]: I1008 18:17:10.578568 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 18:17:10 crc kubenswrapper[4859]: I1008 18:17:10.578578 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 18:17:10 crc kubenswrapper[4859]: I1008 18:17:10.579853 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 18:17:10 crc kubenswrapper[4859]: I1008 18:17:10.579877 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 18:17:10 crc kubenswrapper[4859]: I1008 18:17:10.579889 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 18:17:10 crc kubenswrapper[4859]: I1008 18:17:10.580121 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 18:17:10 crc kubenswrapper[4859]: I1008 18:17:10.580169 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 18:17:10 crc kubenswrapper[4859]: I1008 18:17:10.580187 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 18:17:10 crc kubenswrapper[4859]: I1008 18:17:10.580494 4859 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Oct 08 18:17:10 crc kubenswrapper[4859]: I1008 18:17:10.580547 4859 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 08 18:17:10 crc kubenswrapper[4859]: I1008 18:17:10.581662 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 18:17:10 crc kubenswrapper[4859]: I1008 18:17:10.581683 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 18:17:10 crc kubenswrapper[4859]: I1008 18:17:10.581698 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 18:17:10 crc kubenswrapper[4859]: E1008 18:17:10.606767 4859 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.13:6443: connect: connection refused" interval="400ms" Oct 08 18:17:10 crc kubenswrapper[4859]: I1008 18:17:10.628492 4859 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/3dcd261975c3d6b9a6ad6367fd4facd3-cert-dir\") pod \"openshift-kube-scheduler-crc\" (UID: \"3dcd261975c3d6b9a6ad6367fd4facd3\") " pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Oct 08 18:17:10 crc kubenswrapper[4859]: I1008 18:17:10.628540 4859 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-cert-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Oct 08 18:17:10 crc kubenswrapper[4859]: I1008 18:17:10.628587 4859 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/f614b9022728cf315e60c057852e563e-cert-dir\") pod \"kube-controller-manager-crc\" (UID: \"f614b9022728cf315e60c057852e563e\") " pod="openshift-kube-controller-manager/kube-controller-manager-crc" Oct 08 18:17:10 crc kubenswrapper[4859]: I1008 18:17:10.628722 4859 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/3dcd261975c3d6b9a6ad6367fd4facd3-resource-dir\") pod \"openshift-kube-scheduler-crc\" (UID: \"3dcd261975c3d6b9a6ad6367fd4facd3\") " pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Oct 08 18:17:10 crc kubenswrapper[4859]: I1008 18:17:10.628762 4859 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-cert-dir\") pod \"kube-apiserver-crc\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Oct 08 18:17:10 crc kubenswrapper[4859]: I1008 18:17:10.628782 4859 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-audit-dir\") pod \"kube-apiserver-crc\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Oct 08 18:17:10 crc kubenswrapper[4859]: I1008 18:17:10.628800 4859 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-lib-kubelet\" (UniqueName: \"kubernetes.io/host-path/d1b160f5dda77d281dd8e69ec8d817f9-var-lib-kubelet\") pod \"kube-rbac-proxy-crio-crc\" (UID: \"d1b160f5dda77d281dd8e69ec8d817f9\") " pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" Oct 08 18:17:10 crc kubenswrapper[4859]: I1008 18:17:10.628819 4859 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"data-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-data-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Oct 08 18:17:10 crc kubenswrapper[4859]: I1008 18:17:10.628838 4859 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-resource-dir\") pod \"kube-apiserver-crc\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Oct 08 18:17:10 crc kubenswrapper[4859]: I1008 18:17:10.628853 4859 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"usr-local-bin\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-usr-local-bin\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Oct 08 18:17:10 crc kubenswrapper[4859]: I1008 18:17:10.628869 4859 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-log-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Oct 08 18:17:10 crc kubenswrapper[4859]: I1008 18:17:10.628885 4859 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f614b9022728cf315e60c057852e563e-resource-dir\") pod \"kube-controller-manager-crc\" (UID: \"f614b9022728cf315e60c057852e563e\") " pod="openshift-kube-controller-manager/kube-controller-manager-crc" Oct 08 18:17:10 crc kubenswrapper[4859]: I1008 18:17:10.628902 4859 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-kube\" (UniqueName: \"kubernetes.io/host-path/d1b160f5dda77d281dd8e69ec8d817f9-etc-kube\") pod \"kube-rbac-proxy-crio-crc\" (UID: \"d1b160f5dda77d281dd8e69ec8d817f9\") " pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" Oct 08 18:17:10 crc kubenswrapper[4859]: I1008 18:17:10.628918 4859 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"static-pod-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-static-pod-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Oct 08 18:17:10 crc kubenswrapper[4859]: I1008 18:17:10.628977 4859 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-resource-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Oct 08 18:17:10 crc kubenswrapper[4859]: I1008 18:17:10.637237 4859 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 08 18:17:10 crc kubenswrapper[4859]: I1008 18:17:10.639448 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 18:17:10 crc kubenswrapper[4859]: I1008 18:17:10.639507 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 18:17:10 crc kubenswrapper[4859]: I1008 18:17:10.639521 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 18:17:10 crc kubenswrapper[4859]: I1008 18:17:10.639557 4859 kubelet_node_status.go:76] "Attempting to register node" node="crc" Oct 08 18:17:10 crc kubenswrapper[4859]: E1008 18:17:10.640295 4859 kubelet_node_status.go:99] "Unable to register node with API server" err="Post \"https://api-int.crc.testing:6443/api/v1/nodes\": dial tcp 38.102.83.13:6443: connect: connection refused" node="crc" Oct 08 18:17:10 crc kubenswrapper[4859]: I1008 18:17:10.730800 4859 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-cert-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Oct 08 18:17:10 crc kubenswrapper[4859]: I1008 18:17:10.730880 4859 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/f614b9022728cf315e60c057852e563e-cert-dir\") pod \"kube-controller-manager-crc\" (UID: \"f614b9022728cf315e60c057852e563e\") " pod="openshift-kube-controller-manager/kube-controller-manager-crc" Oct 08 18:17:10 crc kubenswrapper[4859]: I1008 18:17:10.730903 4859 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/3dcd261975c3d6b9a6ad6367fd4facd3-resource-dir\") pod \"openshift-kube-scheduler-crc\" (UID: \"3dcd261975c3d6b9a6ad6367fd4facd3\") " pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Oct 08 18:17:10 crc kubenswrapper[4859]: I1008 18:17:10.730921 4859 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/3dcd261975c3d6b9a6ad6367fd4facd3-cert-dir\") pod \"openshift-kube-scheduler-crc\" (UID: \"3dcd261975c3d6b9a6ad6367fd4facd3\") " pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Oct 08 18:17:10 crc kubenswrapper[4859]: I1008 18:17:10.730941 4859 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-lib-kubelet\" (UniqueName: \"kubernetes.io/host-path/d1b160f5dda77d281dd8e69ec8d817f9-var-lib-kubelet\") pod \"kube-rbac-proxy-crio-crc\" (UID: \"d1b160f5dda77d281dd8e69ec8d817f9\") " pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" Oct 08 18:17:10 crc kubenswrapper[4859]: I1008 18:17:10.730965 4859 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"data-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-data-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Oct 08 18:17:10 crc kubenswrapper[4859]: I1008 18:17:10.730983 4859 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-resource-dir\") pod \"kube-apiserver-crc\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Oct 08 18:17:10 crc kubenswrapper[4859]: I1008 18:17:10.730997 4859 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-cert-dir\") pod \"kube-apiserver-crc\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Oct 08 18:17:10 crc kubenswrapper[4859]: I1008 18:17:10.731012 4859 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-audit-dir\") pod \"kube-apiserver-crc\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Oct 08 18:17:10 crc kubenswrapper[4859]: I1008 18:17:10.731027 4859 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-kube\" (UniqueName: \"kubernetes.io/host-path/d1b160f5dda77d281dd8e69ec8d817f9-etc-kube\") pod \"kube-rbac-proxy-crio-crc\" (UID: \"d1b160f5dda77d281dd8e69ec8d817f9\") " pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" Oct 08 18:17:10 crc kubenswrapper[4859]: I1008 18:17:10.731043 4859 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"static-pod-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-static-pod-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Oct 08 18:17:10 crc kubenswrapper[4859]: I1008 18:17:10.731058 4859 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-resource-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Oct 08 18:17:10 crc kubenswrapper[4859]: I1008 18:17:10.731074 4859 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"usr-local-bin\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-usr-local-bin\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Oct 08 18:17:10 crc kubenswrapper[4859]: I1008 18:17:10.731088 4859 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-log-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Oct 08 18:17:10 crc kubenswrapper[4859]: I1008 18:17:10.731073 4859 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/3dcd261975c3d6b9a6ad6367fd4facd3-cert-dir\") pod \"openshift-kube-scheduler-crc\" (UID: \"3dcd261975c3d6b9a6ad6367fd4facd3\") " pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Oct 08 18:17:10 crc kubenswrapper[4859]: I1008 18:17:10.731136 4859 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f614b9022728cf315e60c057852e563e-resource-dir\") pod \"kube-controller-manager-crc\" (UID: \"f614b9022728cf315e60c057852e563e\") " pod="openshift-kube-controller-manager/kube-controller-manager-crc" Oct 08 18:17:10 crc kubenswrapper[4859]: I1008 18:17:10.731142 4859 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/3dcd261975c3d6b9a6ad6367fd4facd3-resource-dir\") pod \"openshift-kube-scheduler-crc\" (UID: \"3dcd261975c3d6b9a6ad6367fd4facd3\") " pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Oct 08 18:17:10 crc kubenswrapper[4859]: I1008 18:17:10.731223 4859 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"usr-local-bin\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-usr-local-bin\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Oct 08 18:17:10 crc kubenswrapper[4859]: I1008 18:17:10.731103 4859 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f614b9022728cf315e60c057852e563e-resource-dir\") pod \"kube-controller-manager-crc\" (UID: \"f614b9022728cf315e60c057852e563e\") " pod="openshift-kube-controller-manager/kube-controller-manager-crc" Oct 08 18:17:10 crc kubenswrapper[4859]: I1008 18:17:10.731244 4859 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-kube\" (UniqueName: \"kubernetes.io/host-path/d1b160f5dda77d281dd8e69ec8d817f9-etc-kube\") pod \"kube-rbac-proxy-crio-crc\" (UID: \"d1b160f5dda77d281dd8e69ec8d817f9\") " pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" Oct 08 18:17:10 crc kubenswrapper[4859]: I1008 18:17:10.731252 4859 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-resource-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Oct 08 18:17:10 crc kubenswrapper[4859]: I1008 18:17:10.731309 4859 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"static-pod-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-static-pod-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Oct 08 18:17:10 crc kubenswrapper[4859]: I1008 18:17:10.731316 4859 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"data-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-data-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Oct 08 18:17:10 crc kubenswrapper[4859]: I1008 18:17:10.731340 4859 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-lib-kubelet\" (UniqueName: \"kubernetes.io/host-path/d1b160f5dda77d281dd8e69ec8d817f9-var-lib-kubelet\") pod \"kube-rbac-proxy-crio-crc\" (UID: \"d1b160f5dda77d281dd8e69ec8d817f9\") " pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" Oct 08 18:17:10 crc kubenswrapper[4859]: I1008 18:17:10.731351 4859 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-resource-dir\") pod \"kube-apiserver-crc\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Oct 08 18:17:10 crc kubenswrapper[4859]: I1008 18:17:10.731347 4859 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-log-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Oct 08 18:17:10 crc kubenswrapper[4859]: I1008 18:17:10.731379 4859 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-audit-dir\") pod \"kube-apiserver-crc\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Oct 08 18:17:10 crc kubenswrapper[4859]: I1008 18:17:10.731405 4859 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-cert-dir\") pod \"kube-apiserver-crc\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Oct 08 18:17:10 crc kubenswrapper[4859]: I1008 18:17:10.731436 4859 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-cert-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Oct 08 18:17:10 crc kubenswrapper[4859]: I1008 18:17:10.731438 4859 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/f614b9022728cf315e60c057852e563e-cert-dir\") pod \"kube-controller-manager-crc\" (UID: \"f614b9022728cf315e60c057852e563e\") " pod="openshift-kube-controller-manager/kube-controller-manager-crc" Oct 08 18:17:10 crc kubenswrapper[4859]: I1008 18:17:10.841432 4859 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 08 18:17:10 crc kubenswrapper[4859]: I1008 18:17:10.842967 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 18:17:10 crc kubenswrapper[4859]: I1008 18:17:10.843023 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 18:17:10 crc kubenswrapper[4859]: I1008 18:17:10.843049 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 18:17:10 crc kubenswrapper[4859]: I1008 18:17:10.843078 4859 kubelet_node_status.go:76] "Attempting to register node" node="crc" Oct 08 18:17:10 crc kubenswrapper[4859]: E1008 18:17:10.843753 4859 kubelet_node_status.go:99] "Unable to register node with API server" err="Post \"https://api-int.crc.testing:6443/api/v1/nodes\": dial tcp 38.102.83.13:6443: connect: connection refused" node="crc" Oct 08 18:17:10 crc kubenswrapper[4859]: I1008 18:17:10.902474 4859 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" Oct 08 18:17:10 crc kubenswrapper[4859]: I1008 18:17:10.909226 4859 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-etcd/etcd-crc" Oct 08 18:17:10 crc kubenswrapper[4859]: I1008 18:17:10.934428 4859 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/kube-apiserver-crc" Oct 08 18:17:10 crc kubenswrapper[4859]: I1008 18:17:10.950044 4859 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Oct 08 18:17:10 crc kubenswrapper[4859]: W1008 18:17:10.950752 4859 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podd1b160f5dda77d281dd8e69ec8d817f9.slice/crio-f3c262728c803c954266d7185d873dd91f6756027359b21e941b18d351fa4fad WatchSource:0}: Error finding container f3c262728c803c954266d7185d873dd91f6756027359b21e941b18d351fa4fad: Status 404 returned error can't find the container with id f3c262728c803c954266d7185d873dd91f6756027359b21e941b18d351fa4fad Oct 08 18:17:10 crc kubenswrapper[4859]: W1008 18:17:10.951251 4859 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod2139d3e2895fc6797b9c76a1b4c9886d.slice/crio-5e23dbb910fe75669413df14077df09d30580cda656f56c2a2bfe0b5b87f44e9 WatchSource:0}: Error finding container 5e23dbb910fe75669413df14077df09d30580cda656f56c2a2bfe0b5b87f44e9: Status 404 returned error can't find the container with id 5e23dbb910fe75669413df14077df09d30580cda656f56c2a2bfe0b5b87f44e9 Oct 08 18:17:10 crc kubenswrapper[4859]: I1008 18:17:10.959660 4859 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Oct 08 18:17:10 crc kubenswrapper[4859]: W1008 18:17:10.964833 4859 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podf4b27818a5e8e43d0dc095d08835c792.slice/crio-27945c515600d808b489295f41a552520abd6fe3f9f504d302422474ebafc8a2 WatchSource:0}: Error finding container 27945c515600d808b489295f41a552520abd6fe3f9f504d302422474ebafc8a2: Status 404 returned error can't find the container with id 27945c515600d808b489295f41a552520abd6fe3f9f504d302422474ebafc8a2 Oct 08 18:17:10 crc kubenswrapper[4859]: W1008 18:17:10.970484 4859 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podf614b9022728cf315e60c057852e563e.slice/crio-a64e2c8e19c4c04f22aeb259135b096131d3255fb309e1aa6511ba913fbf9e83 WatchSource:0}: Error finding container a64e2c8e19c4c04f22aeb259135b096131d3255fb309e1aa6511ba913fbf9e83: Status 404 returned error can't find the container with id a64e2c8e19c4c04f22aeb259135b096131d3255fb309e1aa6511ba913fbf9e83 Oct 08 18:17:10 crc kubenswrapper[4859]: W1008 18:17:10.980288 4859 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod3dcd261975c3d6b9a6ad6367fd4facd3.slice/crio-24b9c05b0f109ad38e7b1a5bfea8f84dcf964415bb221aaaf266196284b14ee1 WatchSource:0}: Error finding container 24b9c05b0f109ad38e7b1a5bfea8f84dcf964415bb221aaaf266196284b14ee1: Status 404 returned error can't find the container with id 24b9c05b0f109ad38e7b1a5bfea8f84dcf964415bb221aaaf266196284b14ee1 Oct 08 18:17:11 crc kubenswrapper[4859]: E1008 18:17:11.008959 4859 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.13:6443: connect: connection refused" interval="800ms" Oct 08 18:17:11 crc kubenswrapper[4859]: I1008 18:17:11.244858 4859 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 08 18:17:11 crc kubenswrapper[4859]: I1008 18:17:11.246455 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 18:17:11 crc kubenswrapper[4859]: I1008 18:17:11.246507 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 18:17:11 crc kubenswrapper[4859]: I1008 18:17:11.246521 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 18:17:11 crc kubenswrapper[4859]: I1008 18:17:11.246551 4859 kubelet_node_status.go:76] "Attempting to register node" node="crc" Oct 08 18:17:11 crc kubenswrapper[4859]: E1008 18:17:11.247177 4859 kubelet_node_status.go:99] "Unable to register node with API server" err="Post \"https://api-int.crc.testing:6443/api/v1/nodes\": dial tcp 38.102.83.13:6443: connect: connection refused" node="crc" Oct 08 18:17:11 crc kubenswrapper[4859]: W1008 18:17:11.261090 4859 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.CSIDriver: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csidrivers?limit=500&resourceVersion=0": dial tcp 38.102.83.13:6443: connect: connection refused Oct 08 18:17:11 crc kubenswrapper[4859]: E1008 18:17:11.261194 4859 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.CSIDriver: failed to list *v1.CSIDriver: Get \"https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csidrivers?limit=500&resourceVersion=0\": dial tcp 38.102.83.13:6443: connect: connection refused" logger="UnhandledError" Oct 08 18:17:11 crc kubenswrapper[4859]: I1008 18:17:11.394256 4859 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csinodes/crc?resourceVersion=0": dial tcp 38.102.83.13:6443: connect: connection refused Oct 08 18:17:11 crc kubenswrapper[4859]: I1008 18:17:11.473913 4859 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" event={"ID":"f614b9022728cf315e60c057852e563e","Type":"ContainerStarted","Data":"a64e2c8e19c4c04f22aeb259135b096131d3255fb309e1aa6511ba913fbf9e83"} Oct 08 18:17:11 crc kubenswrapper[4859]: I1008 18:17:11.474721 4859 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerStarted","Data":"27945c515600d808b489295f41a552520abd6fe3f9f504d302422474ebafc8a2"} Oct 08 18:17:11 crc kubenswrapper[4859]: I1008 18:17:11.475272 4859 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd/etcd-crc" event={"ID":"2139d3e2895fc6797b9c76a1b4c9886d","Type":"ContainerStarted","Data":"5e23dbb910fe75669413df14077df09d30580cda656f56c2a2bfe0b5b87f44e9"} Oct 08 18:17:11 crc kubenswrapper[4859]: I1008 18:17:11.475920 4859 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" event={"ID":"d1b160f5dda77d281dd8e69ec8d817f9","Type":"ContainerStarted","Data":"f3c262728c803c954266d7185d873dd91f6756027359b21e941b18d351fa4fad"} Oct 08 18:17:11 crc kubenswrapper[4859]: I1008 18:17:11.476660 4859 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" event={"ID":"3dcd261975c3d6b9a6ad6367fd4facd3","Type":"ContainerStarted","Data":"24b9c05b0f109ad38e7b1a5bfea8f84dcf964415bb221aaaf266196284b14ee1"} Oct 08 18:17:11 crc kubenswrapper[4859]: W1008 18:17:11.641453 4859 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.Node: Get "https://api-int.crc.testing:6443/api/v1/nodes?fieldSelector=metadata.name%3Dcrc&limit=500&resourceVersion=0": dial tcp 38.102.83.13:6443: connect: connection refused Oct 08 18:17:11 crc kubenswrapper[4859]: E1008 18:17:11.641994 4859 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.Node: failed to list *v1.Node: Get \"https://api-int.crc.testing:6443/api/v1/nodes?fieldSelector=metadata.name%3Dcrc&limit=500&resourceVersion=0\": dial tcp 38.102.83.13:6443: connect: connection refused" logger="UnhandledError" Oct 08 18:17:11 crc kubenswrapper[4859]: E1008 18:17:11.810790 4859 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.13:6443: connect: connection refused" interval="1.6s" Oct 08 18:17:11 crc kubenswrapper[4859]: W1008 18:17:11.976637 4859 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.Service: Get "https://api-int.crc.testing:6443/api/v1/services?fieldSelector=spec.clusterIP%21%3DNone&limit=500&resourceVersion=0": dial tcp 38.102.83.13:6443: connect: connection refused Oct 08 18:17:11 crc kubenswrapper[4859]: E1008 18:17:11.976767 4859 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.Service: failed to list *v1.Service: Get \"https://api-int.crc.testing:6443/api/v1/services?fieldSelector=spec.clusterIP%21%3DNone&limit=500&resourceVersion=0\": dial tcp 38.102.83.13:6443: connect: connection refused" logger="UnhandledError" Oct 08 18:17:12 crc kubenswrapper[4859]: W1008 18:17:12.002750 4859 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.RuntimeClass: Get "https://api-int.crc.testing:6443/apis/node.k8s.io/v1/runtimeclasses?limit=500&resourceVersion=0": dial tcp 38.102.83.13:6443: connect: connection refused Oct 08 18:17:12 crc kubenswrapper[4859]: E1008 18:17:12.002818 4859 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.RuntimeClass: failed to list *v1.RuntimeClass: Get \"https://api-int.crc.testing:6443/apis/node.k8s.io/v1/runtimeclasses?limit=500&resourceVersion=0\": dial tcp 38.102.83.13:6443: connect: connection refused" logger="UnhandledError" Oct 08 18:17:12 crc kubenswrapper[4859]: I1008 18:17:12.047317 4859 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 08 18:17:12 crc kubenswrapper[4859]: I1008 18:17:12.049255 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 18:17:12 crc kubenswrapper[4859]: I1008 18:17:12.049305 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 18:17:12 crc kubenswrapper[4859]: I1008 18:17:12.049321 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 18:17:12 crc kubenswrapper[4859]: I1008 18:17:12.049356 4859 kubelet_node_status.go:76] "Attempting to register node" node="crc" Oct 08 18:17:12 crc kubenswrapper[4859]: E1008 18:17:12.049746 4859 kubelet_node_status.go:99] "Unable to register node with API server" err="Post \"https://api-int.crc.testing:6443/api/v1/nodes\": dial tcp 38.102.83.13:6443: connect: connection refused" node="crc" Oct 08 18:17:12 crc kubenswrapper[4859]: I1008 18:17:12.394933 4859 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csinodes/crc?resourceVersion=0": dial tcp 38.102.83.13:6443: connect: connection refused Oct 08 18:17:12 crc kubenswrapper[4859]: I1008 18:17:12.482399 4859 generic.go:334] "Generic (PLEG): container finished" podID="3dcd261975c3d6b9a6ad6367fd4facd3" containerID="008757fce552591f7c58193f605eb60792761d5943927cbd63184549618d2b43" exitCode=0 Oct 08 18:17:12 crc kubenswrapper[4859]: I1008 18:17:12.482533 4859 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" event={"ID":"3dcd261975c3d6b9a6ad6367fd4facd3","Type":"ContainerDied","Data":"008757fce552591f7c58193f605eb60792761d5943927cbd63184549618d2b43"} Oct 08 18:17:12 crc kubenswrapper[4859]: I1008 18:17:12.482773 4859 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 08 18:17:12 crc kubenswrapper[4859]: I1008 18:17:12.484416 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 18:17:12 crc kubenswrapper[4859]: I1008 18:17:12.484504 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 18:17:12 crc kubenswrapper[4859]: I1008 18:17:12.484522 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 18:17:12 crc kubenswrapper[4859]: I1008 18:17:12.490127 4859 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" event={"ID":"f614b9022728cf315e60c057852e563e","Type":"ContainerStarted","Data":"6411d1e7e18d5ceadaaceed5a8a6b20a0cacc55cede793ca9fb4dcad70c64fc6"} Oct 08 18:17:12 crc kubenswrapper[4859]: I1008 18:17:12.490168 4859 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" event={"ID":"f614b9022728cf315e60c057852e563e","Type":"ContainerStarted","Data":"bae6ecfd8b446ef04d393d823e5ffe37fe749337c0ca17fde79c5f88d5e31393"} Oct 08 18:17:12 crc kubenswrapper[4859]: I1008 18:17:12.490181 4859 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" event={"ID":"f614b9022728cf315e60c057852e563e","Type":"ContainerStarted","Data":"aa8f8ea895272cd79a08d1c47cd7863a1ec10e21bb60d12aa10862a56c99e4bf"} Oct 08 18:17:12 crc kubenswrapper[4859]: I1008 18:17:12.490192 4859 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" event={"ID":"f614b9022728cf315e60c057852e563e","Type":"ContainerStarted","Data":"5c4e95d8a712ab219455564539a90286ae48bbc46ddc322a5bfd96c1a8268110"} Oct 08 18:17:12 crc kubenswrapper[4859]: I1008 18:17:12.490269 4859 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 08 18:17:12 crc kubenswrapper[4859]: I1008 18:17:12.491629 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 18:17:12 crc kubenswrapper[4859]: I1008 18:17:12.491694 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 18:17:12 crc kubenswrapper[4859]: I1008 18:17:12.491736 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 18:17:12 crc kubenswrapper[4859]: I1008 18:17:12.496204 4859 generic.go:334] "Generic (PLEG): container finished" podID="f4b27818a5e8e43d0dc095d08835c792" containerID="58c7bba4ace28215b062653a960e561b1dc1f890f2baf450145638a065af038b" exitCode=0 Oct 08 18:17:12 crc kubenswrapper[4859]: I1008 18:17:12.496368 4859 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerDied","Data":"58c7bba4ace28215b062653a960e561b1dc1f890f2baf450145638a065af038b"} Oct 08 18:17:12 crc kubenswrapper[4859]: I1008 18:17:12.496628 4859 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 08 18:17:12 crc kubenswrapper[4859]: I1008 18:17:12.498451 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 18:17:12 crc kubenswrapper[4859]: I1008 18:17:12.498523 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 18:17:12 crc kubenswrapper[4859]: I1008 18:17:12.498546 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 18:17:12 crc kubenswrapper[4859]: I1008 18:17:12.500280 4859 generic.go:334] "Generic (PLEG): container finished" podID="2139d3e2895fc6797b9c76a1b4c9886d" containerID="777599aefe39cd1f4576751573cac8df3ec9c3d55dcefda3334eaaa2d578643c" exitCode=0 Oct 08 18:17:12 crc kubenswrapper[4859]: I1008 18:17:12.500537 4859 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd/etcd-crc" event={"ID":"2139d3e2895fc6797b9c76a1b4c9886d","Type":"ContainerDied","Data":"777599aefe39cd1f4576751573cac8df3ec9c3d55dcefda3334eaaa2d578643c"} Oct 08 18:17:12 crc kubenswrapper[4859]: I1008 18:17:12.500614 4859 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 08 18:17:12 crc kubenswrapper[4859]: I1008 18:17:12.500984 4859 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 08 18:17:12 crc kubenswrapper[4859]: I1008 18:17:12.503726 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 18:17:12 crc kubenswrapper[4859]: I1008 18:17:12.503760 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 18:17:12 crc kubenswrapper[4859]: I1008 18:17:12.503773 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 18:17:12 crc kubenswrapper[4859]: I1008 18:17:12.503734 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 18:17:12 crc kubenswrapper[4859]: I1008 18:17:12.503855 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 18:17:12 crc kubenswrapper[4859]: I1008 18:17:12.503872 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 18:17:12 crc kubenswrapper[4859]: I1008 18:17:12.505202 4859 generic.go:334] "Generic (PLEG): container finished" podID="d1b160f5dda77d281dd8e69ec8d817f9" containerID="3ab3bc5e8e2fd4941f30cbf8a90d9a9078b90fab91063c07da2cd48b63eef657" exitCode=0 Oct 08 18:17:12 crc kubenswrapper[4859]: I1008 18:17:12.505292 4859 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 08 18:17:12 crc kubenswrapper[4859]: I1008 18:17:12.505326 4859 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" event={"ID":"d1b160f5dda77d281dd8e69ec8d817f9","Type":"ContainerDied","Data":"3ab3bc5e8e2fd4941f30cbf8a90d9a9078b90fab91063c07da2cd48b63eef657"} Oct 08 18:17:12 crc kubenswrapper[4859]: I1008 18:17:12.506399 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 18:17:12 crc kubenswrapper[4859]: I1008 18:17:12.506469 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 18:17:12 crc kubenswrapper[4859]: I1008 18:17:12.506486 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 18:17:12 crc kubenswrapper[4859]: I1008 18:17:12.513028 4859 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Oct 08 18:17:13 crc kubenswrapper[4859]: I1008 18:17:13.135240 4859 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Oct 08 18:17:13 crc kubenswrapper[4859]: I1008 18:17:13.394235 4859 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csinodes/crc?resourceVersion=0": dial tcp 38.102.83.13:6443: connect: connection refused Oct 08 18:17:13 crc kubenswrapper[4859]: W1008 18:17:13.408656 4859 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.CSIDriver: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csidrivers?limit=500&resourceVersion=0": dial tcp 38.102.83.13:6443: connect: connection refused Oct 08 18:17:13 crc kubenswrapper[4859]: E1008 18:17:13.408818 4859 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.CSIDriver: failed to list *v1.CSIDriver: Get \"https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csidrivers?limit=500&resourceVersion=0\": dial tcp 38.102.83.13:6443: connect: connection refused" logger="UnhandledError" Oct 08 18:17:13 crc kubenswrapper[4859]: E1008 18:17:13.411777 4859 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.13:6443: connect: connection refused" interval="3.2s" Oct 08 18:17:13 crc kubenswrapper[4859]: I1008 18:17:13.513246 4859 generic.go:334] "Generic (PLEG): container finished" podID="2139d3e2895fc6797b9c76a1b4c9886d" containerID="c988c5fe9b425e35e0fb77cf5918463e4b51b3b47848fb31f0d34e5a7c49ce45" exitCode=0 Oct 08 18:17:13 crc kubenswrapper[4859]: I1008 18:17:13.513297 4859 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd/etcd-crc" event={"ID":"2139d3e2895fc6797b9c76a1b4c9886d","Type":"ContainerDied","Data":"c988c5fe9b425e35e0fb77cf5918463e4b51b3b47848fb31f0d34e5a7c49ce45"} Oct 08 18:17:13 crc kubenswrapper[4859]: I1008 18:17:13.513471 4859 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 08 18:17:13 crc kubenswrapper[4859]: I1008 18:17:13.515137 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 18:17:13 crc kubenswrapper[4859]: I1008 18:17:13.515183 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 18:17:13 crc kubenswrapper[4859]: I1008 18:17:13.515197 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 18:17:13 crc kubenswrapper[4859]: I1008 18:17:13.516783 4859 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 08 18:17:13 crc kubenswrapper[4859]: I1008 18:17:13.516774 4859 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" event={"ID":"d1b160f5dda77d281dd8e69ec8d817f9","Type":"ContainerStarted","Data":"a54a54b9a831ead1fbe6e03abbb622242dd16b4081a4482d559e03d47aebd2fe"} Oct 08 18:17:13 crc kubenswrapper[4859]: I1008 18:17:13.517506 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 18:17:13 crc kubenswrapper[4859]: I1008 18:17:13.518148 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 18:17:13 crc kubenswrapper[4859]: I1008 18:17:13.518189 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 18:17:13 crc kubenswrapper[4859]: I1008 18:17:13.520226 4859 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" event={"ID":"3dcd261975c3d6b9a6ad6367fd4facd3","Type":"ContainerStarted","Data":"a41399a37b1ad86a063001224232ecb8f7ea56aeb5686d3858bc08102a0c8f49"} Oct 08 18:17:13 crc kubenswrapper[4859]: I1008 18:17:13.520257 4859 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" event={"ID":"3dcd261975c3d6b9a6ad6367fd4facd3","Type":"ContainerStarted","Data":"428cbe63daefbfbcb2ec76dad0fe638cdc65934bac80b791cebe0b36b2d6909a"} Oct 08 18:17:13 crc kubenswrapper[4859]: I1008 18:17:13.520268 4859 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" event={"ID":"3dcd261975c3d6b9a6ad6367fd4facd3","Type":"ContainerStarted","Data":"f587e149d57ea53626ed3beb83e421c3b89ba1fec9e949705e8ccb5b4f7cc79a"} Oct 08 18:17:13 crc kubenswrapper[4859]: I1008 18:17:13.520347 4859 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 08 18:17:13 crc kubenswrapper[4859]: I1008 18:17:13.525185 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 18:17:13 crc kubenswrapper[4859]: I1008 18:17:13.525213 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 18:17:13 crc kubenswrapper[4859]: I1008 18:17:13.525226 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 18:17:13 crc kubenswrapper[4859]: I1008 18:17:13.531207 4859 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerStarted","Data":"c65b66f62d325ca10ef8d5d31f04529ccf4e691cf3303d231e7a414fb400e4d6"} Oct 08 18:17:13 crc kubenswrapper[4859]: I1008 18:17:13.531282 4859 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 08 18:17:13 crc kubenswrapper[4859]: I1008 18:17:13.531282 4859 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerStarted","Data":"183d481502ed5108d5e19e483b714c289f355ea3f8b0891883e5335b802c83b6"} Oct 08 18:17:13 crc kubenswrapper[4859]: I1008 18:17:13.531407 4859 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerStarted","Data":"ec848c46fd5dde0469b00d3cab14ba4a652533a265490fadbee7a8b8843ef249"} Oct 08 18:17:13 crc kubenswrapper[4859]: I1008 18:17:13.531419 4859 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerStarted","Data":"d4786a8fb8abd96fb22e60733c974f599017407d589ed815fb8541fd3b64f696"} Oct 08 18:17:13 crc kubenswrapper[4859]: I1008 18:17:13.538280 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 18:17:13 crc kubenswrapper[4859]: I1008 18:17:13.538331 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 18:17:13 crc kubenswrapper[4859]: I1008 18:17:13.539008 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 18:17:13 crc kubenswrapper[4859]: I1008 18:17:13.650957 4859 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 08 18:17:13 crc kubenswrapper[4859]: I1008 18:17:13.654456 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 18:17:13 crc kubenswrapper[4859]: I1008 18:17:13.654501 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 18:17:13 crc kubenswrapper[4859]: I1008 18:17:13.654517 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 18:17:13 crc kubenswrapper[4859]: I1008 18:17:13.654549 4859 kubelet_node_status.go:76] "Attempting to register node" node="crc" Oct 08 18:17:13 crc kubenswrapper[4859]: E1008 18:17:13.655048 4859 kubelet_node_status.go:99] "Unable to register node with API server" err="Post \"https://api-int.crc.testing:6443/api/v1/nodes\": dial tcp 38.102.83.13:6443: connect: connection refused" node="crc" Oct 08 18:17:13 crc kubenswrapper[4859]: W1008 18:17:13.913361 4859 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.RuntimeClass: Get "https://api-int.crc.testing:6443/apis/node.k8s.io/v1/runtimeclasses?limit=500&resourceVersion=0": dial tcp 38.102.83.13:6443: connect: connection refused Oct 08 18:17:13 crc kubenswrapper[4859]: E1008 18:17:13.913496 4859 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.RuntimeClass: failed to list *v1.RuntimeClass: Get \"https://api-int.crc.testing:6443/apis/node.k8s.io/v1/runtimeclasses?limit=500&resourceVersion=0\": dial tcp 38.102.83.13:6443: connect: connection refused" logger="UnhandledError" Oct 08 18:17:14 crc kubenswrapper[4859]: I1008 18:17:14.406767 4859 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Oct 08 18:17:14 crc kubenswrapper[4859]: I1008 18:17:14.537516 4859 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerStarted","Data":"2e700b2a20c81b414617f5b5c018a738278a3fa44e6c6d14b2170051d0725431"} Oct 08 18:17:14 crc kubenswrapper[4859]: I1008 18:17:14.537604 4859 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 08 18:17:14 crc kubenswrapper[4859]: I1008 18:17:14.538778 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 18:17:14 crc kubenswrapper[4859]: I1008 18:17:14.538827 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 18:17:14 crc kubenswrapper[4859]: I1008 18:17:14.538842 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 18:17:14 crc kubenswrapper[4859]: I1008 18:17:14.539851 4859 generic.go:334] "Generic (PLEG): container finished" podID="2139d3e2895fc6797b9c76a1b4c9886d" containerID="347ac1bb615d3a1f5cad12ef19c019378b214e160a0f198a1733e815a2e13b5b" exitCode=0 Oct 08 18:17:14 crc kubenswrapper[4859]: I1008 18:17:14.539902 4859 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd/etcd-crc" event={"ID":"2139d3e2895fc6797b9c76a1b4c9886d","Type":"ContainerDied","Data":"347ac1bb615d3a1f5cad12ef19c019378b214e160a0f198a1733e815a2e13b5b"} Oct 08 18:17:14 crc kubenswrapper[4859]: I1008 18:17:14.539969 4859 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" Oct 08 18:17:14 crc kubenswrapper[4859]: I1008 18:17:14.540025 4859 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 08 18:17:14 crc kubenswrapper[4859]: I1008 18:17:14.540032 4859 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 08 18:17:14 crc kubenswrapper[4859]: I1008 18:17:14.540658 4859 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" Oct 08 18:17:14 crc kubenswrapper[4859]: I1008 18:17:14.540696 4859 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 08 18:17:14 crc kubenswrapper[4859]: I1008 18:17:14.540731 4859 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 08 18:17:14 crc kubenswrapper[4859]: I1008 18:17:14.541159 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 18:17:14 crc kubenswrapper[4859]: I1008 18:17:14.541183 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 18:17:14 crc kubenswrapper[4859]: I1008 18:17:14.541192 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 18:17:14 crc kubenswrapper[4859]: I1008 18:17:14.541287 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 18:17:14 crc kubenswrapper[4859]: I1008 18:17:14.541305 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 18:17:14 crc kubenswrapper[4859]: I1008 18:17:14.541315 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 18:17:14 crc kubenswrapper[4859]: I1008 18:17:14.541747 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 18:17:14 crc kubenswrapper[4859]: I1008 18:17:14.541773 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 18:17:14 crc kubenswrapper[4859]: I1008 18:17:14.541787 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 18:17:14 crc kubenswrapper[4859]: I1008 18:17:14.541836 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 18:17:14 crc kubenswrapper[4859]: I1008 18:17:14.541853 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 18:17:14 crc kubenswrapper[4859]: I1008 18:17:14.541863 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 18:17:14 crc kubenswrapper[4859]: I1008 18:17:14.643325 4859 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-kube-apiserver/kube-apiserver-crc" Oct 08 18:17:14 crc kubenswrapper[4859]: I1008 18:17:14.916630 4859 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-kube-apiserver/kube-apiserver-crc" Oct 08 18:17:15 crc kubenswrapper[4859]: I1008 18:17:15.549165 4859 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd/etcd-crc" event={"ID":"2139d3e2895fc6797b9c76a1b4c9886d","Type":"ContainerStarted","Data":"38269c513677f9c7bcde44d6464abc0003c0b6eff080a51ebb6441bf1ad2acdd"} Oct 08 18:17:15 crc kubenswrapper[4859]: I1008 18:17:15.549232 4859 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd/etcd-crc" event={"ID":"2139d3e2895fc6797b9c76a1b4c9886d","Type":"ContainerStarted","Data":"1a6185c9750945e744f9cb0c825c6c109718bf54403c697fd58a68a8f61fea5a"} Oct 08 18:17:15 crc kubenswrapper[4859]: I1008 18:17:15.549245 4859 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd/etcd-crc" event={"ID":"2139d3e2895fc6797b9c76a1b4c9886d","Type":"ContainerStarted","Data":"e4a7345e8f2f370b0112447387d2d164407070c64932503334cc1ad461d6279b"} Oct 08 18:17:15 crc kubenswrapper[4859]: I1008 18:17:15.549254 4859 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd/etcd-crc" event={"ID":"2139d3e2895fc6797b9c76a1b4c9886d","Type":"ContainerStarted","Data":"1c78b72b960f59cba122b5abe6712507ec21e822a4afa3d6de88e57adbfa2853"} Oct 08 18:17:15 crc kubenswrapper[4859]: I1008 18:17:15.549252 4859 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" Oct 08 18:17:15 crc kubenswrapper[4859]: I1008 18:17:15.549302 4859 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 08 18:17:15 crc kubenswrapper[4859]: I1008 18:17:15.549376 4859 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-kube-apiserver/kube-apiserver-crc" Oct 08 18:17:15 crc kubenswrapper[4859]: I1008 18:17:15.549319 4859 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 08 18:17:15 crc kubenswrapper[4859]: I1008 18:17:15.550940 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 18:17:15 crc kubenswrapper[4859]: I1008 18:17:15.550968 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 18:17:15 crc kubenswrapper[4859]: I1008 18:17:15.550976 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 18:17:15 crc kubenswrapper[4859]: I1008 18:17:15.551062 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 18:17:15 crc kubenswrapper[4859]: I1008 18:17:15.551091 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 18:17:15 crc kubenswrapper[4859]: I1008 18:17:15.551104 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 18:17:16 crc kubenswrapper[4859]: I1008 18:17:16.561480 4859 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 08 18:17:16 crc kubenswrapper[4859]: I1008 18:17:16.562631 4859 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd/etcd-crc" event={"ID":"2139d3e2895fc6797b9c76a1b4c9886d","Type":"ContainerStarted","Data":"19fefb46db7f81e2b87c7619dbe072c3a47cf86fe177decb835fba27d3fdbd49"} Oct 08 18:17:16 crc kubenswrapper[4859]: I1008 18:17:16.562679 4859 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 08 18:17:16 crc kubenswrapper[4859]: I1008 18:17:16.563191 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 18:17:16 crc kubenswrapper[4859]: I1008 18:17:16.563264 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 18:17:16 crc kubenswrapper[4859]: I1008 18:17:16.563287 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 18:17:16 crc kubenswrapper[4859]: I1008 18:17:16.564732 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 18:17:16 crc kubenswrapper[4859]: I1008 18:17:16.564817 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 18:17:16 crc kubenswrapper[4859]: I1008 18:17:16.564843 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 18:17:16 crc kubenswrapper[4859]: I1008 18:17:16.856191 4859 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 08 18:17:16 crc kubenswrapper[4859]: I1008 18:17:16.858366 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 18:17:16 crc kubenswrapper[4859]: I1008 18:17:16.858447 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 18:17:16 crc kubenswrapper[4859]: I1008 18:17:16.858473 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 18:17:16 crc kubenswrapper[4859]: I1008 18:17:16.858521 4859 kubelet_node_status.go:76] "Attempting to register node" node="crc" Oct 08 18:17:17 crc kubenswrapper[4859]: I1008 18:17:17.407911 4859 patch_prober.go:28] interesting pod/kube-controller-manager-crc container/cluster-policy-controller namespace/openshift-kube-controller-manager: Startup probe status=failure output="Get \"https://192.168.126.11:10357/healthz\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" start-of-body= Oct 08 18:17:17 crc kubenswrapper[4859]: I1008 18:17:17.408103 4859 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-kube-controller-manager/kube-controller-manager-crc" podUID="f614b9022728cf315e60c057852e563e" containerName="cluster-policy-controller" probeResult="failure" output="Get \"https://192.168.126.11:10357/healthz\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Oct 08 18:17:17 crc kubenswrapper[4859]: I1008 18:17:17.565360 4859 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 08 18:17:17 crc kubenswrapper[4859]: I1008 18:17:17.566777 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 18:17:17 crc kubenswrapper[4859]: I1008 18:17:17.566842 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 18:17:17 crc kubenswrapper[4859]: I1008 18:17:17.566865 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 18:17:18 crc kubenswrapper[4859]: I1008 18:17:18.243630 4859 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Oct 08 18:17:18 crc kubenswrapper[4859]: I1008 18:17:18.244017 4859 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 08 18:17:18 crc kubenswrapper[4859]: I1008 18:17:18.246051 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 18:17:18 crc kubenswrapper[4859]: I1008 18:17:18.246113 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 18:17:18 crc kubenswrapper[4859]: I1008 18:17:18.246140 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 18:17:18 crc kubenswrapper[4859]: I1008 18:17:18.714173 4859 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-etcd/etcd-crc" Oct 08 18:17:18 crc kubenswrapper[4859]: I1008 18:17:18.714471 4859 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 08 18:17:18 crc kubenswrapper[4859]: I1008 18:17:18.716104 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 18:17:18 crc kubenswrapper[4859]: I1008 18:17:18.716165 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 18:17:18 crc kubenswrapper[4859]: I1008 18:17:18.716178 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 18:17:20 crc kubenswrapper[4859]: E1008 18:17:20.551824 4859 eviction_manager.go:285] "Eviction manager: failed to get summary stats" err="failed to get node info: node \"crc\" not found" Oct 08 18:17:20 crc kubenswrapper[4859]: I1008 18:17:20.951850 4859 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Oct 08 18:17:20 crc kubenswrapper[4859]: I1008 18:17:20.952167 4859 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 08 18:17:20 crc kubenswrapper[4859]: I1008 18:17:20.953923 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 18:17:20 crc kubenswrapper[4859]: I1008 18:17:20.953978 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 18:17:20 crc kubenswrapper[4859]: I1008 18:17:20.953993 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 18:17:22 crc kubenswrapper[4859]: I1008 18:17:22.054811 4859 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Oct 08 18:17:22 crc kubenswrapper[4859]: I1008 18:17:22.054993 4859 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 08 18:17:22 crc kubenswrapper[4859]: I1008 18:17:22.056404 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 18:17:22 crc kubenswrapper[4859]: I1008 18:17:22.056456 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 18:17:22 crc kubenswrapper[4859]: I1008 18:17:22.056472 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 18:17:22 crc kubenswrapper[4859]: I1008 18:17:22.059627 4859 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Oct 08 18:17:22 crc kubenswrapper[4859]: I1008 18:17:22.578052 4859 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 08 18:17:22 crc kubenswrapper[4859]: I1008 18:17:22.579497 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 18:17:22 crc kubenswrapper[4859]: I1008 18:17:22.579564 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 18:17:22 crc kubenswrapper[4859]: I1008 18:17:22.579580 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 18:17:24 crc kubenswrapper[4859]: W1008 18:17:24.255453 4859 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.Node: Get "https://api-int.crc.testing:6443/api/v1/nodes?fieldSelector=metadata.name%3Dcrc&limit=500&resourceVersion=0": net/http: TLS handshake timeout Oct 08 18:17:24 crc kubenswrapper[4859]: I1008 18:17:24.256054 4859 trace.go:236] Trace[1565283426]: "Reflector ListAndWatch" name:k8s.io/client-go/informers/factory.go:160 (08-Oct-2025 18:17:14.254) (total time: 10001ms): Oct 08 18:17:24 crc kubenswrapper[4859]: Trace[1565283426]: ---"Objects listed" error:Get "https://api-int.crc.testing:6443/api/v1/nodes?fieldSelector=metadata.name%3Dcrc&limit=500&resourceVersion=0": net/http: TLS handshake timeout 10001ms (18:17:24.255) Oct 08 18:17:24 crc kubenswrapper[4859]: Trace[1565283426]: [10.001728128s] [10.001728128s] END Oct 08 18:17:24 crc kubenswrapper[4859]: E1008 18:17:24.256084 4859 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.Node: failed to list *v1.Node: Get \"https://api-int.crc.testing:6443/api/v1/nodes?fieldSelector=metadata.name%3Dcrc&limit=500&resourceVersion=0\": net/http: TLS handshake timeout" logger="UnhandledError" Oct 08 18:17:24 crc kubenswrapper[4859]: I1008 18:17:24.295229 4859 patch_prober.go:28] interesting pod/kube-apiserver-crc container/kube-apiserver-check-endpoints namespace/openshift-kube-apiserver: Liveness probe status=failure output="Get \"https://192.168.126.11:17697/healthz\": read tcp 192.168.126.11:52296->192.168.126.11:17697: read: connection reset by peer" start-of-body= Oct 08 18:17:24 crc kubenswrapper[4859]: I1008 18:17:24.295361 4859 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-check-endpoints" probeResult="failure" output="Get \"https://192.168.126.11:17697/healthz\": read tcp 192.168.126.11:52296->192.168.126.11:17697: read: connection reset by peer" Oct 08 18:17:24 crc kubenswrapper[4859]: I1008 18:17:24.394883 4859 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csinodes/crc?resourceVersion=0": net/http: TLS handshake timeout Oct 08 18:17:24 crc kubenswrapper[4859]: I1008 18:17:24.584750 4859 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_kube-apiserver-crc_f4b27818a5e8e43d0dc095d08835c792/kube-apiserver-check-endpoints/0.log" Oct 08 18:17:24 crc kubenswrapper[4859]: I1008 18:17:24.587036 4859 generic.go:334] "Generic (PLEG): container finished" podID="f4b27818a5e8e43d0dc095d08835c792" containerID="2e700b2a20c81b414617f5b5c018a738278a3fa44e6c6d14b2170051d0725431" exitCode=255 Oct 08 18:17:24 crc kubenswrapper[4859]: I1008 18:17:24.587104 4859 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerDied","Data":"2e700b2a20c81b414617f5b5c018a738278a3fa44e6c6d14b2170051d0725431"} Oct 08 18:17:24 crc kubenswrapper[4859]: I1008 18:17:24.587324 4859 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 08 18:17:24 crc kubenswrapper[4859]: I1008 18:17:24.588335 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 18:17:24 crc kubenswrapper[4859]: I1008 18:17:24.588373 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 18:17:24 crc kubenswrapper[4859]: I1008 18:17:24.588383 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 18:17:24 crc kubenswrapper[4859]: I1008 18:17:24.588976 4859 scope.go:117] "RemoveContainer" containerID="2e700b2a20c81b414617f5b5c018a738278a3fa44e6c6d14b2170051d0725431" Oct 08 18:17:24 crc kubenswrapper[4859]: I1008 18:17:24.916903 4859 patch_prober.go:28] interesting pod/kube-apiserver-crc container/kube-apiserver namespace/openshift-kube-apiserver: Startup probe status=failure output="Get \"https://192.168.126.11:6443/livez\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" start-of-body= Oct 08 18:17:24 crc kubenswrapper[4859]: I1008 18:17:24.917089 4859 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver" probeResult="failure" output="Get \"https://192.168.126.11:6443/livez\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" Oct 08 18:17:25 crc kubenswrapper[4859]: W1008 18:17:25.005534 4859 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.Service: Get "https://api-int.crc.testing:6443/api/v1/services?fieldSelector=spec.clusterIP%21%3DNone&limit=500&resourceVersion=0": net/http: TLS handshake timeout Oct 08 18:17:25 crc kubenswrapper[4859]: I1008 18:17:25.005655 4859 trace.go:236] Trace[91810705]: "Reflector ListAndWatch" name:k8s.io/client-go/informers/factory.go:160 (08-Oct-2025 18:17:15.003) (total time: 10002ms): Oct 08 18:17:25 crc kubenswrapper[4859]: Trace[91810705]: ---"Objects listed" error:Get "https://api-int.crc.testing:6443/api/v1/services?fieldSelector=spec.clusterIP%21%3DNone&limit=500&resourceVersion=0": net/http: TLS handshake timeout 10001ms (18:17:25.005) Oct 08 18:17:25 crc kubenswrapper[4859]: Trace[91810705]: [10.002030039s] [10.002030039s] END Oct 08 18:17:25 crc kubenswrapper[4859]: E1008 18:17:25.005698 4859 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.Service: failed to list *v1.Service: Get \"https://api-int.crc.testing:6443/api/v1/services?fieldSelector=spec.clusterIP%21%3DNone&limit=500&resourceVersion=0\": net/http: TLS handshake timeout" logger="UnhandledError" Oct 08 18:17:25 crc kubenswrapper[4859]: I1008 18:17:25.201079 4859 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-etcd/etcd-crc" Oct 08 18:17:25 crc kubenswrapper[4859]: I1008 18:17:25.201305 4859 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 08 18:17:25 crc kubenswrapper[4859]: I1008 18:17:25.202758 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 18:17:25 crc kubenswrapper[4859]: I1008 18:17:25.202842 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 18:17:25 crc kubenswrapper[4859]: I1008 18:17:25.202865 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 18:17:25 crc kubenswrapper[4859]: I1008 18:17:25.245867 4859 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-etcd/etcd-crc" Oct 08 18:17:25 crc kubenswrapper[4859]: I1008 18:17:25.468117 4859 patch_prober.go:28] interesting pod/kube-apiserver-crc container/kube-apiserver namespace/openshift-kube-apiserver: Startup probe status=failure output="HTTP probe failed with statuscode: 403" start-of-body={"kind":"Status","apiVersion":"v1","metadata":{},"status":"Failure","message":"forbidden: User \"system:anonymous\" cannot get path \"/livez\"","reason":"Forbidden","details":{},"code":403} Oct 08 18:17:25 crc kubenswrapper[4859]: I1008 18:17:25.468202 4859 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver" probeResult="failure" output="HTTP probe failed with statuscode: 403" Oct 08 18:17:25 crc kubenswrapper[4859]: I1008 18:17:25.591992 4859 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_kube-apiserver-crc_f4b27818a5e8e43d0dc095d08835c792/kube-apiserver-check-endpoints/0.log" Oct 08 18:17:25 crc kubenswrapper[4859]: I1008 18:17:25.593432 4859 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerStarted","Data":"4fe1828b4d63ff8106bfcd15e1dc1c1bd1ba7f77777cdb23ff1b291d0418774c"} Oct 08 18:17:25 crc kubenswrapper[4859]: I1008 18:17:25.593523 4859 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 08 18:17:25 crc kubenswrapper[4859]: I1008 18:17:25.593605 4859 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 08 18:17:25 crc kubenswrapper[4859]: I1008 18:17:25.594536 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 18:17:25 crc kubenswrapper[4859]: I1008 18:17:25.594572 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 18:17:25 crc kubenswrapper[4859]: I1008 18:17:25.594583 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 18:17:25 crc kubenswrapper[4859]: I1008 18:17:25.594533 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 18:17:25 crc kubenswrapper[4859]: I1008 18:17:25.594836 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 18:17:25 crc kubenswrapper[4859]: I1008 18:17:25.594918 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 18:17:25 crc kubenswrapper[4859]: I1008 18:17:25.607081 4859 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-etcd/etcd-crc" Oct 08 18:17:26 crc kubenswrapper[4859]: I1008 18:17:26.596350 4859 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 08 18:17:26 crc kubenswrapper[4859]: I1008 18:17:26.600076 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 18:17:26 crc kubenswrapper[4859]: I1008 18:17:26.600272 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 18:17:26 crc kubenswrapper[4859]: I1008 18:17:26.600343 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 18:17:27 crc kubenswrapper[4859]: I1008 18:17:27.407740 4859 patch_prober.go:28] interesting pod/kube-controller-manager-crc container/cluster-policy-controller namespace/openshift-kube-controller-manager: Startup probe status=failure output="Get \"https://192.168.126.11:10357/healthz\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" start-of-body= Oct 08 18:17:27 crc kubenswrapper[4859]: I1008 18:17:27.407852 4859 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-kube-controller-manager/kube-controller-manager-crc" podUID="f614b9022728cf315e60c057852e563e" containerName="cluster-policy-controller" probeResult="failure" output="Get \"https://192.168.126.11:10357/healthz\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" Oct 08 18:17:29 crc kubenswrapper[4859]: I1008 18:17:29.031733 4859 reflector.go:368] Caches populated for *v1.Node from k8s.io/client-go/informers/factory.go:160 Oct 08 18:17:29 crc kubenswrapper[4859]: I1008 18:17:29.389509 4859 apiserver.go:52] "Watching apiserver" Oct 08 18:17:29 crc kubenswrapper[4859]: I1008 18:17:29.396549 4859 reflector.go:368] Caches populated for *v1.Pod from pkg/kubelet/config/apiserver.go:66 Oct 08 18:17:29 crc kubenswrapper[4859]: I1008 18:17:29.396855 4859 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-network-operator/network-operator-58b4c7f79c-55gtf","openshift-network-console/networking-console-plugin-85b44fc459-gdk6g","openshift-network-diagnostics/network-check-source-55646444c4-trplf","openshift-network-diagnostics/network-check-target-xd92c","openshift-network-node-identity/network-node-identity-vrzqb","openshift-network-operator/iptables-alerter-4ln5h"] Oct 08 18:17:29 crc kubenswrapper[4859]: I1008 18:17:29.397302 4859 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" Oct 08 18:17:29 crc kubenswrapper[4859]: I1008 18:17:29.397419 4859 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 08 18:17:29 crc kubenswrapper[4859]: I1008 18:17:29.397470 4859 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 08 18:17:29 crc kubenswrapper[4859]: I1008 18:17:29.397612 4859 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-node-identity/network-node-identity-vrzqb" Oct 08 18:17:29 crc kubenswrapper[4859]: E1008 18:17:29.397764 4859 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 08 18:17:29 crc kubenswrapper[4859]: E1008 18:17:29.397834 4859 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 08 18:17:29 crc kubenswrapper[4859]: I1008 18:17:29.397991 4859 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 08 18:17:29 crc kubenswrapper[4859]: I1008 18:17:29.398008 4859 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-operator/iptables-alerter-4ln5h" Oct 08 18:17:29 crc kubenswrapper[4859]: E1008 18:17:29.398047 4859 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 08 18:17:29 crc kubenswrapper[4859]: I1008 18:17:29.401029 4859 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-node-identity"/"openshift-service-ca.crt" Oct 08 18:17:29 crc kubenswrapper[4859]: I1008 18:17:29.401224 4859 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-node-identity"/"env-overrides" Oct 08 18:17:29 crc kubenswrapper[4859]: I1008 18:17:29.401404 4859 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-node-identity"/"kube-root-ca.crt" Oct 08 18:17:29 crc kubenswrapper[4859]: I1008 18:17:29.402624 4859 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-operator"/"openshift-service-ca.crt" Oct 08 18:17:29 crc kubenswrapper[4859]: I1008 18:17:29.402659 4859 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-network-operator"/"metrics-tls" Oct 08 18:17:29 crc kubenswrapper[4859]: I1008 18:17:29.402775 4859 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-node-identity"/"ovnkube-identity-cm" Oct 08 18:17:29 crc kubenswrapper[4859]: I1008 18:17:29.403118 4859 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-network-node-identity"/"network-node-identity-cert" Oct 08 18:17:29 crc kubenswrapper[4859]: I1008 18:17:29.403779 4859 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-operator"/"kube-root-ca.crt" Oct 08 18:17:29 crc kubenswrapper[4859]: I1008 18:17:29.405109 4859 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-operator"/"iptables-alerter-script" Oct 08 18:17:29 crc kubenswrapper[4859]: I1008 18:17:29.441403 4859 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-08T18:17:29Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T18:17:29Z\\\",\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":false,\\\"restartCount\\\":5,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Oct 08 18:17:29 crc kubenswrapper[4859]: I1008 18:17:29.463629 4859 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-08T18:17:29Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T18:17:29Z\\\",\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":false,\\\"restartCount\\\":5,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Oct 08 18:17:29 crc kubenswrapper[4859]: I1008 18:17:29.478237 4859 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-08T18:17:29Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T18:17:29Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Oct 08 18:17:29 crc kubenswrapper[4859]: I1008 18:17:29.490378 4859 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-08T18:17:29Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T18:17:29Z\\\",\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Oct 08 18:17:29 crc kubenswrapper[4859]: I1008 18:17:29.500140 4859 desired_state_of_world_populator.go:154] "Finished populating initial desired state of world" Oct 08 18:17:29 crc kubenswrapper[4859]: I1008 18:17:29.502979 4859 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-08T18:17:29Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T18:17:29Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Oct 08 18:17:29 crc kubenswrapper[4859]: I1008 18:17:29.514232 4859 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-08T18:17:29Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [webhook approver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T18:17:29Z\\\",\\\"message\\\":\\\"containers with unready status: [webhook approver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":false,\\\"restartCount\\\":6,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Oct 08 18:17:29 crc kubenswrapper[4859]: I1008 18:17:29.527318 4859 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-08T18:17:29Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T18:17:29Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Oct 08 18:17:29 crc kubenswrapper[4859]: I1008 18:17:29.921827 4859 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-kube-apiserver/kube-apiserver-crc" Oct 08 18:17:29 crc kubenswrapper[4859]: I1008 18:17:29.922587 4859 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-kube-apiserver/kube-apiserver-crc" Oct 08 18:17:29 crc kubenswrapper[4859]: I1008 18:17:29.926885 4859 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-kube-apiserver/kube-apiserver-crc" Oct 08 18:17:29 crc kubenswrapper[4859]: I1008 18:17:29.934515 4859 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-08T18:17:29Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T18:17:29Z\\\",\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":false,\\\"restartCount\\\":5,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Oct 08 18:17:29 crc kubenswrapper[4859]: I1008 18:17:29.935440 4859 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-apiserver/kube-apiserver-crc"] Oct 08 18:17:29 crc kubenswrapper[4859]: I1008 18:17:29.947298 4859 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-08T18:17:29Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T18:17:29Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Oct 08 18:17:29 crc kubenswrapper[4859]: I1008 18:17:29.957560 4859 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-08T18:17:29Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T18:17:29Z\\\",\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Oct 08 18:17:29 crc kubenswrapper[4859]: I1008 18:17:29.967756 4859 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-08T18:17:29Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T18:17:29Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Oct 08 18:17:29 crc kubenswrapper[4859]: I1008 18:17:29.982500 4859 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-08T18:17:29Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [webhook approver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T18:17:29Z\\\",\\\"message\\\":\\\"containers with unready status: [webhook approver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":false,\\\"restartCount\\\":6,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Oct 08 18:17:29 crc kubenswrapper[4859]: I1008 18:17:29.993473 4859 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-08T18:17:29Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T18:17:29Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Oct 08 18:17:30 crc kubenswrapper[4859]: I1008 18:17:30.006478 4859 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"63a623de-5608-4228-b1c2-77188c276986\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:17:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:17:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:17:10Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:17:10Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:17:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d4786a8fb8abd96fb22e60733c974f599017407d589ed815fb8541fd3b64f696\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T18:17:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://183d481502ed5108d5e19e483b714c289f355ea3f8b0891883e5335b802c83b6\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T18:17:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ec848c46fd5dde0469b00d3cab14ba4a652533a265490fadbee7a8b8843ef249\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T18:17:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4fe1828b4d63ff8106bfcd15e1dc1c1bd1ba7f77777cdb23ff1b291d0418774c\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2e700b2a20c81b414617f5b5c018a738278a3fa44e6c6d14b2170051d0725431\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-08T18:17:24Z\\\",\\\"message\\\":\\\"W1008 18:17:13.670625 1 cmd.go:257] Using insecure, self-signed certificates\\\\nI1008 18:17:13.671202 1 crypto.go:601] Generating new CA for check-endpoints-signer@1759947433 cert, and key in /tmp/serving-cert-3867218267/serving-signer.crt, /tmp/serving-cert-3867218267/serving-signer.key\\\\nI1008 18:17:13.944463 1 observer_polling.go:159] Starting file observer\\\\nW1008 18:17:13.951958 1 builder.go:272] unable to get owner reference (falling back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": dial tcp [::1]:6443: connect: connection refused\\\\nI1008 18:17:13.952088 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1008 18:17:13.956755 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3867218267/tls.crt::/tmp/serving-cert-3867218267/tls.key\\\\\\\"\\\\nF1008 18:17:24.290991 1 cmd.go:182] error initializing delegating authentication: unable to load configmap based request-header-client-ca-file: Get \\\\\\\"https://localhost:6443/api/v1/namespaces/kube-system/configmaps/extension-apiserver-authentication\\\\\\\": net/http: TLS handshake timeout\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-08T18:17:13Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T18:17:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c65b66f62d325ca10ef8d5d31f04529ccf4e691cf3303d231e7a414fb400e4d6\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T18:17:13Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://58c7bba4ace28215b062653a960e561b1dc1f890f2baf450145638a065af038b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://58c7bba4ace28215b062653a960e561b1dc1f890f2baf450145638a065af038b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-08T18:17:11Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-08T18:17:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T18:17:10Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Oct 08 18:17:30 crc kubenswrapper[4859]: I1008 18:17:30.022391 4859 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-08T18:17:29Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [webhook approver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T18:17:29Z\\\",\\\"message\\\":\\\"containers with unready status: [webhook approver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":false,\\\"restartCount\\\":6,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Oct 08 18:17:30 crc kubenswrapper[4859]: I1008 18:17:30.033169 4859 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-08T18:17:29Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T18:17:29Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Oct 08 18:17:30 crc kubenswrapper[4859]: I1008 18:17:30.045557 4859 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-08T18:17:29Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T18:17:29Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Oct 08 18:17:30 crc kubenswrapper[4859]: I1008 18:17:30.057369 4859 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-08T18:17:29Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T18:17:29Z\\\",\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":false,\\\"restartCount\\\":5,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Oct 08 18:17:30 crc kubenswrapper[4859]: I1008 18:17:30.071067 4859 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-08T18:17:29Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T18:17:29Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Oct 08 18:17:30 crc kubenswrapper[4859]: I1008 18:17:30.083163 4859 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-08T18:17:29Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T18:17:29Z\\\",\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Oct 08 18:17:30 crc kubenswrapper[4859]: E1008 18:17:30.465934 4859 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": context deadline exceeded" interval="6.4s" Oct 08 18:17:30 crc kubenswrapper[4859]: I1008 18:17:30.467422 4859 trace.go:236] Trace[1442369786]: "Reflector ListAndWatch" name:k8s.io/client-go/informers/factory.go:160 (08-Oct-2025 18:17:17.920) (total time: 12546ms): Oct 08 18:17:30 crc kubenswrapper[4859]: Trace[1442369786]: ---"Objects listed" error: 12546ms (18:17:30.467) Oct 08 18:17:30 crc kubenswrapper[4859]: Trace[1442369786]: [12.546829614s] [12.546829614s] END Oct 08 18:17:30 crc kubenswrapper[4859]: I1008 18:17:30.467576 4859 reflector.go:368] Caches populated for *v1.CSIDriver from k8s.io/client-go/informers/factory.go:160 Oct 08 18:17:30 crc kubenswrapper[4859]: I1008 18:17:30.467905 4859 trace.go:236] Trace[1331705987]: "Reflector ListAndWatch" name:k8s.io/client-go/informers/factory.go:160 (08-Oct-2025 18:17:17.651) (total time: 12816ms): Oct 08 18:17:30 crc kubenswrapper[4859]: Trace[1331705987]: ---"Objects listed" error: 12816ms (18:17:30.467) Oct 08 18:17:30 crc kubenswrapper[4859]: Trace[1331705987]: [12.816811211s] [12.816811211s] END Oct 08 18:17:30 crc kubenswrapper[4859]: I1008 18:17:30.467931 4859 reflector.go:368] Caches populated for *v1.RuntimeClass from k8s.io/client-go/informers/factory.go:160 Oct 08 18:17:30 crc kubenswrapper[4859]: I1008 18:17:30.468742 4859 reconstruct.go:205] "DevicePaths of reconstructed volumes updated" Oct 08 18:17:30 crc kubenswrapper[4859]: E1008 18:17:30.468785 4859 kubelet_node_status.go:99] "Unable to register node with API server" err="nodes \"crc\" is forbidden: autoscaling.openshift.io/ManagedNode infra config cache not synchronized" node="crc" Oct 08 18:17:30 crc kubenswrapper[4859]: I1008 18:17:30.487864 4859 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"63a623de-5608-4228-b1c2-77188c276986\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:17:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:17:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:17:10Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:17:10Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:17:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d4786a8fb8abd96fb22e60733c974f599017407d589ed815fb8541fd3b64f696\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T18:17:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://183d481502ed5108d5e19e483b714c289f355ea3f8b0891883e5335b802c83b6\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T18:17:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ec848c46fd5dde0469b00d3cab14ba4a652533a265490fadbee7a8b8843ef249\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T18:17:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4fe1828b4d63ff8106bfcd15e1dc1c1bd1ba7f77777cdb23ff1b291d0418774c\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2e700b2a20c81b414617f5b5c018a738278a3fa44e6c6d14b2170051d0725431\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-08T18:17:24Z\\\",\\\"message\\\":\\\"W1008 18:17:13.670625 1 cmd.go:257] Using insecure, self-signed certificates\\\\nI1008 18:17:13.671202 1 crypto.go:601] Generating new CA for check-endpoints-signer@1759947433 cert, and key in /tmp/serving-cert-3867218267/serving-signer.crt, /tmp/serving-cert-3867218267/serving-signer.key\\\\nI1008 18:17:13.944463 1 observer_polling.go:159] Starting file observer\\\\nW1008 18:17:13.951958 1 builder.go:272] unable to get owner reference (falling back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": dial tcp [::1]:6443: connect: connection refused\\\\nI1008 18:17:13.952088 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1008 18:17:13.956755 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3867218267/tls.crt::/tmp/serving-cert-3867218267/tls.key\\\\\\\"\\\\nF1008 18:17:24.290991 1 cmd.go:182] error initializing delegating authentication: unable to load configmap based request-header-client-ca-file: Get \\\\\\\"https://localhost:6443/api/v1/namespaces/kube-system/configmaps/extension-apiserver-authentication\\\\\\\": net/http: TLS handshake timeout\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-08T18:17:13Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T18:17:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c65b66f62d325ca10ef8d5d31f04529ccf4e691cf3303d231e7a414fb400e4d6\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T18:17:13Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://58c7bba4ace28215b062653a960e561b1dc1f890f2baf450145638a065af038b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://58c7bba4ace28215b062653a960e561b1dc1f890f2baf450145638a065af038b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-08T18:17:11Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-08T18:17:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T18:17:10Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Oct 08 18:17:30 crc kubenswrapper[4859]: I1008 18:17:30.504934 4859 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-08T18:17:29Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [webhook approver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T18:17:29Z\\\",\\\"message\\\":\\\"containers with unready status: [webhook approver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":false,\\\"restartCount\\\":6,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Oct 08 18:17:30 crc kubenswrapper[4859]: I1008 18:17:30.515587 4859 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-08T18:17:29Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T18:17:29Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Oct 08 18:17:30 crc kubenswrapper[4859]: I1008 18:17:30.524618 4859 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-08T18:17:29Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T18:17:29Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Oct 08 18:17:30 crc kubenswrapper[4859]: I1008 18:17:30.536631 4859 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-08T18:17:29Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T18:17:29Z\\\",\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":false,\\\"restartCount\\\":5,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Oct 08 18:17:30 crc kubenswrapper[4859]: I1008 18:17:30.550243 4859 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-08T18:17:29Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T18:17:29Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Oct 08 18:17:30 crc kubenswrapper[4859]: I1008 18:17:30.569216 4859 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-htfz6\" (UniqueName: \"kubernetes.io/projected/6ea678ab-3438-413e-bfe3-290ae7725660-kube-api-access-htfz6\") pod \"6ea678ab-3438-413e-bfe3-290ae7725660\" (UID: \"6ea678ab-3438-413e-bfe3-290ae7725660\") " Oct 08 18:17:30 crc kubenswrapper[4859]: I1008 18:17:30.569281 4859 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-catalog-content\") pod \"b11524ee-3fca-4b1b-9cdf-6da289fdbc7d\" (UID: \"b11524ee-3fca-4b1b-9cdf-6da289fdbc7d\") " Oct 08 18:17:30 crc kubenswrapper[4859]: I1008 18:17:30.569324 4859 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-marketplace-trusted-ca\") pod \"b6cd30de-2eeb-49a2-ab40-9167f4560ff5\" (UID: \"b6cd30de-2eeb-49a2-ab40-9167f4560ff5\") " Oct 08 18:17:30 crc kubenswrapper[4859]: I1008 18:17:30.569350 4859 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/7583ce53-e0fe-4a16-9e4d-50516596a136-serving-cert\") pod \"7583ce53-e0fe-4a16-9e4d-50516596a136\" (UID: \"7583ce53-e0fe-4a16-9e4d-50516596a136\") " Oct 08 18:17:30 crc kubenswrapper[4859]: I1008 18:17:30.569374 4859 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-dbsvg\" (UniqueName: \"kubernetes.io/projected/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-kube-api-access-dbsvg\") pod \"f88749ec-7931-4ee7-b3fc-1ec5e11f92e9\" (UID: \"f88749ec-7931-4ee7-b3fc-1ec5e11f92e9\") " Oct 08 18:17:30 crc kubenswrapper[4859]: I1008 18:17:30.569405 4859 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-jkwtn\" (UniqueName: \"kubernetes.io/projected/5b88f790-22fa-440e-b583-365168c0b23d-kube-api-access-jkwtn\") pod \"5b88f790-22fa-440e-b583-365168c0b23d\" (UID: \"5b88f790-22fa-440e-b583-365168c0b23d\") " Oct 08 18:17:30 crc kubenswrapper[4859]: I1008 18:17:30.569450 4859 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-trusted-ca-bundle\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Oct 08 18:17:30 crc kubenswrapper[4859]: I1008 18:17:30.569477 4859 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/09efc573-dbb6-4249-bd59-9b87aba8dd28-serving-cert\") pod \"09efc573-dbb6-4249-bd59-9b87aba8dd28\" (UID: \"09efc573-dbb6-4249-bd59-9b87aba8dd28\") " Oct 08 18:17:30 crc kubenswrapper[4859]: I1008 18:17:30.569500 4859 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-serving-cert\") pod \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\" (UID: \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\") " Oct 08 18:17:30 crc kubenswrapper[4859]: I1008 18:17:30.569522 4859 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-idp-0-file-data\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Oct 08 18:17:30 crc kubenswrapper[4859]: I1008 18:17:30.569549 4859 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-xcgwh\" (UniqueName: \"kubernetes.io/projected/fda69060-fa79-4696-b1a6-7980f124bf7c-kube-api-access-xcgwh\") pod \"fda69060-fa79-4696-b1a6-7980f124bf7c\" (UID: \"fda69060-fa79-4696-b1a6-7980f124bf7c\") " Oct 08 18:17:30 crc kubenswrapper[4859]: I1008 18:17:30.569571 4859 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-fqsjt\" (UniqueName: \"kubernetes.io/projected/efdd0498-1daa-4136-9a4a-3b948c2293fc-kube-api-access-fqsjt\") pod \"efdd0498-1daa-4136-9a4a-3b948c2293fc\" (UID: \"efdd0498-1daa-4136-9a4a-3b948c2293fc\") " Oct 08 18:17:30 crc kubenswrapper[4859]: I1008 18:17:30.569594 4859 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/8f668bae-612b-4b75-9490-919e737c6a3b-ca-trust-extracted\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 08 18:17:30 crc kubenswrapper[4859]: I1008 18:17:30.569653 4859 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-marketplace-operator-metrics\") pod \"b6cd30de-2eeb-49a2-ab40-9167f4560ff5\" (UID: \"b6cd30de-2eeb-49a2-ab40-9167f4560ff5\") " Oct 08 18:17:30 crc kubenswrapper[4859]: I1008 18:17:30.569656 4859 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6ea678ab-3438-413e-bfe3-290ae7725660-kube-api-access-htfz6" (OuterVolumeSpecName: "kube-api-access-htfz6") pod "6ea678ab-3438-413e-bfe3-290ae7725660" (UID: "6ea678ab-3438-413e-bfe3-290ae7725660"). InnerVolumeSpecName "kube-api-access-htfz6". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 08 18:17:30 crc kubenswrapper[4859]: I1008 18:17:30.569679 4859 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-nzwt7\" (UniqueName: \"kubernetes.io/projected/96b93a3a-6083-4aea-8eab-fe1aa8245ad9-kube-api-access-nzwt7\") pod \"96b93a3a-6083-4aea-8eab-fe1aa8245ad9\" (UID: \"96b93a3a-6083-4aea-8eab-fe1aa8245ad9\") " Oct 08 18:17:30 crc kubenswrapper[4859]: I1008 18:17:30.569794 4859 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/87cf06ed-a83f-41a7-828d-70653580a8cb-config-volume\") pod \"87cf06ed-a83f-41a7-828d-70653580a8cb\" (UID: \"87cf06ed-a83f-41a7-828d-70653580a8cb\") " Oct 08 18:17:30 crc kubenswrapper[4859]: I1008 18:17:30.569824 4859 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-webhook-cert\") pod \"308be0ea-9f5f-4b29-aeb1-5abd31a0b17b\" (UID: \"308be0ea-9f5f-4b29-aeb1-5abd31a0b17b\") " Oct 08 18:17:30 crc kubenswrapper[4859]: I1008 18:17:30.569845 4859 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-fcqwp\" (UniqueName: \"kubernetes.io/projected/5fe579f8-e8a6-4643-bce5-a661393c4dde-kube-api-access-fcqwp\") pod \"5fe579f8-e8a6-4643-bce5-a661393c4dde\" (UID: \"5fe579f8-e8a6-4643-bce5-a661393c4dde\") " Oct 08 18:17:30 crc kubenswrapper[4859]: I1008 18:17:30.569880 4859 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-249nr\" (UniqueName: \"kubernetes.io/projected/b6312bbd-5731-4ea0-a20f-81d5a57df44a-kube-api-access-249nr\") pod \"b6312bbd-5731-4ea0-a20f-81d5a57df44a\" (UID: \"b6312bbd-5731-4ea0-a20f-81d5a57df44a\") " Oct 08 18:17:30 crc kubenswrapper[4859]: I1008 18:17:30.569902 4859 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/57a731c4-ef35-47a8-b875-bfb08a7f8011-catalog-content\") pod \"57a731c4-ef35-47a8-b875-bfb08a7f8011\" (UID: \"57a731c4-ef35-47a8-b875-bfb08a7f8011\") " Oct 08 18:17:30 crc kubenswrapper[4859]: I1008 18:17:30.569921 4859 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-service-ca\") pod \"43509403-f426-496e-be36-56cef71462f5\" (UID: \"43509403-f426-496e-be36-56cef71462f5\") " Oct 08 18:17:30 crc kubenswrapper[4859]: I1008 18:17:30.569941 4859 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/1d611f23-29be-4491-8495-bee1670e935f-utilities\") pod \"1d611f23-29be-4491-8495-bee1670e935f\" (UID: \"1d611f23-29be-4491-8495-bee1670e935f\") " Oct 08 18:17:30 crc kubenswrapper[4859]: I1008 18:17:30.569965 4859 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/e7e6199b-1264-4501-8953-767f51328d08-config\") pod \"e7e6199b-1264-4501-8953-767f51328d08\" (UID: \"e7e6199b-1264-4501-8953-767f51328d08\") " Oct 08 18:17:30 crc kubenswrapper[4859]: I1008 18:17:30.569985 4859 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"tmpfs\" (UniqueName: \"kubernetes.io/empty-dir/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-tmpfs\") pod \"308be0ea-9f5f-4b29-aeb1-5abd31a0b17b\" (UID: \"308be0ea-9f5f-4b29-aeb1-5abd31a0b17b\") " Oct 08 18:17:30 crc kubenswrapper[4859]: I1008 18:17:30.570004 4859 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/7539238d-5fe0-46ed-884e-1c3b566537ec-serving-cert\") pod \"7539238d-5fe0-46ed-884e-1c3b566537ec\" (UID: \"7539238d-5fe0-46ed-884e-1c3b566537ec\") " Oct 08 18:17:30 crc kubenswrapper[4859]: I1008 18:17:30.570021 4859 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-login\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Oct 08 18:17:30 crc kubenswrapper[4859]: I1008 18:17:30.570040 4859 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"available-featuregates\" (UniqueName: \"kubernetes.io/empty-dir/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-available-featuregates\") pod \"bc5039c0-ea34-426b-a2b7-fbbc87b49a6d\" (UID: \"bc5039c0-ea34-426b-a2b7-fbbc87b49a6d\") " Oct 08 18:17:30 crc kubenswrapper[4859]: I1008 18:17:30.570057 4859 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-config\") pod \"7583ce53-e0fe-4a16-9e4d-50516596a136\" (UID: \"7583ce53-e0fe-4a16-9e4d-50516596a136\") " Oct 08 18:17:30 crc kubenswrapper[4859]: I1008 18:17:30.570074 4859 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/6509e943-70c6-444c-bc41-48a544e36fbd-serving-cert\") pod \"6509e943-70c6-444c-bc41-48a544e36fbd\" (UID: \"6509e943-70c6-444c-bc41-48a544e36fbd\") " Oct 08 18:17:30 crc kubenswrapper[4859]: I1008 18:17:30.570076 4859 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/96b93a3a-6083-4aea-8eab-fe1aa8245ad9-kube-api-access-nzwt7" (OuterVolumeSpecName: "kube-api-access-nzwt7") pod "96b93a3a-6083-4aea-8eab-fe1aa8245ad9" (UID: "96b93a3a-6083-4aea-8eab-fe1aa8245ad9"). InnerVolumeSpecName "kube-api-access-nzwt7". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 08 18:17:30 crc kubenswrapper[4859]: I1008 18:17:30.570091 4859 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-mg5zb\" (UniqueName: \"kubernetes.io/projected/6402fda4-df10-493c-b4e5-d0569419652d-kube-api-access-mg5zb\") pod \"6402fda4-df10-493c-b4e5-d0569419652d\" (UID: \"6402fda4-df10-493c-b4e5-d0569419652d\") " Oct 08 18:17:30 crc kubenswrapper[4859]: I1008 18:17:30.570113 4859 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/efdd0498-1daa-4136-9a4a-3b948c2293fc-webhook-certs\") pod \"efdd0498-1daa-4136-9a4a-3b948c2293fc\" (UID: \"efdd0498-1daa-4136-9a4a-3b948c2293fc\") " Oct 08 18:17:30 crc kubenswrapper[4859]: I1008 18:17:30.570130 4859 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/e7e6199b-1264-4501-8953-767f51328d08-serving-cert\") pod \"e7e6199b-1264-4501-8953-767f51328d08\" (UID: \"e7e6199b-1264-4501-8953-767f51328d08\") " Oct 08 18:17:30 crc kubenswrapper[4859]: I1008 18:17:30.570151 4859 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"machine-approver-tls\" (UniqueName: \"kubernetes.io/secret/22c825df-677d-4ca6-82db-3454ed06e783-machine-approver-tls\") pod \"22c825df-677d-4ca6-82db-3454ed06e783\" (UID: \"22c825df-677d-4ca6-82db-3454ed06e783\") " Oct 08 18:17:30 crc kubenswrapper[4859]: I1008 18:17:30.570169 4859 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-wxkg8\" (UniqueName: \"kubernetes.io/projected/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59-kube-api-access-wxkg8\") pod \"3cb93b32-e0ae-4377-b9c8-fdb9842c6d59\" (UID: \"3cb93b32-e0ae-4377-b9c8-fdb9842c6d59\") " Oct 08 18:17:30 crc kubenswrapper[4859]: I1008 18:17:30.570186 4859 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"default-certificate\" (UniqueName: \"kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-default-certificate\") pod \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\" (UID: \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\") " Oct 08 18:17:30 crc kubenswrapper[4859]: I1008 18:17:30.570209 4859 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-utilities\") pod \"b11524ee-3fca-4b1b-9cdf-6da289fdbc7d\" (UID: \"b11524ee-3fca-4b1b-9cdf-6da289fdbc7d\") " Oct 08 18:17:30 crc kubenswrapper[4859]: I1008 18:17:30.570234 4859 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/925f1c65-6136-48ba-85aa-3a3b50560753-ovnkube-config\") pod \"925f1c65-6136-48ba-85aa-3a3b50560753\" (UID: \"925f1c65-6136-48ba-85aa-3a3b50560753\") " Oct 08 18:17:30 crc kubenswrapper[4859]: I1008 18:17:30.570253 4859 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-metrics-certs\") pod \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\" (UID: \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\") " Oct 08 18:17:30 crc kubenswrapper[4859]: I1008 18:17:30.570274 4859 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"machine-api-operator-tls\" (UniqueName: \"kubernetes.io/secret/6402fda4-df10-493c-b4e5-d0569419652d-machine-api-operator-tls\") pod \"6402fda4-df10-493c-b4e5-d0569419652d\" (UID: \"6402fda4-df10-493c-b4e5-d0569419652d\") " Oct 08 18:17:30 crc kubenswrapper[4859]: I1008 18:17:30.570294 4859 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-2w9zh\" (UniqueName: \"kubernetes.io/projected/4bb40260-dbaa-4fb0-84df-5e680505d512-kube-api-access-2w9zh\") pod \"4bb40260-dbaa-4fb0-84df-5e680505d512\" (UID: \"4bb40260-dbaa-4fb0-84df-5e680505d512\") " Oct 08 18:17:30 crc kubenswrapper[4859]: I1008 18:17:30.570313 4859 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"control-plane-machine-set-operator-tls\" (UniqueName: \"kubernetes.io/secret/6731426b-95fe-49ff-bb5f-40441049fde2-control-plane-machine-set-operator-tls\") pod \"6731426b-95fe-49ff-bb5f-40441049fde2\" (UID: \"6731426b-95fe-49ff-bb5f-40441049fde2\") " Oct 08 18:17:30 crc kubenswrapper[4859]: I1008 18:17:30.570318 4859 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/7583ce53-e0fe-4a16-9e4d-50516596a136-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "7583ce53-e0fe-4a16-9e4d-50516596a136" (UID: "7583ce53-e0fe-4a16-9e4d-50516596a136"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 08 18:17:30 crc kubenswrapper[4859]: I1008 18:17:30.570354 4859 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/0b78653f-4ff9-4508-8672-245ed9b561e3-kube-api-access\") pod \"0b78653f-4ff9-4508-8672-245ed9b561e3\" (UID: \"0b78653f-4ff9-4508-8672-245ed9b561e3\") " Oct 08 18:17:30 crc kubenswrapper[4859]: I1008 18:17:30.570374 4859 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-qg5z5\" (UniqueName: \"kubernetes.io/projected/43509403-f426-496e-be36-56cef71462f5-kube-api-access-qg5z5\") pod \"43509403-f426-496e-be36-56cef71462f5\" (UID: \"43509403-f426-496e-be36-56cef71462f5\") " Oct 08 18:17:30 crc kubenswrapper[4859]: I1008 18:17:30.570393 4859 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-2d4wz\" (UniqueName: \"kubernetes.io/projected/5441d097-087c-4d9a-baa8-b210afa90fc9-kube-api-access-2d4wz\") pod \"5441d097-087c-4d9a-baa8-b210afa90fc9\" (UID: \"5441d097-087c-4d9a-baa8-b210afa90fc9\") " Oct 08 18:17:30 crc kubenswrapper[4859]: I1008 18:17:30.570411 4859 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-v47cf\" (UniqueName: \"kubernetes.io/projected/c03ee662-fb2f-4fc4-a2c1-af487c19d254-kube-api-access-v47cf\") pod \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\" (UID: \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\") " Oct 08 18:17:30 crc kubenswrapper[4859]: I1008 18:17:30.570430 4859 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-8tdtz\" (UniqueName: \"kubernetes.io/projected/09efc573-dbb6-4249-bd59-9b87aba8dd28-kube-api-access-8tdtz\") pod \"09efc573-dbb6-4249-bd59-9b87aba8dd28\" (UID: \"09efc573-dbb6-4249-bd59-9b87aba8dd28\") " Oct 08 18:17:30 crc kubenswrapper[4859]: I1008 18:17:30.570451 4859 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-error\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Oct 08 18:17:30 crc kubenswrapper[4859]: I1008 18:17:30.570468 4859 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/5441d097-087c-4d9a-baa8-b210afa90fc9-serving-cert\") pod \"5441d097-087c-4d9a-baa8-b210afa90fc9\" (UID: \"5441d097-087c-4d9a-baa8-b210afa90fc9\") " Oct 08 18:17:30 crc kubenswrapper[4859]: I1008 18:17:30.570484 4859 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/e7e6199b-1264-4501-8953-767f51328d08-kube-api-access\") pod \"e7e6199b-1264-4501-8953-767f51328d08\" (UID: \"e7e6199b-1264-4501-8953-767f51328d08\") " Oct 08 18:17:30 crc kubenswrapper[4859]: I1008 18:17:30.570500 4859 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/0b574797-001e-440a-8f4e-c0be86edad0f-proxy-tls\") pod \"0b574797-001e-440a-8f4e-c0be86edad0f\" (UID: \"0b574797-001e-440a-8f4e-c0be86edad0f\") " Oct 08 18:17:30 crc kubenswrapper[4859]: I1008 18:17:30.570540 4859 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/c03ee662-fb2f-4fc4-a2c1-af487c19d254-service-ca-bundle\") pod \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\" (UID: \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\") " Oct 08 18:17:30 crc kubenswrapper[4859]: I1008 18:17:30.570559 4859 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/5441d097-087c-4d9a-baa8-b210afa90fc9-client-ca\") pod \"5441d097-087c-4d9a-baa8-b210afa90fc9\" (UID: \"5441d097-087c-4d9a-baa8-b210afa90fc9\") " Oct 08 18:17:30 crc kubenswrapper[4859]: I1008 18:17:30.570564 4859 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-kube-api-access-dbsvg" (OuterVolumeSpecName: "kube-api-access-dbsvg") pod "f88749ec-7931-4ee7-b3fc-1ec5e11f92e9" (UID: "f88749ec-7931-4ee7-b3fc-1ec5e11f92e9"). InnerVolumeSpecName "kube-api-access-dbsvg". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 08 18:17:30 crc kubenswrapper[4859]: I1008 18:17:30.570575 4859 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-9xfj7\" (UniqueName: \"kubernetes.io/projected/5225d0e4-402f-4861-b410-819f433b1803-kube-api-access-9xfj7\") pod \"5225d0e4-402f-4861-b410-819f433b1803\" (UID: \"5225d0e4-402f-4861-b410-819f433b1803\") " Oct 08 18:17:30 crc kubenswrapper[4859]: I1008 18:17:30.570597 4859 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-mnrrd\" (UniqueName: \"kubernetes.io/projected/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-kube-api-access-mnrrd\") pod \"bc5039c0-ea34-426b-a2b7-fbbc87b49a6d\" (UID: \"bc5039c0-ea34-426b-a2b7-fbbc87b49a6d\") " Oct 08 18:17:30 crc kubenswrapper[4859]: I1008 18:17:30.570585 4859 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-marketplace-trusted-ca" (OuterVolumeSpecName: "marketplace-trusted-ca") pod "b6cd30de-2eeb-49a2-ab40-9167f4560ff5" (UID: "b6cd30de-2eeb-49a2-ab40-9167f4560ff5"). InnerVolumeSpecName "marketplace-trusted-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 08 18:17:30 crc kubenswrapper[4859]: I1008 18:17:30.570619 4859 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-w7l8j\" (UniqueName: \"kubernetes.io/projected/01ab3dd5-8196-46d0-ad33-122e2ca51def-kube-api-access-w7l8j\") pod \"01ab3dd5-8196-46d0-ad33-122e2ca51def\" (UID: \"01ab3dd5-8196-46d0-ad33-122e2ca51def\") " Oct 08 18:17:30 crc kubenswrapper[4859]: I1008 18:17:30.570638 4859 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-x4zgh\" (UniqueName: \"kubernetes.io/projected/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-kube-api-access-x4zgh\") pod \"b11524ee-3fca-4b1b-9cdf-6da289fdbc7d\" (UID: \"b11524ee-3fca-4b1b-9cdf-6da289fdbc7d\") " Oct 08 18:17:30 crc kubenswrapper[4859]: I1008 18:17:30.570656 4859 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/fda69060-fa79-4696-b1a6-7980f124bf7c-proxy-tls\") pod \"fda69060-fa79-4696-b1a6-7980f124bf7c\" (UID: \"fda69060-fa79-4696-b1a6-7980f124bf7c\") " Oct 08 18:17:30 crc kubenswrapper[4859]: I1008 18:17:30.570674 4859 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/925f1c65-6136-48ba-85aa-3a3b50560753-env-overrides\") pod \"925f1c65-6136-48ba-85aa-3a3b50560753\" (UID: \"925f1c65-6136-48ba-85aa-3a3b50560753\") " Oct 08 18:17:30 crc kubenswrapper[4859]: I1008 18:17:30.570707 4859 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-trusted-ca-bundle\") pod \"1bf7eb37-55a3-4c65-b768-a94c82151e69\" (UID: \"1bf7eb37-55a3-4c65-b768-a94c82151e69\") " Oct 08 18:17:30 crc kubenswrapper[4859]: I1008 18:17:30.570726 4859 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-zkvpv\" (UniqueName: \"kubernetes.io/projected/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-kube-api-access-zkvpv\") pod \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\" (UID: \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\") " Oct 08 18:17:30 crc kubenswrapper[4859]: I1008 18:17:30.570743 4859 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"node-bootstrap-token\" (UniqueName: \"kubernetes.io/secret/5fe579f8-e8a6-4643-bce5-a661393c4dde-node-bootstrap-token\") pod \"5fe579f8-e8a6-4643-bce5-a661393c4dde\" (UID: \"5fe579f8-e8a6-4643-bce5-a661393c4dde\") " Oct 08 18:17:30 crc kubenswrapper[4859]: I1008 18:17:30.570759 4859 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-jhbk2\" (UniqueName: \"kubernetes.io/projected/bd23aa5c-e532-4e53-bccf-e79f130c5ae8-kube-api-access-jhbk2\") pod \"bd23aa5c-e532-4e53-bccf-e79f130c5ae8\" (UID: \"bd23aa5c-e532-4e53-bccf-e79f130c5ae8\") " Oct 08 18:17:30 crc kubenswrapper[4859]: I1008 18:17:30.570778 4859 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-pj782\" (UniqueName: \"kubernetes.io/projected/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-kube-api-access-pj782\") pod \"b6cd30de-2eeb-49a2-ab40-9167f4560ff5\" (UID: \"b6cd30de-2eeb-49a2-ab40-9167f4560ff5\") " Oct 08 18:17:30 crc kubenswrapper[4859]: I1008 18:17:30.570822 4859 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/20b0d48f-5fd6-431c-a545-e3c800c7b866-cert\") pod \"20b0d48f-5fd6-431c-a545-e3c800c7b866\" (UID: \"20b0d48f-5fd6-431c-a545-e3c800c7b866\") " Oct 08 18:17:30 crc kubenswrapper[4859]: I1008 18:17:30.570824 4859 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5b88f790-22fa-440e-b583-365168c0b23d-kube-api-access-jkwtn" (OuterVolumeSpecName: "kube-api-access-jkwtn") pod "5b88f790-22fa-440e-b583-365168c0b23d" (UID: "5b88f790-22fa-440e-b583-365168c0b23d"). InnerVolumeSpecName "kube-api-access-jkwtn". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 08 18:17:30 crc kubenswrapper[4859]: I1008 18:17:30.570839 4859 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/7bb08738-c794-4ee8-9972-3a62ca171029-cni-binary-copy\") pod \"7bb08738-c794-4ee8-9972-3a62ca171029\" (UID: \"7bb08738-c794-4ee8-9972-3a62ca171029\") " Oct 08 18:17:30 crc kubenswrapper[4859]: I1008 18:17:30.570857 4859 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-audit-policies\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Oct 08 18:17:30 crc kubenswrapper[4859]: I1008 18:17:30.570879 4859 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-srv-cert\") pod \"f88749ec-7931-4ee7-b3fc-1ec5e11f92e9\" (UID: \"f88749ec-7931-4ee7-b3fc-1ec5e11f92e9\") " Oct 08 18:17:30 crc kubenswrapper[4859]: I1008 18:17:30.570897 4859 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-router-certs\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Oct 08 18:17:30 crc kubenswrapper[4859]: I1008 18:17:30.570916 4859 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-encryption-config\") pod \"1bf7eb37-55a3-4c65-b768-a94c82151e69\" (UID: \"1bf7eb37-55a3-4c65-b768-a94c82151e69\") " Oct 08 18:17:30 crc kubenswrapper[4859]: I1008 18:17:30.570937 4859 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-session\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Oct 08 18:17:30 crc kubenswrapper[4859]: I1008 18:17:30.570994 4859 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-4d4hj\" (UniqueName: \"kubernetes.io/projected/3ab1a177-2de0-46d9-b765-d0d0649bb42e-kube-api-access-4d4hj\") pod \"3ab1a177-2de0-46d9-b765-d0d0649bb42e\" (UID: \"3ab1a177-2de0-46d9-b765-d0d0649bb42e\") " Oct 08 18:17:30 crc kubenswrapper[4859]: I1008 18:17:30.571002 4859 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-default-certificate" (OuterVolumeSpecName: "default-certificate") pod "c03ee662-fb2f-4fc4-a2c1-af487c19d254" (UID: "c03ee662-fb2f-4fc4-a2c1-af487c19d254"). InnerVolumeSpecName "default-certificate". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 08 18:17:30 crc kubenswrapper[4859]: I1008 18:17:30.571017 4859 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/8f668bae-612b-4b75-9490-919e737c6a3b-trusted-ca\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 08 18:17:30 crc kubenswrapper[4859]: I1008 18:17:30.571045 4859 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-kfwg7\" (UniqueName: \"kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-kube-api-access-kfwg7\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 08 18:17:30 crc kubenswrapper[4859]: I1008 18:17:30.571069 4859 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/b6312bbd-5731-4ea0-a20f-81d5a57df44a-srv-cert\") pod \"b6312bbd-5731-4ea0-a20f-81d5a57df44a\" (UID: \"b6312bbd-5731-4ea0-a20f-81d5a57df44a\") " Oct 08 18:17:30 crc kubenswrapper[4859]: I1008 18:17:30.571092 4859 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-sb6h7\" (UniqueName: \"kubernetes.io/projected/1bf7eb37-55a3-4c65-b768-a94c82151e69-kube-api-access-sb6h7\") pod \"1bf7eb37-55a3-4c65-b768-a94c82151e69\" (UID: \"1bf7eb37-55a3-4c65-b768-a94c82151e69\") " Oct 08 18:17:30 crc kubenswrapper[4859]: I1008 18:17:30.571123 4859 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/87cf06ed-a83f-41a7-828d-70653580a8cb-metrics-tls\") pod \"87cf06ed-a83f-41a7-828d-70653580a8cb\" (UID: \"87cf06ed-a83f-41a7-828d-70653580a8cb\") " Oct 08 18:17:30 crc kubenswrapper[4859]: I1008 18:17:30.571141 4859 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/1386a44e-36a2-460c-96d0-0359d2b6f0f5-serving-cert\") pod \"1386a44e-36a2-460c-96d0-0359d2b6f0f5\" (UID: \"1386a44e-36a2-460c-96d0-0359d2b6f0f5\") " Oct 08 18:17:30 crc kubenswrapper[4859]: I1008 18:17:30.571161 4859 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"multus-daemon-config\" (UniqueName: \"kubernetes.io/configmap/4bb40260-dbaa-4fb0-84df-5e680505d512-multus-daemon-config\") pod \"4bb40260-dbaa-4fb0-84df-5e680505d512\" (UID: \"4bb40260-dbaa-4fb0-84df-5e680505d512\") " Oct 08 18:17:30 crc kubenswrapper[4859]: I1008 18:17:30.571177 4859 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/a31745f5-9847-4afe-82a5-3161cc66ca93-metrics-tls\") pod \"a31745f5-9847-4afe-82a5-3161cc66ca93\" (UID: \"a31745f5-9847-4afe-82a5-3161cc66ca93\") " Oct 08 18:17:30 crc kubenswrapper[4859]: I1008 18:17:30.571193 4859 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/9d4552c7-cd75-42dd-8880-30dd377c49a4-config\") pod \"9d4552c7-cd75-42dd-8880-30dd377c49a4\" (UID: \"9d4552c7-cd75-42dd-8880-30dd377c49a4\") " Oct 08 18:17:30 crc kubenswrapper[4859]: I1008 18:17:30.571209 4859 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-client\") pod \"09efc573-dbb6-4249-bd59-9b87aba8dd28\" (UID: \"09efc573-dbb6-4249-bd59-9b87aba8dd28\") " Oct 08 18:17:30 crc kubenswrapper[4859]: I1008 18:17:30.571226 4859 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/96b93a3a-6083-4aea-8eab-fe1aa8245ad9-metrics-tls\") pod \"96b93a3a-6083-4aea-8eab-fe1aa8245ad9\" (UID: \"96b93a3a-6083-4aea-8eab-fe1aa8245ad9\") " Oct 08 18:17:30 crc kubenswrapper[4859]: I1008 18:17:30.571253 4859 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/7539238d-5fe0-46ed-884e-1c3b566537ec-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "7539238d-5fe0-46ed-884e-1c3b566537ec" (UID: "7539238d-5fe0-46ed-884e-1c3b566537ec"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 08 18:17:30 crc kubenswrapper[4859]: I1008 18:17:30.571274 4859 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/0b574797-001e-440a-8f4e-c0be86edad0f-proxy-tls" (OuterVolumeSpecName: "proxy-tls") pod "0b574797-001e-440a-8f4e-c0be86edad0f" (UID: "0b574797-001e-440a-8f4e-c0be86edad0f"). InnerVolumeSpecName "proxy-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 08 18:17:30 crc kubenswrapper[4859]: I1008 18:17:30.571274 4859 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"stats-auth\" (UniqueName: \"kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-stats-auth\") pod \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\" (UID: \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\") " Oct 08 18:17:30 crc kubenswrapper[4859]: I1008 18:17:30.571325 4859 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5fe579f8-e8a6-4643-bce5-a661393c4dde-node-bootstrap-token" (OuterVolumeSpecName: "node-bootstrap-token") pod "5fe579f8-e8a6-4643-bce5-a661393c4dde" (UID: "5fe579f8-e8a6-4643-bce5-a661393c4dde"). InnerVolumeSpecName "node-bootstrap-token". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 08 18:17:30 crc kubenswrapper[4859]: I1008 18:17:30.571381 4859 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/4bb40260-dbaa-4fb0-84df-5e680505d512-cni-binary-copy\") pod \"4bb40260-dbaa-4fb0-84df-5e680505d512\" (UID: \"4bb40260-dbaa-4fb0-84df-5e680505d512\") " Oct 08 18:17:30 crc kubenswrapper[4859]: I1008 18:17:30.571416 4859 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/8f668bae-612b-4b75-9490-919e737c6a3b-registry-certificates\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 08 18:17:30 crc kubenswrapper[4859]: I1008 18:17:30.571448 4859 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-qs4fp\" (UniqueName: \"kubernetes.io/projected/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-kube-api-access-qs4fp\") pod \"210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c\" (UID: \"210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c\") " Oct 08 18:17:30 crc kubenswrapper[4859]: I1008 18:17:30.571479 4859 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-etcd-client\") pod \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\" (UID: \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\") " Oct 08 18:17:30 crc kubenswrapper[4859]: I1008 18:17:30.571488 4859 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-stats-auth" (OuterVolumeSpecName: "stats-auth") pod "c03ee662-fb2f-4fc4-a2c1-af487c19d254" (UID: "c03ee662-fb2f-4fc4-a2c1-af487c19d254"). InnerVolumeSpecName "stats-auth". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 08 18:17:30 crc kubenswrapper[4859]: I1008 18:17:30.571469 4859 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-login" (OuterVolumeSpecName: "v4-0-config-user-template-login") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-user-template-login". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 08 18:17:30 crc kubenswrapper[4859]: I1008 18:17:30.571504 4859 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-ocp-branding-template\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Oct 08 18:17:30 crc kubenswrapper[4859]: I1008 18:17:30.571605 4859 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/31d8b7a1-420e-4252-a5b7-eebe8a111292-auth-proxy-config\") pod \"31d8b7a1-420e-4252-a5b7-eebe8a111292\" (UID: \"31d8b7a1-420e-4252-a5b7-eebe8a111292\") " Oct 08 18:17:30 crc kubenswrapper[4859]: I1008 18:17:30.571637 4859 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/31d8b7a1-420e-4252-a5b7-eebe8a111292-proxy-tls\") pod \"31d8b7a1-420e-4252-a5b7-eebe8a111292\" (UID: \"31d8b7a1-420e-4252-a5b7-eebe8a111292\") " Oct 08 18:17:30 crc kubenswrapper[4859]: I1008 18:17:30.571657 4859 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-bound-sa-token\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 08 18:17:30 crc kubenswrapper[4859]: I1008 18:17:30.571677 4859 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-config\") pod \"1bf7eb37-55a3-4c65-b768-a94c82151e69\" (UID: \"1bf7eb37-55a3-4c65-b768-a94c82151e69\") " Oct 08 18:17:30 crc kubenswrapper[4859]: I1008 18:17:30.571678 4859 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/09efc573-dbb6-4249-bd59-9b87aba8dd28-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "09efc573-dbb6-4249-bd59-9b87aba8dd28" (UID: "09efc573-dbb6-4249-bd59-9b87aba8dd28"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 08 18:17:30 crc kubenswrapper[4859]: I1008 18:17:30.571717 4859 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/bd23aa5c-e532-4e53-bccf-e79f130c5ae8-kube-api-access-jhbk2" (OuterVolumeSpecName: "kube-api-access-jhbk2") pod "bd23aa5c-e532-4e53-bccf-e79f130c5ae8" (UID: "bd23aa5c-e532-4e53-bccf-e79f130c5ae8"). InnerVolumeSpecName "kube-api-access-jhbk2". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 08 18:17:30 crc kubenswrapper[4859]: I1008 18:17:30.571797 4859 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-ocp-branding-template" (OuterVolumeSpecName: "v4-0-config-system-ocp-branding-template") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-system-ocp-branding-template". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 08 18:17:30 crc kubenswrapper[4859]: I1008 18:17:30.571969 4859 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-kube-api-access-pj782" (OuterVolumeSpecName: "kube-api-access-pj782") pod "b6cd30de-2eeb-49a2-ab40-9167f4560ff5" (UID: "b6cd30de-2eeb-49a2-ab40-9167f4560ff5"). InnerVolumeSpecName "kube-api-access-pj782". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 08 18:17:30 crc kubenswrapper[4859]: I1008 18:17:30.572029 4859 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "09ae3b1a-e8e7-4524-b54b-61eab6f9239a" (UID: "09ae3b1a-e8e7-4524-b54b-61eab6f9239a"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 08 18:17:30 crc kubenswrapper[4859]: I1008 18:17:30.572073 4859 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/c03ee662-fb2f-4fc4-a2c1-af487c19d254-service-ca-bundle" (OuterVolumeSpecName: "service-ca-bundle") pod "c03ee662-fb2f-4fc4-a2c1-af487c19d254" (UID: "c03ee662-fb2f-4fc4-a2c1-af487c19d254"). InnerVolumeSpecName "service-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 08 18:17:30 crc kubenswrapper[4859]: I1008 18:17:30.572116 4859 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-trusted-ca-bundle" (OuterVolumeSpecName: "v4-0-config-system-trusted-ca-bundle") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-system-trusted-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 08 18:17:30 crc kubenswrapper[4859]: I1008 18:17:30.572135 4859 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/87cf06ed-a83f-41a7-828d-70653580a8cb-config-volume" (OuterVolumeSpecName: "config-volume") pod "87cf06ed-a83f-41a7-828d-70653580a8cb" (UID: "87cf06ed-a83f-41a7-828d-70653580a8cb"). InnerVolumeSpecName "config-volume". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 08 18:17:30 crc kubenswrapper[4859]: I1008 18:17:30.572136 4859 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-utilities" (OuterVolumeSpecName: "utilities") pod "b11524ee-3fca-4b1b-9cdf-6da289fdbc7d" (UID: "b11524ee-3fca-4b1b-9cdf-6da289fdbc7d"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 08 18:17:30 crc kubenswrapper[4859]: I1008 18:17:30.574101 4859 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/20b0d48f-5fd6-431c-a545-e3c800c7b866-cert" (OuterVolumeSpecName: "cert") pod "20b0d48f-5fd6-431c-a545-e3c800c7b866" (UID: "20b0d48f-5fd6-431c-a545-e3c800c7b866"). InnerVolumeSpecName "cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 08 18:17:30 crc kubenswrapper[4859]: I1008 18:17:30.574493 4859 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-idp-0-file-data" (OuterVolumeSpecName: "v4-0-config-user-idp-0-file-data") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-user-idp-0-file-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 08 18:17:30 crc kubenswrapper[4859]: I1008 18:17:30.574535 4859 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b6312bbd-5731-4ea0-a20f-81d5a57df44a-srv-cert" (OuterVolumeSpecName: "srv-cert") pod "b6312bbd-5731-4ea0-a20f-81d5a57df44a" (UID: "b6312bbd-5731-4ea0-a20f-81d5a57df44a"). InnerVolumeSpecName "srv-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 08 18:17:30 crc kubenswrapper[4859]: I1008 18:17:30.574765 4859 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/1bf7eb37-55a3-4c65-b768-a94c82151e69-kube-api-access-sb6h7" (OuterVolumeSpecName: "kube-api-access-sb6h7") pod "1bf7eb37-55a3-4c65-b768-a94c82151e69" (UID: "1bf7eb37-55a3-4c65-b768-a94c82151e69"). InnerVolumeSpecName "kube-api-access-sb6h7". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 08 18:17:30 crc kubenswrapper[4859]: I1008 18:17:30.571721 4859 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"signing-cabundle\" (UniqueName: \"kubernetes.io/configmap/25e176fe-21b4-4974-b1ed-c8b94f112a7f-signing-cabundle\") pod \"25e176fe-21b4-4974-b1ed-c8b94f112a7f\" (UID: \"25e176fe-21b4-4974-b1ed-c8b94f112a7f\") " Oct 08 18:17:30 crc kubenswrapper[4859]: I1008 18:17:30.574775 4859 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-marketplace-operator-metrics" (OuterVolumeSpecName: "marketplace-operator-metrics") pod "b6cd30de-2eeb-49a2-ab40-9167f4560ff5" (UID: "b6cd30de-2eeb-49a2-ab40-9167f4560ff5"). InnerVolumeSpecName "marketplace-operator-metrics". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 08 18:17:30 crc kubenswrapper[4859]: I1008 18:17:30.574846 4859 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/9d4552c7-cd75-42dd-8880-30dd377c49a4-serving-cert\") pod \"9d4552c7-cd75-42dd-8880-30dd377c49a4\" (UID: \"9d4552c7-cd75-42dd-8880-30dd377c49a4\") " Oct 08 18:17:30 crc kubenswrapper[4859]: I1008 18:17:30.574876 4859 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-serving-cert\") pod \"210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c\" (UID: \"210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c\") " Oct 08 18:17:30 crc kubenswrapper[4859]: I1008 18:17:30.574907 4859 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-pjr6v\" (UniqueName: \"kubernetes.io/projected/49ef4625-1d3a-4a9f-b595-c2433d32326d-kube-api-access-pjr6v\") pod \"49ef4625-1d3a-4a9f-b595-c2433d32326d\" (UID: \"49ef4625-1d3a-4a9f-b595-c2433d32326d\") " Oct 08 18:17:30 crc kubenswrapper[4859]: I1008 18:17:30.574938 4859 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/22c825df-677d-4ca6-82db-3454ed06e783-config\") pod \"22c825df-677d-4ca6-82db-3454ed06e783\" (UID: \"22c825df-677d-4ca6-82db-3454ed06e783\") " Oct 08 18:17:30 crc kubenswrapper[4859]: I1008 18:17:30.574997 4859 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-proxy-ca-bundles\") pod \"7583ce53-e0fe-4a16-9e4d-50516596a136\" (UID: \"7583ce53-e0fe-4a16-9e4d-50516596a136\") " Oct 08 18:17:30 crc kubenswrapper[4859]: I1008 18:17:30.575035 4859 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-trusted-ca-bundle\") pod \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\" (UID: \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\") " Oct 08 18:17:30 crc kubenswrapper[4859]: I1008 18:17:30.575068 4859 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serviceca\" (UniqueName: \"kubernetes.io/configmap/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59-serviceca\") pod \"3cb93b32-e0ae-4377-b9c8-fdb9842c6d59\" (UID: \"3cb93b32-e0ae-4377-b9c8-fdb9842c6d59\") " Oct 08 18:17:30 crc kubenswrapper[4859]: I1008 18:17:30.575093 4859 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7539238d-5fe0-46ed-884e-1c3b566537ec-config\") pod \"7539238d-5fe0-46ed-884e-1c3b566537ec\" (UID: \"7539238d-5fe0-46ed-884e-1c3b566537ec\") " Oct 08 18:17:30 crc kubenswrapper[4859]: I1008 18:17:30.575125 4859 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-etcd-serving-ca\") pod \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\" (UID: \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\") " Oct 08 18:17:30 crc kubenswrapper[4859]: I1008 18:17:30.575152 4859 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-serving-cert\") pod \"8cea82b4-6893-4ddc-af9f-1bb5ae425c5b\" (UID: \"8cea82b4-6893-4ddc-af9f-1bb5ae425c5b\") " Oct 08 18:17:30 crc kubenswrapper[4859]: I1008 18:17:30.575185 4859 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/8f668bae-612b-4b75-9490-919e737c6a3b-installation-pull-secrets\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 08 18:17:30 crc kubenswrapper[4859]: I1008 18:17:30.575218 4859 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"package-server-manager-serving-cert\" (UniqueName: \"kubernetes.io/secret/3ab1a177-2de0-46d9-b765-d0d0649bb42e-package-server-manager-serving-cert\") pod \"3ab1a177-2de0-46d9-b765-d0d0649bb42e\" (UID: \"3ab1a177-2de0-46d9-b765-d0d0649bb42e\") " Oct 08 18:17:30 crc kubenswrapper[4859]: I1008 18:17:30.575248 4859 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/1d611f23-29be-4491-8495-bee1670e935f-catalog-content\") pod \"1d611f23-29be-4491-8495-bee1670e935f\" (UID: \"1d611f23-29be-4491-8495-bee1670e935f\") " Oct 08 18:17:30 crc kubenswrapper[4859]: I1008 18:17:30.575280 4859 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"samples-operator-tls\" (UniqueName: \"kubernetes.io/secret/a0128f3a-b052-44ed-a84e-c4c8aaf17c13-samples-operator-tls\") pod \"a0128f3a-b052-44ed-a84e-c4c8aaf17c13\" (UID: \"a0128f3a-b052-44ed-a84e-c4c8aaf17c13\") " Oct 08 18:17:30 crc kubenswrapper[4859]: I1008 18:17:30.575310 4859 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/43509403-f426-496e-be36-56cef71462f5-console-serving-cert\") pod \"43509403-f426-496e-be36-56cef71462f5\" (UID: \"43509403-f426-496e-be36-56cef71462f5\") " Oct 08 18:17:30 crc kubenswrapper[4859]: I1008 18:17:30.575341 4859 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-gf66m\" (UniqueName: \"kubernetes.io/projected/a0128f3a-b052-44ed-a84e-c4c8aaf17c13-kube-api-access-gf66m\") pod \"a0128f3a-b052-44ed-a84e-c4c8aaf17c13\" (UID: \"a0128f3a-b052-44ed-a84e-c4c8aaf17c13\") " Oct 08 18:17:30 crc kubenswrapper[4859]: I1008 18:17:30.575373 4859 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-service-ca-bundle\") pod \"6509e943-70c6-444c-bc41-48a544e36fbd\" (UID: \"6509e943-70c6-444c-bc41-48a544e36fbd\") " Oct 08 18:17:30 crc kubenswrapper[4859]: I1008 18:17:30.575404 4859 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-rnphk\" (UniqueName: \"kubernetes.io/projected/bf126b07-da06-4140-9a57-dfd54fc6b486-kube-api-access-rnphk\") pod \"bf126b07-da06-4140-9a57-dfd54fc6b486\" (UID: \"bf126b07-da06-4140-9a57-dfd54fc6b486\") " Oct 08 18:17:30 crc kubenswrapper[4859]: I1008 18:17:30.575428 4859 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-d6qdx\" (UniqueName: \"kubernetes.io/projected/87cf06ed-a83f-41a7-828d-70653580a8cb-kube-api-access-d6qdx\") pod \"87cf06ed-a83f-41a7-828d-70653580a8cb\" (UID: \"87cf06ed-a83f-41a7-828d-70653580a8cb\") " Oct 08 18:17:30 crc kubenswrapper[4859]: I1008 18:17:30.575448 4859 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-provider-selection\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Oct 08 18:17:30 crc kubenswrapper[4859]: I1008 18:17:30.575479 4859 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-etcd-serving-ca\") pod \"1bf7eb37-55a3-4c65-b768-a94c82151e69\" (UID: \"1bf7eb37-55a3-4c65-b768-a94c82151e69\") " Oct 08 18:17:30 crc kubenswrapper[4859]: I1008 18:17:30.575502 4859 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-ovnkube-script-lib\") pod \"6ea678ab-3438-413e-bfe3-290ae7725660\" (UID: \"6ea678ab-3438-413e-bfe3-290ae7725660\") " Oct 08 18:17:30 crc kubenswrapper[4859]: I1008 18:17:30.575521 4859 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/a31745f5-9847-4afe-82a5-3161cc66ca93-bound-sa-token\") pod \"a31745f5-9847-4afe-82a5-3161cc66ca93\" (UID: \"a31745f5-9847-4afe-82a5-3161cc66ca93\") " Oct 08 18:17:30 crc kubenswrapper[4859]: I1008 18:17:30.575538 4859 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/5225d0e4-402f-4861-b410-819f433b1803-catalog-content\") pod \"5225d0e4-402f-4861-b410-819f433b1803\" (UID: \"5225d0e4-402f-4861-b410-819f433b1803\") " Oct 08 18:17:30 crc kubenswrapper[4859]: I1008 18:17:30.575555 4859 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/496e6271-fb68-4057-954e-a0d97a4afa3f-kube-api-access\") pod \"496e6271-fb68-4057-954e-a0d97a4afa3f\" (UID: \"496e6271-fb68-4057-954e-a0d97a4afa3f\") " Oct 08 18:17:30 crc kubenswrapper[4859]: I1008 18:17:30.575574 4859 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-w9rds\" (UniqueName: \"kubernetes.io/projected/20b0d48f-5fd6-431c-a545-e3c800c7b866-kube-api-access-w9rds\") pod \"20b0d48f-5fd6-431c-a545-e3c800c7b866\" (UID: \"20b0d48f-5fd6-431c-a545-e3c800c7b866\") " Oct 08 18:17:30 crc kubenswrapper[4859]: I1008 18:17:30.575592 4859 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/1386a44e-36a2-460c-96d0-0359d2b6f0f5-kube-api-access\") pod \"1386a44e-36a2-460c-96d0-0359d2b6f0f5\" (UID: \"1386a44e-36a2-460c-96d0-0359d2b6f0f5\") " Oct 08 18:17:30 crc kubenswrapper[4859]: I1008 18:17:30.575612 4859 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/0b78653f-4ff9-4508-8672-245ed9b561e3-serving-cert\") pod \"0b78653f-4ff9-4508-8672-245ed9b561e3\" (UID: \"0b78653f-4ff9-4508-8672-245ed9b561e3\") " Oct 08 18:17:30 crc kubenswrapper[4859]: I1008 18:17:30.575629 4859 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"audit\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-audit\") pod \"1bf7eb37-55a3-4c65-b768-a94c82151e69\" (UID: \"1bf7eb37-55a3-4c65-b768-a94c82151e69\") " Oct 08 18:17:30 crc kubenswrapper[4859]: I1008 18:17:30.575648 4859 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/22c825df-677d-4ca6-82db-3454ed06e783-auth-proxy-config\") pod \"22c825df-677d-4ca6-82db-3454ed06e783\" (UID: \"22c825df-677d-4ca6-82db-3454ed06e783\") " Oct 08 18:17:30 crc kubenswrapper[4859]: I1008 18:17:30.575667 4859 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-7c4vf\" (UniqueName: \"kubernetes.io/projected/22c825df-677d-4ca6-82db-3454ed06e783-kube-api-access-7c4vf\") pod \"22c825df-677d-4ca6-82db-3454ed06e783\" (UID: \"22c825df-677d-4ca6-82db-3454ed06e783\") " Oct 08 18:17:30 crc kubenswrapper[4859]: I1008 18:17:30.575703 4859 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"images\" (UniqueName: \"kubernetes.io/configmap/31d8b7a1-420e-4252-a5b7-eebe8a111292-images\") pod \"31d8b7a1-420e-4252-a5b7-eebe8a111292\" (UID: \"31d8b7a1-420e-4252-a5b7-eebe8a111292\") " Oct 08 18:17:30 crc kubenswrapper[4859]: I1008 18:17:30.575726 4859 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/43509403-f426-496e-be36-56cef71462f5-console-oauth-config\") pod \"43509403-f426-496e-be36-56cef71462f5\" (UID: \"43509403-f426-496e-be36-56cef71462f5\") " Oct 08 18:17:30 crc kubenswrapper[4859]: I1008 18:17:30.575752 4859 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-etcd-client\") pod \"1bf7eb37-55a3-4c65-b768-a94c82151e69\" (UID: \"1bf7eb37-55a3-4c65-b768-a94c82151e69\") " Oct 08 18:17:30 crc kubenswrapper[4859]: I1008 18:17:30.575772 4859 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-s4n52\" (UniqueName: \"kubernetes.io/projected/925f1c65-6136-48ba-85aa-3a3b50560753-kube-api-access-s4n52\") pod \"925f1c65-6136-48ba-85aa-3a3b50560753\" (UID: \"925f1c65-6136-48ba-85aa-3a3b50560753\") " Oct 08 18:17:30 crc kubenswrapper[4859]: I1008 18:17:30.575792 4859 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/9d4552c7-cd75-42dd-8880-30dd377c49a4-trusted-ca\") pod \"9d4552c7-cd75-42dd-8880-30dd377c49a4\" (UID: \"9d4552c7-cd75-42dd-8880-30dd377c49a4\") " Oct 08 18:17:30 crc kubenswrapper[4859]: I1008 18:17:30.575813 4859 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/0b78653f-4ff9-4508-8672-245ed9b561e3-service-ca\") pod \"0b78653f-4ff9-4508-8672-245ed9b561e3\" (UID: \"0b78653f-4ff9-4508-8672-245ed9b561e3\") " Oct 08 18:17:30 crc kubenswrapper[4859]: I1008 18:17:30.575835 4859 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"mcc-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/0b574797-001e-440a-8f4e-c0be86edad0f-mcc-auth-proxy-config\") pod \"0b574797-001e-440a-8f4e-c0be86edad0f\" (UID: \"0b574797-001e-440a-8f4e-c0be86edad0f\") " Oct 08 18:17:30 crc kubenswrapper[4859]: I1008 18:17:30.575854 4859 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-serving-cert\") pod \"bc5039c0-ea34-426b-a2b7-fbbc87b49a6d\" (UID: \"bc5039c0-ea34-426b-a2b7-fbbc87b49a6d\") " Oct 08 18:17:30 crc kubenswrapper[4859]: I1008 18:17:30.575873 4859 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-cliconfig\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Oct 08 18:17:30 crc kubenswrapper[4859]: I1008 18:17:30.575895 4859 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-encryption-config\") pod \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\" (UID: \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\") " Oct 08 18:17:30 crc kubenswrapper[4859]: I1008 18:17:30.575919 4859 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-config\") pod \"6509e943-70c6-444c-bc41-48a544e36fbd\" (UID: \"6509e943-70c6-444c-bc41-48a544e36fbd\") " Oct 08 18:17:30 crc kubenswrapper[4859]: I1008 18:17:30.575943 4859 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-bf2bz\" (UniqueName: \"kubernetes.io/projected/1d611f23-29be-4491-8495-bee1670e935f-kube-api-access-bf2bz\") pod \"1d611f23-29be-4491-8495-bee1670e935f\" (UID: \"1d611f23-29be-4491-8495-bee1670e935f\") " Oct 08 18:17:30 crc kubenswrapper[4859]: I1008 18:17:30.575962 4859 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-config\") pod \"210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c\" (UID: \"210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c\") " Oct 08 18:17:30 crc kubenswrapper[4859]: I1008 18:17:30.575981 4859 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-service-ca\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Oct 08 18:17:30 crc kubenswrapper[4859]: I1008 18:17:30.576000 4859 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-tk88c\" (UniqueName: \"kubernetes.io/projected/7539238d-5fe0-46ed-884e-1c3b566537ec-kube-api-access-tk88c\") pod \"7539238d-5fe0-46ed-884e-1c3b566537ec\" (UID: \"7539238d-5fe0-46ed-884e-1c3b566537ec\") " Oct 08 18:17:30 crc kubenswrapper[4859]: I1008 18:17:30.576022 4859 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/496e6271-fb68-4057-954e-a0d97a4afa3f-serving-cert\") pod \"496e6271-fb68-4057-954e-a0d97a4afa3f\" (UID: \"496e6271-fb68-4057-954e-a0d97a4afa3f\") " Oct 08 18:17:30 crc kubenswrapper[4859]: I1008 18:17:30.576040 4859 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-6g6sz\" (UniqueName: \"kubernetes.io/projected/6509e943-70c6-444c-bc41-48a544e36fbd-kube-api-access-6g6sz\") pod \"6509e943-70c6-444c-bc41-48a544e36fbd\" (UID: \"6509e943-70c6-444c-bc41-48a544e36fbd\") " Oct 08 18:17:30 crc kubenswrapper[4859]: I1008 18:17:30.576060 4859 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"signing-key\" (UniqueName: \"kubernetes.io/secret/25e176fe-21b4-4974-b1ed-c8b94f112a7f-signing-key\") pod \"25e176fe-21b4-4974-b1ed-c8b94f112a7f\" (UID: \"25e176fe-21b4-4974-b1ed-c8b94f112a7f\") " Oct 08 18:17:30 crc kubenswrapper[4859]: I1008 18:17:30.576084 4859 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etcd-ca\" (UniqueName: \"kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-ca\") pod \"09efc573-dbb6-4249-bd59-9b87aba8dd28\" (UID: \"09efc573-dbb6-4249-bd59-9b87aba8dd28\") " Oct 08 18:17:30 crc kubenswrapper[4859]: I1008 18:17:30.576111 4859 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/496e6271-fb68-4057-954e-a0d97a4afa3f-config\") pod \"496e6271-fb68-4057-954e-a0d97a4afa3f\" (UID: \"496e6271-fb68-4057-954e-a0d97a4afa3f\") " Oct 08 18:17:30 crc kubenswrapper[4859]: I1008 18:17:30.576130 4859 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-6ccd8\" (UniqueName: \"kubernetes.io/projected/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-kube-api-access-6ccd8\") pod \"308be0ea-9f5f-4b29-aeb1-5abd31a0b17b\" (UID: \"308be0ea-9f5f-4b29-aeb1-5abd31a0b17b\") " Oct 08 18:17:30 crc kubenswrapper[4859]: I1008 18:17:30.576149 4859 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/01ab3dd5-8196-46d0-ad33-122e2ca51def-config\") pod \"01ab3dd5-8196-46d0-ad33-122e2ca51def\" (UID: \"01ab3dd5-8196-46d0-ad33-122e2ca51def\") " Oct 08 18:17:30 crc kubenswrapper[4859]: I1008 18:17:30.576167 4859 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/5b88f790-22fa-440e-b583-365168c0b23d-metrics-certs\") pod \"5b88f790-22fa-440e-b583-365168c0b23d\" (UID: \"5b88f790-22fa-440e-b583-365168c0b23d\") " Oct 08 18:17:30 crc kubenswrapper[4859]: I1008 18:17:30.576184 4859 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-vt5rc\" (UniqueName: \"kubernetes.io/projected/44663579-783b-4372-86d6-acf235a62d72-kube-api-access-vt5rc\") pod \"44663579-783b-4372-86d6-acf235a62d72\" (UID: \"44663579-783b-4372-86d6-acf235a62d72\") " Oct 08 18:17:30 crc kubenswrapper[4859]: I1008 18:17:30.576202 4859 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"image-import-ca\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-image-import-ca\") pod \"1bf7eb37-55a3-4c65-b768-a94c82151e69\" (UID: \"1bf7eb37-55a3-4c65-b768-a94c82151e69\") " Oct 08 18:17:30 crc kubenswrapper[4859]: I1008 18:17:30.576222 4859 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/6402fda4-df10-493c-b4e5-d0569419652d-config\") pod \"6402fda4-df10-493c-b4e5-d0569419652d\" (UID: \"6402fda4-df10-493c-b4e5-d0569419652d\") " Oct 08 18:17:30 crc kubenswrapper[4859]: I1008 18:17:30.576244 4859 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-trusted-ca-bundle\") pod \"43509403-f426-496e-be36-56cef71462f5\" (UID: \"43509403-f426-496e-be36-56cef71462f5\") " Oct 08 18:17:30 crc kubenswrapper[4859]: I1008 18:17:30.576269 4859 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/a31745f5-9847-4afe-82a5-3161cc66ca93-trusted-ca\") pod \"a31745f5-9847-4afe-82a5-3161cc66ca93\" (UID: \"a31745f5-9847-4afe-82a5-3161cc66ca93\") " Oct 08 18:17:30 crc kubenswrapper[4859]: I1008 18:17:30.576295 4859 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-serving-cert\") pod \"1bf7eb37-55a3-4c65-b768-a94c82151e69\" (UID: \"1bf7eb37-55a3-4c65-b768-a94c82151e69\") " Oct 08 18:17:30 crc kubenswrapper[4859]: I1008 18:17:30.576318 4859 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-profile-collector-cert\") pod \"f88749ec-7931-4ee7-b3fc-1ec5e11f92e9\" (UID: \"f88749ec-7931-4ee7-b3fc-1ec5e11f92e9\") " Oct 08 18:17:30 crc kubenswrapper[4859]: I1008 18:17:30.576340 4859 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-cfbct\" (UniqueName: \"kubernetes.io/projected/57a731c4-ef35-47a8-b875-bfb08a7f8011-kube-api-access-cfbct\") pod \"57a731c4-ef35-47a8-b875-bfb08a7f8011\" (UID: \"57a731c4-ef35-47a8-b875-bfb08a7f8011\") " Oct 08 18:17:30 crc kubenswrapper[4859]: I1008 18:17:30.576358 4859 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"certs\" (UniqueName: \"kubernetes.io/secret/5fe579f8-e8a6-4643-bce5-a661393c4dde-certs\") pod \"5fe579f8-e8a6-4643-bce5-a661393c4dde\" (UID: \"5fe579f8-e8a6-4643-bce5-a661393c4dde\") " Oct 08 18:17:30 crc kubenswrapper[4859]: I1008 18:17:30.576376 4859 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"mcd-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/fda69060-fa79-4696-b1a6-7980f124bf7c-mcd-auth-proxy-config\") pod \"fda69060-fa79-4696-b1a6-7980f124bf7c\" (UID: \"fda69060-fa79-4696-b1a6-7980f124bf7c\") " Oct 08 18:17:30 crc kubenswrapper[4859]: I1008 18:17:30.576398 4859 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/5225d0e4-402f-4861-b410-819f433b1803-utilities\") pod \"5225d0e4-402f-4861-b410-819f433b1803\" (UID: \"5225d0e4-402f-4861-b410-819f433b1803\") " Oct 08 18:17:30 crc kubenswrapper[4859]: I1008 18:17:30.576423 4859 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/bf126b07-da06-4140-9a57-dfd54fc6b486-bound-sa-token\") pod \"bf126b07-da06-4140-9a57-dfd54fc6b486\" (UID: \"bf126b07-da06-4140-9a57-dfd54fc6b486\") " Oct 08 18:17:30 crc kubenswrapper[4859]: I1008 18:17:30.576449 4859 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-trusted-ca-bundle\") pod \"6509e943-70c6-444c-bc41-48a544e36fbd\" (UID: \"6509e943-70c6-444c-bc41-48a544e36fbd\") " Oct 08 18:17:30 crc kubenswrapper[4859]: I1008 18:17:30.576476 4859 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-zgdk5\" (UniqueName: \"kubernetes.io/projected/31d8b7a1-420e-4252-a5b7-eebe8a111292-kube-api-access-zgdk5\") pod \"31d8b7a1-420e-4252-a5b7-eebe8a111292\" (UID: \"31d8b7a1-420e-4252-a5b7-eebe8a111292\") " Oct 08 18:17:30 crc kubenswrapper[4859]: I1008 18:17:30.576505 4859 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-lzf88\" (UniqueName: \"kubernetes.io/projected/0b574797-001e-440a-8f4e-c0be86edad0f-kube-api-access-lzf88\") pod \"0b574797-001e-440a-8f4e-c0be86edad0f\" (UID: \"0b574797-001e-440a-8f4e-c0be86edad0f\") " Oct 08 18:17:30 crc kubenswrapper[4859]: I1008 18:17:30.576534 4859 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/6ea678ab-3438-413e-bfe3-290ae7725660-ovn-node-metrics-cert\") pod \"6ea678ab-3438-413e-bfe3-290ae7725660\" (UID: \"6ea678ab-3438-413e-bfe3-290ae7725660\") " Oct 08 18:17:30 crc kubenswrapper[4859]: I1008 18:17:30.576555 4859 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-w4xd4\" (UniqueName: \"kubernetes.io/projected/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-kube-api-access-w4xd4\") pod \"8cea82b4-6893-4ddc-af9f-1bb5ae425c5b\" (UID: \"8cea82b4-6893-4ddc-af9f-1bb5ae425c5b\") " Oct 08 18:17:30 crc kubenswrapper[4859]: I1008 18:17:30.576580 4859 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-x7zkh\" (UniqueName: \"kubernetes.io/projected/6731426b-95fe-49ff-bb5f-40441049fde2-kube-api-access-x7zkh\") pod \"6731426b-95fe-49ff-bb5f-40441049fde2\" (UID: \"6731426b-95fe-49ff-bb5f-40441049fde2\") " Oct 08 18:17:30 crc kubenswrapper[4859]: I1008 18:17:30.576610 4859 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovn-control-plane-metrics-cert\" (UniqueName: \"kubernetes.io/secret/925f1c65-6136-48ba-85aa-3a3b50560753-ovn-control-plane-metrics-cert\") pod \"925f1c65-6136-48ba-85aa-3a3b50560753\" (UID: \"925f1c65-6136-48ba-85aa-3a3b50560753\") " Oct 08 18:17:30 crc kubenswrapper[4859]: I1008 18:17:30.576634 4859 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/1386a44e-36a2-460c-96d0-0359d2b6f0f5-config\") pod \"1386a44e-36a2-460c-96d0-0359d2b6f0f5\" (UID: \"1386a44e-36a2-460c-96d0-0359d2b6f0f5\") " Oct 08 18:17:30 crc kubenswrapper[4859]: I1008 18:17:30.576656 4859 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-console-config\") pod \"43509403-f426-496e-be36-56cef71462f5\" (UID: \"43509403-f426-496e-be36-56cef71462f5\") " Oct 08 18:17:30 crc kubenswrapper[4859]: I1008 18:17:30.576686 4859 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/b6312bbd-5731-4ea0-a20f-81d5a57df44a-profile-collector-cert\") pod \"b6312bbd-5731-4ea0-a20f-81d5a57df44a\" (UID: \"b6312bbd-5731-4ea0-a20f-81d5a57df44a\") " Oct 08 18:17:30 crc kubenswrapper[4859]: I1008 18:17:30.576727 4859 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-oauth-serving-cert\") pod \"43509403-f426-496e-be36-56cef71462f5\" (UID: \"43509403-f426-496e-be36-56cef71462f5\") " Oct 08 18:17:30 crc kubenswrapper[4859]: I1008 18:17:30.576753 4859 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-279lb\" (UniqueName: \"kubernetes.io/projected/7bb08738-c794-4ee8-9972-3a62ca171029-kube-api-access-279lb\") pod \"7bb08738-c794-4ee8-9972-3a62ca171029\" (UID: \"7bb08738-c794-4ee8-9972-3a62ca171029\") " Oct 08 18:17:30 crc kubenswrapper[4859]: I1008 18:17:30.576815 4859 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-config\") pod \"8cea82b4-6893-4ddc-af9f-1bb5ae425c5b\" (UID: \"8cea82b4-6893-4ddc-af9f-1bb5ae425c5b\") " Oct 08 18:17:30 crc kubenswrapper[4859]: I1008 18:17:30.576840 4859 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etcd-service-ca\" (UniqueName: \"kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-service-ca\") pod \"09efc573-dbb6-4249-bd59-9b87aba8dd28\" (UID: \"09efc573-dbb6-4249-bd59-9b87aba8dd28\") " Oct 08 18:17:30 crc kubenswrapper[4859]: I1008 18:17:30.576864 4859 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-registry-tls\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 08 18:17:30 crc kubenswrapper[4859]: I1008 18:17:30.576888 4859 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-apiservice-cert\") pod \"308be0ea-9f5f-4b29-aeb1-5abd31a0b17b\" (UID: \"308be0ea-9f5f-4b29-aeb1-5abd31a0b17b\") " Oct 08 18:17:30 crc kubenswrapper[4859]: I1008 18:17:30.576914 4859 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/5441d097-087c-4d9a-baa8-b210afa90fc9-config\") pod \"5441d097-087c-4d9a-baa8-b210afa90fc9\" (UID: \"5441d097-087c-4d9a-baa8-b210afa90fc9\") " Oct 08 18:17:30 crc kubenswrapper[4859]: I1008 18:17:30.576938 4859 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-lz9wn\" (UniqueName: \"kubernetes.io/projected/a31745f5-9847-4afe-82a5-3161cc66ca93-kube-api-access-lz9wn\") pod \"a31745f5-9847-4afe-82a5-3161cc66ca93\" (UID: \"a31745f5-9847-4afe-82a5-3161cc66ca93\") " Oct 08 18:17:30 crc kubenswrapper[4859]: I1008 18:17:30.576962 4859 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-ovnkube-config\") pod \"6ea678ab-3438-413e-bfe3-290ae7725660\" (UID: \"6ea678ab-3438-413e-bfe3-290ae7725660\") " Oct 08 18:17:30 crc kubenswrapper[4859]: I1008 18:17:30.576986 4859 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-config\") pod \"09efc573-dbb6-4249-bd59-9b87aba8dd28\" (UID: \"09efc573-dbb6-4249-bd59-9b87aba8dd28\") " Oct 08 18:17:30 crc kubenswrapper[4859]: I1008 18:17:30.577009 4859 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-xcphl\" (UniqueName: \"kubernetes.io/projected/7583ce53-e0fe-4a16-9e4d-50516596a136-kube-api-access-xcphl\") pod \"7583ce53-e0fe-4a16-9e4d-50516596a136\" (UID: \"7583ce53-e0fe-4a16-9e4d-50516596a136\") " Oct 08 18:17:30 crc kubenswrapper[4859]: I1008 18:17:30.577039 4859 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-x2m85\" (UniqueName: \"kubernetes.io/projected/cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d-kube-api-access-x2m85\") pod \"cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d\" (UID: \"cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d\") " Oct 08 18:17:30 crc kubenswrapper[4859]: I1008 18:17:30.577064 4859 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-audit-policies\") pod \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\" (UID: \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\") " Oct 08 18:17:30 crc kubenswrapper[4859]: I1008 18:17:30.577167 4859 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 08 18:17:30 crc kubenswrapper[4859]: I1008 18:17:30.577194 4859 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"images\" (UniqueName: \"kubernetes.io/configmap/6402fda4-df10-493c-b4e5-d0569419652d-images\") pod \"6402fda4-df10-493c-b4e5-d0569419652d\" (UID: \"6402fda4-df10-493c-b4e5-d0569419652d\") " Oct 08 18:17:30 crc kubenswrapper[4859]: I1008 18:17:30.577217 4859 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"cni-sysctl-allowlist\" (UniqueName: \"kubernetes.io/configmap/7bb08738-c794-4ee8-9972-3a62ca171029-cni-sysctl-allowlist\") pod \"7bb08738-c794-4ee8-9972-3a62ca171029\" (UID: \"7bb08738-c794-4ee8-9972-3a62ca171029\") " Oct 08 18:17:30 crc kubenswrapper[4859]: I1008 18:17:30.577242 4859 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/bf126b07-da06-4140-9a57-dfd54fc6b486-trusted-ca\") pod \"bf126b07-da06-4140-9a57-dfd54fc6b486\" (UID: \"bf126b07-da06-4140-9a57-dfd54fc6b486\") " Oct 08 18:17:30 crc kubenswrapper[4859]: I1008 18:17:30.577267 4859 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-d4lsv\" (UniqueName: \"kubernetes.io/projected/25e176fe-21b4-4974-b1ed-c8b94f112a7f-kube-api-access-d4lsv\") pod \"25e176fe-21b4-4974-b1ed-c8b94f112a7f\" (UID: \"25e176fe-21b4-4974-b1ed-c8b94f112a7f\") " Oct 08 18:17:30 crc kubenswrapper[4859]: I1008 18:17:30.577292 4859 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-ngvvp\" (UniqueName: \"kubernetes.io/projected/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-kube-api-access-ngvvp\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Oct 08 18:17:30 crc kubenswrapper[4859]: I1008 18:17:30.577316 4859 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/57a731c4-ef35-47a8-b875-bfb08a7f8011-utilities\") pod \"57a731c4-ef35-47a8-b875-bfb08a7f8011\" (UID: \"57a731c4-ef35-47a8-b875-bfb08a7f8011\") " Oct 08 18:17:30 crc kubenswrapper[4859]: I1008 18:17:30.577339 4859 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-serving-cert\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Oct 08 18:17:30 crc kubenswrapper[4859]: I1008 18:17:30.577362 4859 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-env-overrides\") pod \"6ea678ab-3438-413e-bfe3-290ae7725660\" (UID: \"6ea678ab-3438-413e-bfe3-290ae7725660\") " Oct 08 18:17:30 crc kubenswrapper[4859]: I1008 18:17:30.577384 4859 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-client-ca\") pod \"7583ce53-e0fe-4a16-9e4d-50516596a136\" (UID: \"7583ce53-e0fe-4a16-9e4d-50516596a136\") " Oct 08 18:17:30 crc kubenswrapper[4859]: I1008 18:17:30.577406 4859 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"image-registry-operator-tls\" (UniqueName: \"kubernetes.io/secret/bf126b07-da06-4140-9a57-dfd54fc6b486-image-registry-operator-tls\") pod \"bf126b07-da06-4140-9a57-dfd54fc6b486\" (UID: \"bf126b07-da06-4140-9a57-dfd54fc6b486\") " Oct 08 18:17:30 crc kubenswrapper[4859]: I1008 18:17:30.577428 4859 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/01ab3dd5-8196-46d0-ad33-122e2ca51def-serving-cert\") pod \"01ab3dd5-8196-46d0-ad33-122e2ca51def\" (UID: \"01ab3dd5-8196-46d0-ad33-122e2ca51def\") " Oct 08 18:17:30 crc kubenswrapper[4859]: I1008 18:17:30.577448 4859 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-pcxfs\" (UniqueName: \"kubernetes.io/projected/9d4552c7-cd75-42dd-8880-30dd377c49a4-kube-api-access-pcxfs\") pod \"9d4552c7-cd75-42dd-8880-30dd377c49a4\" (UID: \"9d4552c7-cd75-42dd-8880-30dd377c49a4\") " Oct 08 18:17:30 crc kubenswrapper[4859]: I1008 18:17:30.577502 4859 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-etc-kube\" (UniqueName: \"kubernetes.io/host-path/37a5e44f-9a88-4405-be8a-b645485e7312-host-etc-kube\") pod \"network-operator-58b4c7f79c-55gtf\" (UID: \"37a5e44f-9a88-4405-be8a-b645485e7312\") " pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" Oct 08 18:17:30 crc kubenswrapper[4859]: I1008 18:17:30.577535 4859 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"iptables-alerter-script\" (UniqueName: \"kubernetes.io/configmap/d75a4c96-2883-4a0b-bab2-0fab2b6c0b49-iptables-alerter-script\") pod \"iptables-alerter-4ln5h\" (UID: \"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\") " pod="openshift-network-operator/iptables-alerter-4ln5h" Oct 08 18:17:30 crc kubenswrapper[4859]: I1008 18:17:30.577559 4859 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovnkube-identity-cm\" (UniqueName: \"kubernetes.io/configmap/ef543e1b-8068-4ea3-b32a-61027b32e95d-ovnkube-identity-cm\") pod \"network-node-identity-vrzqb\" (UID: \"ef543e1b-8068-4ea3-b32a-61027b32e95d\") " pod="openshift-network-node-identity/network-node-identity-vrzqb" Oct 08 18:17:30 crc kubenswrapper[4859]: I1008 18:17:30.577608 4859 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/ef543e1b-8068-4ea3-b32a-61027b32e95d-env-overrides\") pod \"network-node-identity-vrzqb\" (UID: \"ef543e1b-8068-4ea3-b32a-61027b32e95d\") " pod="openshift-network-node-identity/network-node-identity-vrzqb" Oct 08 18:17:30 crc kubenswrapper[4859]: I1008 18:17:30.577634 4859 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 08 18:17:30 crc kubenswrapper[4859]: I1008 18:17:30.577657 4859 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 08 18:17:30 crc kubenswrapper[4859]: I1008 18:17:30.577677 4859 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 08 18:17:30 crc kubenswrapper[4859]: I1008 18:17:30.577785 4859 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/ef543e1b-8068-4ea3-b32a-61027b32e95d-webhook-cert\") pod \"network-node-identity-vrzqb\" (UID: \"ef543e1b-8068-4ea3-b32a-61027b32e95d\") " pod="openshift-network-node-identity/network-node-identity-vrzqb" Oct 08 18:17:30 crc kubenswrapper[4859]: I1008 18:17:30.577814 4859 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/d75a4c96-2883-4a0b-bab2-0fab2b6c0b49-host-slash\") pod \"iptables-alerter-4ln5h\" (UID: \"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\") " pod="openshift-network-operator/iptables-alerter-4ln5h" Oct 08 18:17:30 crc kubenswrapper[4859]: I1008 18:17:30.577839 4859 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/37a5e44f-9a88-4405-be8a-b645485e7312-metrics-tls\") pod \"network-operator-58b4c7f79c-55gtf\" (UID: \"37a5e44f-9a88-4405-be8a-b645485e7312\") " pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" Oct 08 18:17:30 crc kubenswrapper[4859]: I1008 18:17:30.577860 4859 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 08 18:17:30 crc kubenswrapper[4859]: I1008 18:17:30.577887 4859 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rczfb\" (UniqueName: \"kubernetes.io/projected/d75a4c96-2883-4a0b-bab2-0fab2b6c0b49-kube-api-access-rczfb\") pod \"iptables-alerter-4ln5h\" (UID: \"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\") " pod="openshift-network-operator/iptables-alerter-4ln5h" Oct 08 18:17:30 crc kubenswrapper[4859]: I1008 18:17:30.577909 4859 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2kz5\" (UniqueName: \"kubernetes.io/projected/ef543e1b-8068-4ea3-b32a-61027b32e95d-kube-api-access-s2kz5\") pod \"network-node-identity-vrzqb\" (UID: \"ef543e1b-8068-4ea3-b32a-61027b32e95d\") " pod="openshift-network-node-identity/network-node-identity-vrzqb" Oct 08 18:17:30 crc kubenswrapper[4859]: I1008 18:17:30.577933 4859 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rdwmf\" (UniqueName: \"kubernetes.io/projected/37a5e44f-9a88-4405-be8a-b645485e7312-kube-api-access-rdwmf\") pod \"network-operator-58b4c7f79c-55gtf\" (UID: \"37a5e44f-9a88-4405-be8a-b645485e7312\") " pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" Oct 08 18:17:30 crc kubenswrapper[4859]: I1008 18:17:30.578306 4859 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-htfz6\" (UniqueName: \"kubernetes.io/projected/6ea678ab-3438-413e-bfe3-290ae7725660-kube-api-access-htfz6\") on node \"crc\" DevicePath \"\"" Oct 08 18:17:30 crc kubenswrapper[4859]: I1008 18:17:30.578328 4859 reconciler_common.go:293] "Volume detached for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-marketplace-trusted-ca\") on node \"crc\" DevicePath \"\"" Oct 08 18:17:30 crc kubenswrapper[4859]: I1008 18:17:30.578342 4859 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/7583ce53-e0fe-4a16-9e4d-50516596a136-serving-cert\") on node \"crc\" DevicePath \"\"" Oct 08 18:17:30 crc kubenswrapper[4859]: I1008 18:17:30.578354 4859 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-dbsvg\" (UniqueName: \"kubernetes.io/projected/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-kube-api-access-dbsvg\") on node \"crc\" DevicePath \"\"" Oct 08 18:17:30 crc kubenswrapper[4859]: I1008 18:17:30.578372 4859 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-jkwtn\" (UniqueName: \"kubernetes.io/projected/5b88f790-22fa-440e-b583-365168c0b23d-kube-api-access-jkwtn\") on node \"crc\" DevicePath \"\"" Oct 08 18:17:30 crc kubenswrapper[4859]: I1008 18:17:30.578386 4859 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-trusted-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 08 18:17:30 crc kubenswrapper[4859]: I1008 18:17:30.578401 4859 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/09efc573-dbb6-4249-bd59-9b87aba8dd28-serving-cert\") on node \"crc\" DevicePath \"\"" Oct 08 18:17:30 crc kubenswrapper[4859]: I1008 18:17:30.578415 4859 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-serving-cert\") on node \"crc\" DevicePath \"\"" Oct 08 18:17:30 crc kubenswrapper[4859]: I1008 18:17:30.578428 4859 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-idp-0-file-data\") on node \"crc\" DevicePath \"\"" Oct 08 18:17:30 crc kubenswrapper[4859]: I1008 18:17:30.578442 4859 reconciler_common.go:293] "Volume detached for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-marketplace-operator-metrics\") on node \"crc\" DevicePath \"\"" Oct 08 18:17:30 crc kubenswrapper[4859]: I1008 18:17:30.578457 4859 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-nzwt7\" (UniqueName: \"kubernetes.io/projected/96b93a3a-6083-4aea-8eab-fe1aa8245ad9-kube-api-access-nzwt7\") on node \"crc\" DevicePath \"\"" Oct 08 18:17:30 crc kubenswrapper[4859]: I1008 18:17:30.578469 4859 reconciler_common.go:293] "Volume detached for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/87cf06ed-a83f-41a7-828d-70653580a8cb-config-volume\") on node \"crc\" DevicePath \"\"" Oct 08 18:17:30 crc kubenswrapper[4859]: I1008 18:17:30.578482 4859 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/7539238d-5fe0-46ed-884e-1c3b566537ec-serving-cert\") on node \"crc\" DevicePath \"\"" Oct 08 18:17:30 crc kubenswrapper[4859]: I1008 18:17:30.578494 4859 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-login\") on node \"crc\" DevicePath \"\"" Oct 08 18:17:30 crc kubenswrapper[4859]: I1008 18:17:30.578535 4859 reconciler_common.go:293] "Volume detached for volume \"default-certificate\" (UniqueName: \"kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-default-certificate\") on node \"crc\" DevicePath \"\"" Oct 08 18:17:30 crc kubenswrapper[4859]: I1008 18:17:30.578549 4859 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-utilities\") on node \"crc\" DevicePath \"\"" Oct 08 18:17:30 crc kubenswrapper[4859]: I1008 18:17:30.578561 4859 reconciler_common.go:293] "Volume detached for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/0b574797-001e-440a-8f4e-c0be86edad0f-proxy-tls\") on node \"crc\" DevicePath \"\"" Oct 08 18:17:30 crc kubenswrapper[4859]: I1008 18:17:30.578575 4859 reconciler_common.go:293] "Volume detached for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/c03ee662-fb2f-4fc4-a2c1-af487c19d254-service-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 08 18:17:30 crc kubenswrapper[4859]: I1008 18:17:30.578587 4859 reconciler_common.go:293] "Volume detached for volume \"node-bootstrap-token\" (UniqueName: \"kubernetes.io/secret/5fe579f8-e8a6-4643-bce5-a661393c4dde-node-bootstrap-token\") on node \"crc\" DevicePath \"\"" Oct 08 18:17:30 crc kubenswrapper[4859]: I1008 18:17:30.578601 4859 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-jhbk2\" (UniqueName: \"kubernetes.io/projected/bd23aa5c-e532-4e53-bccf-e79f130c5ae8-kube-api-access-jhbk2\") on node \"crc\" DevicePath \"\"" Oct 08 18:17:30 crc kubenswrapper[4859]: I1008 18:17:30.578615 4859 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-pj782\" (UniqueName: \"kubernetes.io/projected/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-kube-api-access-pj782\") on node \"crc\" DevicePath \"\"" Oct 08 18:17:30 crc kubenswrapper[4859]: I1008 18:17:30.578627 4859 reconciler_common.go:293] "Volume detached for volume \"cert\" (UniqueName: \"kubernetes.io/secret/20b0d48f-5fd6-431c-a545-e3c800c7b866-cert\") on node \"crc\" DevicePath \"\"" Oct 08 18:17:30 crc kubenswrapper[4859]: I1008 18:17:30.578638 4859 reconciler_common.go:293] "Volume detached for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/b6312bbd-5731-4ea0-a20f-81d5a57df44a-srv-cert\") on node \"crc\" DevicePath \"\"" Oct 08 18:17:30 crc kubenswrapper[4859]: I1008 18:17:30.578650 4859 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-sb6h7\" (UniqueName: \"kubernetes.io/projected/1bf7eb37-55a3-4c65-b768-a94c82151e69-kube-api-access-sb6h7\") on node \"crc\" DevicePath \"\"" Oct 08 18:17:30 crc kubenswrapper[4859]: I1008 18:17:30.578674 4859 reconciler_common.go:293] "Volume detached for volume \"stats-auth\" (UniqueName: \"kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-stats-auth\") on node \"crc\" DevicePath \"\"" Oct 08 18:17:30 crc kubenswrapper[4859]: I1008 18:17:30.578706 4859 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-ocp-branding-template\") on node \"crc\" DevicePath \"\"" Oct 08 18:17:30 crc kubenswrapper[4859]: I1008 18:17:30.582836 4859 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-08T18:17:29Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T18:17:29Z\\\",\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Oct 08 18:17:30 crc kubenswrapper[4859]: I1008 18:17:30.574939 4859 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/87cf06ed-a83f-41a7-828d-70653580a8cb-metrics-tls" (OuterVolumeSpecName: "metrics-tls") pod "87cf06ed-a83f-41a7-828d-70653580a8cb" (UID: "87cf06ed-a83f-41a7-828d-70653580a8cb"). InnerVolumeSpecName "metrics-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 08 18:17:30 crc kubenswrapper[4859]: I1008 18:17:30.575123 4859 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1386a44e-36a2-460c-96d0-0359d2b6f0f5-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "1386a44e-36a2-460c-96d0-0359d2b6f0f5" (UID: "1386a44e-36a2-460c-96d0-0359d2b6f0f5"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 08 18:17:30 crc kubenswrapper[4859]: I1008 18:17:30.575439 4859 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-webhook-cert" (OuterVolumeSpecName: "webhook-cert") pod "308be0ea-9f5f-4b29-aeb1-5abd31a0b17b" (UID: "308be0ea-9f5f-4b29-aeb1-5abd31a0b17b"). InnerVolumeSpecName "webhook-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 08 18:17:30 crc kubenswrapper[4859]: I1008 18:17:30.575483 4859 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e7e6199b-1264-4501-8953-767f51328d08-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "e7e6199b-1264-4501-8953-767f51328d08" (UID: "e7e6199b-1264-4501-8953-767f51328d08"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 08 18:17:30 crc kubenswrapper[4859]: I1008 18:17:30.575583 4859 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/5441d097-087c-4d9a-baa8-b210afa90fc9-client-ca" (OuterVolumeSpecName: "client-ca") pod "5441d097-087c-4d9a-baa8-b210afa90fc9" (UID: "5441d097-087c-4d9a-baa8-b210afa90fc9"). InnerVolumeSpecName "client-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 08 18:17:30 crc kubenswrapper[4859]: I1008 18:17:30.575760 4859 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/925f1c65-6136-48ba-85aa-3a3b50560753-ovnkube-config" (OuterVolumeSpecName: "ovnkube-config") pod "925f1c65-6136-48ba-85aa-3a3b50560753" (UID: "925f1c65-6136-48ba-85aa-3a3b50560753"). InnerVolumeSpecName "ovnkube-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 08 18:17:30 crc kubenswrapper[4859]: I1008 18:17:30.575794 4859 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/25e176fe-21b4-4974-b1ed-c8b94f112a7f-signing-cabundle" (OuterVolumeSpecName: "signing-cabundle") pod "25e176fe-21b4-4974-b1ed-c8b94f112a7f" (UID: "25e176fe-21b4-4974-b1ed-c8b94f112a7f"). InnerVolumeSpecName "signing-cabundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 08 18:17:30 crc kubenswrapper[4859]: I1008 18:17:30.575812 4859 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/1d611f23-29be-4491-8495-bee1670e935f-utilities" (OuterVolumeSpecName: "utilities") pod "1d611f23-29be-4491-8495-bee1670e935f" (UID: "1d611f23-29be-4491-8495-bee1670e935f"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 08 18:17:30 crc kubenswrapper[4859]: I1008 18:17:30.575968 4859 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-metrics-certs" (OuterVolumeSpecName: "metrics-certs") pod "c03ee662-fb2f-4fc4-a2c1-af487c19d254" (UID: "c03ee662-fb2f-4fc4-a2c1-af487c19d254"). InnerVolumeSpecName "metrics-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 08 18:17:30 crc kubenswrapper[4859]: I1008 18:17:30.575982 4859 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5225d0e4-402f-4861-b410-819f433b1803-kube-api-access-9xfj7" (OuterVolumeSpecName: "kube-api-access-9xfj7") pod "5225d0e4-402f-4861-b410-819f433b1803" (UID: "5225d0e4-402f-4861-b410-819f433b1803"). InnerVolumeSpecName "kube-api-access-9xfj7". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 08 18:17:30 crc kubenswrapper[4859]: I1008 18:17:30.575998 4859 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/22c825df-677d-4ca6-82db-3454ed06e783-machine-approver-tls" (OuterVolumeSpecName: "machine-approver-tls") pod "22c825df-677d-4ca6-82db-3454ed06e783" (UID: "22c825df-677d-4ca6-82db-3454ed06e783"). InnerVolumeSpecName "machine-approver-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 08 18:17:30 crc kubenswrapper[4859]: I1008 18:17:30.576128 4859 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-available-featuregates" (OuterVolumeSpecName: "available-featuregates") pod "bc5039c0-ea34-426b-a2b7-fbbc87b49a6d" (UID: "bc5039c0-ea34-426b-a2b7-fbbc87b49a6d"). InnerVolumeSpecName "available-featuregates". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 08 18:17:30 crc kubenswrapper[4859]: I1008 18:17:30.576201 4859 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59-kube-api-access-wxkg8" (OuterVolumeSpecName: "kube-api-access-wxkg8") pod "3cb93b32-e0ae-4377-b9c8-fdb9842c6d59" (UID: "3cb93b32-e0ae-4377-b9c8-fdb9842c6d59"). InnerVolumeSpecName "kube-api-access-wxkg8". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 08 18:17:30 crc kubenswrapper[4859]: I1008 18:17:30.576235 4859 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6402fda4-df10-493c-b4e5-d0569419652d-machine-api-operator-tls" (OuterVolumeSpecName: "machine-api-operator-tls") pod "6402fda4-df10-493c-b4e5-d0569419652d" (UID: "6402fda4-df10-493c-b4e5-d0569419652d"). InnerVolumeSpecName "machine-api-operator-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 08 18:17:30 crc kubenswrapper[4859]: I1008 18:17:30.576268 4859 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-kube-api-access-mnrrd" (OuterVolumeSpecName: "kube-api-access-mnrrd") pod "bc5039c0-ea34-426b-a2b7-fbbc87b49a6d" (UID: "bc5039c0-ea34-426b-a2b7-fbbc87b49a6d"). InnerVolumeSpecName "kube-api-access-mnrrd". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 08 18:17:30 crc kubenswrapper[4859]: I1008 18:17:30.576402 4859 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/e7e6199b-1264-4501-8953-767f51328d08-config" (OuterVolumeSpecName: "config") pod "e7e6199b-1264-4501-8953-767f51328d08" (UID: "e7e6199b-1264-4501-8953-767f51328d08"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 08 18:17:30 crc kubenswrapper[4859]: I1008 18:17:30.576463 4859 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/4bb40260-dbaa-4fb0-84df-5e680505d512-kube-api-access-2w9zh" (OuterVolumeSpecName: "kube-api-access-2w9zh") pod "4bb40260-dbaa-4fb0-84df-5e680505d512" (UID: "4bb40260-dbaa-4fb0-84df-5e680505d512"). InnerVolumeSpecName "kube-api-access-2w9zh". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 08 18:17:30 crc kubenswrapper[4859]: I1008 18:17:30.576523 4859 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/01ab3dd5-8196-46d0-ad33-122e2ca51def-kube-api-access-w7l8j" (OuterVolumeSpecName: "kube-api-access-w7l8j") pod "01ab3dd5-8196-46d0-ad33-122e2ca51def" (UID: "01ab3dd5-8196-46d0-ad33-122e2ca51def"). InnerVolumeSpecName "kube-api-access-w7l8j". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 08 18:17:30 crc kubenswrapper[4859]: I1008 18:17:30.576667 4859 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/31d8b7a1-420e-4252-a5b7-eebe8a111292-auth-proxy-config" (OuterVolumeSpecName: "auth-proxy-config") pod "31d8b7a1-420e-4252-a5b7-eebe8a111292" (UID: "31d8b7a1-420e-4252-a5b7-eebe8a111292"). InnerVolumeSpecName "auth-proxy-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 08 18:17:30 crc kubenswrapper[4859]: I1008 18:17:30.576674 4859 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6731426b-95fe-49ff-bb5f-40441049fde2-control-plane-machine-set-operator-tls" (OuterVolumeSpecName: "control-plane-machine-set-operator-tls") pod "6731426b-95fe-49ff-bb5f-40441049fde2" (UID: "6731426b-95fe-49ff-bb5f-40441049fde2"). InnerVolumeSpecName "control-plane-machine-set-operator-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 08 18:17:30 crc kubenswrapper[4859]: I1008 18:17:30.576786 4859 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-kube-api-access-x4zgh" (OuterVolumeSpecName: "kube-api-access-x4zgh") pod "b11524ee-3fca-4b1b-9cdf-6da289fdbc7d" (UID: "b11524ee-3fca-4b1b-9cdf-6da289fdbc7d"). InnerVolumeSpecName "kube-api-access-x4zgh". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 08 18:17:30 crc kubenswrapper[4859]: I1008 18:17:30.576891 4859 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/31d8b7a1-420e-4252-a5b7-eebe8a111292-proxy-tls" (OuterVolumeSpecName: "proxy-tls") pod "31d8b7a1-420e-4252-a5b7-eebe8a111292" (UID: "31d8b7a1-420e-4252-a5b7-eebe8a111292"). InnerVolumeSpecName "proxy-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 08 18:17:30 crc kubenswrapper[4859]: I1008 18:17:30.576950 4859 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/0b78653f-4ff9-4508-8672-245ed9b561e3-kube-api-access" (OuterVolumeSpecName: "kube-api-access") pod "0b78653f-4ff9-4508-8672-245ed9b561e3" (UID: "0b78653f-4ff9-4508-8672-245ed9b561e3"). InnerVolumeSpecName "kube-api-access". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 08 18:17:30 crc kubenswrapper[4859]: I1008 18:17:30.577105 4859 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/fda69060-fa79-4696-b1a6-7980f124bf7c-proxy-tls" (OuterVolumeSpecName: "proxy-tls") pod "fda69060-fa79-4696-b1a6-7980f124bf7c" (UID: "fda69060-fa79-4696-b1a6-7980f124bf7c"). InnerVolumeSpecName "proxy-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 08 18:17:30 crc kubenswrapper[4859]: I1008 18:17:30.578003 4859 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-config" (OuterVolumeSpecName: "config") pod "7583ce53-e0fe-4a16-9e4d-50516596a136" (UID: "7583ce53-e0fe-4a16-9e4d-50516596a136"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 08 18:17:30 crc kubenswrapper[4859]: I1008 18:17:30.578129 4859 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-tmpfs" (OuterVolumeSpecName: "tmpfs") pod "308be0ea-9f5f-4b29-aeb1-5abd31a0b17b" (UID: "308be0ea-9f5f-4b29-aeb1-5abd31a0b17b"). InnerVolumeSpecName "tmpfs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 08 18:17:30 crc kubenswrapper[4859]: I1008 18:17:30.578349 4859 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6509e943-70c6-444c-bc41-48a544e36fbd-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "6509e943-70c6-444c-bc41-48a544e36fbd" (UID: "6509e943-70c6-444c-bc41-48a544e36fbd"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 08 18:17:30 crc kubenswrapper[4859]: I1008 18:17:30.578459 4859 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/4bb40260-dbaa-4fb0-84df-5e680505d512-multus-daemon-config" (OuterVolumeSpecName: "multus-daemon-config") pod "4bb40260-dbaa-4fb0-84df-5e680505d512" (UID: "4bb40260-dbaa-4fb0-84df-5e680505d512"). InnerVolumeSpecName "multus-daemon-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 08 18:17:30 crc kubenswrapper[4859]: I1008 18:17:30.578953 4859 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-client" (OuterVolumeSpecName: "etcd-client") pod "09efc573-dbb6-4249-bd59-9b87aba8dd28" (UID: "09efc573-dbb6-4249-bd59-9b87aba8dd28"). InnerVolumeSpecName "etcd-client". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 08 18:17:30 crc kubenswrapper[4859]: I1008 18:17:30.581677 4859 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5441d097-087c-4d9a-baa8-b210afa90fc9-kube-api-access-2d4wz" (OuterVolumeSpecName: "kube-api-access-2d4wz") pod "5441d097-087c-4d9a-baa8-b210afa90fc9" (UID: "5441d097-087c-4d9a-baa8-b210afa90fc9"). InnerVolumeSpecName "kube-api-access-2d4wz". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 08 18:17:30 crc kubenswrapper[4859]: I1008 18:17:30.581821 4859 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/9d4552c7-cd75-42dd-8880-30dd377c49a4-config" (OuterVolumeSpecName: "config") pod "9d4552c7-cd75-42dd-8880-30dd377c49a4" (UID: "9d4552c7-cd75-42dd-8880-30dd377c49a4"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 08 18:17:30 crc kubenswrapper[4859]: I1008 18:17:30.581975 4859 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6402fda4-df10-493c-b4e5-d0569419652d-kube-api-access-mg5zb" (OuterVolumeSpecName: "kube-api-access-mg5zb") pod "6402fda4-df10-493c-b4e5-d0569419652d" (UID: "6402fda4-df10-493c-b4e5-d0569419652d"). InnerVolumeSpecName "kube-api-access-mg5zb". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 08 18:17:30 crc kubenswrapper[4859]: I1008 18:17:30.582032 4859 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/96b93a3a-6083-4aea-8eab-fe1aa8245ad9-metrics-tls" (OuterVolumeSpecName: "metrics-tls") pod "96b93a3a-6083-4aea-8eab-fe1aa8245ad9" (UID: "96b93a3a-6083-4aea-8eab-fe1aa8245ad9"). InnerVolumeSpecName "metrics-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 08 18:17:30 crc kubenswrapper[4859]: I1008 18:17:30.582066 4859 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/c03ee662-fb2f-4fc4-a2c1-af487c19d254-kube-api-access-v47cf" (OuterVolumeSpecName: "kube-api-access-v47cf") pod "c03ee662-fb2f-4fc4-a2c1-af487c19d254" (UID: "c03ee662-fb2f-4fc4-a2c1-af487c19d254"). InnerVolumeSpecName "kube-api-access-v47cf". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 08 18:17:30 crc kubenswrapper[4859]: I1008 18:17:30.582202 4859 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-service-ca" (OuterVolumeSpecName: "service-ca") pod "43509403-f426-496e-be36-56cef71462f5" (UID: "43509403-f426-496e-be36-56cef71462f5"). InnerVolumeSpecName "service-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 08 18:17:30 crc kubenswrapper[4859]: I1008 18:17:30.582211 4859 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/925f1c65-6136-48ba-85aa-3a3b50560753-env-overrides" (OuterVolumeSpecName: "env-overrides") pod "925f1c65-6136-48ba-85aa-3a3b50560753" (UID: "925f1c65-6136-48ba-85aa-3a3b50560753"). InnerVolumeSpecName "env-overrides". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 08 18:17:30 crc kubenswrapper[4859]: I1008 18:17:30.582347 4859 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/efdd0498-1daa-4136-9a4a-3b948c2293fc-webhook-certs" (OuterVolumeSpecName: "webhook-certs") pod "efdd0498-1daa-4136-9a4a-3b948c2293fc" (UID: "efdd0498-1daa-4136-9a4a-3b948c2293fc"). InnerVolumeSpecName "webhook-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 08 18:17:30 crc kubenswrapper[4859]: I1008 18:17:30.582504 4859 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/09efc573-dbb6-4249-bd59-9b87aba8dd28-kube-api-access-8tdtz" (OuterVolumeSpecName: "kube-api-access-8tdtz") pod "09efc573-dbb6-4249-bd59-9b87aba8dd28" (UID: "09efc573-dbb6-4249-bd59-9b87aba8dd28"). InnerVolumeSpecName "kube-api-access-8tdtz". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 08 18:17:30 crc kubenswrapper[4859]: I1008 18:17:30.582759 4859 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-error" (OuterVolumeSpecName: "v4-0-config-user-template-error") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-user-template-error". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 08 18:17:30 crc kubenswrapper[4859]: I1008 18:17:30.594005 4859 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5441d097-087c-4d9a-baa8-b210afa90fc9-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "5441d097-087c-4d9a-baa8-b210afa90fc9" (UID: "5441d097-087c-4d9a-baa8-b210afa90fc9"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 08 18:17:30 crc kubenswrapper[4859]: I1008 18:17:30.583200 4859 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-kube-api-access-zkvpv" (OuterVolumeSpecName: "kube-api-access-zkvpv") pod "09ae3b1a-e8e7-4524-b54b-61eab6f9239a" (UID: "09ae3b1a-e8e7-4524-b54b-61eab6f9239a"). InnerVolumeSpecName "kube-api-access-zkvpv". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 08 18:17:30 crc kubenswrapper[4859]: I1008 18:17:30.583418 4859 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b6312bbd-5731-4ea0-a20f-81d5a57df44a-kube-api-access-249nr" (OuterVolumeSpecName: "kube-api-access-249nr") pod "b6312bbd-5731-4ea0-a20f-81d5a57df44a" (UID: "b6312bbd-5731-4ea0-a20f-81d5a57df44a"). InnerVolumeSpecName "kube-api-access-249nr". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 08 18:17:30 crc kubenswrapper[4859]: I1008 18:17:30.583754 4859 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-trusted-ca-bundle" (OuterVolumeSpecName: "trusted-ca-bundle") pod "43509403-f426-496e-be36-56cef71462f5" (UID: "43509403-f426-496e-be36-56cef71462f5"). InnerVolumeSpecName "trusted-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 08 18:17:30 crc kubenswrapper[4859]: I1008 18:17:30.595150 4859 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "1bf7eb37-55a3-4c65-b768-a94c82151e69" (UID: "1bf7eb37-55a3-4c65-b768-a94c82151e69"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 08 18:17:30 crc kubenswrapper[4859]: I1008 18:17:30.584861 4859 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/fda69060-fa79-4696-b1a6-7980f124bf7c-kube-api-access-xcgwh" (OuterVolumeSpecName: "kube-api-access-xcgwh") pod "fda69060-fa79-4696-b1a6-7980f124bf7c" (UID: "fda69060-fa79-4696-b1a6-7980f124bf7c"). InnerVolumeSpecName "kube-api-access-xcgwh". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 08 18:17:30 crc kubenswrapper[4859]: I1008 18:17:30.584883 4859 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c" (UID: "210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 08 18:17:30 crc kubenswrapper[4859]: I1008 18:17:30.585165 4859 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/49ef4625-1d3a-4a9f-b595-c2433d32326d-kube-api-access-pjr6v" (OuterVolumeSpecName: "kube-api-access-pjr6v") pod "49ef4625-1d3a-4a9f-b595-c2433d32326d" (UID: "49ef4625-1d3a-4a9f-b595-c2433d32326d"). InnerVolumeSpecName "kube-api-access-pjr6v". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 08 18:17:30 crc kubenswrapper[4859]: I1008 18:17:30.585566 4859 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7bb08738-c794-4ee8-9972-3a62ca171029-cni-binary-copy" (OuterVolumeSpecName: "cni-binary-copy") pod "7bb08738-c794-4ee8-9972-3a62ca171029" (UID: "7bb08738-c794-4ee8-9972-3a62ca171029"). InnerVolumeSpecName "cni-binary-copy". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 08 18:17:30 crc kubenswrapper[4859]: I1008 18:17:30.585808 4859 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a31745f5-9847-4afe-82a5-3161cc66ca93-metrics-tls" (OuterVolumeSpecName: "metrics-tls") pod "a31745f5-9847-4afe-82a5-3161cc66ca93" (UID: "a31745f5-9847-4afe-82a5-3161cc66ca93"). InnerVolumeSpecName "metrics-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 08 18:17:30 crc kubenswrapper[4859]: I1008 18:17:30.595278 4859 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/bf126b07-da06-4140-9a57-dfd54fc6b486-kube-api-access-rnphk" (OuterVolumeSpecName: "kube-api-access-rnphk") pod "bf126b07-da06-4140-9a57-dfd54fc6b486" (UID: "bf126b07-da06-4140-9a57-dfd54fc6b486"). InnerVolumeSpecName "kube-api-access-rnphk". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 08 18:17:30 crc kubenswrapper[4859]: I1008 18:17:30.585824 4859 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-bound-sa-token" (OuterVolumeSpecName: "bound-sa-token") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b"). InnerVolumeSpecName "bound-sa-token". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 08 18:17:30 crc kubenswrapper[4859]: I1008 18:17:30.585841 4859 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/efdd0498-1daa-4136-9a4a-3b948c2293fc-kube-api-access-fqsjt" (OuterVolumeSpecName: "kube-api-access-fqsjt") pod "efdd0498-1daa-4136-9a4a-3b948c2293fc" (UID: "efdd0498-1daa-4136-9a4a-3b948c2293fc"). InnerVolumeSpecName "kube-api-access-fqsjt". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 08 18:17:30 crc kubenswrapper[4859]: I1008 18:17:30.586974 4859 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/a31745f5-9847-4afe-82a5-3161cc66ca93-trusted-ca" (OuterVolumeSpecName: "trusted-ca") pod "a31745f5-9847-4afe-82a5-3161cc66ca93" (UID: "a31745f5-9847-4afe-82a5-3161cc66ca93"). InnerVolumeSpecName "trusted-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 08 18:17:30 crc kubenswrapper[4859]: I1008 18:17:30.592600 4859 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/9d4552c7-cd75-42dd-8880-30dd377c49a4-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "9d4552c7-cd75-42dd-8880-30dd377c49a4" (UID: "9d4552c7-cd75-42dd-8880-30dd377c49a4"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 08 18:17:30 crc kubenswrapper[4859]: I1008 18:17:30.593744 4859 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-audit-policies" (OuterVolumeSpecName: "audit-policies") pod "09ae3b1a-e8e7-4524-b54b-61eab6f9239a" (UID: "09ae3b1a-e8e7-4524-b54b-61eab6f9239a"). InnerVolumeSpecName "audit-policies". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 08 18:17:30 crc kubenswrapper[4859]: I1008 18:17:30.594256 4859 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a0128f3a-b052-44ed-a84e-c4c8aaf17c13-kube-api-access-gf66m" (OuterVolumeSpecName: "kube-api-access-gf66m") pod "a0128f3a-b052-44ed-a84e-c4c8aaf17c13" (UID: "a0128f3a-b052-44ed-a84e-c4c8aaf17c13"). InnerVolumeSpecName "kube-api-access-gf66m". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 08 18:17:30 crc kubenswrapper[4859]: I1008 18:17:30.594583 4859 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/4bb40260-dbaa-4fb0-84df-5e680505d512-cni-binary-copy" (OuterVolumeSpecName: "cni-binary-copy") pod "4bb40260-dbaa-4fb0-84df-5e680505d512" (UID: "4bb40260-dbaa-4fb0-84df-5e680505d512"). InnerVolumeSpecName "cni-binary-copy". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 08 18:17:30 crc kubenswrapper[4859]: I1008 18:17:30.594998 4859 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/22c825df-677d-4ca6-82db-3454ed06e783-kube-api-access-7c4vf" (OuterVolumeSpecName: "kube-api-access-7c4vf") pod "22c825df-677d-4ca6-82db-3454ed06e783" (UID: "22c825df-677d-4ca6-82db-3454ed06e783"). InnerVolumeSpecName "kube-api-access-7c4vf". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 08 18:17:30 crc kubenswrapper[4859]: I1008 18:17:30.595937 4859 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-kube-api-access-qs4fp" (OuterVolumeSpecName: "kube-api-access-qs4fp") pod "210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c" (UID: "210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c"). InnerVolumeSpecName "kube-api-access-qs4fp". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 08 18:17:30 crc kubenswrapper[4859]: I1008 18:17:30.596274 4859 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-etcd-client" (OuterVolumeSpecName: "etcd-client") pod "09ae3b1a-e8e7-4524-b54b-61eab6f9239a" (UID: "09ae3b1a-e8e7-4524-b54b-61eab6f9239a"). InnerVolumeSpecName "etcd-client". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 08 18:17:30 crc kubenswrapper[4859]: E1008 18:17:30.596831 4859 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-08 18:17:31.096803552 +0000 UTC m=+21.343642931 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 08 18:17:30 crc kubenswrapper[4859]: I1008 18:17:30.596893 4859 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-kube-api-access-kfwg7" (OuterVolumeSpecName: "kube-api-access-kfwg7") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b"). InnerVolumeSpecName "kube-api-access-kfwg7". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 08 18:17:30 crc kubenswrapper[4859]: I1008 18:17:30.597864 4859 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7bb08738-c794-4ee8-9972-3a62ca171029-cni-sysctl-allowlist" (OuterVolumeSpecName: "cni-sysctl-allowlist") pod "7bb08738-c794-4ee8-9972-3a62ca171029" (UID: "7bb08738-c794-4ee8-9972-3a62ca171029"). InnerVolumeSpecName "cni-sysctl-allowlist". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 08 18:17:30 crc kubenswrapper[4859]: I1008 18:17:30.597963 4859 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/496e6271-fb68-4057-954e-a0d97a4afa3f-kube-api-access" (OuterVolumeSpecName: "kube-api-access") pod "496e6271-fb68-4057-954e-a0d97a4afa3f" (UID: "496e6271-fb68-4057-954e-a0d97a4afa3f"). InnerVolumeSpecName "kube-api-access". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 08 18:17:30 crc kubenswrapper[4859]: I1008 18:17:30.598647 4859 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/bf126b07-da06-4140-9a57-dfd54fc6b486-trusted-ca" (OuterVolumeSpecName: "trusted-ca") pod "bf126b07-da06-4140-9a57-dfd54fc6b486" (UID: "bf126b07-da06-4140-9a57-dfd54fc6b486"). InnerVolumeSpecName "trusted-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 08 18:17:30 crc kubenswrapper[4859]: I1008 18:17:30.600509 4859 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/87cf06ed-a83f-41a7-828d-70653580a8cb-kube-api-access-d6qdx" (OuterVolumeSpecName: "kube-api-access-d6qdx") pod "87cf06ed-a83f-41a7-828d-70653580a8cb" (UID: "87cf06ed-a83f-41a7-828d-70653580a8cb"). InnerVolumeSpecName "kube-api-access-d6qdx". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 08 18:17:30 crc kubenswrapper[4859]: I1008 18:17:30.600962 4859 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-provider-selection" (OuterVolumeSpecName: "v4-0-config-user-template-provider-selection") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-user-template-provider-selection". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 08 18:17:30 crc kubenswrapper[4859]: I1008 18:17:30.601600 4859 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/43509403-f426-496e-be36-56cef71462f5-console-oauth-config" (OuterVolumeSpecName: "console-oauth-config") pod "43509403-f426-496e-be36-56cef71462f5" (UID: "43509403-f426-496e-be36-56cef71462f5"). InnerVolumeSpecName "console-oauth-config". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 08 18:17:30 crc kubenswrapper[4859]: I1008 18:17:30.601776 4859 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-ovnkube-script-lib" (OuterVolumeSpecName: "ovnkube-script-lib") pod "6ea678ab-3438-413e-bfe3-290ae7725660" (UID: "6ea678ab-3438-413e-bfe3-290ae7725660"). InnerVolumeSpecName "ovnkube-script-lib". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 08 18:17:30 crc kubenswrapper[4859]: I1008 18:17:30.602081 4859 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a31745f5-9847-4afe-82a5-3161cc66ca93-bound-sa-token" (OuterVolumeSpecName: "bound-sa-token") pod "a31745f5-9847-4afe-82a5-3161cc66ca93" (UID: "a31745f5-9847-4afe-82a5-3161cc66ca93"). InnerVolumeSpecName "bound-sa-token". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 08 18:17:30 crc kubenswrapper[4859]: I1008 18:17:30.602284 4859 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/496e6271-fb68-4057-954e-a0d97a4afa3f-config" (OuterVolumeSpecName: "config") pod "496e6271-fb68-4057-954e-a0d97a4afa3f" (UID: "496e6271-fb68-4057-954e-a0d97a4afa3f"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 08 18:17:30 crc kubenswrapper[4859]: I1008 18:17:30.602649 4859 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-router-certs" (OuterVolumeSpecName: "v4-0-config-system-router-certs") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-system-router-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 08 18:17:30 crc kubenswrapper[4859]: I1008 18:17:30.602825 4859 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-trusted-ca-bundle" (OuterVolumeSpecName: "trusted-ca-bundle") pod "1bf7eb37-55a3-4c65-b768-a94c82151e69" (UID: "1bf7eb37-55a3-4c65-b768-a94c82151e69"). InnerVolumeSpecName "trusted-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 08 18:17:30 crc kubenswrapper[4859]: I1008 18:17:30.603009 4859 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7539238d-5fe0-46ed-884e-1c3b566537ec-config" (OuterVolumeSpecName: "config") pod "7539238d-5fe0-46ed-884e-1c3b566537ec" (UID: "7539238d-5fe0-46ed-884e-1c3b566537ec"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 08 18:17:30 crc kubenswrapper[4859]: I1008 18:17:30.603116 4859 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/8f668bae-612b-4b75-9490-919e737c6a3b-registry-certificates" (OuterVolumeSpecName: "registry-certificates") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b"). InnerVolumeSpecName "registry-certificates". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 08 18:17:30 crc kubenswrapper[4859]: I1008 18:17:30.603364 4859 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-kube-api-access-ngvvp" (OuterVolumeSpecName: "kube-api-access-ngvvp") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "kube-api-access-ngvvp". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 08 18:17:30 crc kubenswrapper[4859]: I1008 18:17:30.603387 4859 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/925f1c65-6136-48ba-85aa-3a3b50560753-kube-api-access-s4n52" (OuterVolumeSpecName: "kube-api-access-s4n52") pod "925f1c65-6136-48ba-85aa-3a3b50560753" (UID: "925f1c65-6136-48ba-85aa-3a3b50560753"). InnerVolumeSpecName "kube-api-access-s4n52". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 08 18:17:30 crc kubenswrapper[4859]: I1008 18:17:30.603565 4859 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1386a44e-36a2-460c-96d0-0359d2b6f0f5-config" (OuterVolumeSpecName: "config") pod "1386a44e-36a2-460c-96d0-0359d2b6f0f5" (UID: "1386a44e-36a2-460c-96d0-0359d2b6f0f5"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 08 18:17:30 crc kubenswrapper[4859]: I1008 18:17:30.603760 4859 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/57a731c4-ef35-47a8-b875-bfb08a7f8011-utilities" (OuterVolumeSpecName: "utilities") pod "57a731c4-ef35-47a8-b875-bfb08a7f8011" (UID: "57a731c4-ef35-47a8-b875-bfb08a7f8011"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 08 18:17:30 crc kubenswrapper[4859]: I1008 18:17:30.603993 4859 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "8cea82b4-6893-4ddc-af9f-1bb5ae425c5b" (UID: "8cea82b4-6893-4ddc-af9f-1bb5ae425c5b"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 08 18:17:30 crc kubenswrapper[4859]: I1008 18:17:30.604219 4859 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-serving-cert" (OuterVolumeSpecName: "v4-0-config-system-serving-cert") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-system-serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 08 18:17:30 crc kubenswrapper[4859]: I1008 18:17:30.604242 4859 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/01ab3dd5-8196-46d0-ad33-122e2ca51def-config" (OuterVolumeSpecName: "config") pod "01ab3dd5-8196-46d0-ad33-122e2ca51def" (UID: "01ab3dd5-8196-46d0-ad33-122e2ca51def"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 08 18:17:30 crc kubenswrapper[4859]: I1008 18:17:30.606895 4859 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/25e176fe-21b4-4974-b1ed-c8b94f112a7f-kube-api-access-d4lsv" (OuterVolumeSpecName: "kube-api-access-d4lsv") pod "25e176fe-21b4-4974-b1ed-c8b94f112a7f" (UID: "25e176fe-21b4-4974-b1ed-c8b94f112a7f"). InnerVolumeSpecName "kube-api-access-d4lsv". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 08 18:17:30 crc kubenswrapper[4859]: I1008 18:17:30.606977 4859 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/01ab3dd5-8196-46d0-ad33-122e2ca51def-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "01ab3dd5-8196-46d0-ad33-122e2ca51def" (UID: "01ab3dd5-8196-46d0-ad33-122e2ca51def"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 08 18:17:30 crc kubenswrapper[4859]: I1008 18:17:30.607192 4859 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-trusted-ca-bundle" (OuterVolumeSpecName: "trusted-ca-bundle") pod "09ae3b1a-e8e7-4524-b54b-61eab6f9239a" (UID: "09ae3b1a-e8e7-4524-b54b-61eab6f9239a"). InnerVolumeSpecName "trusted-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 08 18:17:30 crc kubenswrapper[4859]: I1008 18:17:30.607701 4859 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/9d4552c7-cd75-42dd-8880-30dd377c49a4-kube-api-access-pcxfs" (OuterVolumeSpecName: "kube-api-access-pcxfs") pod "9d4552c7-cd75-42dd-8880-30dd377c49a4" (UID: "9d4552c7-cd75-42dd-8880-30dd377c49a4"). InnerVolumeSpecName "kube-api-access-pcxfs". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 08 18:17:30 crc kubenswrapper[4859]: I1008 18:17:30.607818 4859 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-etcd-serving-ca" (OuterVolumeSpecName: "etcd-serving-ca") pod "09ae3b1a-e8e7-4524-b54b-61eab6f9239a" (UID: "09ae3b1a-e8e7-4524-b54b-61eab6f9239a"). InnerVolumeSpecName "etcd-serving-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 08 18:17:30 crc kubenswrapper[4859]: I1008 18:17:30.607420 4859 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-etcd-serving-ca" (OuterVolumeSpecName: "etcd-serving-ca") pod "1bf7eb37-55a3-4c65-b768-a94c82151e69" (UID: "1bf7eb37-55a3-4c65-b768-a94c82151e69"). InnerVolumeSpecName "etcd-serving-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 08 18:17:30 crc kubenswrapper[4859]: I1008 18:17:30.608925 4859 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"iptables-alerter-script\" (UniqueName: \"kubernetes.io/configmap/d75a4c96-2883-4a0b-bab2-0fab2b6c0b49-iptables-alerter-script\") pod \"iptables-alerter-4ln5h\" (UID: \"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\") " pod="openshift-network-operator/iptables-alerter-4ln5h" Oct 08 18:17:30 crc kubenswrapper[4859]: I1008 18:17:30.609051 4859 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/0b78653f-4ff9-4508-8672-245ed9b561e3-service-ca" (OuterVolumeSpecName: "service-ca") pod "0b78653f-4ff9-4508-8672-245ed9b561e3" (UID: "0b78653f-4ff9-4508-8672-245ed9b561e3"). InnerVolumeSpecName "service-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 08 18:17:30 crc kubenswrapper[4859]: I1008 18:17:30.609213 4859 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/9d4552c7-cd75-42dd-8880-30dd377c49a4-trusted-ca" (OuterVolumeSpecName: "trusted-ca") pod "9d4552c7-cd75-42dd-8880-30dd377c49a4" (UID: "9d4552c7-cd75-42dd-8880-30dd377c49a4"). InnerVolumeSpecName "trusted-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 08 18:17:30 crc kubenswrapper[4859]: I1008 18:17:30.609275 4859 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59-serviceca" (OuterVolumeSpecName: "serviceca") pod "3cb93b32-e0ae-4377-b9c8-fdb9842c6d59" (UID: "3cb93b32-e0ae-4377-b9c8-fdb9842c6d59"). InnerVolumeSpecName "serviceca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 08 18:17:30 crc kubenswrapper[4859]: I1008 18:17:30.609313 4859 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/ef543e1b-8068-4ea3-b32a-61027b32e95d-env-overrides\") pod \"network-node-identity-vrzqb\" (UID: \"ef543e1b-8068-4ea3-b32a-61027b32e95d\") " pod="openshift-network-node-identity/network-node-identity-vrzqb" Oct 08 18:17:30 crc kubenswrapper[4859]: I1008 18:17:30.597478 4859 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6402fda4-df10-493c-b4e5-d0569419652d-images" (OuterVolumeSpecName: "images") pod "6402fda4-df10-493c-b4e5-d0569419652d" (UID: "6402fda4-df10-493c-b4e5-d0569419652d"). InnerVolumeSpecName "images". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 08 18:17:30 crc kubenswrapper[4859]: E1008 18:17:30.609909 4859 configmap.go:193] Couldn't get configMap openshift-network-console/networking-console-plugin: object "openshift-network-console"/"networking-console-plugin" not registered Oct 08 18:17:30 crc kubenswrapper[4859]: E1008 18:17:30.609975 4859 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-10-08 18:17:31.109955772 +0000 UTC m=+21.356795151 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "nginx-conf" (UniqueName: "kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin" not registered Oct 08 18:17:30 crc kubenswrapper[4859]: I1008 18:17:30.610549 4859 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-env-overrides" (OuterVolumeSpecName: "env-overrides") pod "6ea678ab-3438-413e-bfe3-290ae7725660" (UID: "6ea678ab-3438-413e-bfe3-290ae7725660"). InnerVolumeSpecName "env-overrides". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 08 18:17:30 crc kubenswrapper[4859]: I1008 18:17:30.611510 4859 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a0128f3a-b052-44ed-a84e-c4c8aaf17c13-samples-operator-tls" (OuterVolumeSpecName: "samples-operator-tls") pod "a0128f3a-b052-44ed-a84e-c4c8aaf17c13" (UID: "a0128f3a-b052-44ed-a84e-c4c8aaf17c13"). InnerVolumeSpecName "samples-operator-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 08 18:17:30 crc kubenswrapper[4859]: I1008 18:17:30.611814 4859 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/43509403-f426-496e-be36-56cef71462f5-console-serving-cert" (OuterVolumeSpecName: "console-serving-cert") pod "43509403-f426-496e-be36-56cef71462f5" (UID: "43509403-f426-496e-be36-56cef71462f5"). InnerVolumeSpecName "console-serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 08 18:17:30 crc kubenswrapper[4859]: I1008 18:17:30.612055 4859 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5fe579f8-e8a6-4643-bce5-a661393c4dde-kube-api-access-fcqwp" (OuterVolumeSpecName: "kube-api-access-fcqwp") pod "5fe579f8-e8a6-4643-bce5-a661393c4dde" (UID: "5fe579f8-e8a6-4643-bce5-a661393c4dde"). InnerVolumeSpecName "kube-api-access-fcqwp". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 08 18:17:30 crc kubenswrapper[4859]: I1008 18:17:30.612485 4859 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/25e176fe-21b4-4974-b1ed-c8b94f112a7f-signing-key" (OuterVolumeSpecName: "signing-key") pod "25e176fe-21b4-4974-b1ed-c8b94f112a7f" (UID: "25e176fe-21b4-4974-b1ed-c8b94f112a7f"). InnerVolumeSpecName "signing-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 08 18:17:30 crc kubenswrapper[4859]: I1008 18:17:30.613009 4859 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/43509403-f426-496e-be36-56cef71462f5-kube-api-access-qg5z5" (OuterVolumeSpecName: "kube-api-access-qg5z5") pod "43509403-f426-496e-be36-56cef71462f5" (UID: "43509403-f426-496e-be36-56cef71462f5"). InnerVolumeSpecName "kube-api-access-qg5z5". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 08 18:17:30 crc kubenswrapper[4859]: I1008 18:17:30.613347 4859 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-ca" (OuterVolumeSpecName: "etcd-ca") pod "09efc573-dbb6-4249-bd59-9b87aba8dd28" (UID: "09efc573-dbb6-4249-bd59-9b87aba8dd28"). InnerVolumeSpecName "etcd-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 08 18:17:30 crc kubenswrapper[4859]: I1008 18:17:30.614629 4859 swap_util.go:74] "error creating dir to test if tmpfs noswap is enabled. Assuming not supported" mount path="" error="stat /var/lib/kubelet/plugins/kubernetes.io/empty-dir: no such file or directory" Oct 08 18:17:30 crc kubenswrapper[4859]: I1008 18:17:30.614655 4859 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/44663579-783b-4372-86d6-acf235a62d72-kube-api-access-vt5rc" (OuterVolumeSpecName: "kube-api-access-vt5rc") pod "44663579-783b-4372-86d6-acf235a62d72" (UID: "44663579-783b-4372-86d6-acf235a62d72"). InnerVolumeSpecName "kube-api-access-vt5rc". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 08 18:17:30 crc kubenswrapper[4859]: I1008 18:17:30.615771 4859 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovnkube-identity-cm\" (UniqueName: \"kubernetes.io/configmap/ef543e1b-8068-4ea3-b32a-61027b32e95d-ovnkube-identity-cm\") pod \"network-node-identity-vrzqb\" (UID: \"ef543e1b-8068-4ea3-b32a-61027b32e95d\") " pod="openshift-network-node-identity/network-node-identity-vrzqb" Oct 08 18:17:30 crc kubenswrapper[4859]: I1008 18:17:30.616118 4859 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "bc5039c0-ea34-426b-a2b7-fbbc87b49a6d" (UID: "bc5039c0-ea34-426b-a2b7-fbbc87b49a6d"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 08 18:17:30 crc kubenswrapper[4859]: I1008 18:17:30.616801 4859 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-encryption-config" (OuterVolumeSpecName: "encryption-config") pod "09ae3b1a-e8e7-4524-b54b-61eab6f9239a" (UID: "09ae3b1a-e8e7-4524-b54b-61eab6f9239a"). InnerVolumeSpecName "encryption-config". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 08 18:17:30 crc kubenswrapper[4859]: I1008 18:17:30.617041 4859 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/22c825df-677d-4ca6-82db-3454ed06e783-auth-proxy-config" (OuterVolumeSpecName: "auth-proxy-config") pod "22c825df-677d-4ca6-82db-3454ed06e783" (UID: "22c825df-677d-4ca6-82db-3454ed06e783"). InnerVolumeSpecName "auth-proxy-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 08 18:17:30 crc kubenswrapper[4859]: I1008 18:17:30.617626 4859 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/7bb08738-c794-4ee8-9972-3a62ca171029-kube-api-access-279lb" (OuterVolumeSpecName: "kube-api-access-279lb") pod "7bb08738-c794-4ee8-9972-3a62ca171029" (UID: "7bb08738-c794-4ee8-9972-3a62ca171029"). InnerVolumeSpecName "kube-api-access-279lb". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 08 18:17:30 crc kubenswrapper[4859]: I1008 18:17:30.618336 4859 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-oauth-serving-cert" (OuterVolumeSpecName: "oauth-serving-cert") pod "43509403-f426-496e-be36-56cef71462f5" (UID: "43509403-f426-496e-be36-56cef71462f5"). InnerVolumeSpecName "oauth-serving-cert". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 08 18:17:30 crc kubenswrapper[4859]: I1008 18:17:30.618532 4859 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-audit-policies" (OuterVolumeSpecName: "audit-policies") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "audit-policies". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 08 18:17:30 crc kubenswrapper[4859]: I1008 18:17:30.619173 4859 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-config" (OuterVolumeSpecName: "config") pod "210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c" (UID: "210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 08 18:17:30 crc kubenswrapper[4859]: I1008 18:17:30.620492 4859 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/fda69060-fa79-4696-b1a6-7980f124bf7c-mcd-auth-proxy-config" (OuterVolumeSpecName: "mcd-auth-proxy-config") pod "fda69060-fa79-4696-b1a6-7980f124bf7c" (UID: "fda69060-fa79-4696-b1a6-7980f124bf7c"). InnerVolumeSpecName "mcd-auth-proxy-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 08 18:17:30 crc kubenswrapper[4859]: E1008 18:17:30.620943 4859 secret.go:188] Couldn't get secret openshift-network-console/networking-console-plugin-cert: object "openshift-network-console"/"networking-console-plugin-cert" not registered Oct 08 18:17:30 crc kubenswrapper[4859]: E1008 18:17:30.621135 4859 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-10-08 18:17:31.121111208 +0000 UTC m=+21.367950587 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "networking-console-plugin-cert" (UniqueName: "kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin-cert" not registered Oct 08 18:17:30 crc kubenswrapper[4859]: I1008 18:17:30.621304 4859 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/7583ce53-e0fe-4a16-9e4d-50516596a136-kube-api-access-xcphl" (OuterVolumeSpecName: "kube-api-access-xcphl") pod "7583ce53-e0fe-4a16-9e4d-50516596a136" (UID: "7583ce53-e0fe-4a16-9e4d-50516596a136"). InnerVolumeSpecName "kube-api-access-xcphl". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 08 18:17:30 crc kubenswrapper[4859]: I1008 18:17:30.621456 4859 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/e7e6199b-1264-4501-8953-767f51328d08-kube-api-access" (OuterVolumeSpecName: "kube-api-access") pod "e7e6199b-1264-4501-8953-767f51328d08" (UID: "e7e6199b-1264-4501-8953-767f51328d08"). InnerVolumeSpecName "kube-api-access". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 08 18:17:30 crc kubenswrapper[4859]: I1008 18:17:30.621467 4859 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/0b574797-001e-440a-8f4e-c0be86edad0f-kube-api-access-lzf88" (OuterVolumeSpecName: "kube-api-access-lzf88") pod "0b574797-001e-440a-8f4e-c0be86edad0f" (UID: "0b574797-001e-440a-8f4e-c0be86edad0f"). InnerVolumeSpecName "kube-api-access-lzf88". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 08 18:17:30 crc kubenswrapper[4859]: I1008 18:17:30.621464 4859 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/5225d0e4-402f-4861-b410-819f433b1803-utilities" (OuterVolumeSpecName: "utilities") pod "5225d0e4-402f-4861-b410-819f433b1803" (UID: "5225d0e4-402f-4861-b410-819f433b1803"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 08 18:17:30 crc kubenswrapper[4859]: I1008 18:17:30.621587 4859 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-service-ca" (OuterVolumeSpecName: "etcd-service-ca") pod "09efc573-dbb6-4249-bd59-9b87aba8dd28" (UID: "09efc573-dbb6-4249-bd59-9b87aba8dd28"). InnerVolumeSpecName "etcd-service-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 08 18:17:30 crc kubenswrapper[4859]: I1008 18:17:30.622376 4859 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6402fda4-df10-493c-b4e5-d0569419652d-config" (OuterVolumeSpecName: "config") pod "6402fda4-df10-493c-b4e5-d0569419652d" (UID: "6402fda4-df10-493c-b4e5-d0569419652d"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 08 18:17:30 crc kubenswrapper[4859]: I1008 18:17:30.622555 4859 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5fe579f8-e8a6-4643-bce5-a661393c4dde-certs" (OuterVolumeSpecName: "certs") pod "5fe579f8-e8a6-4643-bce5-a661393c4dde" (UID: "5fe579f8-e8a6-4643-bce5-a661393c4dde"). InnerVolumeSpecName "certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 08 18:17:30 crc kubenswrapper[4859]: I1008 18:17:30.623633 4859 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a31745f5-9847-4afe-82a5-3161cc66ca93-kube-api-access-lz9wn" (OuterVolumeSpecName: "kube-api-access-lz9wn") pod "a31745f5-9847-4afe-82a5-3161cc66ca93" (UID: "a31745f5-9847-4afe-82a5-3161cc66ca93"). InnerVolumeSpecName "kube-api-access-lz9wn". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 08 18:17:30 crc kubenswrapper[4859]: I1008 18:17:30.624166 4859 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-kube-api-access-w4xd4" (OuterVolumeSpecName: "kube-api-access-w4xd4") pod "8cea82b4-6893-4ddc-af9f-1bb5ae425c5b" (UID: "8cea82b4-6893-4ddc-af9f-1bb5ae425c5b"). InnerVolumeSpecName "kube-api-access-w4xd4". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 08 18:17:30 crc kubenswrapper[4859]: I1008 18:17:30.624455 4859 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-service-ca" (OuterVolumeSpecName: "v4-0-config-system-service-ca") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-system-service-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 08 18:17:30 crc kubenswrapper[4859]: I1008 18:17:30.624577 4859 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/22c825df-677d-4ca6-82db-3454ed06e783-config" (OuterVolumeSpecName: "config") pod "22c825df-677d-4ca6-82db-3454ed06e783" (UID: "22c825df-677d-4ca6-82db-3454ed06e783"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 08 18:17:30 crc kubenswrapper[4859]: E1008 18:17:30.629456 4859 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Oct 08 18:17:30 crc kubenswrapper[4859]: E1008 18:17:30.629498 4859 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Oct 08 18:17:30 crc kubenswrapper[4859]: E1008 18:17:30.629513 4859 projected.go:194] Error preparing data for projected volume kube-api-access-s2dwl for pod openshift-network-diagnostics/network-check-source-55646444c4-trplf: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Oct 08 18:17:30 crc kubenswrapper[4859]: E1008 18:17:30.629593 4859 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl podName:9d751cbb-f2e2-430d-9754-c882a5e924a5 nodeName:}" failed. No retries permitted until 2025-10-08 18:17:31.129575983 +0000 UTC m=+21.376415362 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "kube-api-access-s2dwl" (UniqueName: "kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl") pod "network-check-source-55646444c4-trplf" (UID: "9d751cbb-f2e2-430d-9754-c882a5e924a5") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Oct 08 18:17:30 crc kubenswrapper[4859]: I1008 18:17:30.629895 4859 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/5441d097-087c-4d9a-baa8-b210afa90fc9-config" (OuterVolumeSpecName: "config") pod "5441d097-087c-4d9a-baa8-b210afa90fc9" (UID: "5441d097-087c-4d9a-baa8-b210afa90fc9"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 08 18:17:30 crc kubenswrapper[4859]: I1008 18:17:30.630236 4859 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/925f1c65-6136-48ba-85aa-3a3b50560753-ovn-control-plane-metrics-cert" (OuterVolumeSpecName: "ovn-control-plane-metrics-cert") pod "925f1c65-6136-48ba-85aa-3a3b50560753" (UID: "925f1c65-6136-48ba-85aa-3a3b50560753"). InnerVolumeSpecName "ovn-control-plane-metrics-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 08 18:17:30 crc kubenswrapper[4859]: I1008 18:17:30.630469 4859 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/ef543e1b-8068-4ea3-b32a-61027b32e95d-webhook-cert\") pod \"network-node-identity-vrzqb\" (UID: \"ef543e1b-8068-4ea3-b32a-61027b32e95d\") " pod="openshift-network-node-identity/network-node-identity-vrzqb" Oct 08 18:17:30 crc kubenswrapper[4859]: I1008 18:17:30.631034 4859 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-apiservice-cert" (OuterVolumeSpecName: "apiservice-cert") pod "308be0ea-9f5f-4b29-aeb1-5abd31a0b17b" (UID: "308be0ea-9f5f-4b29-aeb1-5abd31a0b17b"). InnerVolumeSpecName "apiservice-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 08 18:17:30 crc kubenswrapper[4859]: I1008 18:17:30.635861 4859 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6731426b-95fe-49ff-bb5f-40441049fde2-kube-api-access-x7zkh" (OuterVolumeSpecName: "kube-api-access-x7zkh") pod "6731426b-95fe-49ff-bb5f-40441049fde2" (UID: "6731426b-95fe-49ff-bb5f-40441049fde2"). InnerVolumeSpecName "kube-api-access-x7zkh". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 08 18:17:30 crc kubenswrapper[4859]: I1008 18:17:30.636036 4859 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/1d611f23-29be-4491-8495-bee1670e935f-kube-api-access-bf2bz" (OuterVolumeSpecName: "kube-api-access-bf2bz") pod "1d611f23-29be-4491-8495-bee1670e935f" (UID: "1d611f23-29be-4491-8495-bee1670e935f"). InnerVolumeSpecName "kube-api-access-bf2bz". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 08 18:17:30 crc kubenswrapper[4859]: E1008 18:17:30.636106 4859 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Oct 08 18:17:30 crc kubenswrapper[4859]: E1008 18:17:30.636142 4859 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Oct 08 18:17:30 crc kubenswrapper[4859]: E1008 18:17:30.636158 4859 projected.go:194] Error preparing data for projected volume kube-api-access-cqllr for pod openshift-network-diagnostics/network-check-target-xd92c: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Oct 08 18:17:30 crc kubenswrapper[4859]: E1008 18:17:30.636231 4859 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr podName:3b6479f0-333b-4a96-9adf-2099afdc2447 nodeName:}" failed. No retries permitted until 2025-10-08 18:17:31.136206809 +0000 UTC m=+21.383046188 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "kube-api-access-cqllr" (UniqueName: "kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr") pod "network-check-target-xd92c" (UID: "3b6479f0-333b-4a96-9adf-2099afdc2447") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Oct 08 18:17:30 crc kubenswrapper[4859]: I1008 18:17:30.639077 4859 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/0b574797-001e-440a-8f4e-c0be86edad0f-mcc-auth-proxy-config" (OuterVolumeSpecName: "mcc-auth-proxy-config") pod "0b574797-001e-440a-8f4e-c0be86edad0f" (UID: "0b574797-001e-440a-8f4e-c0be86edad0f"). InnerVolumeSpecName "mcc-auth-proxy-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 08 18:17:30 crc kubenswrapper[4859]: I1008 18:17:30.642736 4859 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-s2kz5\" (UniqueName: \"kubernetes.io/projected/ef543e1b-8068-4ea3-b32a-61027b32e95d-kube-api-access-s2kz5\") pod \"network-node-identity-vrzqb\" (UID: \"ef543e1b-8068-4ea3-b32a-61027b32e95d\") " pod="openshift-network-node-identity/network-node-identity-vrzqb" Oct 08 18:17:30 crc kubenswrapper[4859]: I1008 18:17:30.648995 4859 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-profile-collector-cert" (OuterVolumeSpecName: "profile-collector-cert") pod "f88749ec-7931-4ee7-b3fc-1ec5e11f92e9" (UID: "f88749ec-7931-4ee7-b3fc-1ec5e11f92e9"). InnerVolumeSpecName "profile-collector-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 08 18:17:30 crc kubenswrapper[4859]: I1008 18:17:30.649427 4859 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-service-ca-bundle" (OuterVolumeSpecName: "service-ca-bundle") pod "6509e943-70c6-444c-bc41-48a544e36fbd" (UID: "6509e943-70c6-444c-bc41-48a544e36fbd"). InnerVolumeSpecName "service-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 08 18:17:30 crc kubenswrapper[4859]: I1008 18:17:30.649681 4859 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5b88f790-22fa-440e-b583-365168c0b23d-metrics-certs" (OuterVolumeSpecName: "metrics-certs") pod "5b88f790-22fa-440e-b583-365168c0b23d" (UID: "5b88f790-22fa-440e-b583-365168c0b23d"). InnerVolumeSpecName "metrics-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 08 18:17:30 crc kubenswrapper[4859]: I1008 18:17:30.650621 4859 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rczfb\" (UniqueName: \"kubernetes.io/projected/d75a4c96-2883-4a0b-bab2-0fab2b6c0b49-kube-api-access-rczfb\") pod \"iptables-alerter-4ln5h\" (UID: \"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\") " pod="openshift-network-operator/iptables-alerter-4ln5h" Oct 08 18:17:30 crc kubenswrapper[4859]: I1008 18:17:30.655010 4859 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/20b0d48f-5fd6-431c-a545-e3c800c7b866-kube-api-access-w9rds" (OuterVolumeSpecName: "kube-api-access-w9rds") pod "20b0d48f-5fd6-431c-a545-e3c800c7b866" (UID: "20b0d48f-5fd6-431c-a545-e3c800c7b866"). InnerVolumeSpecName "kube-api-access-w9rds". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 08 18:17:30 crc kubenswrapper[4859]: I1008 18:17:30.655062 4859 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rdwmf\" (UniqueName: \"kubernetes.io/projected/37a5e44f-9a88-4405-be8a-b645485e7312-kube-api-access-rdwmf\") pod \"network-operator-58b4c7f79c-55gtf\" (UID: \"37a5e44f-9a88-4405-be8a-b645485e7312\") " pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" Oct 08 18:17:30 crc kubenswrapper[4859]: I1008 18:17:30.655516 4859 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-encryption-config" (OuterVolumeSpecName: "encryption-config") pod "1bf7eb37-55a3-4c65-b768-a94c82151e69" (UID: "1bf7eb37-55a3-4c65-b768-a94c82151e69"). InnerVolumeSpecName "encryption-config". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 08 18:17:30 crc kubenswrapper[4859]: I1008 18:17:30.657315 4859 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/7539238d-5fe0-46ed-884e-1c3b566537ec-kube-api-access-tk88c" (OuterVolumeSpecName: "kube-api-access-tk88c") pod "7539238d-5fe0-46ed-884e-1c3b566537ec" (UID: "7539238d-5fe0-46ed-884e-1c3b566537ec"). InnerVolumeSpecName "kube-api-access-tk88c". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 08 18:17:30 crc kubenswrapper[4859]: I1008 18:17:30.657355 4859 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-client-ca" (OuterVolumeSpecName: "client-ca") pod "7583ce53-e0fe-4a16-9e4d-50516596a136" (UID: "7583ce53-e0fe-4a16-9e4d-50516596a136"). InnerVolumeSpecName "client-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 08 18:17:30 crc kubenswrapper[4859]: I1008 18:17:30.659941 4859 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/3ab1a177-2de0-46d9-b765-d0d0649bb42e-kube-api-access-4d4hj" (OuterVolumeSpecName: "kube-api-access-4d4hj") pod "3ab1a177-2de0-46d9-b765-d0d0649bb42e" (UID: "3ab1a177-2de0-46d9-b765-d0d0649bb42e"). InnerVolumeSpecName "kube-api-access-4d4hj". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 08 18:17:30 crc kubenswrapper[4859]: I1008 18:17:30.660000 4859 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-image-import-ca" (OuterVolumeSpecName: "image-import-ca") pod "1bf7eb37-55a3-4c65-b768-a94c82151e69" (UID: "1bf7eb37-55a3-4c65-b768-a94c82151e69"). InnerVolumeSpecName "image-import-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 08 18:17:30 crc kubenswrapper[4859]: I1008 18:17:30.660081 4859 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/bf126b07-da06-4140-9a57-dfd54fc6b486-image-registry-operator-tls" (OuterVolumeSpecName: "image-registry-operator-tls") pod "bf126b07-da06-4140-9a57-dfd54fc6b486" (UID: "bf126b07-da06-4140-9a57-dfd54fc6b486"). InnerVolumeSpecName "image-registry-operator-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 08 18:17:30 crc kubenswrapper[4859]: I1008 18:17:30.660161 4859 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-session" (OuterVolumeSpecName: "v4-0-config-system-session") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-system-session". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 08 18:17:30 crc kubenswrapper[4859]: I1008 18:17:30.660173 4859 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-cliconfig" (OuterVolumeSpecName: "v4-0-config-system-cliconfig") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-system-cliconfig". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 08 18:17:30 crc kubenswrapper[4859]: I1008 18:17:30.661206 4859 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-audit" (OuterVolumeSpecName: "audit") pod "1bf7eb37-55a3-4c65-b768-a94c82151e69" (UID: "1bf7eb37-55a3-4c65-b768-a94c82151e69"). InnerVolumeSpecName "audit". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 08 18:17:30 crc kubenswrapper[4859]: I1008 18:17:30.664469 4859 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/496e6271-fb68-4057-954e-a0d97a4afa3f-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "496e6271-fb68-4057-954e-a0d97a4afa3f" (UID: "496e6271-fb68-4057-954e-a0d97a4afa3f"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 08 18:17:30 crc kubenswrapper[4859]: I1008 18:17:30.664597 4859 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/3ab1a177-2de0-46d9-b765-d0d0649bb42e-package-server-manager-serving-cert" (OuterVolumeSpecName: "package-server-manager-serving-cert") pod "3ab1a177-2de0-46d9-b765-d0d0649bb42e" (UID: "3ab1a177-2de0-46d9-b765-d0d0649bb42e"). InnerVolumeSpecName "package-server-manager-serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 08 18:17:30 crc kubenswrapper[4859]: I1008 18:17:30.664675 4859 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/57a731c4-ef35-47a8-b875-bfb08a7f8011-kube-api-access-cfbct" (OuterVolumeSpecName: "kube-api-access-cfbct") pod "57a731c4-ef35-47a8-b875-bfb08a7f8011" (UID: "57a731c4-ef35-47a8-b875-bfb08a7f8011"). InnerVolumeSpecName "kube-api-access-cfbct". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 08 18:17:30 crc kubenswrapper[4859]: I1008 18:17:30.664784 4859 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6509e943-70c6-444c-bc41-48a544e36fbd-kube-api-access-6g6sz" (OuterVolumeSpecName: "kube-api-access-6g6sz") pod "6509e943-70c6-444c-bc41-48a544e36fbd" (UID: "6509e943-70c6-444c-bc41-48a544e36fbd"). InnerVolumeSpecName "kube-api-access-6g6sz". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 08 18:17:30 crc kubenswrapper[4859]: I1008 18:17:30.664854 4859 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/1386a44e-36a2-460c-96d0-0359d2b6f0f5-kube-api-access" (OuterVolumeSpecName: "kube-api-access") pod "1386a44e-36a2-460c-96d0-0359d2b6f0f5" (UID: "1386a44e-36a2-460c-96d0-0359d2b6f0f5"). InnerVolumeSpecName "kube-api-access". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 08 18:17:30 crc kubenswrapper[4859]: I1008 18:17:30.664926 4859 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b6312bbd-5731-4ea0-a20f-81d5a57df44a-profile-collector-cert" (OuterVolumeSpecName: "profile-collector-cert") pod "b6312bbd-5731-4ea0-a20f-81d5a57df44a" (UID: "b6312bbd-5731-4ea0-a20f-81d5a57df44a"). InnerVolumeSpecName "profile-collector-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 08 18:17:30 crc kubenswrapper[4859]: I1008 18:17:30.665013 4859 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-etcd-client" (OuterVolumeSpecName: "etcd-client") pod "1bf7eb37-55a3-4c65-b768-a94c82151e69" (UID: "1bf7eb37-55a3-4c65-b768-a94c82151e69"). InnerVolumeSpecName "etcd-client". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 08 18:17:30 crc kubenswrapper[4859]: I1008 18:17:30.664961 4859 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-console-config" (OuterVolumeSpecName: "console-config") pod "43509403-f426-496e-be36-56cef71462f5" (UID: "43509403-f426-496e-be36-56cef71462f5"). InnerVolumeSpecName "console-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 08 18:17:30 crc kubenswrapper[4859]: I1008 18:17:30.665080 4859 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/bf126b07-da06-4140-9a57-dfd54fc6b486-bound-sa-token" (OuterVolumeSpecName: "bound-sa-token") pod "bf126b07-da06-4140-9a57-dfd54fc6b486" (UID: "bf126b07-da06-4140-9a57-dfd54fc6b486"). InnerVolumeSpecName "bound-sa-token". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 08 18:17:30 crc kubenswrapper[4859]: I1008 18:17:30.665178 4859 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/37a5e44f-9a88-4405-be8a-b645485e7312-metrics-tls\") pod \"network-operator-58b4c7f79c-55gtf\" (UID: \"37a5e44f-9a88-4405-be8a-b645485e7312\") " pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" Oct 08 18:17:30 crc kubenswrapper[4859]: I1008 18:17:30.665959 4859 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-config" (OuterVolumeSpecName: "config") pod "6509e943-70c6-444c-bc41-48a544e36fbd" (UID: "6509e943-70c6-444c-bc41-48a544e36fbd"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 08 18:17:30 crc kubenswrapper[4859]: I1008 18:17:30.667735 4859 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-config" (OuterVolumeSpecName: "config") pod "1bf7eb37-55a3-4c65-b768-a94c82151e69" (UID: "1bf7eb37-55a3-4c65-b768-a94c82151e69"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 08 18:17:30 crc kubenswrapper[4859]: I1008 18:17:30.668783 4859 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-trusted-ca-bundle" (OuterVolumeSpecName: "trusted-ca-bundle") pod "6509e943-70c6-444c-bc41-48a544e36fbd" (UID: "6509e943-70c6-444c-bc41-48a544e36fbd"). InnerVolumeSpecName "trusted-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 08 18:17:30 crc kubenswrapper[4859]: I1008 18:17:30.670289 4859 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/8f668bae-612b-4b75-9490-919e737c6a3b-trusted-ca" (OuterVolumeSpecName: "trusted-ca") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b"). InnerVolumeSpecName "trusted-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 08 18:17:30 crc kubenswrapper[4859]: I1008 18:17:30.671606 4859 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8f668bae-612b-4b75-9490-919e737c6a3b-installation-pull-secrets" (OuterVolumeSpecName: "installation-pull-secrets") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b"). InnerVolumeSpecName "installation-pull-secrets". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 08 18:17:30 crc kubenswrapper[4859]: I1008 18:17:30.671726 4859 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-kube-api-access-6ccd8" (OuterVolumeSpecName: "kube-api-access-6ccd8") pod "308be0ea-9f5f-4b29-aeb1-5abd31a0b17b" (UID: "308be0ea-9f5f-4b29-aeb1-5abd31a0b17b"). InnerVolumeSpecName "kube-api-access-6ccd8". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 08 18:17:30 crc kubenswrapper[4859]: I1008 18:17:30.673675 4859 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-config" (OuterVolumeSpecName: "config") pod "8cea82b4-6893-4ddc-af9f-1bb5ae425c5b" (UID: "8cea82b4-6893-4ddc-af9f-1bb5ae425c5b"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 08 18:17:30 crc kubenswrapper[4859]: I1008 18:17:30.673961 4859 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-registry-tls" (OuterVolumeSpecName: "registry-tls") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b"). InnerVolumeSpecName "registry-tls". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 08 18:17:30 crc kubenswrapper[4859]: I1008 18:17:30.674279 4859 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-ovnkube-config" (OuterVolumeSpecName: "ovnkube-config") pod "6ea678ab-3438-413e-bfe3-290ae7725660" (UID: "6ea678ab-3438-413e-bfe3-290ae7725660"). InnerVolumeSpecName "ovnkube-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 08 18:17:30 crc kubenswrapper[4859]: I1008 18:17:30.676136 4859 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-config" (OuterVolumeSpecName: "config") pod "09efc573-dbb6-4249-bd59-9b87aba8dd28" (UID: "09efc573-dbb6-4249-bd59-9b87aba8dd28"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 08 18:17:30 crc kubenswrapper[4859]: I1008 18:17:30.679229 4859 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-srv-cert" (OuterVolumeSpecName: "srv-cert") pod "f88749ec-7931-4ee7-b3fc-1ec5e11f92e9" (UID: "f88749ec-7931-4ee7-b3fc-1ec5e11f92e9"). InnerVolumeSpecName "srv-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 08 18:17:30 crc kubenswrapper[4859]: I1008 18:17:30.693894 4859 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "b11524ee-3fca-4b1b-9cdf-6da289fdbc7d" (UID: "b11524ee-3fca-4b1b-9cdf-6da289fdbc7d"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 08 18:17:30 crc kubenswrapper[4859]: I1008 18:17:30.694068 4859 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6ea678ab-3438-413e-bfe3-290ae7725660-ovn-node-metrics-cert" (OuterVolumeSpecName: "ovn-node-metrics-cert") pod "6ea678ab-3438-413e-bfe3-290ae7725660" (UID: "6ea678ab-3438-413e-bfe3-290ae7725660"). InnerVolumeSpecName "ovn-node-metrics-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 08 18:17:30 crc kubenswrapper[4859]: I1008 18:17:30.699281 4859 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-proxy-ca-bundles" (OuterVolumeSpecName: "proxy-ca-bundles") pod "7583ce53-e0fe-4a16-9e4d-50516596a136" (UID: "7583ce53-e0fe-4a16-9e4d-50516596a136"). InnerVolumeSpecName "proxy-ca-bundles". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 08 18:17:30 crc kubenswrapper[4859]: I1008 18:17:30.699366 4859 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d-kube-api-access-x2m85" (OuterVolumeSpecName: "kube-api-access-x2m85") pod "cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d" (UID: "cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d"). InnerVolumeSpecName "kube-api-access-x2m85". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 08 18:17:30 crc kubenswrapper[4859]: I1008 18:17:30.699386 4859 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/0b78653f-4ff9-4508-8672-245ed9b561e3-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "0b78653f-4ff9-4508-8672-245ed9b561e3" (UID: "0b78653f-4ff9-4508-8672-245ed9b561e3"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 08 18:17:30 crc kubenswrapper[4859]: I1008 18:17:30.699413 4859 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/31d8b7a1-420e-4252-a5b7-eebe8a111292-kube-api-access-zgdk5" (OuterVolumeSpecName: "kube-api-access-zgdk5") pod "31d8b7a1-420e-4252-a5b7-eebe8a111292" (UID: "31d8b7a1-420e-4252-a5b7-eebe8a111292"). InnerVolumeSpecName "kube-api-access-zgdk5". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 08 18:17:30 crc kubenswrapper[4859]: I1008 18:17:30.699443 4859 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/31d8b7a1-420e-4252-a5b7-eebe8a111292-images" (OuterVolumeSpecName: "images") pod "31d8b7a1-420e-4252-a5b7-eebe8a111292" (UID: "31d8b7a1-420e-4252-a5b7-eebe8a111292"). InnerVolumeSpecName "images". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 08 18:17:30 crc kubenswrapper[4859]: I1008 18:17:30.700828 4859 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/d75a4c96-2883-4a0b-bab2-0fab2b6c0b49-host-slash\") pod \"iptables-alerter-4ln5h\" (UID: \"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\") " pod="openshift-network-operator/iptables-alerter-4ln5h" Oct 08 18:17:30 crc kubenswrapper[4859]: I1008 18:17:30.700935 4859 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/d75a4c96-2883-4a0b-bab2-0fab2b6c0b49-host-slash\") pod \"iptables-alerter-4ln5h\" (UID: \"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\") " pod="openshift-network-operator/iptables-alerter-4ln5h" Oct 08 18:17:30 crc kubenswrapper[4859]: I1008 18:17:30.701039 4859 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-etc-kube\" (UniqueName: \"kubernetes.io/host-path/37a5e44f-9a88-4405-be8a-b645485e7312-host-etc-kube\") pod \"network-operator-58b4c7f79c-55gtf\" (UID: \"37a5e44f-9a88-4405-be8a-b645485e7312\") " pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" Oct 08 18:17:30 crc kubenswrapper[4859]: I1008 18:17:30.701144 4859 reflector.go:368] Caches populated for *v1.Service from k8s.io/client-go/informers/factory.go:160 Oct 08 18:17:30 crc kubenswrapper[4859]: I1008 18:17:30.701300 4859 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/1d611f23-29be-4491-8495-bee1670e935f-utilities\") on node \"crc\" DevicePath \"\"" Oct 08 18:17:30 crc kubenswrapper[4859]: I1008 18:17:30.701328 4859 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/e7e6199b-1264-4501-8953-767f51328d08-config\") on node \"crc\" DevicePath \"\"" Oct 08 18:17:30 crc kubenswrapper[4859]: I1008 18:17:30.701353 4859 reconciler_common.go:293] "Volume detached for volume \"tmpfs\" (UniqueName: \"kubernetes.io/empty-dir/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-tmpfs\") on node \"crc\" DevicePath \"\"" Oct 08 18:17:30 crc kubenswrapper[4859]: I1008 18:17:30.701369 4859 reconciler_common.go:293] "Volume detached for volume \"available-featuregates\" (UniqueName: \"kubernetes.io/empty-dir/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-available-featuregates\") on node \"crc\" DevicePath \"\"" Oct 08 18:17:30 crc kubenswrapper[4859]: I1008 18:17:30.701388 4859 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-config\") on node \"crc\" DevicePath \"\"" Oct 08 18:17:30 crc kubenswrapper[4859]: I1008 18:17:30.701401 4859 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/6509e943-70c6-444c-bc41-48a544e36fbd-serving-cert\") on node \"crc\" DevicePath \"\"" Oct 08 18:17:30 crc kubenswrapper[4859]: I1008 18:17:30.701422 4859 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-mg5zb\" (UniqueName: \"kubernetes.io/projected/6402fda4-df10-493c-b4e5-d0569419652d-kube-api-access-mg5zb\") on node \"crc\" DevicePath \"\"" Oct 08 18:17:30 crc kubenswrapper[4859]: I1008 18:17:30.701435 4859 reconciler_common.go:293] "Volume detached for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/efdd0498-1daa-4136-9a4a-3b948c2293fc-webhook-certs\") on node \"crc\" DevicePath \"\"" Oct 08 18:17:30 crc kubenswrapper[4859]: I1008 18:17:30.701449 4859 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/e7e6199b-1264-4501-8953-767f51328d08-serving-cert\") on node \"crc\" DevicePath \"\"" Oct 08 18:17:30 crc kubenswrapper[4859]: I1008 18:17:30.701469 4859 reconciler_common.go:293] "Volume detached for volume \"machine-approver-tls\" (UniqueName: \"kubernetes.io/secret/22c825df-677d-4ca6-82db-3454ed06e783-machine-approver-tls\") on node \"crc\" DevicePath \"\"" Oct 08 18:17:30 crc kubenswrapper[4859]: I1008 18:17:30.701489 4859 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-wxkg8\" (UniqueName: \"kubernetes.io/projected/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59-kube-api-access-wxkg8\") on node \"crc\" DevicePath \"\"" Oct 08 18:17:30 crc kubenswrapper[4859]: I1008 18:17:30.701505 4859 reconciler_common.go:293] "Volume detached for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/925f1c65-6136-48ba-85aa-3a3b50560753-ovnkube-config\") on node \"crc\" DevicePath \"\"" Oct 08 18:17:30 crc kubenswrapper[4859]: I1008 18:17:30.701592 4859 reconciler_common.go:293] "Volume detached for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-metrics-certs\") on node \"crc\" DevicePath \"\"" Oct 08 18:17:30 crc kubenswrapper[4859]: I1008 18:17:30.701612 4859 reconciler_common.go:293] "Volume detached for volume \"machine-api-operator-tls\" (UniqueName: \"kubernetes.io/secret/6402fda4-df10-493c-b4e5-d0569419652d-machine-api-operator-tls\") on node \"crc\" DevicePath \"\"" Oct 08 18:17:30 crc kubenswrapper[4859]: I1008 18:17:30.701628 4859 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-2w9zh\" (UniqueName: \"kubernetes.io/projected/4bb40260-dbaa-4fb0-84df-5e680505d512-kube-api-access-2w9zh\") on node \"crc\" DevicePath \"\"" Oct 08 18:17:30 crc kubenswrapper[4859]: I1008 18:17:30.701643 4859 reconciler_common.go:293] "Volume detached for volume \"control-plane-machine-set-operator-tls\" (UniqueName: \"kubernetes.io/secret/6731426b-95fe-49ff-bb5f-40441049fde2-control-plane-machine-set-operator-tls\") on node \"crc\" DevicePath \"\"" Oct 08 18:17:30 crc kubenswrapper[4859]: I1008 18:17:30.701654 4859 reconciler_common.go:293] "Volume detached for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/0b78653f-4ff9-4508-8672-245ed9b561e3-kube-api-access\") on node \"crc\" DevicePath \"\"" Oct 08 18:17:30 crc kubenswrapper[4859]: I1008 18:17:30.701664 4859 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-qg5z5\" (UniqueName: \"kubernetes.io/projected/43509403-f426-496e-be36-56cef71462f5-kube-api-access-qg5z5\") on node \"crc\" DevicePath \"\"" Oct 08 18:17:30 crc kubenswrapper[4859]: I1008 18:17:30.701608 4859 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-etc-kube\" (UniqueName: \"kubernetes.io/host-path/37a5e44f-9a88-4405-be8a-b645485e7312-host-etc-kube\") pod \"network-operator-58b4c7f79c-55gtf\" (UID: \"37a5e44f-9a88-4405-be8a-b645485e7312\") " pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" Oct 08 18:17:30 crc kubenswrapper[4859]: I1008 18:17:30.701777 4859 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-2d4wz\" (UniqueName: \"kubernetes.io/projected/5441d097-087c-4d9a-baa8-b210afa90fc9-kube-api-access-2d4wz\") on node \"crc\" DevicePath \"\"" Oct 08 18:17:30 crc kubenswrapper[4859]: I1008 18:17:30.701796 4859 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-v47cf\" (UniqueName: \"kubernetes.io/projected/c03ee662-fb2f-4fc4-a2c1-af487c19d254-kube-api-access-v47cf\") on node \"crc\" DevicePath \"\"" Oct 08 18:17:30 crc kubenswrapper[4859]: I1008 18:17:30.701820 4859 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-8tdtz\" (UniqueName: \"kubernetes.io/projected/09efc573-dbb6-4249-bd59-9b87aba8dd28-kube-api-access-8tdtz\") on node \"crc\" DevicePath \"\"" Oct 08 18:17:30 crc kubenswrapper[4859]: I1008 18:17:30.701830 4859 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-error\") on node \"crc\" DevicePath \"\"" Oct 08 18:17:30 crc kubenswrapper[4859]: I1008 18:17:30.701840 4859 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/5441d097-087c-4d9a-baa8-b210afa90fc9-serving-cert\") on node \"crc\" DevicePath \"\"" Oct 08 18:17:30 crc kubenswrapper[4859]: I1008 18:17:30.701853 4859 reconciler_common.go:293] "Volume detached for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/e7e6199b-1264-4501-8953-767f51328d08-kube-api-access\") on node \"crc\" DevicePath \"\"" Oct 08 18:17:30 crc kubenswrapper[4859]: I1008 18:17:30.701864 4859 reconciler_common.go:293] "Volume detached for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/5441d097-087c-4d9a-baa8-b210afa90fc9-client-ca\") on node \"crc\" DevicePath \"\"" Oct 08 18:17:30 crc kubenswrapper[4859]: I1008 18:17:30.701873 4859 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-9xfj7\" (UniqueName: \"kubernetes.io/projected/5225d0e4-402f-4861-b410-819f433b1803-kube-api-access-9xfj7\") on node \"crc\" DevicePath \"\"" Oct 08 18:17:30 crc kubenswrapper[4859]: I1008 18:17:30.701886 4859 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-mnrrd\" (UniqueName: \"kubernetes.io/projected/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-kube-api-access-mnrrd\") on node \"crc\" DevicePath \"\"" Oct 08 18:17:30 crc kubenswrapper[4859]: I1008 18:17:30.701896 4859 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-w7l8j\" (UniqueName: \"kubernetes.io/projected/01ab3dd5-8196-46d0-ad33-122e2ca51def-kube-api-access-w7l8j\") on node \"crc\" DevicePath \"\"" Oct 08 18:17:30 crc kubenswrapper[4859]: I1008 18:17:30.701906 4859 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-x4zgh\" (UniqueName: \"kubernetes.io/projected/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-kube-api-access-x4zgh\") on node \"crc\" DevicePath \"\"" Oct 08 18:17:30 crc kubenswrapper[4859]: I1008 18:17:30.701942 4859 reconciler_common.go:293] "Volume detached for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/fda69060-fa79-4696-b1a6-7980f124bf7c-proxy-tls\") on node \"crc\" DevicePath \"\"" Oct 08 18:17:30 crc kubenswrapper[4859]: I1008 18:17:30.701955 4859 reconciler_common.go:293] "Volume detached for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/925f1c65-6136-48ba-85aa-3a3b50560753-env-overrides\") on node \"crc\" DevicePath \"\"" Oct 08 18:17:30 crc kubenswrapper[4859]: I1008 18:17:30.701965 4859 reconciler_common.go:293] "Volume detached for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-trusted-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 08 18:17:30 crc kubenswrapper[4859]: I1008 18:17:30.701975 4859 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-zkvpv\" (UniqueName: \"kubernetes.io/projected/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-kube-api-access-zkvpv\") on node \"crc\" DevicePath \"\"" Oct 08 18:17:30 crc kubenswrapper[4859]: I1008 18:17:30.701986 4859 reconciler_common.go:293] "Volume detached for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/7bb08738-c794-4ee8-9972-3a62ca171029-cni-binary-copy\") on node \"crc\" DevicePath \"\"" Oct 08 18:17:30 crc kubenswrapper[4859]: I1008 18:17:30.701997 4859 reconciler_common.go:293] "Volume detached for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-audit-policies\") on node \"crc\" DevicePath \"\"" Oct 08 18:17:30 crc kubenswrapper[4859]: I1008 18:17:30.702006 4859 reconciler_common.go:293] "Volume detached for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-srv-cert\") on node \"crc\" DevicePath \"\"" Oct 08 18:17:30 crc kubenswrapper[4859]: I1008 18:17:30.702026 4859 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-router-certs\") on node \"crc\" DevicePath \"\"" Oct 08 18:17:30 crc kubenswrapper[4859]: I1008 18:17:30.702040 4859 reconciler_common.go:293] "Volume detached for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-encryption-config\") on node \"crc\" DevicePath \"\"" Oct 08 18:17:30 crc kubenswrapper[4859]: I1008 18:17:30.702052 4859 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-session\") on node \"crc\" DevicePath \"\"" Oct 08 18:17:30 crc kubenswrapper[4859]: I1008 18:17:30.702065 4859 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-4d4hj\" (UniqueName: \"kubernetes.io/projected/3ab1a177-2de0-46d9-b765-d0d0649bb42e-kube-api-access-4d4hj\") on node \"crc\" DevicePath \"\"" Oct 08 18:17:30 crc kubenswrapper[4859]: I1008 18:17:30.702099 4859 reconciler_common.go:293] "Volume detached for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/8f668bae-612b-4b75-9490-919e737c6a3b-trusted-ca\") on node \"crc\" DevicePath \"\"" Oct 08 18:17:30 crc kubenswrapper[4859]: I1008 18:17:30.702112 4859 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-kfwg7\" (UniqueName: \"kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-kube-api-access-kfwg7\") on node \"crc\" DevicePath \"\"" Oct 08 18:17:30 crc kubenswrapper[4859]: I1008 18:17:30.702145 4859 reconciler_common.go:293] "Volume detached for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/87cf06ed-a83f-41a7-828d-70653580a8cb-metrics-tls\") on node \"crc\" DevicePath \"\"" Oct 08 18:17:30 crc kubenswrapper[4859]: I1008 18:17:30.702155 4859 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/1386a44e-36a2-460c-96d0-0359d2b6f0f5-serving-cert\") on node \"crc\" DevicePath \"\"" Oct 08 18:17:30 crc kubenswrapper[4859]: I1008 18:17:30.702164 4859 reconciler_common.go:293] "Volume detached for volume \"multus-daemon-config\" (UniqueName: \"kubernetes.io/configmap/4bb40260-dbaa-4fb0-84df-5e680505d512-multus-daemon-config\") on node \"crc\" DevicePath \"\"" Oct 08 18:17:30 crc kubenswrapper[4859]: I1008 18:17:30.702176 4859 reconciler_common.go:293] "Volume detached for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/a31745f5-9847-4afe-82a5-3161cc66ca93-metrics-tls\") on node \"crc\" DevicePath \"\"" Oct 08 18:17:30 crc kubenswrapper[4859]: I1008 18:17:30.702185 4859 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/9d4552c7-cd75-42dd-8880-30dd377c49a4-config\") on node \"crc\" DevicePath \"\"" Oct 08 18:17:30 crc kubenswrapper[4859]: I1008 18:17:30.702193 4859 reconciler_common.go:293] "Volume detached for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-client\") on node \"crc\" DevicePath \"\"" Oct 08 18:17:30 crc kubenswrapper[4859]: I1008 18:17:30.702228 4859 reconciler_common.go:293] "Volume detached for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/96b93a3a-6083-4aea-8eab-fe1aa8245ad9-metrics-tls\") on node \"crc\" DevicePath \"\"" Oct 08 18:17:30 crc kubenswrapper[4859]: I1008 18:17:30.702240 4859 reconciler_common.go:293] "Volume detached for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/4bb40260-dbaa-4fb0-84df-5e680505d512-cni-binary-copy\") on node \"crc\" DevicePath \"\"" Oct 08 18:17:30 crc kubenswrapper[4859]: I1008 18:17:30.702249 4859 reconciler_common.go:293] "Volume detached for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/8f668bae-612b-4b75-9490-919e737c6a3b-registry-certificates\") on node \"crc\" DevicePath \"\"" Oct 08 18:17:30 crc kubenswrapper[4859]: I1008 18:17:30.702258 4859 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-qs4fp\" (UniqueName: \"kubernetes.io/projected/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-kube-api-access-qs4fp\") on node \"crc\" DevicePath \"\"" Oct 08 18:17:30 crc kubenswrapper[4859]: I1008 18:17:30.702269 4859 reconciler_common.go:293] "Volume detached for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-etcd-client\") on node \"crc\" DevicePath \"\"" Oct 08 18:17:30 crc kubenswrapper[4859]: I1008 18:17:30.702279 4859 reconciler_common.go:293] "Volume detached for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/31d8b7a1-420e-4252-a5b7-eebe8a111292-auth-proxy-config\") on node \"crc\" DevicePath \"\"" Oct 08 18:17:30 crc kubenswrapper[4859]: I1008 18:17:30.702307 4859 reconciler_common.go:293] "Volume detached for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/31d8b7a1-420e-4252-a5b7-eebe8a111292-proxy-tls\") on node \"crc\" DevicePath \"\"" Oct 08 18:17:30 crc kubenswrapper[4859]: I1008 18:17:30.702317 4859 reconciler_common.go:293] "Volume detached for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-bound-sa-token\") on node \"crc\" DevicePath \"\"" Oct 08 18:17:30 crc kubenswrapper[4859]: I1008 18:17:30.702330 4859 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-config\") on node \"crc\" DevicePath \"\"" Oct 08 18:17:30 crc kubenswrapper[4859]: I1008 18:17:30.702338 4859 reconciler_common.go:293] "Volume detached for volume \"signing-cabundle\" (UniqueName: \"kubernetes.io/configmap/25e176fe-21b4-4974-b1ed-c8b94f112a7f-signing-cabundle\") on node \"crc\" DevicePath \"\"" Oct 08 18:17:30 crc kubenswrapper[4859]: I1008 18:17:30.702348 4859 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/9d4552c7-cd75-42dd-8880-30dd377c49a4-serving-cert\") on node \"crc\" DevicePath \"\"" Oct 08 18:17:30 crc kubenswrapper[4859]: I1008 18:17:30.702357 4859 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-serving-cert\") on node \"crc\" DevicePath \"\"" Oct 08 18:17:30 crc kubenswrapper[4859]: I1008 18:17:30.702401 4859 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-pjr6v\" (UniqueName: \"kubernetes.io/projected/49ef4625-1d3a-4a9f-b595-c2433d32326d-kube-api-access-pjr6v\") on node \"crc\" DevicePath \"\"" Oct 08 18:17:30 crc kubenswrapper[4859]: I1008 18:17:30.702412 4859 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/22c825df-677d-4ca6-82db-3454ed06e783-config\") on node \"crc\" DevicePath \"\"" Oct 08 18:17:30 crc kubenswrapper[4859]: I1008 18:17:30.702422 4859 reconciler_common.go:293] "Volume detached for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-proxy-ca-bundles\") on node \"crc\" DevicePath \"\"" Oct 08 18:17:30 crc kubenswrapper[4859]: I1008 18:17:30.702433 4859 reconciler_common.go:293] "Volume detached for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-trusted-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 08 18:17:30 crc kubenswrapper[4859]: I1008 18:17:30.702442 4859 reconciler_common.go:293] "Volume detached for volume \"serviceca\" (UniqueName: \"kubernetes.io/configmap/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59-serviceca\") on node \"crc\" DevicePath \"\"" Oct 08 18:17:30 crc kubenswrapper[4859]: I1008 18:17:30.702450 4859 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7539238d-5fe0-46ed-884e-1c3b566537ec-config\") on node \"crc\" DevicePath \"\"" Oct 08 18:17:30 crc kubenswrapper[4859]: I1008 18:17:30.702479 4859 reconciler_common.go:293] "Volume detached for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-etcd-serving-ca\") on node \"crc\" DevicePath \"\"" Oct 08 18:17:30 crc kubenswrapper[4859]: I1008 18:17:30.702489 4859 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-serving-cert\") on node \"crc\" DevicePath \"\"" Oct 08 18:17:30 crc kubenswrapper[4859]: I1008 18:17:30.702498 4859 reconciler_common.go:293] "Volume detached for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/8f668bae-612b-4b75-9490-919e737c6a3b-installation-pull-secrets\") on node \"crc\" DevicePath \"\"" Oct 08 18:17:30 crc kubenswrapper[4859]: I1008 18:17:30.702507 4859 reconciler_common.go:293] "Volume detached for volume \"package-server-manager-serving-cert\" (UniqueName: \"kubernetes.io/secret/3ab1a177-2de0-46d9-b765-d0d0649bb42e-package-server-manager-serving-cert\") on node \"crc\" DevicePath \"\"" Oct 08 18:17:30 crc kubenswrapper[4859]: I1008 18:17:30.702774 4859 reconciler_common.go:293] "Volume detached for volume \"samples-operator-tls\" (UniqueName: \"kubernetes.io/secret/a0128f3a-b052-44ed-a84e-c4c8aaf17c13-samples-operator-tls\") on node \"crc\" DevicePath \"\"" Oct 08 18:17:30 crc kubenswrapper[4859]: I1008 18:17:30.702790 4859 reconciler_common.go:293] "Volume detached for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/43509403-f426-496e-be36-56cef71462f5-console-serving-cert\") on node \"crc\" DevicePath \"\"" Oct 08 18:17:30 crc kubenswrapper[4859]: I1008 18:17:30.702822 4859 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-gf66m\" (UniqueName: \"kubernetes.io/projected/a0128f3a-b052-44ed-a84e-c4c8aaf17c13-kube-api-access-gf66m\") on node \"crc\" DevicePath \"\"" Oct 08 18:17:30 crc kubenswrapper[4859]: I1008 18:17:30.702831 4859 reconciler_common.go:293] "Volume detached for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-service-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 08 18:17:30 crc kubenswrapper[4859]: I1008 18:17:30.702843 4859 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-rnphk\" (UniqueName: \"kubernetes.io/projected/bf126b07-da06-4140-9a57-dfd54fc6b486-kube-api-access-rnphk\") on node \"crc\" DevicePath \"\"" Oct 08 18:17:30 crc kubenswrapper[4859]: I1008 18:17:30.702852 4859 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-d6qdx\" (UniqueName: \"kubernetes.io/projected/87cf06ed-a83f-41a7-828d-70653580a8cb-kube-api-access-d6qdx\") on node \"crc\" DevicePath \"\"" Oct 08 18:17:30 crc kubenswrapper[4859]: I1008 18:17:30.702862 4859 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-provider-selection\") on node \"crc\" DevicePath \"\"" Oct 08 18:17:30 crc kubenswrapper[4859]: I1008 18:17:30.702872 4859 reconciler_common.go:293] "Volume detached for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-etcd-serving-ca\") on node \"crc\" DevicePath \"\"" Oct 08 18:17:30 crc kubenswrapper[4859]: I1008 18:17:30.702906 4859 reconciler_common.go:293] "Volume detached for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-ovnkube-script-lib\") on node \"crc\" DevicePath \"\"" Oct 08 18:17:30 crc kubenswrapper[4859]: I1008 18:17:30.702914 4859 reconciler_common.go:293] "Volume detached for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/a31745f5-9847-4afe-82a5-3161cc66ca93-bound-sa-token\") on node \"crc\" DevicePath \"\"" Oct 08 18:17:30 crc kubenswrapper[4859]: I1008 18:17:30.702922 4859 reconciler_common.go:293] "Volume detached for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/496e6271-fb68-4057-954e-a0d97a4afa3f-kube-api-access\") on node \"crc\" DevicePath \"\"" Oct 08 18:17:30 crc kubenswrapper[4859]: I1008 18:17:30.702931 4859 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-w9rds\" (UniqueName: \"kubernetes.io/projected/20b0d48f-5fd6-431c-a545-e3c800c7b866-kube-api-access-w9rds\") on node \"crc\" DevicePath \"\"" Oct 08 18:17:30 crc kubenswrapper[4859]: I1008 18:17:30.702943 4859 reconciler_common.go:293] "Volume detached for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/1386a44e-36a2-460c-96d0-0359d2b6f0f5-kube-api-access\") on node \"crc\" DevicePath \"\"" Oct 08 18:17:30 crc kubenswrapper[4859]: I1008 18:17:30.702951 4859 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/0b78653f-4ff9-4508-8672-245ed9b561e3-serving-cert\") on node \"crc\" DevicePath \"\"" Oct 08 18:17:30 crc kubenswrapper[4859]: I1008 18:17:30.702786 4859 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/1d611f23-29be-4491-8495-bee1670e935f-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "1d611f23-29be-4491-8495-bee1670e935f" (UID: "1d611f23-29be-4491-8495-bee1670e935f"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 08 18:17:30 crc kubenswrapper[4859]: I1008 18:17:30.703002 4859 reconciler_common.go:293] "Volume detached for volume \"audit\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-audit\") on node \"crc\" DevicePath \"\"" Oct 08 18:17:30 crc kubenswrapper[4859]: I1008 18:17:30.703014 4859 reconciler_common.go:293] "Volume detached for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/22c825df-677d-4ca6-82db-3454ed06e783-auth-proxy-config\") on node \"crc\" DevicePath \"\"" Oct 08 18:17:30 crc kubenswrapper[4859]: I1008 18:17:30.703024 4859 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-7c4vf\" (UniqueName: \"kubernetes.io/projected/22c825df-677d-4ca6-82db-3454ed06e783-kube-api-access-7c4vf\") on node \"crc\" DevicePath \"\"" Oct 08 18:17:30 crc kubenswrapper[4859]: I1008 18:17:30.703033 4859 reconciler_common.go:293] "Volume detached for volume \"images\" (UniqueName: \"kubernetes.io/configmap/31d8b7a1-420e-4252-a5b7-eebe8a111292-images\") on node \"crc\" DevicePath \"\"" Oct 08 18:17:30 crc kubenswrapper[4859]: I1008 18:17:30.703063 4859 reconciler_common.go:293] "Volume detached for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/43509403-f426-496e-be36-56cef71462f5-console-oauth-config\") on node \"crc\" DevicePath \"\"" Oct 08 18:17:30 crc kubenswrapper[4859]: I1008 18:17:30.703075 4859 reconciler_common.go:293] "Volume detached for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-etcd-client\") on node \"crc\" DevicePath \"\"" Oct 08 18:17:30 crc kubenswrapper[4859]: I1008 18:17:30.703083 4859 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-s4n52\" (UniqueName: \"kubernetes.io/projected/925f1c65-6136-48ba-85aa-3a3b50560753-kube-api-access-s4n52\") on node \"crc\" DevicePath \"\"" Oct 08 18:17:30 crc kubenswrapper[4859]: I1008 18:17:30.703091 4859 reconciler_common.go:293] "Volume detached for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/9d4552c7-cd75-42dd-8880-30dd377c49a4-trusted-ca\") on node \"crc\" DevicePath \"\"" Oct 08 18:17:30 crc kubenswrapper[4859]: I1008 18:17:30.703100 4859 reconciler_common.go:293] "Volume detached for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/0b78653f-4ff9-4508-8672-245ed9b561e3-service-ca\") on node \"crc\" DevicePath \"\"" Oct 08 18:17:30 crc kubenswrapper[4859]: I1008 18:17:30.703112 4859 reconciler_common.go:293] "Volume detached for volume \"mcc-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/0b574797-001e-440a-8f4e-c0be86edad0f-mcc-auth-proxy-config\") on node \"crc\" DevicePath \"\"" Oct 08 18:17:30 crc kubenswrapper[4859]: I1008 18:17:30.703119 4859 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-serving-cert\") on node \"crc\" DevicePath \"\"" Oct 08 18:17:30 crc kubenswrapper[4859]: I1008 18:17:30.703167 4859 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-cliconfig\") on node \"crc\" DevicePath \"\"" Oct 08 18:17:30 crc kubenswrapper[4859]: I1008 18:17:30.703180 4859 reconciler_common.go:293] "Volume detached for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-encryption-config\") on node \"crc\" DevicePath \"\"" Oct 08 18:17:30 crc kubenswrapper[4859]: I1008 18:17:30.703191 4859 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-config\") on node \"crc\" DevicePath \"\"" Oct 08 18:17:30 crc kubenswrapper[4859]: I1008 18:17:30.703199 4859 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-bf2bz\" (UniqueName: \"kubernetes.io/projected/1d611f23-29be-4491-8495-bee1670e935f-kube-api-access-bf2bz\") on node \"crc\" DevicePath \"\"" Oct 08 18:17:30 crc kubenswrapper[4859]: I1008 18:17:30.703226 4859 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-config\") on node \"crc\" DevicePath \"\"" Oct 08 18:17:30 crc kubenswrapper[4859]: I1008 18:17:30.703238 4859 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-service-ca\") on node \"crc\" DevicePath \"\"" Oct 08 18:17:30 crc kubenswrapper[4859]: I1008 18:17:30.703246 4859 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-tk88c\" (UniqueName: \"kubernetes.io/projected/7539238d-5fe0-46ed-884e-1c3b566537ec-kube-api-access-tk88c\") on node \"crc\" DevicePath \"\"" Oct 08 18:17:30 crc kubenswrapper[4859]: I1008 18:17:30.703256 4859 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/496e6271-fb68-4057-954e-a0d97a4afa3f-serving-cert\") on node \"crc\" DevicePath \"\"" Oct 08 18:17:30 crc kubenswrapper[4859]: I1008 18:17:30.703325 4859 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-6g6sz\" (UniqueName: \"kubernetes.io/projected/6509e943-70c6-444c-bc41-48a544e36fbd-kube-api-access-6g6sz\") on node \"crc\" DevicePath \"\"" Oct 08 18:17:30 crc kubenswrapper[4859]: I1008 18:17:30.703340 4859 reconciler_common.go:293] "Volume detached for volume \"signing-key\" (UniqueName: \"kubernetes.io/secret/25e176fe-21b4-4974-b1ed-c8b94f112a7f-signing-key\") on node \"crc\" DevicePath \"\"" Oct 08 18:17:30 crc kubenswrapper[4859]: I1008 18:17:30.703348 4859 reconciler_common.go:293] "Volume detached for volume \"etcd-ca\" (UniqueName: \"kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-ca\") on node \"crc\" DevicePath \"\"" Oct 08 18:17:30 crc kubenswrapper[4859]: I1008 18:17:30.703357 4859 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/496e6271-fb68-4057-954e-a0d97a4afa3f-config\") on node \"crc\" DevicePath \"\"" Oct 08 18:17:30 crc kubenswrapper[4859]: I1008 18:17:30.703369 4859 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-6ccd8\" (UniqueName: \"kubernetes.io/projected/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-kube-api-access-6ccd8\") on node \"crc\" DevicePath \"\"" Oct 08 18:17:30 crc kubenswrapper[4859]: I1008 18:17:30.703397 4859 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/01ab3dd5-8196-46d0-ad33-122e2ca51def-config\") on node \"crc\" DevicePath \"\"" Oct 08 18:17:30 crc kubenswrapper[4859]: I1008 18:17:30.703406 4859 reconciler_common.go:293] "Volume detached for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/5b88f790-22fa-440e-b583-365168c0b23d-metrics-certs\") on node \"crc\" DevicePath \"\"" Oct 08 18:17:30 crc kubenswrapper[4859]: I1008 18:17:30.703415 4859 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-vt5rc\" (UniqueName: \"kubernetes.io/projected/44663579-783b-4372-86d6-acf235a62d72-kube-api-access-vt5rc\") on node \"crc\" DevicePath \"\"" Oct 08 18:17:30 crc kubenswrapper[4859]: I1008 18:17:30.703425 4859 reconciler_common.go:293] "Volume detached for volume \"image-import-ca\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-image-import-ca\") on node \"crc\" DevicePath \"\"" Oct 08 18:17:30 crc kubenswrapper[4859]: I1008 18:17:30.703435 4859 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/6402fda4-df10-493c-b4e5-d0569419652d-config\") on node \"crc\" DevicePath \"\"" Oct 08 18:17:30 crc kubenswrapper[4859]: I1008 18:17:30.703446 4859 reconciler_common.go:293] "Volume detached for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-trusted-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 08 18:17:30 crc kubenswrapper[4859]: I1008 18:17:30.703482 4859 reconciler_common.go:293] "Volume detached for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/a31745f5-9847-4afe-82a5-3161cc66ca93-trusted-ca\") on node \"crc\" DevicePath \"\"" Oct 08 18:17:30 crc kubenswrapper[4859]: I1008 18:17:30.703494 4859 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-serving-cert\") on node \"crc\" DevicePath \"\"" Oct 08 18:17:30 crc kubenswrapper[4859]: I1008 18:17:30.703503 4859 reconciler_common.go:293] "Volume detached for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-profile-collector-cert\") on node \"crc\" DevicePath \"\"" Oct 08 18:17:30 crc kubenswrapper[4859]: I1008 18:17:30.703513 4859 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-cfbct\" (UniqueName: \"kubernetes.io/projected/57a731c4-ef35-47a8-b875-bfb08a7f8011-kube-api-access-cfbct\") on node \"crc\" DevicePath \"\"" Oct 08 18:17:30 crc kubenswrapper[4859]: I1008 18:17:30.703524 4859 reconciler_common.go:293] "Volume detached for volume \"certs\" (UniqueName: \"kubernetes.io/secret/5fe579f8-e8a6-4643-bce5-a661393c4dde-certs\") on node \"crc\" DevicePath \"\"" Oct 08 18:17:30 crc kubenswrapper[4859]: I1008 18:17:30.703533 4859 reconciler_common.go:293] "Volume detached for volume \"mcd-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/fda69060-fa79-4696-b1a6-7980f124bf7c-mcd-auth-proxy-config\") on node \"crc\" DevicePath \"\"" Oct 08 18:17:30 crc kubenswrapper[4859]: I1008 18:17:30.703561 4859 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/5225d0e4-402f-4861-b410-819f433b1803-utilities\") on node \"crc\" DevicePath \"\"" Oct 08 18:17:30 crc kubenswrapper[4859]: I1008 18:17:30.703570 4859 reconciler_common.go:293] "Volume detached for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/bf126b07-da06-4140-9a57-dfd54fc6b486-bound-sa-token\") on node \"crc\" DevicePath \"\"" Oct 08 18:17:30 crc kubenswrapper[4859]: I1008 18:17:30.703582 4859 reconciler_common.go:293] "Volume detached for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-trusted-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 08 18:17:30 crc kubenswrapper[4859]: I1008 18:17:30.703591 4859 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-zgdk5\" (UniqueName: \"kubernetes.io/projected/31d8b7a1-420e-4252-a5b7-eebe8a111292-kube-api-access-zgdk5\") on node \"crc\" DevicePath \"\"" Oct 08 18:17:30 crc kubenswrapper[4859]: I1008 18:17:30.703599 4859 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-lzf88\" (UniqueName: \"kubernetes.io/projected/0b574797-001e-440a-8f4e-c0be86edad0f-kube-api-access-lzf88\") on node \"crc\" DevicePath \"\"" Oct 08 18:17:30 crc kubenswrapper[4859]: I1008 18:17:30.703608 4859 reconciler_common.go:293] "Volume detached for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/6ea678ab-3438-413e-bfe3-290ae7725660-ovn-node-metrics-cert\") on node \"crc\" DevicePath \"\"" Oct 08 18:17:30 crc kubenswrapper[4859]: I1008 18:17:30.703648 4859 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-w4xd4\" (UniqueName: \"kubernetes.io/projected/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-kube-api-access-w4xd4\") on node \"crc\" DevicePath \"\"" Oct 08 18:17:30 crc kubenswrapper[4859]: I1008 18:17:30.703659 4859 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-x7zkh\" (UniqueName: \"kubernetes.io/projected/6731426b-95fe-49ff-bb5f-40441049fde2-kube-api-access-x7zkh\") on node \"crc\" DevicePath \"\"" Oct 08 18:17:30 crc kubenswrapper[4859]: I1008 18:17:30.703669 4859 reconciler_common.go:293] "Volume detached for volume \"ovn-control-plane-metrics-cert\" (UniqueName: \"kubernetes.io/secret/925f1c65-6136-48ba-85aa-3a3b50560753-ovn-control-plane-metrics-cert\") on node \"crc\" DevicePath \"\"" Oct 08 18:17:30 crc kubenswrapper[4859]: I1008 18:17:30.703680 4859 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/1386a44e-36a2-460c-96d0-0359d2b6f0f5-config\") on node \"crc\" DevicePath \"\"" Oct 08 18:17:30 crc kubenswrapper[4859]: I1008 18:17:30.703703 4859 reconciler_common.go:293] "Volume detached for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-console-config\") on node \"crc\" DevicePath \"\"" Oct 08 18:17:30 crc kubenswrapper[4859]: I1008 18:17:30.703714 4859 reconciler_common.go:293] "Volume detached for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/b6312bbd-5731-4ea0-a20f-81d5a57df44a-profile-collector-cert\") on node \"crc\" DevicePath \"\"" Oct 08 18:17:30 crc kubenswrapper[4859]: I1008 18:17:30.703723 4859 reconciler_common.go:293] "Volume detached for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-oauth-serving-cert\") on node \"crc\" DevicePath \"\"" Oct 08 18:17:30 crc kubenswrapper[4859]: I1008 18:17:30.703764 4859 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-279lb\" (UniqueName: \"kubernetes.io/projected/7bb08738-c794-4ee8-9972-3a62ca171029-kube-api-access-279lb\") on node \"crc\" DevicePath \"\"" Oct 08 18:17:30 crc kubenswrapper[4859]: I1008 18:17:30.703777 4859 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-config\") on node \"crc\" DevicePath \"\"" Oct 08 18:17:30 crc kubenswrapper[4859]: I1008 18:17:30.703786 4859 reconciler_common.go:293] "Volume detached for volume \"etcd-service-ca\" (UniqueName: \"kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-service-ca\") on node \"crc\" DevicePath \"\"" Oct 08 18:17:30 crc kubenswrapper[4859]: I1008 18:17:30.704346 4859 reconciler_common.go:293] "Volume detached for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-registry-tls\") on node \"crc\" DevicePath \"\"" Oct 08 18:17:30 crc kubenswrapper[4859]: I1008 18:17:30.704358 4859 reconciler_common.go:293] "Volume detached for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-apiservice-cert\") on node \"crc\" DevicePath \"\"" Oct 08 18:17:30 crc kubenswrapper[4859]: I1008 18:17:30.704368 4859 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/5441d097-087c-4d9a-baa8-b210afa90fc9-config\") on node \"crc\" DevicePath \"\"" Oct 08 18:17:30 crc kubenswrapper[4859]: I1008 18:17:30.704377 4859 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-lz9wn\" (UniqueName: \"kubernetes.io/projected/a31745f5-9847-4afe-82a5-3161cc66ca93-kube-api-access-lz9wn\") on node \"crc\" DevicePath \"\"" Oct 08 18:17:30 crc kubenswrapper[4859]: I1008 18:17:30.704387 4859 reconciler_common.go:293] "Volume detached for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-ovnkube-config\") on node \"crc\" DevicePath \"\"" Oct 08 18:17:30 crc kubenswrapper[4859]: I1008 18:17:30.704397 4859 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-config\") on node \"crc\" DevicePath \"\"" Oct 08 18:17:30 crc kubenswrapper[4859]: I1008 18:17:30.704406 4859 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-xcphl\" (UniqueName: \"kubernetes.io/projected/7583ce53-e0fe-4a16-9e4d-50516596a136-kube-api-access-xcphl\") on node \"crc\" DevicePath \"\"" Oct 08 18:17:30 crc kubenswrapper[4859]: I1008 18:17:30.704415 4859 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-x2m85\" (UniqueName: \"kubernetes.io/projected/cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d-kube-api-access-x2m85\") on node \"crc\" DevicePath \"\"" Oct 08 18:17:30 crc kubenswrapper[4859]: I1008 18:17:30.704424 4859 reconciler_common.go:293] "Volume detached for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-audit-policies\") on node \"crc\" DevicePath \"\"" Oct 08 18:17:30 crc kubenswrapper[4859]: I1008 18:17:30.704434 4859 reconciler_common.go:293] "Volume detached for volume \"images\" (UniqueName: \"kubernetes.io/configmap/6402fda4-df10-493c-b4e5-d0569419652d-images\") on node \"crc\" DevicePath \"\"" Oct 08 18:17:30 crc kubenswrapper[4859]: I1008 18:17:30.704442 4859 reconciler_common.go:293] "Volume detached for volume \"cni-sysctl-allowlist\" (UniqueName: \"kubernetes.io/configmap/7bb08738-c794-4ee8-9972-3a62ca171029-cni-sysctl-allowlist\") on node \"crc\" DevicePath \"\"" Oct 08 18:17:30 crc kubenswrapper[4859]: I1008 18:17:30.704452 4859 reconciler_common.go:293] "Volume detached for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/bf126b07-da06-4140-9a57-dfd54fc6b486-trusted-ca\") on node \"crc\" DevicePath \"\"" Oct 08 18:17:30 crc kubenswrapper[4859]: I1008 18:17:30.704461 4859 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-d4lsv\" (UniqueName: \"kubernetes.io/projected/25e176fe-21b4-4974-b1ed-c8b94f112a7f-kube-api-access-d4lsv\") on node \"crc\" DevicePath \"\"" Oct 08 18:17:30 crc kubenswrapper[4859]: I1008 18:17:30.704470 4859 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-ngvvp\" (UniqueName: \"kubernetes.io/projected/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-kube-api-access-ngvvp\") on node \"crc\" DevicePath \"\"" Oct 08 18:17:30 crc kubenswrapper[4859]: I1008 18:17:30.704495 4859 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/57a731c4-ef35-47a8-b875-bfb08a7f8011-utilities\") on node \"crc\" DevicePath \"\"" Oct 08 18:17:30 crc kubenswrapper[4859]: I1008 18:17:30.704505 4859 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-serving-cert\") on node \"crc\" DevicePath \"\"" Oct 08 18:17:30 crc kubenswrapper[4859]: I1008 18:17:30.704515 4859 reconciler_common.go:293] "Volume detached for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-env-overrides\") on node \"crc\" DevicePath \"\"" Oct 08 18:17:30 crc kubenswrapper[4859]: I1008 18:17:30.704525 4859 reconciler_common.go:293] "Volume detached for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-client-ca\") on node \"crc\" DevicePath \"\"" Oct 08 18:17:30 crc kubenswrapper[4859]: I1008 18:17:30.704534 4859 reconciler_common.go:293] "Volume detached for volume \"image-registry-operator-tls\" (UniqueName: \"kubernetes.io/secret/bf126b07-da06-4140-9a57-dfd54fc6b486-image-registry-operator-tls\") on node \"crc\" DevicePath \"\"" Oct 08 18:17:30 crc kubenswrapper[4859]: I1008 18:17:30.704543 4859 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/01ab3dd5-8196-46d0-ad33-122e2ca51def-serving-cert\") on node \"crc\" DevicePath \"\"" Oct 08 18:17:30 crc kubenswrapper[4859]: I1008 18:17:30.704551 4859 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-pcxfs\" (UniqueName: \"kubernetes.io/projected/9d4552c7-cd75-42dd-8880-30dd377c49a4-kube-api-access-pcxfs\") on node \"crc\" DevicePath \"\"" Oct 08 18:17:30 crc kubenswrapper[4859]: I1008 18:17:30.704560 4859 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 08 18:17:30 crc kubenswrapper[4859]: I1008 18:17:30.704568 4859 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-xcgwh\" (UniqueName: \"kubernetes.io/projected/fda69060-fa79-4696-b1a6-7980f124bf7c-kube-api-access-xcgwh\") on node \"crc\" DevicePath \"\"" Oct 08 18:17:30 crc kubenswrapper[4859]: I1008 18:17:30.704577 4859 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-fqsjt\" (UniqueName: \"kubernetes.io/projected/efdd0498-1daa-4136-9a4a-3b948c2293fc-kube-api-access-fqsjt\") on node \"crc\" DevicePath \"\"" Oct 08 18:17:30 crc kubenswrapper[4859]: I1008 18:17:30.704586 4859 reconciler_common.go:293] "Volume detached for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-webhook-cert\") on node \"crc\" DevicePath \"\"" Oct 08 18:17:30 crc kubenswrapper[4859]: I1008 18:17:30.704594 4859 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-fcqwp\" (UniqueName: \"kubernetes.io/projected/5fe579f8-e8a6-4643-bce5-a661393c4dde-kube-api-access-fcqwp\") on node \"crc\" DevicePath \"\"" Oct 08 18:17:30 crc kubenswrapper[4859]: I1008 18:17:30.704602 4859 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-249nr\" (UniqueName: \"kubernetes.io/projected/b6312bbd-5731-4ea0-a20f-81d5a57df44a-kube-api-access-249nr\") on node \"crc\" DevicePath \"\"" Oct 08 18:17:30 crc kubenswrapper[4859]: I1008 18:17:30.704613 4859 reconciler_common.go:293] "Volume detached for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-service-ca\") on node \"crc\" DevicePath \"\"" Oct 08 18:17:30 crc kubenswrapper[4859]: I1008 18:17:30.713577 4859 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/8f668bae-612b-4b75-9490-919e737c6a3b-ca-trust-extracted" (OuterVolumeSpecName: "ca-trust-extracted") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b"). InnerVolumeSpecName "ca-trust-extracted". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 08 18:17:30 crc kubenswrapper[4859]: I1008 18:17:30.717838 4859 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/57a731c4-ef35-47a8-b875-bfb08a7f8011-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "57a731c4-ef35-47a8-b875-bfb08a7f8011" (UID: "57a731c4-ef35-47a8-b875-bfb08a7f8011"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 08 18:17:30 crc kubenswrapper[4859]: I1008 18:17:30.733153 4859 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/5225d0e4-402f-4861-b410-819f433b1803-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "5225d0e4-402f-4861-b410-819f433b1803" (UID: "5225d0e4-402f-4861-b410-819f433b1803"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 08 18:17:30 crc kubenswrapper[4859]: I1008 18:17:30.806041 4859 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/1d611f23-29be-4491-8495-bee1670e935f-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 08 18:17:30 crc kubenswrapper[4859]: I1008 18:17:30.806072 4859 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/5225d0e4-402f-4861-b410-819f433b1803-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 08 18:17:30 crc kubenswrapper[4859]: I1008 18:17:30.806081 4859 reconciler_common.go:293] "Volume detached for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/8f668bae-612b-4b75-9490-919e737c6a3b-ca-trust-extracted\") on node \"crc\" DevicePath \"\"" Oct 08 18:17:30 crc kubenswrapper[4859]: I1008 18:17:30.806090 4859 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/57a731c4-ef35-47a8-b875-bfb08a7f8011-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 08 18:17:30 crc kubenswrapper[4859]: I1008 18:17:30.912360 4859 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" Oct 08 18:17:30 crc kubenswrapper[4859]: I1008 18:17:30.919595 4859 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-node-identity/network-node-identity-vrzqb" Oct 08 18:17:30 crc kubenswrapper[4859]: I1008 18:17:30.928660 4859 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-operator/iptables-alerter-4ln5h" Oct 08 18:17:30 crc kubenswrapper[4859]: W1008 18:17:30.937333 4859 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podef543e1b_8068_4ea3_b32a_61027b32e95d.slice/crio-5e612aa49f7e98b7197227ca6f3f9499193c841abdea78d8b278038db4a522f6 WatchSource:0}: Error finding container 5e612aa49f7e98b7197227ca6f3f9499193c841abdea78d8b278038db4a522f6: Status 404 returned error can't find the container with id 5e612aa49f7e98b7197227ca6f3f9499193c841abdea78d8b278038db4a522f6 Oct 08 18:17:31 crc kubenswrapper[4859]: I1008 18:17:31.108758 4859 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 08 18:17:31 crc kubenswrapper[4859]: E1008 18:17:31.109029 4859 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-08 18:17:32.108990257 +0000 UTC m=+22.355829636 (durationBeforeRetry 1s). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 08 18:17:31 crc kubenswrapper[4859]: I1008 18:17:31.209503 4859 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 08 18:17:31 crc kubenswrapper[4859]: I1008 18:17:31.209569 4859 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 08 18:17:31 crc kubenswrapper[4859]: I1008 18:17:31.209602 4859 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 08 18:17:31 crc kubenswrapper[4859]: I1008 18:17:31.209658 4859 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 08 18:17:31 crc kubenswrapper[4859]: E1008 18:17:31.209704 4859 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Oct 08 18:17:31 crc kubenswrapper[4859]: E1008 18:17:31.209746 4859 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Oct 08 18:17:31 crc kubenswrapper[4859]: E1008 18:17:31.209759 4859 projected.go:194] Error preparing data for projected volume kube-api-access-s2dwl for pod openshift-network-diagnostics/network-check-source-55646444c4-trplf: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Oct 08 18:17:31 crc kubenswrapper[4859]: E1008 18:17:31.209818 4859 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl podName:9d751cbb-f2e2-430d-9754-c882a5e924a5 nodeName:}" failed. No retries permitted until 2025-10-08 18:17:32.209797527 +0000 UTC m=+22.456636896 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "kube-api-access-s2dwl" (UniqueName: "kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl") pod "network-check-source-55646444c4-trplf" (UID: "9d751cbb-f2e2-430d-9754-c882a5e924a5") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Oct 08 18:17:31 crc kubenswrapper[4859]: E1008 18:17:31.209822 4859 secret.go:188] Couldn't get secret openshift-network-console/networking-console-plugin-cert: object "openshift-network-console"/"networking-console-plugin-cert" not registered Oct 08 18:17:31 crc kubenswrapper[4859]: E1008 18:17:31.209877 4859 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Oct 08 18:17:31 crc kubenswrapper[4859]: E1008 18:17:31.209888 4859 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Oct 08 18:17:31 crc kubenswrapper[4859]: E1008 18:17:31.209895 4859 projected.go:194] Error preparing data for projected volume kube-api-access-cqllr for pod openshift-network-diagnostics/network-check-target-xd92c: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Oct 08 18:17:31 crc kubenswrapper[4859]: E1008 18:17:31.209902 4859 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-10-08 18:17:32.209880589 +0000 UTC m=+22.456719968 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "networking-console-plugin-cert" (UniqueName: "kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin-cert" not registered Oct 08 18:17:31 crc kubenswrapper[4859]: E1008 18:17:31.209921 4859 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr podName:3b6479f0-333b-4a96-9adf-2099afdc2447 nodeName:}" failed. No retries permitted until 2025-10-08 18:17:32.20991408 +0000 UTC m=+22.456753459 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "kube-api-access-cqllr" (UniqueName: "kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr") pod "network-check-target-xd92c" (UID: "3b6479f0-333b-4a96-9adf-2099afdc2447") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Oct 08 18:17:31 crc kubenswrapper[4859]: E1008 18:17:31.209966 4859 configmap.go:193] Couldn't get configMap openshift-network-console/networking-console-plugin: object "openshift-network-console"/"networking-console-plugin" not registered Oct 08 18:17:31 crc kubenswrapper[4859]: E1008 18:17:31.209997 4859 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-10-08 18:17:32.209990812 +0000 UTC m=+22.456830191 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "nginx-conf" (UniqueName: "kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin" not registered Oct 08 18:17:31 crc kubenswrapper[4859]: I1008 18:17:31.469946 4859 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 08 18:17:31 crc kubenswrapper[4859]: I1008 18:17:31.469946 4859 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 08 18:17:31 crc kubenswrapper[4859]: E1008 18:17:31.470093 4859 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 08 18:17:31 crc kubenswrapper[4859]: E1008 18:17:31.470178 4859 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 08 18:17:31 crc kubenswrapper[4859]: I1008 18:17:31.469968 4859 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 08 18:17:31 crc kubenswrapper[4859]: E1008 18:17:31.470484 4859 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 08 18:17:31 crc kubenswrapper[4859]: I1008 18:17:31.540595 4859 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-dns/node-resolver-btw28"] Oct 08 18:17:31 crc kubenswrapper[4859]: I1008 18:17:31.541109 4859 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-dns/node-resolver-btw28" Oct 08 18:17:31 crc kubenswrapper[4859]: I1008 18:17:31.543096 4859 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-dns"/"node-resolver-dockercfg-kz9s7" Oct 08 18:17:31 crc kubenswrapper[4859]: I1008 18:17:31.543132 4859 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-dns"/"openshift-service-ca.crt" Oct 08 18:17:31 crc kubenswrapper[4859]: I1008 18:17:31.543191 4859 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-dns"/"kube-root-ca.crt" Oct 08 18:17:31 crc kubenswrapper[4859]: I1008 18:17:31.551635 4859 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-btw28" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"97bc5598-3ead-4300-a938-3aebbff1e8c6\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:17:31Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:17:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:17:31Z\\\",\\\"message\\\":\\\"containers with unready status: [dns-node-resolver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:17:31Z\\\",\\\"message\\\":\\\"containers with unready status: [dns-node-resolver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qcxsd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T18:17:31Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-btw28\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Oct 08 18:17:31 crc kubenswrapper[4859]: I1008 18:17:31.563053 4859 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"63a623de-5608-4228-b1c2-77188c276986\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:17:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:17:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:17:10Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:17:10Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:17:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d4786a8fb8abd96fb22e60733c974f599017407d589ed815fb8541fd3b64f696\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T18:17:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://183d481502ed5108d5e19e483b714c289f355ea3f8b0891883e5335b802c83b6\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T18:17:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ec848c46fd5dde0469b00d3cab14ba4a652533a265490fadbee7a8b8843ef249\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T18:17:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4fe1828b4d63ff8106bfcd15e1dc1c1bd1ba7f77777cdb23ff1b291d0418774c\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2e700b2a20c81b414617f5b5c018a738278a3fa44e6c6d14b2170051d0725431\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-08T18:17:24Z\\\",\\\"message\\\":\\\"W1008 18:17:13.670625 1 cmd.go:257] Using insecure, self-signed certificates\\\\nI1008 18:17:13.671202 1 crypto.go:601] Generating new CA for check-endpoints-signer@1759947433 cert, and key in /tmp/serving-cert-3867218267/serving-signer.crt, /tmp/serving-cert-3867218267/serving-signer.key\\\\nI1008 18:17:13.944463 1 observer_polling.go:159] Starting file observer\\\\nW1008 18:17:13.951958 1 builder.go:272] unable to get owner reference (falling back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": dial tcp [::1]:6443: connect: connection refused\\\\nI1008 18:17:13.952088 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1008 18:17:13.956755 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3867218267/tls.crt::/tmp/serving-cert-3867218267/tls.key\\\\\\\"\\\\nF1008 18:17:24.290991 1 cmd.go:182] error initializing delegating authentication: unable to load configmap based request-header-client-ca-file: Get \\\\\\\"https://localhost:6443/api/v1/namespaces/kube-system/configmaps/extension-apiserver-authentication\\\\\\\": net/http: TLS handshake timeout\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-08T18:17:13Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T18:17:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c65b66f62d325ca10ef8d5d31f04529ccf4e691cf3303d231e7a414fb400e4d6\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T18:17:13Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://58c7bba4ace28215b062653a960e561b1dc1f890f2baf450145638a065af038b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://58c7bba4ace28215b062653a960e561b1dc1f890f2baf450145638a065af038b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-08T18:17:11Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-08T18:17:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T18:17:10Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Oct 08 18:17:31 crc kubenswrapper[4859]: I1008 18:17:31.573154 4859 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-08T18:17:29Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [webhook approver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T18:17:29Z\\\",\\\"message\\\":\\\"containers with unready status: [webhook approver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":false,\\\"restartCount\\\":6,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Oct 08 18:17:31 crc kubenswrapper[4859]: I1008 18:17:31.583718 4859 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-08T18:17:29Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T18:17:29Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Oct 08 18:17:31 crc kubenswrapper[4859]: I1008 18:17:31.594254 4859 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-08T18:17:29Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T18:17:29Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Oct 08 18:17:31 crc kubenswrapper[4859]: I1008 18:17:31.605903 4859 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-08T18:17:29Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T18:17:29Z\\\",\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":false,\\\"restartCount\\\":5,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Oct 08 18:17:31 crc kubenswrapper[4859]: I1008 18:17:31.616388 4859 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-08T18:17:29Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T18:17:29Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Oct 08 18:17:31 crc kubenswrapper[4859]: I1008 18:17:31.628098 4859 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-08T18:17:29Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T18:17:29Z\\\",\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Oct 08 18:17:31 crc kubenswrapper[4859]: I1008 18:17:31.632363 4859 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" event={"ID":"37a5e44f-9a88-4405-be8a-b645485e7312","Type":"ContainerStarted","Data":"c6d15905694b6bbd4999d49dc8f7a53eb6bfaa7adb59efb0d7e4973c21d1c7be"} Oct 08 18:17:31 crc kubenswrapper[4859]: I1008 18:17:31.632406 4859 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" event={"ID":"37a5e44f-9a88-4405-be8a-b645485e7312","Type":"ContainerStarted","Data":"15ca52dcc7738558b6b053f684634b015fc1fdd3d1a86ab5e6cc2113d3c43240"} Oct 08 18:17:31 crc kubenswrapper[4859]: I1008 18:17:31.633767 4859 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" event={"ID":"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49","Type":"ContainerStarted","Data":"84bd8a3b6155c25d48460c95a20ae8c0b3366667984eca50baf6d475fdb3bbf4"} Oct 08 18:17:31 crc kubenswrapper[4859]: I1008 18:17:31.635506 4859 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" event={"ID":"ef543e1b-8068-4ea3-b32a-61027b32e95d","Type":"ContainerStarted","Data":"99e2d51bd7757498dae60949eea6d732b5f7b481fa00f808dd5308a530a1daf1"} Oct 08 18:17:31 crc kubenswrapper[4859]: I1008 18:17:31.635539 4859 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" event={"ID":"ef543e1b-8068-4ea3-b32a-61027b32e95d","Type":"ContainerStarted","Data":"e32aaba93699944665660ae41e9b5779178c1354c31e15ad8188bf8c3c8fa500"} Oct 08 18:17:31 crc kubenswrapper[4859]: I1008 18:17:31.635550 4859 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" event={"ID":"ef543e1b-8068-4ea3-b32a-61027b32e95d","Type":"ContainerStarted","Data":"5e612aa49f7e98b7197227ca6f3f9499193c841abdea78d8b278038db4a522f6"} Oct 08 18:17:31 crc kubenswrapper[4859]: I1008 18:17:31.656675 4859 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-08T18:17:29Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T18:17:29Z\\\",\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Oct 08 18:17:31 crc kubenswrapper[4859]: I1008 18:17:31.667468 4859 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-08T18:17:31Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T18:17:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T18:17:31Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c6d15905694b6bbd4999d49dc8f7a53eb6bfaa7adb59efb0d7e4973c21d1c7be\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T18:17:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Oct 08 18:17:31 crc kubenswrapper[4859]: I1008 18:17:31.678871 4859 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-08T18:17:29Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T18:17:29Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Oct 08 18:17:31 crc kubenswrapper[4859]: I1008 18:17:31.687956 4859 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-08T18:17:29Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [webhook approver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T18:17:29Z\\\",\\\"message\\\":\\\"containers with unready status: [webhook approver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":false,\\\"restartCount\\\":6,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Oct 08 18:17:31 crc kubenswrapper[4859]: I1008 18:17:31.697874 4859 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-08T18:17:29Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T18:17:29Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Oct 08 18:17:31 crc kubenswrapper[4859]: I1008 18:17:31.707043 4859 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-08T18:17:29Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T18:17:29Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Oct 08 18:17:31 crc kubenswrapper[4859]: I1008 18:17:31.714806 4859 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-btw28" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"97bc5598-3ead-4300-a938-3aebbff1e8c6\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:17:31Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:17:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:17:31Z\\\",\\\"message\\\":\\\"containers with unready status: [dns-node-resolver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:17:31Z\\\",\\\"message\\\":\\\"containers with unready status: [dns-node-resolver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qcxsd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T18:17:31Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-btw28\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Oct 08 18:17:31 crc kubenswrapper[4859]: I1008 18:17:31.715096 4859 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-qcxsd\" (UniqueName: \"kubernetes.io/projected/97bc5598-3ead-4300-a938-3aebbff1e8c6-kube-api-access-qcxsd\") pod \"node-resolver-btw28\" (UID: \"97bc5598-3ead-4300-a938-3aebbff1e8c6\") " pod="openshift-dns/node-resolver-btw28" Oct 08 18:17:31 crc kubenswrapper[4859]: I1008 18:17:31.715147 4859 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"hosts-file\" (UniqueName: \"kubernetes.io/host-path/97bc5598-3ead-4300-a938-3aebbff1e8c6-hosts-file\") pod \"node-resolver-btw28\" (UID: \"97bc5598-3ead-4300-a938-3aebbff1e8c6\") " pod="openshift-dns/node-resolver-btw28" Oct 08 18:17:31 crc kubenswrapper[4859]: I1008 18:17:31.726608 4859 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"63a623de-5608-4228-b1c2-77188c276986\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:17:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:17:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:17:10Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:17:10Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:17:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d4786a8fb8abd96fb22e60733c974f599017407d589ed815fb8541fd3b64f696\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T18:17:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://183d481502ed5108d5e19e483b714c289f355ea3f8b0891883e5335b802c83b6\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T18:17:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ec848c46fd5dde0469b00d3cab14ba4a652533a265490fadbee7a8b8843ef249\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T18:17:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4fe1828b4d63ff8106bfcd15e1dc1c1bd1ba7f77777cdb23ff1b291d0418774c\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2e700b2a20c81b414617f5b5c018a738278a3fa44e6c6d14b2170051d0725431\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-08T18:17:24Z\\\",\\\"message\\\":\\\"W1008 18:17:13.670625 1 cmd.go:257] Using insecure, self-signed certificates\\\\nI1008 18:17:13.671202 1 crypto.go:601] Generating new CA for check-endpoints-signer@1759947433 cert, and key in /tmp/serving-cert-3867218267/serving-signer.crt, /tmp/serving-cert-3867218267/serving-signer.key\\\\nI1008 18:17:13.944463 1 observer_polling.go:159] Starting file observer\\\\nW1008 18:17:13.951958 1 builder.go:272] unable to get owner reference (falling back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": dial tcp [::1]:6443: connect: connection refused\\\\nI1008 18:17:13.952088 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1008 18:17:13.956755 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3867218267/tls.crt::/tmp/serving-cert-3867218267/tls.key\\\\\\\"\\\\nF1008 18:17:24.290991 1 cmd.go:182] error initializing delegating authentication: unable to load configmap based request-header-client-ca-file: Get \\\\\\\"https://localhost:6443/api/v1/namespaces/kube-system/configmaps/extension-apiserver-authentication\\\\\\\": net/http: TLS handshake timeout\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-08T18:17:13Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T18:17:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c65b66f62d325ca10ef8d5d31f04529ccf4e691cf3303d231e7a414fb400e4d6\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T18:17:13Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://58c7bba4ace28215b062653a960e561b1dc1f890f2baf450145638a065af038b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://58c7bba4ace28215b062653a960e561b1dc1f890f2baf450145638a065af038b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-08T18:17:11Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-08T18:17:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T18:17:10Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Oct 08 18:17:31 crc kubenswrapper[4859]: I1008 18:17:31.737977 4859 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-08T18:17:31Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T18:17:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T18:17:31Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://99e2d51bd7757498dae60949eea6d732b5f7b481fa00f808dd5308a530a1daf1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T18:17:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e32aaba93699944665660ae41e9b5779178c1354c31e15ad8188bf8c3c8fa500\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T18:17:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Oct 08 18:17:31 crc kubenswrapper[4859]: I1008 18:17:31.752626 4859 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-08T18:17:29Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T18:17:29Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T18:17:31Z is after 2025-08-24T17:21:41Z" Oct 08 18:17:31 crc kubenswrapper[4859]: I1008 18:17:31.767348 4859 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-08T18:17:29Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T18:17:29Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T18:17:31Z is after 2025-08-24T17:21:41Z" Oct 08 18:17:31 crc kubenswrapper[4859]: I1008 18:17:31.778933 4859 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-btw28" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"97bc5598-3ead-4300-a938-3aebbff1e8c6\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:17:31Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:17:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:17:31Z\\\",\\\"message\\\":\\\"containers with unready status: [dns-node-resolver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:17:31Z\\\",\\\"message\\\":\\\"containers with unready status: [dns-node-resolver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qcxsd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T18:17:31Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-btw28\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T18:17:31Z is after 2025-08-24T17:21:41Z" Oct 08 18:17:31 crc kubenswrapper[4859]: I1008 18:17:31.791398 4859 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"63a623de-5608-4228-b1c2-77188c276986\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:17:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:17:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:17:10Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:17:10Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:17:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d4786a8fb8abd96fb22e60733c974f599017407d589ed815fb8541fd3b64f696\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T18:17:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://183d481502ed5108d5e19e483b714c289f355ea3f8b0891883e5335b802c83b6\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T18:17:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ec848c46fd5dde0469b00d3cab14ba4a652533a265490fadbee7a8b8843ef249\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T18:17:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4fe1828b4d63ff8106bfcd15e1dc1c1bd1ba7f77777cdb23ff1b291d0418774c\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2e700b2a20c81b414617f5b5c018a738278a3fa44e6c6d14b2170051d0725431\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-08T18:17:24Z\\\",\\\"message\\\":\\\"W1008 18:17:13.670625 1 cmd.go:257] Using insecure, self-signed certificates\\\\nI1008 18:17:13.671202 1 crypto.go:601] Generating new CA for check-endpoints-signer@1759947433 cert, and key in /tmp/serving-cert-3867218267/serving-signer.crt, /tmp/serving-cert-3867218267/serving-signer.key\\\\nI1008 18:17:13.944463 1 observer_polling.go:159] Starting file observer\\\\nW1008 18:17:13.951958 1 builder.go:272] unable to get owner reference (falling back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": dial tcp [::1]:6443: connect: connection refused\\\\nI1008 18:17:13.952088 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1008 18:17:13.956755 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3867218267/tls.crt::/tmp/serving-cert-3867218267/tls.key\\\\\\\"\\\\nF1008 18:17:24.290991 1 cmd.go:182] error initializing delegating authentication: unable to load configmap based request-header-client-ca-file: Get \\\\\\\"https://localhost:6443/api/v1/namespaces/kube-system/configmaps/extension-apiserver-authentication\\\\\\\": net/http: TLS handshake timeout\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-08T18:17:13Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T18:17:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c65b66f62d325ca10ef8d5d31f04529ccf4e691cf3303d231e7a414fb400e4d6\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T18:17:13Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://58c7bba4ace28215b062653a960e561b1dc1f890f2baf450145638a065af038b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://58c7bba4ace28215b062653a960e561b1dc1f890f2baf450145638a065af038b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-08T18:17:11Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-08T18:17:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T18:17:10Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T18:17:31Z is after 2025-08-24T17:21:41Z" Oct 08 18:17:31 crc kubenswrapper[4859]: I1008 18:17:31.803456 4859 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-08T18:17:29Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T18:17:29Z\\\",\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T18:17:31Z is after 2025-08-24T17:21:41Z" Oct 08 18:17:31 crc kubenswrapper[4859]: I1008 18:17:31.816573 4859 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-08T18:17:31Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T18:17:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T18:17:31Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c6d15905694b6bbd4999d49dc8f7a53eb6bfaa7adb59efb0d7e4973c21d1c7be\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T18:17:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T18:17:31Z is after 2025-08-24T17:21:41Z" Oct 08 18:17:31 crc kubenswrapper[4859]: I1008 18:17:31.816793 4859 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-qcxsd\" (UniqueName: \"kubernetes.io/projected/97bc5598-3ead-4300-a938-3aebbff1e8c6-kube-api-access-qcxsd\") pod \"node-resolver-btw28\" (UID: \"97bc5598-3ead-4300-a938-3aebbff1e8c6\") " pod="openshift-dns/node-resolver-btw28" Oct 08 18:17:31 crc kubenswrapper[4859]: I1008 18:17:31.816849 4859 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"hosts-file\" (UniqueName: \"kubernetes.io/host-path/97bc5598-3ead-4300-a938-3aebbff1e8c6-hosts-file\") pod \"node-resolver-btw28\" (UID: \"97bc5598-3ead-4300-a938-3aebbff1e8c6\") " pod="openshift-dns/node-resolver-btw28" Oct 08 18:17:31 crc kubenswrapper[4859]: I1008 18:17:31.817008 4859 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"hosts-file\" (UniqueName: \"kubernetes.io/host-path/97bc5598-3ead-4300-a938-3aebbff1e8c6-hosts-file\") pod \"node-resolver-btw28\" (UID: \"97bc5598-3ead-4300-a938-3aebbff1e8c6\") " pod="openshift-dns/node-resolver-btw28" Oct 08 18:17:31 crc kubenswrapper[4859]: I1008 18:17:31.829393 4859 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-08T18:17:29Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T18:17:29Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T18:17:31Z is after 2025-08-24T17:21:41Z" Oct 08 18:17:31 crc kubenswrapper[4859]: I1008 18:17:31.837583 4859 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-qcxsd\" (UniqueName: \"kubernetes.io/projected/97bc5598-3ead-4300-a938-3aebbff1e8c6-kube-api-access-qcxsd\") pod \"node-resolver-btw28\" (UID: \"97bc5598-3ead-4300-a938-3aebbff1e8c6\") " pod="openshift-dns/node-resolver-btw28" Oct 08 18:17:31 crc kubenswrapper[4859]: I1008 18:17:31.853049 4859 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-dns/node-resolver-btw28" Oct 08 18:17:31 crc kubenswrapper[4859]: W1008 18:17:31.871713 4859 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod97bc5598_3ead_4300_a938_3aebbff1e8c6.slice/crio-ee6847e36adfd6281ee24aa0814dca2ca4c129940f9962d0e030ec56cbe319d4 WatchSource:0}: Error finding container ee6847e36adfd6281ee24aa0814dca2ca4c129940f9962d0e030ec56cbe319d4: Status 404 returned error can't find the container with id ee6847e36adfd6281ee24aa0814dca2ca4c129940f9962d0e030ec56cbe319d4 Oct 08 18:17:31 crc kubenswrapper[4859]: I1008 18:17:31.924637 4859 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-multus/multus-8plkg"] Oct 08 18:17:31 crc kubenswrapper[4859]: I1008 18:17:31.925018 4859 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-multus/multus-additional-cni-plugins-84k8x"] Oct 08 18:17:31 crc kubenswrapper[4859]: I1008 18:17:31.925279 4859 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/multus-8plkg" Oct 08 18:17:31 crc kubenswrapper[4859]: I1008 18:17:31.926246 4859 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/multus-additional-cni-plugins-84k8x" Oct 08 18:17:31 crc kubenswrapper[4859]: I1008 18:17:31.926630 4859 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-machine-config-operator/machine-config-daemon-82s52"] Oct 08 18:17:31 crc kubenswrapper[4859]: I1008 18:17:31.927101 4859 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/machine-config-daemon-82s52" Oct 08 18:17:31 crc kubenswrapper[4859]: I1008 18:17:31.928974 4859 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-multus"/"multus-daemon-config" Oct 08 18:17:31 crc kubenswrapper[4859]: I1008 18:17:31.929103 4859 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"machine-config-daemon-dockercfg-r5tcq" Oct 08 18:17:31 crc kubenswrapper[4859]: I1008 18:17:31.929151 4859 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-multus"/"cni-copy-resources" Oct 08 18:17:31 crc kubenswrapper[4859]: I1008 18:17:31.929172 4859 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-multus"/"kube-root-ca.crt" Oct 08 18:17:31 crc kubenswrapper[4859]: I1008 18:17:31.929416 4859 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-multus"/"default-cni-sysctl-allowlist" Oct 08 18:17:31 crc kubenswrapper[4859]: I1008 18:17:31.929304 4859 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-config-operator"/"kube-rbac-proxy" Oct 08 18:17:31 crc kubenswrapper[4859]: I1008 18:17:31.929881 4859 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-config-operator"/"openshift-service-ca.crt" Oct 08 18:17:31 crc kubenswrapper[4859]: I1008 18:17:31.930006 4859 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"multus-ancillary-tools-dockercfg-vnmsz" Oct 08 18:17:31 crc kubenswrapper[4859]: I1008 18:17:31.930049 4859 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-config-operator"/"kube-root-ca.crt" Oct 08 18:17:31 crc kubenswrapper[4859]: I1008 18:17:31.930098 4859 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"default-dockercfg-2q5b6" Oct 08 18:17:31 crc kubenswrapper[4859]: I1008 18:17:31.930641 4859 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"proxy-tls" Oct 08 18:17:31 crc kubenswrapper[4859]: I1008 18:17:31.930699 4859 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-multus"/"openshift-service-ca.crt" Oct 08 18:17:31 crc kubenswrapper[4859]: I1008 18:17:31.954343 4859 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-08T18:17:29Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T18:17:29Z\\\",\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T18:17:31Z is after 2025-08-24T17:21:41Z" Oct 08 18:17:31 crc kubenswrapper[4859]: I1008 18:17:31.979123 4859 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-08T18:17:31Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T18:17:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T18:17:31Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c6d15905694b6bbd4999d49dc8f7a53eb6bfaa7adb59efb0d7e4973c21d1c7be\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T18:17:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T18:17:31Z is after 2025-08-24T17:21:41Z" Oct 08 18:17:32 crc kubenswrapper[4859]: I1008 18:17:32.019315 4859 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/92b2d7de-31cc-4a91-95a5-ed5ea964b028-host-run-netns\") pod \"multus-8plkg\" (UID: \"92b2d7de-31cc-4a91-95a5-ed5ea964b028\") " pod="openshift-multus/multus-8plkg" Oct 08 18:17:32 crc kubenswrapper[4859]: I1008 18:17:32.019359 4859 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-kubernetes\" (UniqueName: \"kubernetes.io/host-path/92b2d7de-31cc-4a91-95a5-ed5ea964b028-etc-kubernetes\") pod \"multus-8plkg\" (UID: \"92b2d7de-31cc-4a91-95a5-ed5ea964b028\") " pod="openshift-multus/multus-8plkg" Oct 08 18:17:32 crc kubenswrapper[4859]: I1008 18:17:32.019379 4859 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cnibin\" (UniqueName: \"kubernetes.io/host-path/92b2d7de-31cc-4a91-95a5-ed5ea964b028-cnibin\") pod \"multus-8plkg\" (UID: \"92b2d7de-31cc-4a91-95a5-ed5ea964b028\") " pod="openshift-multus/multus-8plkg" Oct 08 18:17:32 crc kubenswrapper[4859]: I1008 18:17:32.019400 4859 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-var-lib-cni-multus\" (UniqueName: \"kubernetes.io/host-path/92b2d7de-31cc-4a91-95a5-ed5ea964b028-host-var-lib-cni-multus\") pod \"multus-8plkg\" (UID: \"92b2d7de-31cc-4a91-95a5-ed5ea964b028\") " pod="openshift-multus/multus-8plkg" Oct 08 18:17:32 crc kubenswrapper[4859]: I1008 18:17:32.019432 4859 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"system-cni-dir\" (UniqueName: \"kubernetes.io/host-path/1c24078c-c9e8-4160-883a-f483b4b3dd23-system-cni-dir\") pod \"multus-additional-cni-plugins-84k8x\" (UID: \"1c24078c-c9e8-4160-883a-f483b4b3dd23\") " pod="openshift-multus/multus-additional-cni-plugins-84k8x" Oct 08 18:17:32 crc kubenswrapper[4859]: I1008 18:17:32.019473 4859 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rootfs\" (UniqueName: \"kubernetes.io/host-path/b23a6a6c-9d92-4e7b-840e-55cfda873a2d-rootfs\") pod \"machine-config-daemon-82s52\" (UID: \"b23a6a6c-9d92-4e7b-840e-55cfda873a2d\") " pod="openshift-machine-config-operator/machine-config-daemon-82s52" Oct 08 18:17:32 crc kubenswrapper[4859]: I1008 18:17:32.019496 4859 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cnibin\" (UniqueName: \"kubernetes.io/host-path/1c24078c-c9e8-4160-883a-f483b4b3dd23-cnibin\") pod \"multus-additional-cni-plugins-84k8x\" (UID: \"1c24078c-c9e8-4160-883a-f483b4b3dd23\") " pod="openshift-multus/multus-additional-cni-plugins-84k8x" Oct 08 18:17:32 crc kubenswrapper[4859]: I1008 18:17:32.019513 4859 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-var-lib-cni-bin\" (UniqueName: \"kubernetes.io/host-path/92b2d7de-31cc-4a91-95a5-ed5ea964b028-host-var-lib-cni-bin\") pod \"multus-8plkg\" (UID: \"92b2d7de-31cc-4a91-95a5-ed5ea964b028\") " pod="openshift-multus/multus-8plkg" Oct 08 18:17:32 crc kubenswrapper[4859]: I1008 18:17:32.019533 4859 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"multus-cni-dir\" (UniqueName: \"kubernetes.io/host-path/92b2d7de-31cc-4a91-95a5-ed5ea964b028-multus-cni-dir\") pod \"multus-8plkg\" (UID: \"92b2d7de-31cc-4a91-95a5-ed5ea964b028\") " pod="openshift-multus/multus-8plkg" Oct 08 18:17:32 crc kubenswrapper[4859]: I1008 18:17:32.019647 4859 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"multus-conf-dir\" (UniqueName: \"kubernetes.io/host-path/92b2d7de-31cc-4a91-95a5-ed5ea964b028-multus-conf-dir\") pod \"multus-8plkg\" (UID: \"92b2d7de-31cc-4a91-95a5-ed5ea964b028\") " pod="openshift-multus/multus-8plkg" Oct 08 18:17:32 crc kubenswrapper[4859]: I1008 18:17:32.019747 4859 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cni-sysctl-allowlist\" (UniqueName: \"kubernetes.io/configmap/1c24078c-c9e8-4160-883a-f483b4b3dd23-cni-sysctl-allowlist\") pod \"multus-additional-cni-plugins-84k8x\" (UID: \"1c24078c-c9e8-4160-883a-f483b4b3dd23\") " pod="openshift-multus/multus-additional-cni-plugins-84k8x" Oct 08 18:17:32 crc kubenswrapper[4859]: I1008 18:17:32.019816 4859 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/92b2d7de-31cc-4a91-95a5-ed5ea964b028-cni-binary-copy\") pod \"multus-8plkg\" (UID: \"92b2d7de-31cc-4a91-95a5-ed5ea964b028\") " pod="openshift-multus/multus-8plkg" Oct 08 18:17:32 crc kubenswrapper[4859]: I1008 18:17:32.019860 4859 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-run-k8s-cni-cncf-io\" (UniqueName: \"kubernetes.io/host-path/92b2d7de-31cc-4a91-95a5-ed5ea964b028-host-run-k8s-cni-cncf-io\") pod \"multus-8plkg\" (UID: \"92b2d7de-31cc-4a91-95a5-ed5ea964b028\") " pod="openshift-multus/multus-8plkg" Oct 08 18:17:32 crc kubenswrapper[4859]: I1008 18:17:32.019894 4859 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-var-lib-kubelet\" (UniqueName: \"kubernetes.io/host-path/92b2d7de-31cc-4a91-95a5-ed5ea964b028-host-var-lib-kubelet\") pod \"multus-8plkg\" (UID: \"92b2d7de-31cc-4a91-95a5-ed5ea964b028\") " pod="openshift-multus/multus-8plkg" Oct 08 18:17:32 crc kubenswrapper[4859]: I1008 18:17:32.019918 4859 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/b23a6a6c-9d92-4e7b-840e-55cfda873a2d-proxy-tls\") pod \"machine-config-daemon-82s52\" (UID: \"b23a6a6c-9d92-4e7b-840e-55cfda873a2d\") " pod="openshift-machine-config-operator/machine-config-daemon-82s52" Oct 08 18:17:32 crc kubenswrapper[4859]: I1008 18:17:32.019944 4859 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/1c24078c-c9e8-4160-883a-f483b4b3dd23-cni-binary-copy\") pod \"multus-additional-cni-plugins-84k8x\" (UID: \"1c24078c-c9e8-4160-883a-f483b4b3dd23\") " pod="openshift-multus/multus-additional-cni-plugins-84k8x" Oct 08 18:17:32 crc kubenswrapper[4859]: I1008 18:17:32.019966 4859 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"multus-socket-dir-parent\" (UniqueName: \"kubernetes.io/host-path/92b2d7de-31cc-4a91-95a5-ed5ea964b028-multus-socket-dir-parent\") pod \"multus-8plkg\" (UID: \"92b2d7de-31cc-4a91-95a5-ed5ea964b028\") " pod="openshift-multus/multus-8plkg" Oct 08 18:17:32 crc kubenswrapper[4859]: I1008 18:17:32.019990 4859 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"hostroot\" (UniqueName: \"kubernetes.io/host-path/92b2d7de-31cc-4a91-95a5-ed5ea964b028-hostroot\") pod \"multus-8plkg\" (UID: \"92b2d7de-31cc-4a91-95a5-ed5ea964b028\") " pod="openshift-multus/multus-8plkg" Oct 08 18:17:32 crc kubenswrapper[4859]: I1008 18:17:32.020031 4859 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-6vx22\" (UniqueName: \"kubernetes.io/projected/b23a6a6c-9d92-4e7b-840e-55cfda873a2d-kube-api-access-6vx22\") pod \"machine-config-daemon-82s52\" (UID: \"b23a6a6c-9d92-4e7b-840e-55cfda873a2d\") " pod="openshift-machine-config-operator/machine-config-daemon-82s52" Oct 08 18:17:32 crc kubenswrapper[4859]: I1008 18:17:32.020060 4859 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"os-release\" (UniqueName: \"kubernetes.io/host-path/1c24078c-c9e8-4160-883a-f483b4b3dd23-os-release\") pod \"multus-additional-cni-plugins-84k8x\" (UID: \"1c24078c-c9e8-4160-883a-f483b4b3dd23\") " pod="openshift-multus/multus-additional-cni-plugins-84k8x" Oct 08 18:17:32 crc kubenswrapper[4859]: I1008 18:17:32.020086 4859 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-hk5xb\" (UniqueName: \"kubernetes.io/projected/1c24078c-c9e8-4160-883a-f483b4b3dd23-kube-api-access-hk5xb\") pod \"multus-additional-cni-plugins-84k8x\" (UID: \"1c24078c-c9e8-4160-883a-f483b4b3dd23\") " pod="openshift-multus/multus-additional-cni-plugins-84k8x" Oct 08 18:17:32 crc kubenswrapper[4859]: I1008 18:17:32.020121 4859 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"system-cni-dir\" (UniqueName: \"kubernetes.io/host-path/92b2d7de-31cc-4a91-95a5-ed5ea964b028-system-cni-dir\") pod \"multus-8plkg\" (UID: \"92b2d7de-31cc-4a91-95a5-ed5ea964b028\") " pod="openshift-multus/multus-8plkg" Oct 08 18:17:32 crc kubenswrapper[4859]: I1008 18:17:32.020149 4859 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"tuning-conf-dir\" (UniqueName: \"kubernetes.io/host-path/1c24078c-c9e8-4160-883a-f483b4b3dd23-tuning-conf-dir\") pod \"multus-additional-cni-plugins-84k8x\" (UID: \"1c24078c-c9e8-4160-883a-f483b4b3dd23\") " pod="openshift-multus/multus-additional-cni-plugins-84k8x" Oct 08 18:17:32 crc kubenswrapper[4859]: I1008 18:17:32.020193 4859 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"os-release\" (UniqueName: \"kubernetes.io/host-path/92b2d7de-31cc-4a91-95a5-ed5ea964b028-os-release\") pod \"multus-8plkg\" (UID: \"92b2d7de-31cc-4a91-95a5-ed5ea964b028\") " pod="openshift-multus/multus-8plkg" Oct 08 18:17:32 crc kubenswrapper[4859]: I1008 18:17:32.020291 4859 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"multus-daemon-config\" (UniqueName: \"kubernetes.io/configmap/92b2d7de-31cc-4a91-95a5-ed5ea964b028-multus-daemon-config\") pod \"multus-8plkg\" (UID: \"92b2d7de-31cc-4a91-95a5-ed5ea964b028\") " pod="openshift-multus/multus-8plkg" Oct 08 18:17:32 crc kubenswrapper[4859]: I1008 18:17:32.020357 4859 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"mcd-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/b23a6a6c-9d92-4e7b-840e-55cfda873a2d-mcd-auth-proxy-config\") pod \"machine-config-daemon-82s52\" (UID: \"b23a6a6c-9d92-4e7b-840e-55cfda873a2d\") " pod="openshift-machine-config-operator/machine-config-daemon-82s52" Oct 08 18:17:32 crc kubenswrapper[4859]: I1008 18:17:32.020385 4859 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-run-multus-certs\" (UniqueName: \"kubernetes.io/host-path/92b2d7de-31cc-4a91-95a5-ed5ea964b028-host-run-multus-certs\") pod \"multus-8plkg\" (UID: \"92b2d7de-31cc-4a91-95a5-ed5ea964b028\") " pod="openshift-multus/multus-8plkg" Oct 08 18:17:32 crc kubenswrapper[4859]: I1008 18:17:32.020438 4859 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-q9d7m\" (UniqueName: \"kubernetes.io/projected/92b2d7de-31cc-4a91-95a5-ed5ea964b028-kube-api-access-q9d7m\") pod \"multus-8plkg\" (UID: \"92b2d7de-31cc-4a91-95a5-ed5ea964b028\") " pod="openshift-multus/multus-8plkg" Oct 08 18:17:32 crc kubenswrapper[4859]: I1008 18:17:32.022201 4859 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-08T18:17:29Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T18:17:29Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T18:17:32Z is after 2025-08-24T17:21:41Z" Oct 08 18:17:32 crc kubenswrapper[4859]: I1008 18:17:32.036896 4859 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-08T18:17:31Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T18:17:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T18:17:31Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://99e2d51bd7757498dae60949eea6d732b5f7b481fa00f808dd5308a530a1daf1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T18:17:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e32aaba93699944665660ae41e9b5779178c1354c31e15ad8188bf8c3c8fa500\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T18:17:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T18:17:32Z is after 2025-08-24T17:21:41Z" Oct 08 18:17:32 crc kubenswrapper[4859]: I1008 18:17:32.053547 4859 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-08T18:17:29Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T18:17:29Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T18:17:32Z is after 2025-08-24T17:21:41Z" Oct 08 18:17:32 crc kubenswrapper[4859]: I1008 18:17:32.068842 4859 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-08T18:17:29Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T18:17:29Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T18:17:32Z is after 2025-08-24T17:21:41Z" Oct 08 18:17:32 crc kubenswrapper[4859]: I1008 18:17:32.087203 4859 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-btw28" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"97bc5598-3ead-4300-a938-3aebbff1e8c6\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:17:31Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:17:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:17:31Z\\\",\\\"message\\\":\\\"containers with unready status: [dns-node-resolver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:17:31Z\\\",\\\"message\\\":\\\"containers with unready status: [dns-node-resolver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qcxsd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T18:17:31Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-btw28\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T18:17:32Z is after 2025-08-24T17:21:41Z" Oct 08 18:17:32 crc kubenswrapper[4859]: I1008 18:17:32.103767 4859 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"63a623de-5608-4228-b1c2-77188c276986\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:17:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:17:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:17:10Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:17:10Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:17:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d4786a8fb8abd96fb22e60733c974f599017407d589ed815fb8541fd3b64f696\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T18:17:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://183d481502ed5108d5e19e483b714c289f355ea3f8b0891883e5335b802c83b6\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T18:17:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ec848c46fd5dde0469b00d3cab14ba4a652533a265490fadbee7a8b8843ef249\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T18:17:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4fe1828b4d63ff8106bfcd15e1dc1c1bd1ba7f77777cdb23ff1b291d0418774c\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2e700b2a20c81b414617f5b5c018a738278a3fa44e6c6d14b2170051d0725431\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-08T18:17:24Z\\\",\\\"message\\\":\\\"W1008 18:17:13.670625 1 cmd.go:257] Using insecure, self-signed certificates\\\\nI1008 18:17:13.671202 1 crypto.go:601] Generating new CA for check-endpoints-signer@1759947433 cert, and key in /tmp/serving-cert-3867218267/serving-signer.crt, /tmp/serving-cert-3867218267/serving-signer.key\\\\nI1008 18:17:13.944463 1 observer_polling.go:159] Starting file observer\\\\nW1008 18:17:13.951958 1 builder.go:272] unable to get owner reference (falling back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": dial tcp [::1]:6443: connect: connection refused\\\\nI1008 18:17:13.952088 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1008 18:17:13.956755 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3867218267/tls.crt::/tmp/serving-cert-3867218267/tls.key\\\\\\\"\\\\nF1008 18:17:24.290991 1 cmd.go:182] error initializing delegating authentication: unable to load configmap based request-header-client-ca-file: Get \\\\\\\"https://localhost:6443/api/v1/namespaces/kube-system/configmaps/extension-apiserver-authentication\\\\\\\": net/http: TLS handshake timeout\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-08T18:17:13Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T18:17:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c65b66f62d325ca10ef8d5d31f04529ccf4e691cf3303d231e7a414fb400e4d6\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T18:17:13Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://58c7bba4ace28215b062653a960e561b1dc1f890f2baf450145638a065af038b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://58c7bba4ace28215b062653a960e561b1dc1f890f2baf450145638a065af038b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-08T18:17:11Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-08T18:17:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T18:17:10Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T18:17:32Z is after 2025-08-24T17:21:41Z" Oct 08 18:17:32 crc kubenswrapper[4859]: I1008 18:17:32.121386 4859 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-8plkg" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"92b2d7de-31cc-4a91-95a5-ed5ea964b028\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:17:31Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:17:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:17:31Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:17:31Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q9d7m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T18:17:31Z\\\"}}\" for pod \"openshift-multus\"/\"multus-8plkg\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T18:17:32Z is after 2025-08-24T17:21:41Z" Oct 08 18:17:32 crc kubenswrapper[4859]: I1008 18:17:32.121622 4859 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 08 18:17:32 crc kubenswrapper[4859]: E1008 18:17:32.121777 4859 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-08 18:17:34.121749929 +0000 UTC m=+24.368589298 (durationBeforeRetry 2s). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 08 18:17:32 crc kubenswrapper[4859]: I1008 18:17:32.121819 4859 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cnibin\" (UniqueName: \"kubernetes.io/host-path/1c24078c-c9e8-4160-883a-f483b4b3dd23-cnibin\") pod \"multus-additional-cni-plugins-84k8x\" (UID: \"1c24078c-c9e8-4160-883a-f483b4b3dd23\") " pod="openshift-multus/multus-additional-cni-plugins-84k8x" Oct 08 18:17:32 crc kubenswrapper[4859]: I1008 18:17:32.121859 4859 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-var-lib-cni-bin\" (UniqueName: \"kubernetes.io/host-path/92b2d7de-31cc-4a91-95a5-ed5ea964b028-host-var-lib-cni-bin\") pod \"multus-8plkg\" (UID: \"92b2d7de-31cc-4a91-95a5-ed5ea964b028\") " pod="openshift-multus/multus-8plkg" Oct 08 18:17:32 crc kubenswrapper[4859]: I1008 18:17:32.121882 4859 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rootfs\" (UniqueName: \"kubernetes.io/host-path/b23a6a6c-9d92-4e7b-840e-55cfda873a2d-rootfs\") pod \"machine-config-daemon-82s52\" (UID: \"b23a6a6c-9d92-4e7b-840e-55cfda873a2d\") " pod="openshift-machine-config-operator/machine-config-daemon-82s52" Oct 08 18:17:32 crc kubenswrapper[4859]: I1008 18:17:32.121909 4859 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"multus-conf-dir\" (UniqueName: \"kubernetes.io/host-path/92b2d7de-31cc-4a91-95a5-ed5ea964b028-multus-conf-dir\") pod \"multus-8plkg\" (UID: \"92b2d7de-31cc-4a91-95a5-ed5ea964b028\") " pod="openshift-multus/multus-8plkg" Oct 08 18:17:32 crc kubenswrapper[4859]: I1008 18:17:32.121932 4859 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cni-sysctl-allowlist\" (UniqueName: \"kubernetes.io/configmap/1c24078c-c9e8-4160-883a-f483b4b3dd23-cni-sysctl-allowlist\") pod \"multus-additional-cni-plugins-84k8x\" (UID: \"1c24078c-c9e8-4160-883a-f483b4b3dd23\") " pod="openshift-multus/multus-additional-cni-plugins-84k8x" Oct 08 18:17:32 crc kubenswrapper[4859]: I1008 18:17:32.121954 4859 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"multus-cni-dir\" (UniqueName: \"kubernetes.io/host-path/92b2d7de-31cc-4a91-95a5-ed5ea964b028-multus-cni-dir\") pod \"multus-8plkg\" (UID: \"92b2d7de-31cc-4a91-95a5-ed5ea964b028\") " pod="openshift-multus/multus-8plkg" Oct 08 18:17:32 crc kubenswrapper[4859]: I1008 18:17:32.121975 4859 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/92b2d7de-31cc-4a91-95a5-ed5ea964b028-cni-binary-copy\") pod \"multus-8plkg\" (UID: \"92b2d7de-31cc-4a91-95a5-ed5ea964b028\") " pod="openshift-multus/multus-8plkg" Oct 08 18:17:32 crc kubenswrapper[4859]: I1008 18:17:32.121990 4859 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"multus-conf-dir\" (UniqueName: \"kubernetes.io/host-path/92b2d7de-31cc-4a91-95a5-ed5ea964b028-multus-conf-dir\") pod \"multus-8plkg\" (UID: \"92b2d7de-31cc-4a91-95a5-ed5ea964b028\") " pod="openshift-multus/multus-8plkg" Oct 08 18:17:32 crc kubenswrapper[4859]: I1008 18:17:32.122019 4859 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-run-k8s-cni-cncf-io\" (UniqueName: \"kubernetes.io/host-path/92b2d7de-31cc-4a91-95a5-ed5ea964b028-host-run-k8s-cni-cncf-io\") pod \"multus-8plkg\" (UID: \"92b2d7de-31cc-4a91-95a5-ed5ea964b028\") " pod="openshift-multus/multus-8plkg" Oct 08 18:17:32 crc kubenswrapper[4859]: I1008 18:17:32.121944 4859 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cnibin\" (UniqueName: \"kubernetes.io/host-path/1c24078c-c9e8-4160-883a-f483b4b3dd23-cnibin\") pod \"multus-additional-cni-plugins-84k8x\" (UID: \"1c24078c-c9e8-4160-883a-f483b4b3dd23\") " pod="openshift-multus/multus-additional-cni-plugins-84k8x" Oct 08 18:17:32 crc kubenswrapper[4859]: I1008 18:17:32.122132 4859 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-var-lib-cni-bin\" (UniqueName: \"kubernetes.io/host-path/92b2d7de-31cc-4a91-95a5-ed5ea964b028-host-var-lib-cni-bin\") pod \"multus-8plkg\" (UID: \"92b2d7de-31cc-4a91-95a5-ed5ea964b028\") " pod="openshift-multus/multus-8plkg" Oct 08 18:17:32 crc kubenswrapper[4859]: I1008 18:17:32.122196 4859 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-var-lib-kubelet\" (UniqueName: \"kubernetes.io/host-path/92b2d7de-31cc-4a91-95a5-ed5ea964b028-host-var-lib-kubelet\") pod \"multus-8plkg\" (UID: \"92b2d7de-31cc-4a91-95a5-ed5ea964b028\") " pod="openshift-multus/multus-8plkg" Oct 08 18:17:32 crc kubenswrapper[4859]: I1008 18:17:32.122271 4859 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/b23a6a6c-9d92-4e7b-840e-55cfda873a2d-proxy-tls\") pod \"machine-config-daemon-82s52\" (UID: \"b23a6a6c-9d92-4e7b-840e-55cfda873a2d\") " pod="openshift-machine-config-operator/machine-config-daemon-82s52" Oct 08 18:17:32 crc kubenswrapper[4859]: I1008 18:17:32.122302 4859 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"multus-socket-dir-parent\" (UniqueName: \"kubernetes.io/host-path/92b2d7de-31cc-4a91-95a5-ed5ea964b028-multus-socket-dir-parent\") pod \"multus-8plkg\" (UID: \"92b2d7de-31cc-4a91-95a5-ed5ea964b028\") " pod="openshift-multus/multus-8plkg" Oct 08 18:17:32 crc kubenswrapper[4859]: I1008 18:17:32.122376 4859 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"hostroot\" (UniqueName: \"kubernetes.io/host-path/92b2d7de-31cc-4a91-95a5-ed5ea964b028-hostroot\") pod \"multus-8plkg\" (UID: \"92b2d7de-31cc-4a91-95a5-ed5ea964b028\") " pod="openshift-multus/multus-8plkg" Oct 08 18:17:32 crc kubenswrapper[4859]: I1008 18:17:32.122417 4859 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/1c24078c-c9e8-4160-883a-f483b4b3dd23-cni-binary-copy\") pod \"multus-additional-cni-plugins-84k8x\" (UID: \"1c24078c-c9e8-4160-883a-f483b4b3dd23\") " pod="openshift-multus/multus-additional-cni-plugins-84k8x" Oct 08 18:17:32 crc kubenswrapper[4859]: I1008 18:17:32.122442 4859 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"os-release\" (UniqueName: \"kubernetes.io/host-path/1c24078c-c9e8-4160-883a-f483b4b3dd23-os-release\") pod \"multus-additional-cni-plugins-84k8x\" (UID: \"1c24078c-c9e8-4160-883a-f483b4b3dd23\") " pod="openshift-multus/multus-additional-cni-plugins-84k8x" Oct 08 18:17:32 crc kubenswrapper[4859]: I1008 18:17:32.122502 4859 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"multus-cni-dir\" (UniqueName: \"kubernetes.io/host-path/92b2d7de-31cc-4a91-95a5-ed5ea964b028-multus-cni-dir\") pod \"multus-8plkg\" (UID: \"92b2d7de-31cc-4a91-95a5-ed5ea964b028\") " pod="openshift-multus/multus-8plkg" Oct 08 18:17:32 crc kubenswrapper[4859]: I1008 18:17:32.122551 4859 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-run-k8s-cni-cncf-io\" (UniqueName: \"kubernetes.io/host-path/92b2d7de-31cc-4a91-95a5-ed5ea964b028-host-run-k8s-cni-cncf-io\") pod \"multus-8plkg\" (UID: \"92b2d7de-31cc-4a91-95a5-ed5ea964b028\") " pod="openshift-multus/multus-8plkg" Oct 08 18:17:32 crc kubenswrapper[4859]: I1008 18:17:32.122636 4859 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-hk5xb\" (UniqueName: \"kubernetes.io/projected/1c24078c-c9e8-4160-883a-f483b4b3dd23-kube-api-access-hk5xb\") pod \"multus-additional-cni-plugins-84k8x\" (UID: \"1c24078c-c9e8-4160-883a-f483b4b3dd23\") " pod="openshift-multus/multus-additional-cni-plugins-84k8x" Oct 08 18:17:32 crc kubenswrapper[4859]: I1008 18:17:32.122740 4859 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"system-cni-dir\" (UniqueName: \"kubernetes.io/host-path/92b2d7de-31cc-4a91-95a5-ed5ea964b028-system-cni-dir\") pod \"multus-8plkg\" (UID: \"92b2d7de-31cc-4a91-95a5-ed5ea964b028\") " pod="openshift-multus/multus-8plkg" Oct 08 18:17:32 crc kubenswrapper[4859]: I1008 18:17:32.122781 4859 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-6vx22\" (UniqueName: \"kubernetes.io/projected/b23a6a6c-9d92-4e7b-840e-55cfda873a2d-kube-api-access-6vx22\") pod \"machine-config-daemon-82s52\" (UID: \"b23a6a6c-9d92-4e7b-840e-55cfda873a2d\") " pod="openshift-machine-config-operator/machine-config-daemon-82s52" Oct 08 18:17:32 crc kubenswrapper[4859]: I1008 18:17:32.122824 4859 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"os-release\" (UniqueName: \"kubernetes.io/host-path/92b2d7de-31cc-4a91-95a5-ed5ea964b028-os-release\") pod \"multus-8plkg\" (UID: \"92b2d7de-31cc-4a91-95a5-ed5ea964b028\") " pod="openshift-multus/multus-8plkg" Oct 08 18:17:32 crc kubenswrapper[4859]: I1008 18:17:32.122861 4859 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"multus-daemon-config\" (UniqueName: \"kubernetes.io/configmap/92b2d7de-31cc-4a91-95a5-ed5ea964b028-multus-daemon-config\") pod \"multus-8plkg\" (UID: \"92b2d7de-31cc-4a91-95a5-ed5ea964b028\") " pod="openshift-multus/multus-8plkg" Oct 08 18:17:32 crc kubenswrapper[4859]: I1008 18:17:32.122981 4859 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cni-sysctl-allowlist\" (UniqueName: \"kubernetes.io/configmap/1c24078c-c9e8-4160-883a-f483b4b3dd23-cni-sysctl-allowlist\") pod \"multus-additional-cni-plugins-84k8x\" (UID: \"1c24078c-c9e8-4160-883a-f483b4b3dd23\") " pod="openshift-multus/multus-additional-cni-plugins-84k8x" Oct 08 18:17:32 crc kubenswrapper[4859]: I1008 18:17:32.123026 4859 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rootfs\" (UniqueName: \"kubernetes.io/host-path/b23a6a6c-9d92-4e7b-840e-55cfda873a2d-rootfs\") pod \"machine-config-daemon-82s52\" (UID: \"b23a6a6c-9d92-4e7b-840e-55cfda873a2d\") " pod="openshift-machine-config-operator/machine-config-daemon-82s52" Oct 08 18:17:32 crc kubenswrapper[4859]: I1008 18:17:32.123083 4859 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"os-release\" (UniqueName: \"kubernetes.io/host-path/1c24078c-c9e8-4160-883a-f483b4b3dd23-os-release\") pod \"multus-additional-cni-plugins-84k8x\" (UID: \"1c24078c-c9e8-4160-883a-f483b4b3dd23\") " pod="openshift-multus/multus-additional-cni-plugins-84k8x" Oct 08 18:17:32 crc kubenswrapper[4859]: I1008 18:17:32.123171 4859 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"multus-socket-dir-parent\" (UniqueName: \"kubernetes.io/host-path/92b2d7de-31cc-4a91-95a5-ed5ea964b028-multus-socket-dir-parent\") pod \"multus-8plkg\" (UID: \"92b2d7de-31cc-4a91-95a5-ed5ea964b028\") " pod="openshift-multus/multus-8plkg" Oct 08 18:17:32 crc kubenswrapper[4859]: I1008 18:17:32.123337 4859 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/1c24078c-c9e8-4160-883a-f483b4b3dd23-cni-binary-copy\") pod \"multus-additional-cni-plugins-84k8x\" (UID: \"1c24078c-c9e8-4160-883a-f483b4b3dd23\") " pod="openshift-multus/multus-additional-cni-plugins-84k8x" Oct 08 18:17:32 crc kubenswrapper[4859]: I1008 18:17:32.123384 4859 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"hostroot\" (UniqueName: \"kubernetes.io/host-path/92b2d7de-31cc-4a91-95a5-ed5ea964b028-hostroot\") pod \"multus-8plkg\" (UID: \"92b2d7de-31cc-4a91-95a5-ed5ea964b028\") " pod="openshift-multus/multus-8plkg" Oct 08 18:17:32 crc kubenswrapper[4859]: I1008 18:17:32.123416 4859 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"os-release\" (UniqueName: \"kubernetes.io/host-path/92b2d7de-31cc-4a91-95a5-ed5ea964b028-os-release\") pod \"multus-8plkg\" (UID: \"92b2d7de-31cc-4a91-95a5-ed5ea964b028\") " pod="openshift-multus/multus-8plkg" Oct 08 18:17:32 crc kubenswrapper[4859]: I1008 18:17:32.123428 4859 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-var-lib-kubelet\" (UniqueName: \"kubernetes.io/host-path/92b2d7de-31cc-4a91-95a5-ed5ea964b028-host-var-lib-kubelet\") pod \"multus-8plkg\" (UID: \"92b2d7de-31cc-4a91-95a5-ed5ea964b028\") " pod="openshift-multus/multus-8plkg" Oct 08 18:17:32 crc kubenswrapper[4859]: I1008 18:17:32.123458 4859 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/92b2d7de-31cc-4a91-95a5-ed5ea964b028-cni-binary-copy\") pod \"multus-8plkg\" (UID: \"92b2d7de-31cc-4a91-95a5-ed5ea964b028\") " pod="openshift-multus/multus-8plkg" Oct 08 18:17:32 crc kubenswrapper[4859]: I1008 18:17:32.123479 4859 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"system-cni-dir\" (UniqueName: \"kubernetes.io/host-path/92b2d7de-31cc-4a91-95a5-ed5ea964b028-system-cni-dir\") pod \"multus-8plkg\" (UID: \"92b2d7de-31cc-4a91-95a5-ed5ea964b028\") " pod="openshift-multus/multus-8plkg" Oct 08 18:17:32 crc kubenswrapper[4859]: I1008 18:17:32.123564 4859 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"tuning-conf-dir\" (UniqueName: \"kubernetes.io/host-path/1c24078c-c9e8-4160-883a-f483b4b3dd23-tuning-conf-dir\") pod \"multus-additional-cni-plugins-84k8x\" (UID: \"1c24078c-c9e8-4160-883a-f483b4b3dd23\") " pod="openshift-multus/multus-additional-cni-plugins-84k8x" Oct 08 18:17:32 crc kubenswrapper[4859]: I1008 18:17:32.123617 4859 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"mcd-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/b23a6a6c-9d92-4e7b-840e-55cfda873a2d-mcd-auth-proxy-config\") pod \"machine-config-daemon-82s52\" (UID: \"b23a6a6c-9d92-4e7b-840e-55cfda873a2d\") " pod="openshift-machine-config-operator/machine-config-daemon-82s52" Oct 08 18:17:32 crc kubenswrapper[4859]: I1008 18:17:32.123677 4859 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-run-multus-certs\" (UniqueName: \"kubernetes.io/host-path/92b2d7de-31cc-4a91-95a5-ed5ea964b028-host-run-multus-certs\") pod \"multus-8plkg\" (UID: \"92b2d7de-31cc-4a91-95a5-ed5ea964b028\") " pod="openshift-multus/multus-8plkg" Oct 08 18:17:32 crc kubenswrapper[4859]: I1008 18:17:32.123730 4859 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-run-multus-certs\" (UniqueName: \"kubernetes.io/host-path/92b2d7de-31cc-4a91-95a5-ed5ea964b028-host-run-multus-certs\") pod \"multus-8plkg\" (UID: \"92b2d7de-31cc-4a91-95a5-ed5ea964b028\") " pod="openshift-multus/multus-8plkg" Oct 08 18:17:32 crc kubenswrapper[4859]: I1008 18:17:32.123814 4859 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-q9d7m\" (UniqueName: \"kubernetes.io/projected/92b2d7de-31cc-4a91-95a5-ed5ea964b028-kube-api-access-q9d7m\") pod \"multus-8plkg\" (UID: \"92b2d7de-31cc-4a91-95a5-ed5ea964b028\") " pod="openshift-multus/multus-8plkg" Oct 08 18:17:32 crc kubenswrapper[4859]: I1008 18:17:32.123847 4859 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/92b2d7de-31cc-4a91-95a5-ed5ea964b028-host-run-netns\") pod \"multus-8plkg\" (UID: \"92b2d7de-31cc-4a91-95a5-ed5ea964b028\") " pod="openshift-multus/multus-8plkg" Oct 08 18:17:32 crc kubenswrapper[4859]: I1008 18:17:32.123874 4859 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cnibin\" (UniqueName: \"kubernetes.io/host-path/92b2d7de-31cc-4a91-95a5-ed5ea964b028-cnibin\") pod \"multus-8plkg\" (UID: \"92b2d7de-31cc-4a91-95a5-ed5ea964b028\") " pod="openshift-multus/multus-8plkg" Oct 08 18:17:32 crc kubenswrapper[4859]: I1008 18:17:32.123900 4859 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-kubernetes\" (UniqueName: \"kubernetes.io/host-path/92b2d7de-31cc-4a91-95a5-ed5ea964b028-etc-kubernetes\") pod \"multus-8plkg\" (UID: \"92b2d7de-31cc-4a91-95a5-ed5ea964b028\") " pod="openshift-multus/multus-8plkg" Oct 08 18:17:32 crc kubenswrapper[4859]: I1008 18:17:32.123924 4859 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"multus-daemon-config\" (UniqueName: \"kubernetes.io/configmap/92b2d7de-31cc-4a91-95a5-ed5ea964b028-multus-daemon-config\") pod \"multus-8plkg\" (UID: \"92b2d7de-31cc-4a91-95a5-ed5ea964b028\") " pod="openshift-multus/multus-8plkg" Oct 08 18:17:32 crc kubenswrapper[4859]: I1008 18:17:32.123929 4859 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"system-cni-dir\" (UniqueName: \"kubernetes.io/host-path/1c24078c-c9e8-4160-883a-f483b4b3dd23-system-cni-dir\") pod \"multus-additional-cni-plugins-84k8x\" (UID: \"1c24078c-c9e8-4160-883a-f483b4b3dd23\") " pod="openshift-multus/multus-additional-cni-plugins-84k8x" Oct 08 18:17:32 crc kubenswrapper[4859]: I1008 18:17:32.123959 4859 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-var-lib-cni-multus\" (UniqueName: \"kubernetes.io/host-path/92b2d7de-31cc-4a91-95a5-ed5ea964b028-host-var-lib-cni-multus\") pod \"multus-8plkg\" (UID: \"92b2d7de-31cc-4a91-95a5-ed5ea964b028\") " pod="openshift-multus/multus-8plkg" Oct 08 18:17:32 crc kubenswrapper[4859]: I1008 18:17:32.123980 4859 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-kubernetes\" (UniqueName: \"kubernetes.io/host-path/92b2d7de-31cc-4a91-95a5-ed5ea964b028-etc-kubernetes\") pod \"multus-8plkg\" (UID: \"92b2d7de-31cc-4a91-95a5-ed5ea964b028\") " pod="openshift-multus/multus-8plkg" Oct 08 18:17:32 crc kubenswrapper[4859]: I1008 18:17:32.123931 4859 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/92b2d7de-31cc-4a91-95a5-ed5ea964b028-host-run-netns\") pod \"multus-8plkg\" (UID: \"92b2d7de-31cc-4a91-95a5-ed5ea964b028\") " pod="openshift-multus/multus-8plkg" Oct 08 18:17:32 crc kubenswrapper[4859]: I1008 18:17:32.124022 4859 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"system-cni-dir\" (UniqueName: \"kubernetes.io/host-path/1c24078c-c9e8-4160-883a-f483b4b3dd23-system-cni-dir\") pod \"multus-additional-cni-plugins-84k8x\" (UID: \"1c24078c-c9e8-4160-883a-f483b4b3dd23\") " pod="openshift-multus/multus-additional-cni-plugins-84k8x" Oct 08 18:17:32 crc kubenswrapper[4859]: I1008 18:17:32.124048 4859 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cnibin\" (UniqueName: \"kubernetes.io/host-path/92b2d7de-31cc-4a91-95a5-ed5ea964b028-cnibin\") pod \"multus-8plkg\" (UID: \"92b2d7de-31cc-4a91-95a5-ed5ea964b028\") " pod="openshift-multus/multus-8plkg" Oct 08 18:17:32 crc kubenswrapper[4859]: I1008 18:17:32.124085 4859 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-var-lib-cni-multus\" (UniqueName: \"kubernetes.io/host-path/92b2d7de-31cc-4a91-95a5-ed5ea964b028-host-var-lib-cni-multus\") pod \"multus-8plkg\" (UID: \"92b2d7de-31cc-4a91-95a5-ed5ea964b028\") " pod="openshift-multus/multus-8plkg" Oct 08 18:17:32 crc kubenswrapper[4859]: I1008 18:17:32.124503 4859 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"mcd-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/b23a6a6c-9d92-4e7b-840e-55cfda873a2d-mcd-auth-proxy-config\") pod \"machine-config-daemon-82s52\" (UID: \"b23a6a6c-9d92-4e7b-840e-55cfda873a2d\") " pod="openshift-machine-config-operator/machine-config-daemon-82s52" Oct 08 18:17:32 crc kubenswrapper[4859]: I1008 18:17:32.127230 4859 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/b23a6a6c-9d92-4e7b-840e-55cfda873a2d-proxy-tls\") pod \"machine-config-daemon-82s52\" (UID: \"b23a6a6c-9d92-4e7b-840e-55cfda873a2d\") " pod="openshift-machine-config-operator/machine-config-daemon-82s52" Oct 08 18:17:32 crc kubenswrapper[4859]: I1008 18:17:32.135482 4859 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"tuning-conf-dir\" (UniqueName: \"kubernetes.io/host-path/1c24078c-c9e8-4160-883a-f483b4b3dd23-tuning-conf-dir\") pod \"multus-additional-cni-plugins-84k8x\" (UID: \"1c24078c-c9e8-4160-883a-f483b4b3dd23\") " pod="openshift-multus/multus-additional-cni-plugins-84k8x" Oct 08 18:17:32 crc kubenswrapper[4859]: I1008 18:17:32.142866 4859 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-08T18:17:31Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T18:17:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T18:17:31Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c6d15905694b6bbd4999d49dc8f7a53eb6bfaa7adb59efb0d7e4973c21d1c7be\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T18:17:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T18:17:32Z is after 2025-08-24T17:21:41Z" Oct 08 18:17:32 crc kubenswrapper[4859]: I1008 18:17:32.145308 4859 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-hk5xb\" (UniqueName: \"kubernetes.io/projected/1c24078c-c9e8-4160-883a-f483b4b3dd23-kube-api-access-hk5xb\") pod \"multus-additional-cni-plugins-84k8x\" (UID: \"1c24078c-c9e8-4160-883a-f483b4b3dd23\") " pod="openshift-multus/multus-additional-cni-plugins-84k8x" Oct 08 18:17:32 crc kubenswrapper[4859]: I1008 18:17:32.145550 4859 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-6vx22\" (UniqueName: \"kubernetes.io/projected/b23a6a6c-9d92-4e7b-840e-55cfda873a2d-kube-api-access-6vx22\") pod \"machine-config-daemon-82s52\" (UID: \"b23a6a6c-9d92-4e7b-840e-55cfda873a2d\") " pod="openshift-machine-config-operator/machine-config-daemon-82s52" Oct 08 18:17:32 crc kubenswrapper[4859]: I1008 18:17:32.146133 4859 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-q9d7m\" (UniqueName: \"kubernetes.io/projected/92b2d7de-31cc-4a91-95a5-ed5ea964b028-kube-api-access-q9d7m\") pod \"multus-8plkg\" (UID: \"92b2d7de-31cc-4a91-95a5-ed5ea964b028\") " pod="openshift-multus/multus-8plkg" Oct 08 18:17:32 crc kubenswrapper[4859]: I1008 18:17:32.157941 4859 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-08T18:17:29Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T18:17:29Z\\\",\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T18:17:32Z is after 2025-08-24T17:21:41Z" Oct 08 18:17:32 crc kubenswrapper[4859]: I1008 18:17:32.175461 4859 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"63a623de-5608-4228-b1c2-77188c276986\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:17:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:17:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:17:10Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:17:10Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:17:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d4786a8fb8abd96fb22e60733c974f599017407d589ed815fb8541fd3b64f696\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T18:17:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://183d481502ed5108d5e19e483b714c289f355ea3f8b0891883e5335b802c83b6\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T18:17:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ec848c46fd5dde0469b00d3cab14ba4a652533a265490fadbee7a8b8843ef249\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T18:17:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4fe1828b4d63ff8106bfcd15e1dc1c1bd1ba7f77777cdb23ff1b291d0418774c\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2e700b2a20c81b414617f5b5c018a738278a3fa44e6c6d14b2170051d0725431\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-08T18:17:24Z\\\",\\\"message\\\":\\\"W1008 18:17:13.670625 1 cmd.go:257] Using insecure, self-signed certificates\\\\nI1008 18:17:13.671202 1 crypto.go:601] Generating new CA for check-endpoints-signer@1759947433 cert, and key in /tmp/serving-cert-3867218267/serving-signer.crt, /tmp/serving-cert-3867218267/serving-signer.key\\\\nI1008 18:17:13.944463 1 observer_polling.go:159] Starting file observer\\\\nW1008 18:17:13.951958 1 builder.go:272] unable to get owner reference (falling back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": dial tcp [::1]:6443: connect: connection refused\\\\nI1008 18:17:13.952088 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1008 18:17:13.956755 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3867218267/tls.crt::/tmp/serving-cert-3867218267/tls.key\\\\\\\"\\\\nF1008 18:17:24.290991 1 cmd.go:182] error initializing delegating authentication: unable to load configmap based request-header-client-ca-file: Get \\\\\\\"https://localhost:6443/api/v1/namespaces/kube-system/configmaps/extension-apiserver-authentication\\\\\\\": net/http: TLS handshake timeout\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-08T18:17:13Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T18:17:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c65b66f62d325ca10ef8d5d31f04529ccf4e691cf3303d231e7a414fb400e4d6\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T18:17:13Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://58c7bba4ace28215b062653a960e561b1dc1f890f2baf450145638a065af038b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://58c7bba4ace28215b062653a960e561b1dc1f890f2baf450145638a065af038b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-08T18:17:11Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-08T18:17:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T18:17:10Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T18:17:32Z is after 2025-08-24T17:21:41Z" Oct 08 18:17:32 crc kubenswrapper[4859]: I1008 18:17:32.190226 4859 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-08T18:17:31Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T18:17:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T18:17:31Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://99e2d51bd7757498dae60949eea6d732b5f7b481fa00f808dd5308a530a1daf1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T18:17:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e32aaba93699944665660ae41e9b5779178c1354c31e15ad8188bf8c3c8fa500\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T18:17:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T18:17:32Z is after 2025-08-24T17:21:41Z" Oct 08 18:17:32 crc kubenswrapper[4859]: I1008 18:17:32.205718 4859 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-08T18:17:29Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T18:17:29Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T18:17:32Z is after 2025-08-24T17:21:41Z" Oct 08 18:17:32 crc kubenswrapper[4859]: I1008 18:17:32.218641 4859 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-82s52" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b23a6a6c-9d92-4e7b-840e-55cfda873a2d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:17:31Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:17:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:17:31Z\\\",\\\"message\\\":\\\"containers with unready status: [machine-config-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:17:31Z\\\",\\\"message\\\":\\\"containers with unready status: [machine-config-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6vx22\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6vx22\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T18:17:31Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-82s52\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T18:17:32Z is after 2025-08-24T17:21:41Z" Oct 08 18:17:32 crc kubenswrapper[4859]: I1008 18:17:32.225079 4859 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 08 18:17:32 crc kubenswrapper[4859]: I1008 18:17:32.225120 4859 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 08 18:17:32 crc kubenswrapper[4859]: I1008 18:17:32.225156 4859 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 08 18:17:32 crc kubenswrapper[4859]: I1008 18:17:32.225194 4859 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 08 18:17:32 crc kubenswrapper[4859]: E1008 18:17:32.225303 4859 configmap.go:193] Couldn't get configMap openshift-network-console/networking-console-plugin: object "openshift-network-console"/"networking-console-plugin" not registered Oct 08 18:17:32 crc kubenswrapper[4859]: E1008 18:17:32.225370 4859 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-10-08 18:17:34.225351783 +0000 UTC m=+24.472191182 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "nginx-conf" (UniqueName: "kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin" not registered Oct 08 18:17:32 crc kubenswrapper[4859]: E1008 18:17:32.225369 4859 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Oct 08 18:17:32 crc kubenswrapper[4859]: E1008 18:17:32.225416 4859 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Oct 08 18:17:32 crc kubenswrapper[4859]: E1008 18:17:32.225359 4859 secret.go:188] Couldn't get secret openshift-network-console/networking-console-plugin-cert: object "openshift-network-console"/"networking-console-plugin-cert" not registered Oct 08 18:17:32 crc kubenswrapper[4859]: E1008 18:17:32.225370 4859 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Oct 08 18:17:32 crc kubenswrapper[4859]: E1008 18:17:32.225575 4859 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Oct 08 18:17:32 crc kubenswrapper[4859]: E1008 18:17:32.225583 4859 projected.go:194] Error preparing data for projected volume kube-api-access-s2dwl for pod openshift-network-diagnostics/network-check-source-55646444c4-trplf: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Oct 08 18:17:32 crc kubenswrapper[4859]: E1008 18:17:32.225433 4859 projected.go:194] Error preparing data for projected volume kube-api-access-cqllr for pod openshift-network-diagnostics/network-check-target-xd92c: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Oct 08 18:17:32 crc kubenswrapper[4859]: E1008 18:17:32.225622 4859 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-10-08 18:17:34.225568269 +0000 UTC m=+24.472407648 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "networking-console-plugin-cert" (UniqueName: "kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin-cert" not registered Oct 08 18:17:32 crc kubenswrapper[4859]: E1008 18:17:32.225675 4859 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl podName:9d751cbb-f2e2-430d-9754-c882a5e924a5 nodeName:}" failed. No retries permitted until 2025-10-08 18:17:34.225657601 +0000 UTC m=+24.472496980 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "kube-api-access-s2dwl" (UniqueName: "kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl") pod "network-check-source-55646444c4-trplf" (UID: "9d751cbb-f2e2-430d-9754-c882a5e924a5") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Oct 08 18:17:32 crc kubenswrapper[4859]: E1008 18:17:32.225710 4859 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr podName:3b6479f0-333b-4a96-9adf-2099afdc2447 nodeName:}" failed. No retries permitted until 2025-10-08 18:17:34.225704482 +0000 UTC m=+24.472543861 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "kube-api-access-cqllr" (UniqueName: "kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr") pod "network-check-target-xd92c" (UID: "3b6479f0-333b-4a96-9adf-2099afdc2447") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Oct 08 18:17:32 crc kubenswrapper[4859]: I1008 18:17:32.243205 4859 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-08T18:17:29Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T18:17:29Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T18:17:32Z is after 2025-08-24T17:21:41Z" Oct 08 18:17:32 crc kubenswrapper[4859]: I1008 18:17:32.258767 4859 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-84k8x" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1c24078c-c9e8-4160-883a-f483b4b3dd23\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:17:31Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:17:31Z\\\",\\\"message\\\":\\\"containers with incomplete status: [egress-router-binary-copy cni-plugins bond-cni-plugin routeoverride-cni whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:17:31Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:17:31Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hk5xb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hk5xb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hk5xb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hk5xb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hk5xb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hk5xb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hk5xb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T18:17:31Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-84k8x\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T18:17:32Z is after 2025-08-24T17:21:41Z" Oct 08 18:17:32 crc kubenswrapper[4859]: I1008 18:17:32.265824 4859 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/multus-8plkg" Oct 08 18:17:32 crc kubenswrapper[4859]: I1008 18:17:32.271888 4859 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-08T18:17:29Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T18:17:29Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T18:17:32Z is after 2025-08-24T17:21:41Z" Oct 08 18:17:32 crc kubenswrapper[4859]: I1008 18:17:32.274106 4859 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/multus-additional-cni-plugins-84k8x" Oct 08 18:17:32 crc kubenswrapper[4859]: I1008 18:17:32.281947 4859 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/machine-config-daemon-82s52" Oct 08 18:17:32 crc kubenswrapper[4859]: I1008 18:17:32.286352 4859 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-btw28" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"97bc5598-3ead-4300-a938-3aebbff1e8c6\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:17:31Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:17:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:17:31Z\\\",\\\"message\\\":\\\"containers with unready status: [dns-node-resolver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:17:31Z\\\",\\\"message\\\":\\\"containers with unready status: [dns-node-resolver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qcxsd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T18:17:31Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-btw28\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T18:17:32Z is after 2025-08-24T17:21:41Z" Oct 08 18:17:32 crc kubenswrapper[4859]: I1008 18:17:32.300734 4859 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-8plkg" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"92b2d7de-31cc-4a91-95a5-ed5ea964b028\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:17:31Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:17:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:17:31Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:17:31Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q9d7m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T18:17:31Z\\\"}}\" for pod \"openshift-multus\"/\"multus-8plkg\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T18:17:32Z is after 2025-08-24T17:21:41Z" Oct 08 18:17:32 crc kubenswrapper[4859]: W1008 18:17:32.301314 4859 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod1c24078c_c9e8_4160_883a_f483b4b3dd23.slice/crio-d9647b0aa2d0467493c8b3dc89c6c576f96b2a5414b70ec0d146dded92ac6f38 WatchSource:0}: Error finding container d9647b0aa2d0467493c8b3dc89c6c576f96b2a5414b70ec0d146dded92ac6f38: Status 404 returned error can't find the container with id d9647b0aa2d0467493c8b3dc89c6c576f96b2a5414b70ec0d146dded92ac6f38 Oct 08 18:17:32 crc kubenswrapper[4859]: W1008 18:17:32.305928 4859 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podb23a6a6c_9d92_4e7b_840e_55cfda873a2d.slice/crio-8e1d5ad46cfd2b72c615daedc67761cd0282b8bb02e2be8046d9ec3f2bc89419 WatchSource:0}: Error finding container 8e1d5ad46cfd2b72c615daedc67761cd0282b8bb02e2be8046d9ec3f2bc89419: Status 404 returned error can't find the container with id 8e1d5ad46cfd2b72c615daedc67761cd0282b8bb02e2be8046d9ec3f2bc89419 Oct 08 18:17:32 crc kubenswrapper[4859]: I1008 18:17:32.329453 4859 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-ovn-kubernetes/ovnkube-node-4g8gf"] Oct 08 18:17:32 crc kubenswrapper[4859]: I1008 18:17:32.330466 4859 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ovn-kubernetes/ovnkube-node-4g8gf" Oct 08 18:17:32 crc kubenswrapper[4859]: I1008 18:17:32.332290 4859 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ovn-kubernetes"/"ovn-kubernetes-node-dockercfg-pwtwl" Oct 08 18:17:32 crc kubenswrapper[4859]: I1008 18:17:32.332540 4859 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ovn-kubernetes"/"ovn-node-metrics-cert" Oct 08 18:17:32 crc kubenswrapper[4859]: I1008 18:17:32.332494 4859 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ovn-kubernetes"/"env-overrides" Oct 08 18:17:32 crc kubenswrapper[4859]: I1008 18:17:32.332680 4859 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ovn-kubernetes"/"kube-root-ca.crt" Oct 08 18:17:32 crc kubenswrapper[4859]: I1008 18:17:32.332802 4859 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ovn-kubernetes"/"ovnkube-config" Oct 08 18:17:32 crc kubenswrapper[4859]: I1008 18:17:32.332806 4859 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ovn-kubernetes"/"openshift-service-ca.crt" Oct 08 18:17:32 crc kubenswrapper[4859]: I1008 18:17:32.346441 4859 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ovn-kubernetes"/"ovnkube-script-lib" Oct 08 18:17:32 crc kubenswrapper[4859]: I1008 18:17:32.351113 4859 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-8plkg" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"92b2d7de-31cc-4a91-95a5-ed5ea964b028\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:17:31Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:17:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:17:31Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:17:31Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q9d7m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T18:17:31Z\\\"}}\" for pod \"openshift-multus\"/\"multus-8plkg\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T18:17:32Z is after 2025-08-24T17:21:41Z" Oct 08 18:17:32 crc kubenswrapper[4859]: I1008 18:17:32.370547 4859 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-08T18:17:29Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T18:17:29Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T18:17:32Z is after 2025-08-24T17:21:41Z" Oct 08 18:17:32 crc kubenswrapper[4859]: I1008 18:17:32.382504 4859 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-btw28" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"97bc5598-3ead-4300-a938-3aebbff1e8c6\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:17:31Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:17:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:17:31Z\\\",\\\"message\\\":\\\"containers with unready status: [dns-node-resolver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:17:31Z\\\",\\\"message\\\":\\\"containers with unready status: [dns-node-resolver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qcxsd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T18:17:31Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-btw28\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T18:17:32Z is after 2025-08-24T17:21:41Z" Oct 08 18:17:32 crc kubenswrapper[4859]: I1008 18:17:32.398323 4859 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-08T18:17:31Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T18:17:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T18:17:31Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c6d15905694b6bbd4999d49dc8f7a53eb6bfaa7adb59efb0d7e4973c21d1c7be\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T18:17:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T18:17:32Z is after 2025-08-24T17:21:41Z" Oct 08 18:17:32 crc kubenswrapper[4859]: I1008 18:17:32.412177 4859 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-08T18:17:29Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T18:17:29Z\\\",\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T18:17:32Z is after 2025-08-24T17:21:41Z" Oct 08 18:17:32 crc kubenswrapper[4859]: I1008 18:17:32.425452 4859 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-08T18:17:31Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T18:17:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T18:17:31Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://99e2d51bd7757498dae60949eea6d732b5f7b481fa00f808dd5308a530a1daf1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T18:17:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e32aaba93699944665660ae41e9b5779178c1354c31e15ad8188bf8c3c8fa500\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T18:17:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T18:17:32Z is after 2025-08-24T17:21:41Z" Oct 08 18:17:32 crc kubenswrapper[4859]: I1008 18:17:32.426941 4859 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/1dff864e-b75d-4e0b-b182-75f710eac8df-host-slash\") pod \"ovnkube-node-4g8gf\" (UID: \"1dff864e-b75d-4e0b-b182-75f710eac8df\") " pod="openshift-ovn-kubernetes/ovnkube-node-4g8gf" Oct 08 18:17:32 crc kubenswrapper[4859]: I1008 18:17:32.427029 4859 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-cni-netd\" (UniqueName: \"kubernetes.io/host-path/1dff864e-b75d-4e0b-b182-75f710eac8df-host-cni-netd\") pod \"ovnkube-node-4g8gf\" (UID: \"1dff864e-b75d-4e0b-b182-75f710eac8df\") " pod="openshift-ovn-kubernetes/ovnkube-node-4g8gf" Oct 08 18:17:32 crc kubenswrapper[4859]: I1008 18:17:32.427057 4859 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-var-lib-cni-networks-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/1dff864e-b75d-4e0b-b182-75f710eac8df-host-var-lib-cni-networks-ovn-kubernetes\") pod \"ovnkube-node-4g8gf\" (UID: \"1dff864e-b75d-4e0b-b182-75f710eac8df\") " pod="openshift-ovn-kubernetes/ovnkube-node-4g8gf" Oct 08 18:17:32 crc kubenswrapper[4859]: I1008 18:17:32.427132 4859 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/1dff864e-b75d-4e0b-b182-75f710eac8df-ovnkube-config\") pod \"ovnkube-node-4g8gf\" (UID: \"1dff864e-b75d-4e0b-b182-75f710eac8df\") " pod="openshift-ovn-kubernetes/ovnkube-node-4g8gf" Oct 08 18:17:32 crc kubenswrapper[4859]: I1008 18:17:32.427153 4859 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/1dff864e-b75d-4e0b-b182-75f710eac8df-ovnkube-script-lib\") pod \"ovnkube-node-4g8gf\" (UID: \"1dff864e-b75d-4e0b-b182-75f710eac8df\") " pod="openshift-ovn-kubernetes/ovnkube-node-4g8gf" Oct 08 18:17:32 crc kubenswrapper[4859]: I1008 18:17:32.427221 4859 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-systemd\" (UniqueName: \"kubernetes.io/host-path/1dff864e-b75d-4e0b-b182-75f710eac8df-run-systemd\") pod \"ovnkube-node-4g8gf\" (UID: \"1dff864e-b75d-4e0b-b182-75f710eac8df\") " pod="openshift-ovn-kubernetes/ovnkube-node-4g8gf" Oct 08 18:17:32 crc kubenswrapper[4859]: I1008 18:17:32.427253 4859 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-kubelet\" (UniqueName: \"kubernetes.io/host-path/1dff864e-b75d-4e0b-b182-75f710eac8df-host-kubelet\") pod \"ovnkube-node-4g8gf\" (UID: \"1dff864e-b75d-4e0b-b182-75f710eac8df\") " pod="openshift-ovn-kubernetes/ovnkube-node-4g8gf" Oct 08 18:17:32 crc kubenswrapper[4859]: I1008 18:17:32.427269 4859 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"systemd-units\" (UniqueName: \"kubernetes.io/host-path/1dff864e-b75d-4e0b-b182-75f710eac8df-systemd-units\") pod \"ovnkube-node-4g8gf\" (UID: \"1dff864e-b75d-4e0b-b182-75f710eac8df\") " pod="openshift-ovn-kubernetes/ovnkube-node-4g8gf" Oct 08 18:17:32 crc kubenswrapper[4859]: I1008 18:17:32.427291 4859 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-openvswitch\" (UniqueName: \"kubernetes.io/host-path/1dff864e-b75d-4e0b-b182-75f710eac8df-etc-openvswitch\") pod \"ovnkube-node-4g8gf\" (UID: \"1dff864e-b75d-4e0b-b182-75f710eac8df\") " pod="openshift-ovn-kubernetes/ovnkube-node-4g8gf" Oct 08 18:17:32 crc kubenswrapper[4859]: I1008 18:17:32.427311 4859 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-openvswitch\" (UniqueName: \"kubernetes.io/host-path/1dff864e-b75d-4e0b-b182-75f710eac8df-run-openvswitch\") pod \"ovnkube-node-4g8gf\" (UID: \"1dff864e-b75d-4e0b-b182-75f710eac8df\") " pod="openshift-ovn-kubernetes/ovnkube-node-4g8gf" Oct 08 18:17:32 crc kubenswrapper[4859]: I1008 18:17:32.427353 4859 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"node-log\" (UniqueName: \"kubernetes.io/host-path/1dff864e-b75d-4e0b-b182-75f710eac8df-node-log\") pod \"ovnkube-node-4g8gf\" (UID: \"1dff864e-b75d-4e0b-b182-75f710eac8df\") " pod="openshift-ovn-kubernetes/ovnkube-node-4g8gf" Oct 08 18:17:32 crc kubenswrapper[4859]: I1008 18:17:32.427511 4859 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-s8jnn\" (UniqueName: \"kubernetes.io/projected/1dff864e-b75d-4e0b-b182-75f710eac8df-kube-api-access-s8jnn\") pod \"ovnkube-node-4g8gf\" (UID: \"1dff864e-b75d-4e0b-b182-75f710eac8df\") " pod="openshift-ovn-kubernetes/ovnkube-node-4g8gf" Oct 08 18:17:32 crc kubenswrapper[4859]: I1008 18:17:32.427590 4859 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-socket\" (UniqueName: \"kubernetes.io/host-path/1dff864e-b75d-4e0b-b182-75f710eac8df-log-socket\") pod \"ovnkube-node-4g8gf\" (UID: \"1dff864e-b75d-4e0b-b182-75f710eac8df\") " pod="openshift-ovn-kubernetes/ovnkube-node-4g8gf" Oct 08 18:17:32 crc kubenswrapper[4859]: I1008 18:17:32.427634 4859 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-run-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/1dff864e-b75d-4e0b-b182-75f710eac8df-host-run-ovn-kubernetes\") pod \"ovnkube-node-4g8gf\" (UID: \"1dff864e-b75d-4e0b-b182-75f710eac8df\") " pod="openshift-ovn-kubernetes/ovnkube-node-4g8gf" Oct 08 18:17:32 crc kubenswrapper[4859]: I1008 18:17:32.427667 4859 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/1dff864e-b75d-4e0b-b182-75f710eac8df-host-run-netns\") pod \"ovnkube-node-4g8gf\" (UID: \"1dff864e-b75d-4e0b-b182-75f710eac8df\") " pod="openshift-ovn-kubernetes/ovnkube-node-4g8gf" Oct 08 18:17:32 crc kubenswrapper[4859]: I1008 18:17:32.427726 4859 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-lib-openvswitch\" (UniqueName: \"kubernetes.io/host-path/1dff864e-b75d-4e0b-b182-75f710eac8df-var-lib-openvswitch\") pod \"ovnkube-node-4g8gf\" (UID: \"1dff864e-b75d-4e0b-b182-75f710eac8df\") " pod="openshift-ovn-kubernetes/ovnkube-node-4g8gf" Oct 08 18:17:32 crc kubenswrapper[4859]: I1008 18:17:32.427750 4859 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-cni-bin\" (UniqueName: \"kubernetes.io/host-path/1dff864e-b75d-4e0b-b182-75f710eac8df-host-cni-bin\") pod \"ovnkube-node-4g8gf\" (UID: \"1dff864e-b75d-4e0b-b182-75f710eac8df\") " pod="openshift-ovn-kubernetes/ovnkube-node-4g8gf" Oct 08 18:17:32 crc kubenswrapper[4859]: I1008 18:17:32.427914 4859 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-ovn\" (UniqueName: \"kubernetes.io/host-path/1dff864e-b75d-4e0b-b182-75f710eac8df-run-ovn\") pod \"ovnkube-node-4g8gf\" (UID: \"1dff864e-b75d-4e0b-b182-75f710eac8df\") " pod="openshift-ovn-kubernetes/ovnkube-node-4g8gf" Oct 08 18:17:32 crc kubenswrapper[4859]: I1008 18:17:32.427969 4859 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/1dff864e-b75d-4e0b-b182-75f710eac8df-env-overrides\") pod \"ovnkube-node-4g8gf\" (UID: \"1dff864e-b75d-4e0b-b182-75f710eac8df\") " pod="openshift-ovn-kubernetes/ovnkube-node-4g8gf" Oct 08 18:17:32 crc kubenswrapper[4859]: I1008 18:17:32.428001 4859 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/1dff864e-b75d-4e0b-b182-75f710eac8df-ovn-node-metrics-cert\") pod \"ovnkube-node-4g8gf\" (UID: \"1dff864e-b75d-4e0b-b182-75f710eac8df\") " pod="openshift-ovn-kubernetes/ovnkube-node-4g8gf" Oct 08 18:17:32 crc kubenswrapper[4859]: I1008 18:17:32.440394 4859 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-08T18:17:29Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T18:17:29Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T18:17:32Z is after 2025-08-24T17:21:41Z" Oct 08 18:17:32 crc kubenswrapper[4859]: I1008 18:17:32.459929 4859 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-82s52" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b23a6a6c-9d92-4e7b-840e-55cfda873a2d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:17:31Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:17:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:17:31Z\\\",\\\"message\\\":\\\"containers with unready status: [machine-config-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:17:31Z\\\",\\\"message\\\":\\\"containers with unready status: [machine-config-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6vx22\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6vx22\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T18:17:31Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-82s52\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T18:17:32Z is after 2025-08-24T17:21:41Z" Oct 08 18:17:32 crc kubenswrapper[4859]: I1008 18:17:32.480829 4859 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-4g8gf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1dff864e-b75d-4e0b-b182-75f710eac8df\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:17:32Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:17:32Z\\\",\\\"message\\\":\\\"containers with incomplete status: [kubecfg-setup]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:17:32Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:17:32Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s8jnn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s8jnn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s8jnn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s8jnn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s8jnn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s8jnn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s8jnn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s8jnn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s8jnn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T18:17:32Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-4g8gf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T18:17:32Z is after 2025-08-24T17:21:41Z" Oct 08 18:17:32 crc kubenswrapper[4859]: I1008 18:17:32.480961 4859 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="01ab3dd5-8196-46d0-ad33-122e2ca51def" path="/var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/volumes" Oct 08 18:17:32 crc kubenswrapper[4859]: I1008 18:17:32.481646 4859 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="09ae3b1a-e8e7-4524-b54b-61eab6f9239a" path="/var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes" Oct 08 18:17:32 crc kubenswrapper[4859]: I1008 18:17:32.482635 4859 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="09efc573-dbb6-4249-bd59-9b87aba8dd28" path="/var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes" Oct 08 18:17:32 crc kubenswrapper[4859]: I1008 18:17:32.483336 4859 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="0b574797-001e-440a-8f4e-c0be86edad0f" path="/var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/volumes" Oct 08 18:17:32 crc kubenswrapper[4859]: I1008 18:17:32.488194 4859 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="0b78653f-4ff9-4508-8672-245ed9b561e3" path="/var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/volumes" Oct 08 18:17:32 crc kubenswrapper[4859]: I1008 18:17:32.488890 4859 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="1386a44e-36a2-460c-96d0-0359d2b6f0f5" path="/var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/volumes" Oct 08 18:17:32 crc kubenswrapper[4859]: I1008 18:17:32.490135 4859 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="1bf7eb37-55a3-4c65-b768-a94c82151e69" path="/var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes" Oct 08 18:17:32 crc kubenswrapper[4859]: I1008 18:17:32.490868 4859 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="1d611f23-29be-4491-8495-bee1670e935f" path="/var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes" Oct 08 18:17:32 crc kubenswrapper[4859]: I1008 18:17:32.493715 4859 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="20b0d48f-5fd6-431c-a545-e3c800c7b866" path="/var/lib/kubelet/pods/20b0d48f-5fd6-431c-a545-e3c800c7b866/volumes" Oct 08 18:17:32 crc kubenswrapper[4859]: I1008 18:17:32.494756 4859 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c" path="/var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/volumes" Oct 08 18:17:32 crc kubenswrapper[4859]: I1008 18:17:32.496268 4859 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="22c825df-677d-4ca6-82db-3454ed06e783" path="/var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/volumes" Oct 08 18:17:32 crc kubenswrapper[4859]: I1008 18:17:32.497298 4859 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="25e176fe-21b4-4974-b1ed-c8b94f112a7f" path="/var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/volumes" Oct 08 18:17:32 crc kubenswrapper[4859]: I1008 18:17:32.497894 4859 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="308be0ea-9f5f-4b29-aeb1-5abd31a0b17b" path="/var/lib/kubelet/pods/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b/volumes" Oct 08 18:17:32 crc kubenswrapper[4859]: I1008 18:17:32.499167 4859 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="31d8b7a1-420e-4252-a5b7-eebe8a111292" path="/var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes" Oct 08 18:17:32 crc kubenswrapper[4859]: I1008 18:17:32.499862 4859 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="3ab1a177-2de0-46d9-b765-d0d0649bb42e" path="/var/lib/kubelet/pods/3ab1a177-2de0-46d9-b765-d0d0649bb42e/volumes" Oct 08 18:17:32 crc kubenswrapper[4859]: I1008 18:17:32.500869 4859 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="3cb93b32-e0ae-4377-b9c8-fdb9842c6d59" path="/var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes" Oct 08 18:17:32 crc kubenswrapper[4859]: I1008 18:17:32.501473 4859 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="43509403-f426-496e-be36-56cef71462f5" path="/var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes" Oct 08 18:17:32 crc kubenswrapper[4859]: I1008 18:17:32.501916 4859 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="44663579-783b-4372-86d6-acf235a62d72" path="/var/lib/kubelet/pods/44663579-783b-4372-86d6-acf235a62d72/volumes" Oct 08 18:17:32 crc kubenswrapper[4859]: I1008 18:17:32.505625 4859 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="496e6271-fb68-4057-954e-a0d97a4afa3f" path="/var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/volumes" Oct 08 18:17:32 crc kubenswrapper[4859]: I1008 18:17:32.507560 4859 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" path="/var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes" Oct 08 18:17:32 crc kubenswrapper[4859]: I1008 18:17:32.508254 4859 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="49ef4625-1d3a-4a9f-b595-c2433d32326d" path="/var/lib/kubelet/pods/49ef4625-1d3a-4a9f-b595-c2433d32326d/volumes" Oct 08 18:17:32 crc kubenswrapper[4859]: I1008 18:17:32.508910 4859 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"63a623de-5608-4228-b1c2-77188c276986\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:17:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:17:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:17:10Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:17:10Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:17:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d4786a8fb8abd96fb22e60733c974f599017407d589ed815fb8541fd3b64f696\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T18:17:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://183d481502ed5108d5e19e483b714c289f355ea3f8b0891883e5335b802c83b6\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T18:17:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ec848c46fd5dde0469b00d3cab14ba4a652533a265490fadbee7a8b8843ef249\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T18:17:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4fe1828b4d63ff8106bfcd15e1dc1c1bd1ba7f77777cdb23ff1b291d0418774c\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2e700b2a20c81b414617f5b5c018a738278a3fa44e6c6d14b2170051d0725431\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-08T18:17:24Z\\\",\\\"message\\\":\\\"W1008 18:17:13.670625 1 cmd.go:257] Using insecure, self-signed certificates\\\\nI1008 18:17:13.671202 1 crypto.go:601] Generating new CA for check-endpoints-signer@1759947433 cert, and key in /tmp/serving-cert-3867218267/serving-signer.crt, /tmp/serving-cert-3867218267/serving-signer.key\\\\nI1008 18:17:13.944463 1 observer_polling.go:159] Starting file observer\\\\nW1008 18:17:13.951958 1 builder.go:272] unable to get owner reference (falling back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": dial tcp [::1]:6443: connect: connection refused\\\\nI1008 18:17:13.952088 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1008 18:17:13.956755 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3867218267/tls.crt::/tmp/serving-cert-3867218267/tls.key\\\\\\\"\\\\nF1008 18:17:24.290991 1 cmd.go:182] error initializing delegating authentication: unable to load configmap based request-header-client-ca-file: Get \\\\\\\"https://localhost:6443/api/v1/namespaces/kube-system/configmaps/extension-apiserver-authentication\\\\\\\": net/http: TLS handshake timeout\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-08T18:17:13Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T18:17:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c65b66f62d325ca10ef8d5d31f04529ccf4e691cf3303d231e7a414fb400e4d6\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T18:17:13Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://58c7bba4ace28215b062653a960e561b1dc1f890f2baf450145638a065af038b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://58c7bba4ace28215b062653a960e561b1dc1f890f2baf450145638a065af038b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-08T18:17:11Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-08T18:17:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T18:17:10Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T18:17:32Z is after 2025-08-24T17:21:41Z" Oct 08 18:17:32 crc kubenswrapper[4859]: I1008 18:17:32.509157 4859 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="4bb40260-dbaa-4fb0-84df-5e680505d512" path="/var/lib/kubelet/pods/4bb40260-dbaa-4fb0-84df-5e680505d512/volumes" Oct 08 18:17:32 crc kubenswrapper[4859]: I1008 18:17:32.510089 4859 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="5225d0e4-402f-4861-b410-819f433b1803" path="/var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes" Oct 08 18:17:32 crc kubenswrapper[4859]: I1008 18:17:32.510765 4859 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="5441d097-087c-4d9a-baa8-b210afa90fc9" path="/var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes" Oct 08 18:17:32 crc kubenswrapper[4859]: I1008 18:17:32.511825 4859 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="57a731c4-ef35-47a8-b875-bfb08a7f8011" path="/var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes" Oct 08 18:17:32 crc kubenswrapper[4859]: I1008 18:17:32.512457 4859 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="5b88f790-22fa-440e-b583-365168c0b23d" path="/var/lib/kubelet/pods/5b88f790-22fa-440e-b583-365168c0b23d/volumes" Oct 08 18:17:32 crc kubenswrapper[4859]: I1008 18:17:32.513555 4859 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="5fe579f8-e8a6-4643-bce5-a661393c4dde" path="/var/lib/kubelet/pods/5fe579f8-e8a6-4643-bce5-a661393c4dde/volumes" Oct 08 18:17:32 crc kubenswrapper[4859]: I1008 18:17:32.514115 4859 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="6402fda4-df10-493c-b4e5-d0569419652d" path="/var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes" Oct 08 18:17:32 crc kubenswrapper[4859]: I1008 18:17:32.515253 4859 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="6509e943-70c6-444c-bc41-48a544e36fbd" path="/var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes" Oct 08 18:17:32 crc kubenswrapper[4859]: I1008 18:17:32.516275 4859 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="6731426b-95fe-49ff-bb5f-40441049fde2" path="/var/lib/kubelet/pods/6731426b-95fe-49ff-bb5f-40441049fde2/volumes" Oct 08 18:17:32 crc kubenswrapper[4859]: I1008 18:17:32.516976 4859 kubelet_volumes.go:152] "Cleaned up orphaned volume subpath from pod" podUID="6ea678ab-3438-413e-bfe3-290ae7725660" path="/var/lib/kubelet/pods/6ea678ab-3438-413e-bfe3-290ae7725660/volume-subpaths/run-systemd/ovnkube-controller/6" Oct 08 18:17:32 crc kubenswrapper[4859]: I1008 18:17:32.517527 4859 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="6ea678ab-3438-413e-bfe3-290ae7725660" path="/var/lib/kubelet/pods/6ea678ab-3438-413e-bfe3-290ae7725660/volumes" Oct 08 18:17:32 crc kubenswrapper[4859]: I1008 18:17:32.519935 4859 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="7539238d-5fe0-46ed-884e-1c3b566537ec" path="/var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/volumes" Oct 08 18:17:32 crc kubenswrapper[4859]: I1008 18:17:32.520747 4859 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="7583ce53-e0fe-4a16-9e4d-50516596a136" path="/var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes" Oct 08 18:17:32 crc kubenswrapper[4859]: I1008 18:17:32.522133 4859 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="7bb08738-c794-4ee8-9972-3a62ca171029" path="/var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes" Oct 08 18:17:32 crc kubenswrapper[4859]: I1008 18:17:32.524042 4859 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="87cf06ed-a83f-41a7-828d-70653580a8cb" path="/var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/volumes" Oct 08 18:17:32 crc kubenswrapper[4859]: I1008 18:17:32.525103 4859 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-08T18:17:29Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T18:17:29Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T18:17:32Z is after 2025-08-24T17:21:41Z" Oct 08 18:17:32 crc kubenswrapper[4859]: I1008 18:17:32.528279 4859 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="8cea82b4-6893-4ddc-af9f-1bb5ae425c5b" path="/var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/volumes" Oct 08 18:17:32 crc kubenswrapper[4859]: I1008 18:17:32.529204 4859 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-systemd\" (UniqueName: \"kubernetes.io/host-path/1dff864e-b75d-4e0b-b182-75f710eac8df-run-systemd\") pod \"ovnkube-node-4g8gf\" (UID: \"1dff864e-b75d-4e0b-b182-75f710eac8df\") " pod="openshift-ovn-kubernetes/ovnkube-node-4g8gf" Oct 08 18:17:32 crc kubenswrapper[4859]: I1008 18:17:32.529250 4859 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-kubelet\" (UniqueName: \"kubernetes.io/host-path/1dff864e-b75d-4e0b-b182-75f710eac8df-host-kubelet\") pod \"ovnkube-node-4g8gf\" (UID: \"1dff864e-b75d-4e0b-b182-75f710eac8df\") " pod="openshift-ovn-kubernetes/ovnkube-node-4g8gf" Oct 08 18:17:32 crc kubenswrapper[4859]: I1008 18:17:32.529272 4859 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"systemd-units\" (UniqueName: \"kubernetes.io/host-path/1dff864e-b75d-4e0b-b182-75f710eac8df-systemd-units\") pod \"ovnkube-node-4g8gf\" (UID: \"1dff864e-b75d-4e0b-b182-75f710eac8df\") " pod="openshift-ovn-kubernetes/ovnkube-node-4g8gf" Oct 08 18:17:32 crc kubenswrapper[4859]: I1008 18:17:32.529297 4859 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-openvswitch\" (UniqueName: \"kubernetes.io/host-path/1dff864e-b75d-4e0b-b182-75f710eac8df-etc-openvswitch\") pod \"ovnkube-node-4g8gf\" (UID: \"1dff864e-b75d-4e0b-b182-75f710eac8df\") " pod="openshift-ovn-kubernetes/ovnkube-node-4g8gf" Oct 08 18:17:32 crc kubenswrapper[4859]: I1008 18:17:32.529316 4859 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-openvswitch\" (UniqueName: \"kubernetes.io/host-path/1dff864e-b75d-4e0b-b182-75f710eac8df-run-openvswitch\") pod \"ovnkube-node-4g8gf\" (UID: \"1dff864e-b75d-4e0b-b182-75f710eac8df\") " pod="openshift-ovn-kubernetes/ovnkube-node-4g8gf" Oct 08 18:17:32 crc kubenswrapper[4859]: I1008 18:17:32.529335 4859 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"node-log\" (UniqueName: \"kubernetes.io/host-path/1dff864e-b75d-4e0b-b182-75f710eac8df-node-log\") pod \"ovnkube-node-4g8gf\" (UID: \"1dff864e-b75d-4e0b-b182-75f710eac8df\") " pod="openshift-ovn-kubernetes/ovnkube-node-4g8gf" Oct 08 18:17:32 crc kubenswrapper[4859]: I1008 18:17:32.529354 4859 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-socket\" (UniqueName: \"kubernetes.io/host-path/1dff864e-b75d-4e0b-b182-75f710eac8df-log-socket\") pod \"ovnkube-node-4g8gf\" (UID: \"1dff864e-b75d-4e0b-b182-75f710eac8df\") " pod="openshift-ovn-kubernetes/ovnkube-node-4g8gf" Oct 08 18:17:32 crc kubenswrapper[4859]: I1008 18:17:32.529371 4859 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s8jnn\" (UniqueName: \"kubernetes.io/projected/1dff864e-b75d-4e0b-b182-75f710eac8df-kube-api-access-s8jnn\") pod \"ovnkube-node-4g8gf\" (UID: \"1dff864e-b75d-4e0b-b182-75f710eac8df\") " pod="openshift-ovn-kubernetes/ovnkube-node-4g8gf" Oct 08 18:17:32 crc kubenswrapper[4859]: I1008 18:17:32.529385 4859 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/1dff864e-b75d-4e0b-b182-75f710eac8df-host-run-netns\") pod \"ovnkube-node-4g8gf\" (UID: \"1dff864e-b75d-4e0b-b182-75f710eac8df\") " pod="openshift-ovn-kubernetes/ovnkube-node-4g8gf" Oct 08 18:17:32 crc kubenswrapper[4859]: I1008 18:17:32.529400 4859 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-lib-openvswitch\" (UniqueName: \"kubernetes.io/host-path/1dff864e-b75d-4e0b-b182-75f710eac8df-var-lib-openvswitch\") pod \"ovnkube-node-4g8gf\" (UID: \"1dff864e-b75d-4e0b-b182-75f710eac8df\") " pod="openshift-ovn-kubernetes/ovnkube-node-4g8gf" Oct 08 18:17:32 crc kubenswrapper[4859]: I1008 18:17:32.529415 4859 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-run-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/1dff864e-b75d-4e0b-b182-75f710eac8df-host-run-ovn-kubernetes\") pod \"ovnkube-node-4g8gf\" (UID: \"1dff864e-b75d-4e0b-b182-75f710eac8df\") " pod="openshift-ovn-kubernetes/ovnkube-node-4g8gf" Oct 08 18:17:32 crc kubenswrapper[4859]: I1008 18:17:32.529442 4859 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-cni-bin\" (UniqueName: \"kubernetes.io/host-path/1dff864e-b75d-4e0b-b182-75f710eac8df-host-cni-bin\") pod \"ovnkube-node-4g8gf\" (UID: \"1dff864e-b75d-4e0b-b182-75f710eac8df\") " pod="openshift-ovn-kubernetes/ovnkube-node-4g8gf" Oct 08 18:17:32 crc kubenswrapper[4859]: I1008 18:17:32.529461 4859 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-ovn\" (UniqueName: \"kubernetes.io/host-path/1dff864e-b75d-4e0b-b182-75f710eac8df-run-ovn\") pod \"ovnkube-node-4g8gf\" (UID: \"1dff864e-b75d-4e0b-b182-75f710eac8df\") " pod="openshift-ovn-kubernetes/ovnkube-node-4g8gf" Oct 08 18:17:32 crc kubenswrapper[4859]: I1008 18:17:32.529477 4859 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/1dff864e-b75d-4e0b-b182-75f710eac8df-env-overrides\") pod \"ovnkube-node-4g8gf\" (UID: \"1dff864e-b75d-4e0b-b182-75f710eac8df\") " pod="openshift-ovn-kubernetes/ovnkube-node-4g8gf" Oct 08 18:17:32 crc kubenswrapper[4859]: I1008 18:17:32.529497 4859 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/1dff864e-b75d-4e0b-b182-75f710eac8df-ovn-node-metrics-cert\") pod \"ovnkube-node-4g8gf\" (UID: \"1dff864e-b75d-4e0b-b182-75f710eac8df\") " pod="openshift-ovn-kubernetes/ovnkube-node-4g8gf" Oct 08 18:17:32 crc kubenswrapper[4859]: I1008 18:17:32.529514 4859 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/1dff864e-b75d-4e0b-b182-75f710eac8df-host-slash\") pod \"ovnkube-node-4g8gf\" (UID: \"1dff864e-b75d-4e0b-b182-75f710eac8df\") " pod="openshift-ovn-kubernetes/ovnkube-node-4g8gf" Oct 08 18:17:32 crc kubenswrapper[4859]: I1008 18:17:32.529529 4859 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/1dff864e-b75d-4e0b-b182-75f710eac8df-ovnkube-config\") pod \"ovnkube-node-4g8gf\" (UID: \"1dff864e-b75d-4e0b-b182-75f710eac8df\") " pod="openshift-ovn-kubernetes/ovnkube-node-4g8gf" Oct 08 18:17:32 crc kubenswrapper[4859]: I1008 18:17:32.529545 4859 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/1dff864e-b75d-4e0b-b182-75f710eac8df-ovnkube-script-lib\") pod \"ovnkube-node-4g8gf\" (UID: \"1dff864e-b75d-4e0b-b182-75f710eac8df\") " pod="openshift-ovn-kubernetes/ovnkube-node-4g8gf" Oct 08 18:17:32 crc kubenswrapper[4859]: I1008 18:17:32.529562 4859 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-cni-netd\" (UniqueName: \"kubernetes.io/host-path/1dff864e-b75d-4e0b-b182-75f710eac8df-host-cni-netd\") pod \"ovnkube-node-4g8gf\" (UID: \"1dff864e-b75d-4e0b-b182-75f710eac8df\") " pod="openshift-ovn-kubernetes/ovnkube-node-4g8gf" Oct 08 18:17:32 crc kubenswrapper[4859]: I1008 18:17:32.529582 4859 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-var-lib-cni-networks-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/1dff864e-b75d-4e0b-b182-75f710eac8df-host-var-lib-cni-networks-ovn-kubernetes\") pod \"ovnkube-node-4g8gf\" (UID: \"1dff864e-b75d-4e0b-b182-75f710eac8df\") " pod="openshift-ovn-kubernetes/ovnkube-node-4g8gf" Oct 08 18:17:32 crc kubenswrapper[4859]: I1008 18:17:32.529649 4859 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-var-lib-cni-networks-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/1dff864e-b75d-4e0b-b182-75f710eac8df-host-var-lib-cni-networks-ovn-kubernetes\") pod \"ovnkube-node-4g8gf\" (UID: \"1dff864e-b75d-4e0b-b182-75f710eac8df\") " pod="openshift-ovn-kubernetes/ovnkube-node-4g8gf" Oct 08 18:17:32 crc kubenswrapper[4859]: I1008 18:17:32.529703 4859 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-systemd\" (UniqueName: \"kubernetes.io/host-path/1dff864e-b75d-4e0b-b182-75f710eac8df-run-systemd\") pod \"ovnkube-node-4g8gf\" (UID: \"1dff864e-b75d-4e0b-b182-75f710eac8df\") " pod="openshift-ovn-kubernetes/ovnkube-node-4g8gf" Oct 08 18:17:32 crc kubenswrapper[4859]: I1008 18:17:32.529726 4859 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-kubelet\" (UniqueName: \"kubernetes.io/host-path/1dff864e-b75d-4e0b-b182-75f710eac8df-host-kubelet\") pod \"ovnkube-node-4g8gf\" (UID: \"1dff864e-b75d-4e0b-b182-75f710eac8df\") " pod="openshift-ovn-kubernetes/ovnkube-node-4g8gf" Oct 08 18:17:32 crc kubenswrapper[4859]: I1008 18:17:32.529749 4859 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"systemd-units\" (UniqueName: \"kubernetes.io/host-path/1dff864e-b75d-4e0b-b182-75f710eac8df-systemd-units\") pod \"ovnkube-node-4g8gf\" (UID: \"1dff864e-b75d-4e0b-b182-75f710eac8df\") " pod="openshift-ovn-kubernetes/ovnkube-node-4g8gf" Oct 08 18:17:32 crc kubenswrapper[4859]: I1008 18:17:32.529772 4859 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-openvswitch\" (UniqueName: \"kubernetes.io/host-path/1dff864e-b75d-4e0b-b182-75f710eac8df-etc-openvswitch\") pod \"ovnkube-node-4g8gf\" (UID: \"1dff864e-b75d-4e0b-b182-75f710eac8df\") " pod="openshift-ovn-kubernetes/ovnkube-node-4g8gf" Oct 08 18:17:32 crc kubenswrapper[4859]: I1008 18:17:32.529794 4859 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-openvswitch\" (UniqueName: \"kubernetes.io/host-path/1dff864e-b75d-4e0b-b182-75f710eac8df-run-openvswitch\") pod \"ovnkube-node-4g8gf\" (UID: \"1dff864e-b75d-4e0b-b182-75f710eac8df\") " pod="openshift-ovn-kubernetes/ovnkube-node-4g8gf" Oct 08 18:17:32 crc kubenswrapper[4859]: I1008 18:17:32.529818 4859 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"node-log\" (UniqueName: \"kubernetes.io/host-path/1dff864e-b75d-4e0b-b182-75f710eac8df-node-log\") pod \"ovnkube-node-4g8gf\" (UID: \"1dff864e-b75d-4e0b-b182-75f710eac8df\") " pod="openshift-ovn-kubernetes/ovnkube-node-4g8gf" Oct 08 18:17:32 crc kubenswrapper[4859]: I1008 18:17:32.529857 4859 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-socket\" (UniqueName: \"kubernetes.io/host-path/1dff864e-b75d-4e0b-b182-75f710eac8df-log-socket\") pod \"ovnkube-node-4g8gf\" (UID: \"1dff864e-b75d-4e0b-b182-75f710eac8df\") " pod="openshift-ovn-kubernetes/ovnkube-node-4g8gf" Oct 08 18:17:32 crc kubenswrapper[4859]: I1008 18:17:32.530049 4859 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/1dff864e-b75d-4e0b-b182-75f710eac8df-host-run-netns\") pod \"ovnkube-node-4g8gf\" (UID: \"1dff864e-b75d-4e0b-b182-75f710eac8df\") " pod="openshift-ovn-kubernetes/ovnkube-node-4g8gf" Oct 08 18:17:32 crc kubenswrapper[4859]: I1008 18:17:32.530075 4859 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-lib-openvswitch\" (UniqueName: \"kubernetes.io/host-path/1dff864e-b75d-4e0b-b182-75f710eac8df-var-lib-openvswitch\") pod \"ovnkube-node-4g8gf\" (UID: \"1dff864e-b75d-4e0b-b182-75f710eac8df\") " pod="openshift-ovn-kubernetes/ovnkube-node-4g8gf" Oct 08 18:17:32 crc kubenswrapper[4859]: I1008 18:17:32.530080 4859 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="925f1c65-6136-48ba-85aa-3a3b50560753" path="/var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/volumes" Oct 08 18:17:32 crc kubenswrapper[4859]: I1008 18:17:32.530095 4859 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-run-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/1dff864e-b75d-4e0b-b182-75f710eac8df-host-run-ovn-kubernetes\") pod \"ovnkube-node-4g8gf\" (UID: \"1dff864e-b75d-4e0b-b182-75f710eac8df\") " pod="openshift-ovn-kubernetes/ovnkube-node-4g8gf" Oct 08 18:17:32 crc kubenswrapper[4859]: I1008 18:17:32.530120 4859 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-cni-bin\" (UniqueName: \"kubernetes.io/host-path/1dff864e-b75d-4e0b-b182-75f710eac8df-host-cni-bin\") pod \"ovnkube-node-4g8gf\" (UID: \"1dff864e-b75d-4e0b-b182-75f710eac8df\") " pod="openshift-ovn-kubernetes/ovnkube-node-4g8gf" Oct 08 18:17:32 crc kubenswrapper[4859]: I1008 18:17:32.530140 4859 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-ovn\" (UniqueName: \"kubernetes.io/host-path/1dff864e-b75d-4e0b-b182-75f710eac8df-run-ovn\") pod \"ovnkube-node-4g8gf\" (UID: \"1dff864e-b75d-4e0b-b182-75f710eac8df\") " pod="openshift-ovn-kubernetes/ovnkube-node-4g8gf" Oct 08 18:17:32 crc kubenswrapper[4859]: I1008 18:17:32.530731 4859 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/1dff864e-b75d-4e0b-b182-75f710eac8df-env-overrides\") pod \"ovnkube-node-4g8gf\" (UID: \"1dff864e-b75d-4e0b-b182-75f710eac8df\") " pod="openshift-ovn-kubernetes/ovnkube-node-4g8gf" Oct 08 18:17:32 crc kubenswrapper[4859]: I1008 18:17:32.531875 4859 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="96b93a3a-6083-4aea-8eab-fe1aa8245ad9" path="/var/lib/kubelet/pods/96b93a3a-6083-4aea-8eab-fe1aa8245ad9/volumes" Oct 08 18:17:32 crc kubenswrapper[4859]: I1008 18:17:32.532004 4859 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/1dff864e-b75d-4e0b-b182-75f710eac8df-host-slash\") pod \"ovnkube-node-4g8gf\" (UID: \"1dff864e-b75d-4e0b-b182-75f710eac8df\") " pod="openshift-ovn-kubernetes/ovnkube-node-4g8gf" Oct 08 18:17:32 crc kubenswrapper[4859]: I1008 18:17:32.532424 4859 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-cni-netd\" (UniqueName: \"kubernetes.io/host-path/1dff864e-b75d-4e0b-b182-75f710eac8df-host-cni-netd\") pod \"ovnkube-node-4g8gf\" (UID: \"1dff864e-b75d-4e0b-b182-75f710eac8df\") " pod="openshift-ovn-kubernetes/ovnkube-node-4g8gf" Oct 08 18:17:32 crc kubenswrapper[4859]: I1008 18:17:32.532778 4859 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/1dff864e-b75d-4e0b-b182-75f710eac8df-ovnkube-config\") pod \"ovnkube-node-4g8gf\" (UID: \"1dff864e-b75d-4e0b-b182-75f710eac8df\") " pod="openshift-ovn-kubernetes/ovnkube-node-4g8gf" Oct 08 18:17:32 crc kubenswrapper[4859]: I1008 18:17:32.533455 4859 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/1dff864e-b75d-4e0b-b182-75f710eac8df-ovnkube-script-lib\") pod \"ovnkube-node-4g8gf\" (UID: \"1dff864e-b75d-4e0b-b182-75f710eac8df\") " pod="openshift-ovn-kubernetes/ovnkube-node-4g8gf" Oct 08 18:17:32 crc kubenswrapper[4859]: I1008 18:17:32.533496 4859 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="9d4552c7-cd75-42dd-8880-30dd377c49a4" path="/var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes" Oct 08 18:17:32 crc kubenswrapper[4859]: I1008 18:17:32.534885 4859 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="a0128f3a-b052-44ed-a84e-c4c8aaf17c13" path="/var/lib/kubelet/pods/a0128f3a-b052-44ed-a84e-c4c8aaf17c13/volumes" Oct 08 18:17:32 crc kubenswrapper[4859]: I1008 18:17:32.536271 4859 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="a31745f5-9847-4afe-82a5-3161cc66ca93" path="/var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/volumes" Oct 08 18:17:32 crc kubenswrapper[4859]: I1008 18:17:32.537328 4859 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="b11524ee-3fca-4b1b-9cdf-6da289fdbc7d" path="/var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes" Oct 08 18:17:32 crc kubenswrapper[4859]: I1008 18:17:32.537473 4859 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/1dff864e-b75d-4e0b-b182-75f710eac8df-ovn-node-metrics-cert\") pod \"ovnkube-node-4g8gf\" (UID: \"1dff864e-b75d-4e0b-b182-75f710eac8df\") " pod="openshift-ovn-kubernetes/ovnkube-node-4g8gf" Oct 08 18:17:32 crc kubenswrapper[4859]: I1008 18:17:32.538727 4859 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="b6312bbd-5731-4ea0-a20f-81d5a57df44a" path="/var/lib/kubelet/pods/b6312bbd-5731-4ea0-a20f-81d5a57df44a/volumes" Oct 08 18:17:32 crc kubenswrapper[4859]: I1008 18:17:32.542456 4859 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="b6cd30de-2eeb-49a2-ab40-9167f4560ff5" path="/var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/volumes" Oct 08 18:17:32 crc kubenswrapper[4859]: I1008 18:17:32.543402 4859 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="bc5039c0-ea34-426b-a2b7-fbbc87b49a6d" path="/var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/volumes" Oct 08 18:17:32 crc kubenswrapper[4859]: I1008 18:17:32.545340 4859 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="bd23aa5c-e532-4e53-bccf-e79f130c5ae8" path="/var/lib/kubelet/pods/bd23aa5c-e532-4e53-bccf-e79f130c5ae8/volumes" Oct 08 18:17:32 crc kubenswrapper[4859]: I1008 18:17:32.546521 4859 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="bf126b07-da06-4140-9a57-dfd54fc6b486" path="/var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/volumes" Oct 08 18:17:32 crc kubenswrapper[4859]: I1008 18:17:32.547891 4859 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="c03ee662-fb2f-4fc4-a2c1-af487c19d254" path="/var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/volumes" Oct 08 18:17:32 crc kubenswrapper[4859]: I1008 18:17:32.549089 4859 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d" path="/var/lib/kubelet/pods/cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d/volumes" Oct 08 18:17:32 crc kubenswrapper[4859]: I1008 18:17:32.551132 4859 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-s8jnn\" (UniqueName: \"kubernetes.io/projected/1dff864e-b75d-4e0b-b182-75f710eac8df-kube-api-access-s8jnn\") pod \"ovnkube-node-4g8gf\" (UID: \"1dff864e-b75d-4e0b-b182-75f710eac8df\") " pod="openshift-ovn-kubernetes/ovnkube-node-4g8gf" Oct 08 18:17:32 crc kubenswrapper[4859]: I1008 18:17:32.551254 4859 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="e7e6199b-1264-4501-8953-767f51328d08" path="/var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/volumes" Oct 08 18:17:32 crc kubenswrapper[4859]: I1008 18:17:32.552236 4859 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="efdd0498-1daa-4136-9a4a-3b948c2293fc" path="/var/lib/kubelet/pods/efdd0498-1daa-4136-9a4a-3b948c2293fc/volumes" Oct 08 18:17:32 crc kubenswrapper[4859]: I1008 18:17:32.552937 4859 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="f88749ec-7931-4ee7-b3fc-1ec5e11f92e9" path="/var/lib/kubelet/pods/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9/volumes" Oct 08 18:17:32 crc kubenswrapper[4859]: I1008 18:17:32.556402 4859 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="fda69060-fa79-4696-b1a6-7980f124bf7c" path="/var/lib/kubelet/pods/fda69060-fa79-4696-b1a6-7980f124bf7c/volumes" Oct 08 18:17:32 crc kubenswrapper[4859]: I1008 18:17:32.558804 4859 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-84k8x" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1c24078c-c9e8-4160-883a-f483b4b3dd23\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:17:31Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:17:31Z\\\",\\\"message\\\":\\\"containers with incomplete status: [egress-router-binary-copy cni-plugins bond-cni-plugin routeoverride-cni whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:17:31Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:17:31Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hk5xb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hk5xb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hk5xb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hk5xb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hk5xb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hk5xb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hk5xb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T18:17:31Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-84k8x\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T18:17:32Z is after 2025-08-24T17:21:41Z" Oct 08 18:17:32 crc kubenswrapper[4859]: I1008 18:17:32.639548 4859 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-8plkg" event={"ID":"92b2d7de-31cc-4a91-95a5-ed5ea964b028","Type":"ContainerStarted","Data":"fa00b8345d40b74a917ddc23cc1b2c5c684a9ca61b6119d090c92cc1439b71c4"} Oct 08 18:17:32 crc kubenswrapper[4859]: I1008 18:17:32.639611 4859 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-8plkg" event={"ID":"92b2d7de-31cc-4a91-95a5-ed5ea964b028","Type":"ContainerStarted","Data":"0826832b6e8d8cd050d90d66cee4d56f725d7e4cbbc670056992bdbc45ca17a5"} Oct 08 18:17:32 crc kubenswrapper[4859]: I1008 18:17:32.641921 4859 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-dns/node-resolver-btw28" event={"ID":"97bc5598-3ead-4300-a938-3aebbff1e8c6","Type":"ContainerStarted","Data":"d6d4d6ba771cb5fde637449bba05660f634df7e833a0c8dad2b2c7f3a422bb2b"} Oct 08 18:17:32 crc kubenswrapper[4859]: I1008 18:17:32.641971 4859 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-dns/node-resolver-btw28" event={"ID":"97bc5598-3ead-4300-a938-3aebbff1e8c6","Type":"ContainerStarted","Data":"ee6847e36adfd6281ee24aa0814dca2ca4c129940f9962d0e030ec56cbe319d4"} Oct 08 18:17:32 crc kubenswrapper[4859]: I1008 18:17:32.645613 4859 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-84k8x" event={"ID":"1c24078c-c9e8-4160-883a-f483b4b3dd23","Type":"ContainerStarted","Data":"18b291eb3344fb7c00ef5679dca9a6f7ed9d08cea684a803ff7dad2e8a18c585"} Oct 08 18:17:32 crc kubenswrapper[4859]: I1008 18:17:32.645649 4859 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-84k8x" event={"ID":"1c24078c-c9e8-4160-883a-f483b4b3dd23","Type":"ContainerStarted","Data":"d9647b0aa2d0467493c8b3dc89c6c576f96b2a5414b70ec0d146dded92ac6f38"} Oct 08 18:17:32 crc kubenswrapper[4859]: I1008 18:17:32.647723 4859 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-82s52" event={"ID":"b23a6a6c-9d92-4e7b-840e-55cfda873a2d","Type":"ContainerStarted","Data":"c23ea3580998adf90621a0155c20ee4bd590c512629812174e95c28857bcfef9"} Oct 08 18:17:32 crc kubenswrapper[4859]: I1008 18:17:32.647763 4859 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-82s52" event={"ID":"b23a6a6c-9d92-4e7b-840e-55cfda873a2d","Type":"ContainerStarted","Data":"8e1d5ad46cfd2b72c615daedc67761cd0282b8bb02e2be8046d9ec3f2bc89419"} Oct 08 18:17:32 crc kubenswrapper[4859]: I1008 18:17:32.656863 4859 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-08T18:17:29Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T18:17:29Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T18:17:32Z is after 2025-08-24T17:21:41Z" Oct 08 18:17:32 crc kubenswrapper[4859]: I1008 18:17:32.677235 4859 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-btw28" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"97bc5598-3ead-4300-a938-3aebbff1e8c6\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:17:31Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:17:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:17:31Z\\\",\\\"message\\\":\\\"containers with unready status: [dns-node-resolver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:17:31Z\\\",\\\"message\\\":\\\"containers with unready status: [dns-node-resolver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qcxsd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T18:17:31Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-btw28\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T18:17:32Z is after 2025-08-24T17:21:41Z" Oct 08 18:17:32 crc kubenswrapper[4859]: I1008 18:17:32.692795 4859 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-8plkg" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"92b2d7de-31cc-4a91-95a5-ed5ea964b028\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:17:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:17:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:17:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:17:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fa00b8345d40b74a917ddc23cc1b2c5c684a9ca61b6119d090c92cc1439b71c4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T18:17:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q9d7m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T18:17:31Z\\\"}}\" for pod \"openshift-multus\"/\"multus-8plkg\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T18:17:32Z is after 2025-08-24T17:21:41Z" Oct 08 18:17:32 crc kubenswrapper[4859]: I1008 18:17:32.705287 4859 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ovn-kubernetes/ovnkube-node-4g8gf" Oct 08 18:17:32 crc kubenswrapper[4859]: I1008 18:17:32.738474 4859 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-08T18:17:31Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T18:17:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T18:17:31Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c6d15905694b6bbd4999d49dc8f7a53eb6bfaa7adb59efb0d7e4973c21d1c7be\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T18:17:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T18:17:32Z is after 2025-08-24T17:21:41Z" Oct 08 18:17:32 crc kubenswrapper[4859]: I1008 18:17:32.756838 4859 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-08T18:17:29Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T18:17:29Z\\\",\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T18:17:32Z is after 2025-08-24T17:21:41Z" Oct 08 18:17:32 crc kubenswrapper[4859]: I1008 18:17:32.777501 4859 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"63a623de-5608-4228-b1c2-77188c276986\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:17:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:17:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:17:10Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:17:10Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:17:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d4786a8fb8abd96fb22e60733c974f599017407d589ed815fb8541fd3b64f696\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T18:17:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://183d481502ed5108d5e19e483b714c289f355ea3f8b0891883e5335b802c83b6\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T18:17:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ec848c46fd5dde0469b00d3cab14ba4a652533a265490fadbee7a8b8843ef249\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T18:17:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4fe1828b4d63ff8106bfcd15e1dc1c1bd1ba7f77777cdb23ff1b291d0418774c\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2e700b2a20c81b414617f5b5c018a738278a3fa44e6c6d14b2170051d0725431\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-08T18:17:24Z\\\",\\\"message\\\":\\\"W1008 18:17:13.670625 1 cmd.go:257] Using insecure, self-signed certificates\\\\nI1008 18:17:13.671202 1 crypto.go:601] Generating new CA for check-endpoints-signer@1759947433 cert, and key in /tmp/serving-cert-3867218267/serving-signer.crt, /tmp/serving-cert-3867218267/serving-signer.key\\\\nI1008 18:17:13.944463 1 observer_polling.go:159] Starting file observer\\\\nW1008 18:17:13.951958 1 builder.go:272] unable to get owner reference (falling back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": dial tcp [::1]:6443: connect: connection refused\\\\nI1008 18:17:13.952088 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1008 18:17:13.956755 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3867218267/tls.crt::/tmp/serving-cert-3867218267/tls.key\\\\\\\"\\\\nF1008 18:17:24.290991 1 cmd.go:182] error initializing delegating authentication: unable to load configmap based request-header-client-ca-file: Get \\\\\\\"https://localhost:6443/api/v1/namespaces/kube-system/configmaps/extension-apiserver-authentication\\\\\\\": net/http: TLS handshake timeout\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-08T18:17:13Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T18:17:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c65b66f62d325ca10ef8d5d31f04529ccf4e691cf3303d231e7a414fb400e4d6\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T18:17:13Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://58c7bba4ace28215b062653a960e561b1dc1f890f2baf450145638a065af038b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://58c7bba4ace28215b062653a960e561b1dc1f890f2baf450145638a065af038b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-08T18:17:11Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-08T18:17:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T18:17:10Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T18:17:32Z is after 2025-08-24T17:21:41Z" Oct 08 18:17:32 crc kubenswrapper[4859]: I1008 18:17:32.795167 4859 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-08T18:17:31Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T18:17:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T18:17:31Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://99e2d51bd7757498dae60949eea6d732b5f7b481fa00f808dd5308a530a1daf1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T18:17:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e32aaba93699944665660ae41e9b5779178c1354c31e15ad8188bf8c3c8fa500\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T18:17:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T18:17:32Z is after 2025-08-24T17:21:41Z" Oct 08 18:17:32 crc kubenswrapper[4859]: I1008 18:17:32.811972 4859 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-08T18:17:29Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T18:17:29Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T18:17:32Z is after 2025-08-24T17:21:41Z" Oct 08 18:17:32 crc kubenswrapper[4859]: I1008 18:17:32.825614 4859 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-82s52" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b23a6a6c-9d92-4e7b-840e-55cfda873a2d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:17:31Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:17:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:17:31Z\\\",\\\"message\\\":\\\"containers with unready status: [machine-config-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:17:31Z\\\",\\\"message\\\":\\\"containers with unready status: [machine-config-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6vx22\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6vx22\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T18:17:31Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-82s52\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T18:17:32Z is after 2025-08-24T17:21:41Z" Oct 08 18:17:32 crc kubenswrapper[4859]: I1008 18:17:32.846574 4859 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-4g8gf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1dff864e-b75d-4e0b-b182-75f710eac8df\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:17:32Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:17:32Z\\\",\\\"message\\\":\\\"containers with incomplete status: [kubecfg-setup]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:17:32Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:17:32Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s8jnn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s8jnn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s8jnn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s8jnn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s8jnn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s8jnn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s8jnn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s8jnn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s8jnn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T18:17:32Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-4g8gf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T18:17:32Z is after 2025-08-24T17:21:41Z" Oct 08 18:17:32 crc kubenswrapper[4859]: I1008 18:17:32.859914 4859 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-08T18:17:29Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T18:17:29Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T18:17:32Z is after 2025-08-24T17:21:41Z" Oct 08 18:17:32 crc kubenswrapper[4859]: I1008 18:17:32.875542 4859 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-84k8x" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1c24078c-c9e8-4160-883a-f483b4b3dd23\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:17:31Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:17:31Z\\\",\\\"message\\\":\\\"containers with incomplete status: [egress-router-binary-copy cni-plugins bond-cni-plugin routeoverride-cni whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:17:31Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:17:31Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hk5xb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hk5xb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hk5xb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hk5xb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hk5xb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hk5xb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hk5xb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T18:17:31Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-84k8x\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T18:17:32Z is after 2025-08-24T17:21:41Z" Oct 08 18:17:32 crc kubenswrapper[4859]: I1008 18:17:32.891628 4859 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-08T18:17:31Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T18:17:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T18:17:31Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c6d15905694b6bbd4999d49dc8f7a53eb6bfaa7adb59efb0d7e4973c21d1c7be\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T18:17:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T18:17:32Z is after 2025-08-24T17:21:41Z" Oct 08 18:17:32 crc kubenswrapper[4859]: I1008 18:17:32.907271 4859 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-08T18:17:29Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T18:17:29Z\\\",\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T18:17:32Z is after 2025-08-24T17:21:41Z" Oct 08 18:17:32 crc kubenswrapper[4859]: I1008 18:17:32.932038 4859 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-08T18:17:29Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T18:17:29Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T18:17:32Z is after 2025-08-24T17:21:41Z" Oct 08 18:17:32 crc kubenswrapper[4859]: I1008 18:17:32.998832 4859 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-82s52" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b23a6a6c-9d92-4e7b-840e-55cfda873a2d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:17:31Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:17:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:17:31Z\\\",\\\"message\\\":\\\"containers with unready status: [machine-config-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:17:31Z\\\",\\\"message\\\":\\\"containers with unready status: [machine-config-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6vx22\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6vx22\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T18:17:31Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-82s52\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T18:17:32Z is after 2025-08-24T17:21:41Z" Oct 08 18:17:33 crc kubenswrapper[4859]: I1008 18:17:33.028120 4859 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-4g8gf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1dff864e-b75d-4e0b-b182-75f710eac8df\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:17:32Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:17:32Z\\\",\\\"message\\\":\\\"containers with incomplete status: [kubecfg-setup]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:17:32Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:17:32Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s8jnn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s8jnn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s8jnn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s8jnn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s8jnn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s8jnn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s8jnn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s8jnn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s8jnn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T18:17:32Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-4g8gf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T18:17:33Z is after 2025-08-24T17:21:41Z" Oct 08 18:17:33 crc kubenswrapper[4859]: I1008 18:17:33.068057 4859 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"63a623de-5608-4228-b1c2-77188c276986\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:17:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:17:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:17:10Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:17:10Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:17:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d4786a8fb8abd96fb22e60733c974f599017407d589ed815fb8541fd3b64f696\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T18:17:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://183d481502ed5108d5e19e483b714c289f355ea3f8b0891883e5335b802c83b6\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T18:17:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ec848c46fd5dde0469b00d3cab14ba4a652533a265490fadbee7a8b8843ef249\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T18:17:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4fe1828b4d63ff8106bfcd15e1dc1c1bd1ba7f77777cdb23ff1b291d0418774c\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2e700b2a20c81b414617f5b5c018a738278a3fa44e6c6d14b2170051d0725431\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-08T18:17:24Z\\\",\\\"message\\\":\\\"W1008 18:17:13.670625 1 cmd.go:257] Using insecure, self-signed certificates\\\\nI1008 18:17:13.671202 1 crypto.go:601] Generating new CA for check-endpoints-signer@1759947433 cert, and key in /tmp/serving-cert-3867218267/serving-signer.crt, /tmp/serving-cert-3867218267/serving-signer.key\\\\nI1008 18:17:13.944463 1 observer_polling.go:159] Starting file observer\\\\nW1008 18:17:13.951958 1 builder.go:272] unable to get owner reference (falling back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": dial tcp [::1]:6443: connect: connection refused\\\\nI1008 18:17:13.952088 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1008 18:17:13.956755 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3867218267/tls.crt::/tmp/serving-cert-3867218267/tls.key\\\\\\\"\\\\nF1008 18:17:24.290991 1 cmd.go:182] error initializing delegating authentication: unable to load configmap based request-header-client-ca-file: Get \\\\\\\"https://localhost:6443/api/v1/namespaces/kube-system/configmaps/extension-apiserver-authentication\\\\\\\": net/http: TLS handshake timeout\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-08T18:17:13Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T18:17:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c65b66f62d325ca10ef8d5d31f04529ccf4e691cf3303d231e7a414fb400e4d6\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T18:17:13Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://58c7bba4ace28215b062653a960e561b1dc1f890f2baf450145638a065af038b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://58c7bba4ace28215b062653a960e561b1dc1f890f2baf450145638a065af038b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-08T18:17:11Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-08T18:17:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T18:17:10Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T18:17:33Z is after 2025-08-24T17:21:41Z" Oct 08 18:17:33 crc kubenswrapper[4859]: I1008 18:17:33.092512 4859 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-08T18:17:31Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T18:17:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T18:17:31Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://99e2d51bd7757498dae60949eea6d732b5f7b481fa00f808dd5308a530a1daf1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T18:17:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e32aaba93699944665660ae41e9b5779178c1354c31e15ad8188bf8c3c8fa500\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T18:17:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T18:17:33Z is after 2025-08-24T17:21:41Z" Oct 08 18:17:33 crc kubenswrapper[4859]: I1008 18:17:33.133945 4859 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-08T18:17:29Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T18:17:29Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T18:17:33Z is after 2025-08-24T17:21:41Z" Oct 08 18:17:33 crc kubenswrapper[4859]: I1008 18:17:33.180718 4859 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-84k8x" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1c24078c-c9e8-4160-883a-f483b4b3dd23\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:17:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:17:31Z\\\",\\\"message\\\":\\\"containers with incomplete status: [egress-router-binary-copy cni-plugins bond-cni-plugin routeoverride-cni whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:17:31Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:17:31Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hk5xb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://18b291eb3344fb7c00ef5679dca9a6f7ed9d08cea684a803ff7dad2e8a18c585\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T18:17:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hk5xb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hk5xb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hk5xb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hk5xb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hk5xb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hk5xb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T18:17:31Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-84k8x\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T18:17:33Z is after 2025-08-24T17:21:41Z" Oct 08 18:17:33 crc kubenswrapper[4859]: I1008 18:17:33.215358 4859 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-08T18:17:29Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T18:17:29Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T18:17:33Z is after 2025-08-24T17:21:41Z" Oct 08 18:17:33 crc kubenswrapper[4859]: I1008 18:17:33.251584 4859 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-btw28" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"97bc5598-3ead-4300-a938-3aebbff1e8c6\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:17:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:17:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:17:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:17:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d6d4d6ba771cb5fde637449bba05660f634df7e833a0c8dad2b2c7f3a422bb2b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T18:17:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qcxsd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T18:17:31Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-btw28\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T18:17:33Z is after 2025-08-24T17:21:41Z" Oct 08 18:17:33 crc kubenswrapper[4859]: I1008 18:17:33.294957 4859 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-8plkg" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"92b2d7de-31cc-4a91-95a5-ed5ea964b028\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:17:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:17:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:17:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:17:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fa00b8345d40b74a917ddc23cc1b2c5c684a9ca61b6119d090c92cc1439b71c4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T18:17:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q9d7m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T18:17:31Z\\\"}}\" for pod \"openshift-multus\"/\"multus-8plkg\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T18:17:33Z is after 2025-08-24T17:21:41Z" Oct 08 18:17:33 crc kubenswrapper[4859]: I1008 18:17:33.469652 4859 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 08 18:17:33 crc kubenswrapper[4859]: E1008 18:17:33.469818 4859 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 08 18:17:33 crc kubenswrapper[4859]: I1008 18:17:33.470184 4859 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 08 18:17:33 crc kubenswrapper[4859]: E1008 18:17:33.470257 4859 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 08 18:17:33 crc kubenswrapper[4859]: I1008 18:17:33.470312 4859 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 08 18:17:33 crc kubenswrapper[4859]: E1008 18:17:33.470368 4859 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 08 18:17:33 crc kubenswrapper[4859]: I1008 18:17:33.652860 4859 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" event={"ID":"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49","Type":"ContainerStarted","Data":"5d99e071bd083914418d9c128f38955f26295c08b1520f2dc774c99b1c843693"} Oct 08 18:17:33 crc kubenswrapper[4859]: I1008 18:17:33.657293 4859 generic.go:334] "Generic (PLEG): container finished" podID="1c24078c-c9e8-4160-883a-f483b4b3dd23" containerID="18b291eb3344fb7c00ef5679dca9a6f7ed9d08cea684a803ff7dad2e8a18c585" exitCode=0 Oct 08 18:17:33 crc kubenswrapper[4859]: I1008 18:17:33.657372 4859 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-84k8x" event={"ID":"1c24078c-c9e8-4160-883a-f483b4b3dd23","Type":"ContainerDied","Data":"18b291eb3344fb7c00ef5679dca9a6f7ed9d08cea684a803ff7dad2e8a18c585"} Oct 08 18:17:33 crc kubenswrapper[4859]: I1008 18:17:33.659214 4859 generic.go:334] "Generic (PLEG): container finished" podID="1dff864e-b75d-4e0b-b182-75f710eac8df" containerID="7a211d3da02b173867ff3a912eecd3b379483c7cc77e6b91122edd1ffd30fd53" exitCode=0 Oct 08 18:17:33 crc kubenswrapper[4859]: I1008 18:17:33.659272 4859 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-4g8gf" event={"ID":"1dff864e-b75d-4e0b-b182-75f710eac8df","Type":"ContainerDied","Data":"7a211d3da02b173867ff3a912eecd3b379483c7cc77e6b91122edd1ffd30fd53"} Oct 08 18:17:33 crc kubenswrapper[4859]: I1008 18:17:33.659349 4859 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-4g8gf" event={"ID":"1dff864e-b75d-4e0b-b182-75f710eac8df","Type":"ContainerStarted","Data":"9cde21bcc626ceb4fd1a09cfe04cba4c327c7e4ad650edc972e0765907465fba"} Oct 08 18:17:33 crc kubenswrapper[4859]: I1008 18:17:33.661814 4859 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-82s52" event={"ID":"b23a6a6c-9d92-4e7b-840e-55cfda873a2d","Type":"ContainerStarted","Data":"71302c766485b9f55743d85f97440ff578cd44ba1e5a6f5b1c024c0e808072d2"} Oct 08 18:17:33 crc kubenswrapper[4859]: I1008 18:17:33.673528 4859 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-08T18:17:33Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T18:17:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T18:17:33Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5d99e071bd083914418d9c128f38955f26295c08b1520f2dc774c99b1c843693\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T18:17:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T18:17:33Z is after 2025-08-24T17:21:41Z" Oct 08 18:17:33 crc kubenswrapper[4859]: I1008 18:17:33.688858 4859 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-08T18:17:31Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T18:17:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T18:17:31Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c6d15905694b6bbd4999d49dc8f7a53eb6bfaa7adb59efb0d7e4973c21d1c7be\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T18:17:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T18:17:33Z is after 2025-08-24T17:21:41Z" Oct 08 18:17:33 crc kubenswrapper[4859]: I1008 18:17:33.703815 4859 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"63a623de-5608-4228-b1c2-77188c276986\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:17:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:17:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:17:10Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:17:10Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:17:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d4786a8fb8abd96fb22e60733c974f599017407d589ed815fb8541fd3b64f696\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T18:17:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://183d481502ed5108d5e19e483b714c289f355ea3f8b0891883e5335b802c83b6\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T18:17:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ec848c46fd5dde0469b00d3cab14ba4a652533a265490fadbee7a8b8843ef249\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T18:17:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4fe1828b4d63ff8106bfcd15e1dc1c1bd1ba7f77777cdb23ff1b291d0418774c\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2e700b2a20c81b414617f5b5c018a738278a3fa44e6c6d14b2170051d0725431\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-08T18:17:24Z\\\",\\\"message\\\":\\\"W1008 18:17:13.670625 1 cmd.go:257] Using insecure, self-signed certificates\\\\nI1008 18:17:13.671202 1 crypto.go:601] Generating new CA for check-endpoints-signer@1759947433 cert, and key in /tmp/serving-cert-3867218267/serving-signer.crt, /tmp/serving-cert-3867218267/serving-signer.key\\\\nI1008 18:17:13.944463 1 observer_polling.go:159] Starting file observer\\\\nW1008 18:17:13.951958 1 builder.go:272] unable to get owner reference (falling back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": dial tcp [::1]:6443: connect: connection refused\\\\nI1008 18:17:13.952088 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1008 18:17:13.956755 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3867218267/tls.crt::/tmp/serving-cert-3867218267/tls.key\\\\\\\"\\\\nF1008 18:17:24.290991 1 cmd.go:182] error initializing delegating authentication: unable to load configmap based request-header-client-ca-file: Get \\\\\\\"https://localhost:6443/api/v1/namespaces/kube-system/configmaps/extension-apiserver-authentication\\\\\\\": net/http: TLS handshake timeout\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-08T18:17:13Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T18:17:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c65b66f62d325ca10ef8d5d31f04529ccf4e691cf3303d231e7a414fb400e4d6\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T18:17:13Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://58c7bba4ace28215b062653a960e561b1dc1f890f2baf450145638a065af038b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://58c7bba4ace28215b062653a960e561b1dc1f890f2baf450145638a065af038b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-08T18:17:11Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-08T18:17:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T18:17:10Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T18:17:33Z is after 2025-08-24T17:21:41Z" Oct 08 18:17:33 crc kubenswrapper[4859]: I1008 18:17:33.722764 4859 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-08T18:17:31Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T18:17:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T18:17:31Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://99e2d51bd7757498dae60949eea6d732b5f7b481fa00f808dd5308a530a1daf1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T18:17:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e32aaba93699944665660ae41e9b5779178c1354c31e15ad8188bf8c3c8fa500\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T18:17:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T18:17:33Z is after 2025-08-24T17:21:41Z" Oct 08 18:17:33 crc kubenswrapper[4859]: I1008 18:17:33.740352 4859 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-08T18:17:29Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T18:17:29Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T18:17:33Z is after 2025-08-24T17:21:41Z" Oct 08 18:17:33 crc kubenswrapper[4859]: I1008 18:17:33.755368 4859 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-82s52" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b23a6a6c-9d92-4e7b-840e-55cfda873a2d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:17:31Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:17:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:17:31Z\\\",\\\"message\\\":\\\"containers with unready status: [machine-config-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:17:31Z\\\",\\\"message\\\":\\\"containers with unready status: [machine-config-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6vx22\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6vx22\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T18:17:31Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-82s52\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T18:17:33Z is after 2025-08-24T17:21:41Z" Oct 08 18:17:33 crc kubenswrapper[4859]: I1008 18:17:33.778936 4859 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-4g8gf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1dff864e-b75d-4e0b-b182-75f710eac8df\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:17:32Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:17:32Z\\\",\\\"message\\\":\\\"containers with incomplete status: [kubecfg-setup]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:17:32Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:17:32Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s8jnn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s8jnn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s8jnn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s8jnn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s8jnn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s8jnn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s8jnn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s8jnn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s8jnn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T18:17:32Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-4g8gf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T18:17:33Z is after 2025-08-24T17:21:41Z" Oct 08 18:17:33 crc kubenswrapper[4859]: I1008 18:17:33.792850 4859 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-08T18:17:29Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T18:17:29Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T18:17:33Z is after 2025-08-24T17:21:41Z" Oct 08 18:17:33 crc kubenswrapper[4859]: I1008 18:17:33.808388 4859 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-84k8x" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1c24078c-c9e8-4160-883a-f483b4b3dd23\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:17:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:17:31Z\\\",\\\"message\\\":\\\"containers with incomplete status: [egress-router-binary-copy cni-plugins bond-cni-plugin routeoverride-cni whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:17:31Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:17:31Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hk5xb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://18b291eb3344fb7c00ef5679dca9a6f7ed9d08cea684a803ff7dad2e8a18c585\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T18:17:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hk5xb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hk5xb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hk5xb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hk5xb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hk5xb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hk5xb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T18:17:31Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-84k8x\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T18:17:33Z is after 2025-08-24T17:21:41Z" Oct 08 18:17:33 crc kubenswrapper[4859]: I1008 18:17:33.823208 4859 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-08T18:17:29Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T18:17:29Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T18:17:33Z is after 2025-08-24T17:21:41Z" Oct 08 18:17:33 crc kubenswrapper[4859]: I1008 18:17:33.835896 4859 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-btw28" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"97bc5598-3ead-4300-a938-3aebbff1e8c6\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:17:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:17:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:17:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:17:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d6d4d6ba771cb5fde637449bba05660f634df7e833a0c8dad2b2c7f3a422bb2b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T18:17:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qcxsd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T18:17:31Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-btw28\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T18:17:33Z is after 2025-08-24T17:21:41Z" Oct 08 18:17:33 crc kubenswrapper[4859]: I1008 18:17:33.853888 4859 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-8plkg" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"92b2d7de-31cc-4a91-95a5-ed5ea964b028\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:17:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:17:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:17:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:17:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fa00b8345d40b74a917ddc23cc1b2c5c684a9ca61b6119d090c92cc1439b71c4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T18:17:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q9d7m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T18:17:31Z\\\"}}\" for pod \"openshift-multus\"/\"multus-8plkg\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T18:17:33Z is after 2025-08-24T17:21:41Z" Oct 08 18:17:33 crc kubenswrapper[4859]: I1008 18:17:33.875940 4859 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-84k8x" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1c24078c-c9e8-4160-883a-f483b4b3dd23\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:17:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:17:31Z\\\",\\\"message\\\":\\\"containers with incomplete status: [cni-plugins bond-cni-plugin routeoverride-cni whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:17:31Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:17:31Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hk5xb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://18b291eb3344fb7c00ef5679dca9a6f7ed9d08cea684a803ff7dad2e8a18c585\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://18b291eb3344fb7c00ef5679dca9a6f7ed9d08cea684a803ff7dad2e8a18c585\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-08T18:17:32Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-08T18:17:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hk5xb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hk5xb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hk5xb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hk5xb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hk5xb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hk5xb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T18:17:31Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-84k8x\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T18:17:33Z is after 2025-08-24T17:21:41Z" Oct 08 18:17:33 crc kubenswrapper[4859]: I1008 18:17:33.892138 4859 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-08T18:17:29Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T18:17:29Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T18:17:33Z is after 2025-08-24T17:21:41Z" Oct 08 18:17:33 crc kubenswrapper[4859]: I1008 18:17:33.910518 4859 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-btw28" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"97bc5598-3ead-4300-a938-3aebbff1e8c6\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:17:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:17:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:17:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:17:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d6d4d6ba771cb5fde637449bba05660f634df7e833a0c8dad2b2c7f3a422bb2b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T18:17:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qcxsd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T18:17:31Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-btw28\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T18:17:33Z is after 2025-08-24T17:21:41Z" Oct 08 18:17:33 crc kubenswrapper[4859]: I1008 18:17:33.935148 4859 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-8plkg" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"92b2d7de-31cc-4a91-95a5-ed5ea964b028\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:17:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:17:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:17:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:17:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fa00b8345d40b74a917ddc23cc1b2c5c684a9ca61b6119d090c92cc1439b71c4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T18:17:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q9d7m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T18:17:31Z\\\"}}\" for pod \"openshift-multus\"/\"multus-8plkg\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T18:17:33Z is after 2025-08-24T17:21:41Z" Oct 08 18:17:33 crc kubenswrapper[4859]: I1008 18:17:33.974070 4859 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-08T18:17:29Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T18:17:29Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T18:17:33Z is after 2025-08-24T17:21:41Z" Oct 08 18:17:34 crc kubenswrapper[4859]: I1008 18:17:34.016510 4859 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-08T18:17:31Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T18:17:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T18:17:31Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c6d15905694b6bbd4999d49dc8f7a53eb6bfaa7adb59efb0d7e4973c21d1c7be\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T18:17:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T18:17:34Z is after 2025-08-24T17:21:41Z" Oct 08 18:17:34 crc kubenswrapper[4859]: I1008 18:17:34.054339 4859 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-08T18:17:33Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T18:17:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T18:17:33Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5d99e071bd083914418d9c128f38955f26295c08b1520f2dc774c99b1c843693\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T18:17:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T18:17:34Z is after 2025-08-24T17:21:41Z" Oct 08 18:17:34 crc kubenswrapper[4859]: I1008 18:17:34.094966 4859 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"63a623de-5608-4228-b1c2-77188c276986\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:17:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:17:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:17:10Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:17:10Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:17:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d4786a8fb8abd96fb22e60733c974f599017407d589ed815fb8541fd3b64f696\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T18:17:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://183d481502ed5108d5e19e483b714c289f355ea3f8b0891883e5335b802c83b6\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T18:17:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ec848c46fd5dde0469b00d3cab14ba4a652533a265490fadbee7a8b8843ef249\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T18:17:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4fe1828b4d63ff8106bfcd15e1dc1c1bd1ba7f77777cdb23ff1b291d0418774c\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2e700b2a20c81b414617f5b5c018a738278a3fa44e6c6d14b2170051d0725431\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-08T18:17:24Z\\\",\\\"message\\\":\\\"W1008 18:17:13.670625 1 cmd.go:257] Using insecure, self-signed certificates\\\\nI1008 18:17:13.671202 1 crypto.go:601] Generating new CA for check-endpoints-signer@1759947433 cert, and key in /tmp/serving-cert-3867218267/serving-signer.crt, /tmp/serving-cert-3867218267/serving-signer.key\\\\nI1008 18:17:13.944463 1 observer_polling.go:159] Starting file observer\\\\nW1008 18:17:13.951958 1 builder.go:272] unable to get owner reference (falling back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": dial tcp [::1]:6443: connect: connection refused\\\\nI1008 18:17:13.952088 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1008 18:17:13.956755 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3867218267/tls.crt::/tmp/serving-cert-3867218267/tls.key\\\\\\\"\\\\nF1008 18:17:24.290991 1 cmd.go:182] error initializing delegating authentication: unable to load configmap based request-header-client-ca-file: Get \\\\\\\"https://localhost:6443/api/v1/namespaces/kube-system/configmaps/extension-apiserver-authentication\\\\\\\": net/http: TLS handshake timeout\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-08T18:17:13Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T18:17:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c65b66f62d325ca10ef8d5d31f04529ccf4e691cf3303d231e7a414fb400e4d6\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T18:17:13Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://58c7bba4ace28215b062653a960e561b1dc1f890f2baf450145638a065af038b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://58c7bba4ace28215b062653a960e561b1dc1f890f2baf450145638a065af038b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-08T18:17:11Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-08T18:17:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T18:17:10Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T18:17:34Z is after 2025-08-24T17:21:41Z" Oct 08 18:17:34 crc kubenswrapper[4859]: I1008 18:17:34.133504 4859 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-08T18:17:31Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T18:17:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T18:17:31Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://99e2d51bd7757498dae60949eea6d732b5f7b481fa00f808dd5308a530a1daf1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T18:17:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e32aaba93699944665660ae41e9b5779178c1354c31e15ad8188bf8c3c8fa500\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T18:17:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T18:17:34Z is after 2025-08-24T17:21:41Z" Oct 08 18:17:34 crc kubenswrapper[4859]: I1008 18:17:34.147806 4859 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 08 18:17:34 crc kubenswrapper[4859]: E1008 18:17:34.147982 4859 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-08 18:17:38.14795715 +0000 UTC m=+28.394796539 (durationBeforeRetry 4s). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 08 18:17:34 crc kubenswrapper[4859]: I1008 18:17:34.172776 4859 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-08T18:17:29Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T18:17:29Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T18:17:34Z is after 2025-08-24T17:21:41Z" Oct 08 18:17:34 crc kubenswrapper[4859]: I1008 18:17:34.215371 4859 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-82s52" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b23a6a6c-9d92-4e7b-840e-55cfda873a2d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:17:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:17:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:17:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:17:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://71302c766485b9f55743d85f97440ff578cd44ba1e5a6f5b1c024c0e808072d2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T18:17:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6vx22\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c23ea3580998adf90621a0155c20ee4bd590c512629812174e95c28857bcfef9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T18:17:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6vx22\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T18:17:31Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-82s52\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T18:17:34Z is after 2025-08-24T17:21:41Z" Oct 08 18:17:34 crc kubenswrapper[4859]: I1008 18:17:34.249459 4859 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 08 18:17:34 crc kubenswrapper[4859]: I1008 18:17:34.249559 4859 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 08 18:17:34 crc kubenswrapper[4859]: E1008 18:17:34.249673 4859 configmap.go:193] Couldn't get configMap openshift-network-console/networking-console-plugin: object "openshift-network-console"/"networking-console-plugin" not registered Oct 08 18:17:34 crc kubenswrapper[4859]: E1008 18:17:34.249776 4859 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Oct 08 18:17:34 crc kubenswrapper[4859]: E1008 18:17:34.249803 4859 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Oct 08 18:17:34 crc kubenswrapper[4859]: E1008 18:17:34.249818 4859 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-10-08 18:17:38.249792927 +0000 UTC m=+28.496632306 (durationBeforeRetry 4s). Error: MountVolume.SetUp failed for volume "nginx-conf" (UniqueName: "kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin" not registered Oct 08 18:17:34 crc kubenswrapper[4859]: E1008 18:17:34.249821 4859 projected.go:194] Error preparing data for projected volume kube-api-access-s2dwl for pod openshift-network-diagnostics/network-check-source-55646444c4-trplf: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Oct 08 18:17:34 crc kubenswrapper[4859]: E1008 18:17:34.249863 4859 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl podName:9d751cbb-f2e2-430d-9754-c882a5e924a5 nodeName:}" failed. No retries permitted until 2025-10-08 18:17:38.249851718 +0000 UTC m=+28.496691317 (durationBeforeRetry 4s). Error: MountVolume.SetUp failed for volume "kube-api-access-s2dwl" (UniqueName: "kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl") pod "network-check-source-55646444c4-trplf" (UID: "9d751cbb-f2e2-430d-9754-c882a5e924a5") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Oct 08 18:17:34 crc kubenswrapper[4859]: E1008 18:17:34.249915 4859 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Oct 08 18:17:34 crc kubenswrapper[4859]: E1008 18:17:34.249966 4859 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Oct 08 18:17:34 crc kubenswrapper[4859]: E1008 18:17:34.249988 4859 projected.go:194] Error preparing data for projected volume kube-api-access-cqllr for pod openshift-network-diagnostics/network-check-target-xd92c: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Oct 08 18:17:34 crc kubenswrapper[4859]: E1008 18:17:34.250130 4859 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr podName:3b6479f0-333b-4a96-9adf-2099afdc2447 nodeName:}" failed. No retries permitted until 2025-10-08 18:17:38.250101975 +0000 UTC m=+28.496941494 (durationBeforeRetry 4s). Error: MountVolume.SetUp failed for volume "kube-api-access-cqllr" (UniqueName: "kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr") pod "network-check-target-xd92c" (UID: "3b6479f0-333b-4a96-9adf-2099afdc2447") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Oct 08 18:17:34 crc kubenswrapper[4859]: I1008 18:17:34.249680 4859 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 08 18:17:34 crc kubenswrapper[4859]: I1008 18:17:34.250277 4859 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 08 18:17:34 crc kubenswrapper[4859]: E1008 18:17:34.250405 4859 secret.go:188] Couldn't get secret openshift-network-console/networking-console-plugin-cert: object "openshift-network-console"/"networking-console-plugin-cert" not registered Oct 08 18:17:34 crc kubenswrapper[4859]: E1008 18:17:34.250458 4859 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-10-08 18:17:38.250448444 +0000 UTC m=+28.497287823 (durationBeforeRetry 4s). Error: MountVolume.SetUp failed for volume "networking-console-plugin-cert" (UniqueName: "kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin-cert" not registered Oct 08 18:17:34 crc kubenswrapper[4859]: I1008 18:17:34.263282 4859 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-4g8gf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1dff864e-b75d-4e0b-b182-75f710eac8df\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:17:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:17:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:17:32Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:17:32Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s8jnn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s8jnn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s8jnn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s8jnn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s8jnn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s8jnn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s8jnn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s8jnn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7a211d3da02b173867ff3a912eecd3b379483c7cc77e6b91122edd1ffd30fd53\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7a211d3da02b173867ff3a912eecd3b379483c7cc77e6b91122edd1ffd30fd53\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-08T18:17:32Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-08T18:17:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s8jnn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T18:17:32Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-4g8gf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T18:17:34Z is after 2025-08-24T17:21:41Z" Oct 08 18:17:34 crc kubenswrapper[4859]: I1008 18:17:34.411963 4859 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Oct 08 18:17:34 crc kubenswrapper[4859]: I1008 18:17:34.418274 4859 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Oct 08 18:17:34 crc kubenswrapper[4859]: I1008 18:17:34.422189 4859 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-controller-manager/kube-controller-manager-crc"] Oct 08 18:17:34 crc kubenswrapper[4859]: I1008 18:17:34.428962 4859 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-08T18:17:31Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T18:17:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T18:17:31Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://99e2d51bd7757498dae60949eea6d732b5f7b481fa00f808dd5308a530a1daf1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T18:17:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e32aaba93699944665660ae41e9b5779178c1354c31e15ad8188bf8c3c8fa500\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T18:17:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T18:17:34Z is after 2025-08-24T17:21:41Z" Oct 08 18:17:34 crc kubenswrapper[4859]: I1008 18:17:34.448133 4859 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-08T18:17:29Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T18:17:29Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T18:17:34Z is after 2025-08-24T17:21:41Z" Oct 08 18:17:34 crc kubenswrapper[4859]: I1008 18:17:34.463455 4859 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-82s52" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b23a6a6c-9d92-4e7b-840e-55cfda873a2d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:17:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:17:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:17:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:17:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://71302c766485b9f55743d85f97440ff578cd44ba1e5a6f5b1c024c0e808072d2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T18:17:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6vx22\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c23ea3580998adf90621a0155c20ee4bd590c512629812174e95c28857bcfef9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T18:17:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6vx22\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T18:17:31Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-82s52\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T18:17:34Z is after 2025-08-24T17:21:41Z" Oct 08 18:17:34 crc kubenswrapper[4859]: I1008 18:17:34.488407 4859 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-4g8gf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1dff864e-b75d-4e0b-b182-75f710eac8df\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:17:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:17:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:17:32Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:17:32Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s8jnn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s8jnn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s8jnn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s8jnn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s8jnn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s8jnn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s8jnn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s8jnn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7a211d3da02b173867ff3a912eecd3b379483c7cc77e6b91122edd1ffd30fd53\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7a211d3da02b173867ff3a912eecd3b379483c7cc77e6b91122edd1ffd30fd53\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-08T18:17:32Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-08T18:17:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s8jnn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T18:17:32Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-4g8gf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T18:17:34Z is after 2025-08-24T17:21:41Z" Oct 08 18:17:34 crc kubenswrapper[4859]: I1008 18:17:34.506923 4859 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"63a623de-5608-4228-b1c2-77188c276986\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:17:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:17:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:17:10Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:17:10Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:17:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d4786a8fb8abd96fb22e60733c974f599017407d589ed815fb8541fd3b64f696\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T18:17:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://183d481502ed5108d5e19e483b714c289f355ea3f8b0891883e5335b802c83b6\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T18:17:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ec848c46fd5dde0469b00d3cab14ba4a652533a265490fadbee7a8b8843ef249\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T18:17:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4fe1828b4d63ff8106bfcd15e1dc1c1bd1ba7f77777cdb23ff1b291d0418774c\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2e700b2a20c81b414617f5b5c018a738278a3fa44e6c6d14b2170051d0725431\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-08T18:17:24Z\\\",\\\"message\\\":\\\"W1008 18:17:13.670625 1 cmd.go:257] Using insecure, self-signed certificates\\\\nI1008 18:17:13.671202 1 crypto.go:601] Generating new CA for check-endpoints-signer@1759947433 cert, and key in /tmp/serving-cert-3867218267/serving-signer.crt, /tmp/serving-cert-3867218267/serving-signer.key\\\\nI1008 18:17:13.944463 1 observer_polling.go:159] Starting file observer\\\\nW1008 18:17:13.951958 1 builder.go:272] unable to get owner reference (falling back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": dial tcp [::1]:6443: connect: connection refused\\\\nI1008 18:17:13.952088 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1008 18:17:13.956755 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3867218267/tls.crt::/tmp/serving-cert-3867218267/tls.key\\\\\\\"\\\\nF1008 18:17:24.290991 1 cmd.go:182] error initializing delegating authentication: unable to load configmap based request-header-client-ca-file: Get \\\\\\\"https://localhost:6443/api/v1/namespaces/kube-system/configmaps/extension-apiserver-authentication\\\\\\\": net/http: TLS handshake timeout\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-08T18:17:13Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T18:17:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c65b66f62d325ca10ef8d5d31f04529ccf4e691cf3303d231e7a414fb400e4d6\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T18:17:13Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://58c7bba4ace28215b062653a960e561b1dc1f890f2baf450145638a065af038b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://58c7bba4ace28215b062653a960e561b1dc1f890f2baf450145638a065af038b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-08T18:17:11Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-08T18:17:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T18:17:10Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T18:17:34Z is after 2025-08-24T17:21:41Z" Oct 08 18:17:34 crc kubenswrapper[4859]: I1008 18:17:34.520496 4859 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-08T18:17:29Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T18:17:29Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T18:17:34Z is after 2025-08-24T17:21:41Z" Oct 08 18:17:34 crc kubenswrapper[4859]: I1008 18:17:34.556972 4859 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-84k8x" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1c24078c-c9e8-4160-883a-f483b4b3dd23\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:17:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:17:31Z\\\",\\\"message\\\":\\\"containers with incomplete status: [cni-plugins bond-cni-plugin routeoverride-cni whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:17:31Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:17:31Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hk5xb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://18b291eb3344fb7c00ef5679dca9a6f7ed9d08cea684a803ff7dad2e8a18c585\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://18b291eb3344fb7c00ef5679dca9a6f7ed9d08cea684a803ff7dad2e8a18c585\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-08T18:17:32Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-08T18:17:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hk5xb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hk5xb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hk5xb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hk5xb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hk5xb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hk5xb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T18:17:31Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-84k8x\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T18:17:34Z is after 2025-08-24T17:21:41Z" Oct 08 18:17:34 crc kubenswrapper[4859]: I1008 18:17:34.595422 4859 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-8plkg" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"92b2d7de-31cc-4a91-95a5-ed5ea964b028\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:17:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:17:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:17:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:17:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fa00b8345d40b74a917ddc23cc1b2c5c684a9ca61b6119d090c92cc1439b71c4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T18:17:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q9d7m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T18:17:31Z\\\"}}\" for pod \"openshift-multus\"/\"multus-8plkg\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T18:17:34Z is after 2025-08-24T17:21:41Z" Oct 08 18:17:34 crc kubenswrapper[4859]: I1008 18:17:34.632583 4859 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-08T18:17:29Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T18:17:29Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T18:17:34Z is after 2025-08-24T17:21:41Z" Oct 08 18:17:34 crc kubenswrapper[4859]: I1008 18:17:34.669128 4859 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-4g8gf" event={"ID":"1dff864e-b75d-4e0b-b182-75f710eac8df","Type":"ContainerStarted","Data":"b8009a42cc4c9a816457a0678bdb2f7b7c53d0c8604e007a6fb141f66d06a516"} Oct 08 18:17:34 crc kubenswrapper[4859]: I1008 18:17:34.669453 4859 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-4g8gf" event={"ID":"1dff864e-b75d-4e0b-b182-75f710eac8df","Type":"ContainerStarted","Data":"676511a11ed4102084b57ebdd5f3fda4969741b2bc8443b415c4bf94f39ae4c9"} Oct 08 18:17:34 crc kubenswrapper[4859]: I1008 18:17:34.669553 4859 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-4g8gf" event={"ID":"1dff864e-b75d-4e0b-b182-75f710eac8df","Type":"ContainerStarted","Data":"669f5f3c6d3c1ec0c4b4870e090b0a853adc01727c23383e595715b476c32668"} Oct 08 18:17:34 crc kubenswrapper[4859]: I1008 18:17:34.669631 4859 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-4g8gf" event={"ID":"1dff864e-b75d-4e0b-b182-75f710eac8df","Type":"ContainerStarted","Data":"768ca274dedcc6847fc21d15234d3665fa9ccae371b5b6a882cfcb9b182f527d"} Oct 08 18:17:34 crc kubenswrapper[4859]: I1008 18:17:34.669721 4859 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-4g8gf" event={"ID":"1dff864e-b75d-4e0b-b182-75f710eac8df","Type":"ContainerStarted","Data":"4a74c1f0d21c90cdd7e776ee14e1f48757a04e2220fbd1f884449181ec6d1c3a"} Oct 08 18:17:34 crc kubenswrapper[4859]: I1008 18:17:34.669810 4859 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-4g8gf" event={"ID":"1dff864e-b75d-4e0b-b182-75f710eac8df","Type":"ContainerStarted","Data":"6ddc0e842bb8f81663406070023b0e5ed9cc99c98db0123b1c84f20da59fb5b5"} Oct 08 18:17:34 crc kubenswrapper[4859]: I1008 18:17:34.671726 4859 generic.go:334] "Generic (PLEG): container finished" podID="1c24078c-c9e8-4160-883a-f483b4b3dd23" containerID="160163f1a91d981583452db3c7a9244182d1c8ced45749a61b721d0d45c1dfe4" exitCode=0 Oct 08 18:17:34 crc kubenswrapper[4859]: I1008 18:17:34.671838 4859 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-84k8x" event={"ID":"1c24078c-c9e8-4160-883a-f483b4b3dd23","Type":"ContainerDied","Data":"160163f1a91d981583452db3c7a9244182d1c8ced45749a61b721d0d45c1dfe4"} Oct 08 18:17:34 crc kubenswrapper[4859]: I1008 18:17:34.672798 4859 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-btw28" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"97bc5598-3ead-4300-a938-3aebbff1e8c6\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:17:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:17:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:17:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:17:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d6d4d6ba771cb5fde637449bba05660f634df7e833a0c8dad2b2c7f3a422bb2b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T18:17:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qcxsd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T18:17:31Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-btw28\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T18:17:34Z is after 2025-08-24T17:21:41Z" Oct 08 18:17:34 crc kubenswrapper[4859]: I1008 18:17:34.719340 4859 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-08T18:17:31Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T18:17:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T18:17:31Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c6d15905694b6bbd4999d49dc8f7a53eb6bfaa7adb59efb0d7e4973c21d1c7be\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T18:17:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T18:17:34Z is after 2025-08-24T17:21:41Z" Oct 08 18:17:34 crc kubenswrapper[4859]: I1008 18:17:34.753947 4859 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-08T18:17:33Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T18:17:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T18:17:33Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5d99e071bd083914418d9c128f38955f26295c08b1520f2dc774c99b1c843693\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T18:17:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T18:17:34Z is after 2025-08-24T17:21:41Z" Oct 08 18:17:34 crc kubenswrapper[4859]: I1008 18:17:34.794737 4859 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-08T18:17:29Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T18:17:29Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T18:17:34Z is after 2025-08-24T17:21:41Z" Oct 08 18:17:34 crc kubenswrapper[4859]: I1008 18:17:34.834261 4859 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-btw28" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"97bc5598-3ead-4300-a938-3aebbff1e8c6\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:17:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:17:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:17:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:17:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d6d4d6ba771cb5fde637449bba05660f634df7e833a0c8dad2b2c7f3a422bb2b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T18:17:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qcxsd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T18:17:31Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-btw28\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T18:17:34Z is after 2025-08-24T17:21:41Z" Oct 08 18:17:34 crc kubenswrapper[4859]: I1008 18:17:34.876194 4859 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-8plkg" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"92b2d7de-31cc-4a91-95a5-ed5ea964b028\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:17:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:17:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:17:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:17:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fa00b8345d40b74a917ddc23cc1b2c5c684a9ca61b6119d090c92cc1439b71c4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T18:17:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q9d7m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T18:17:31Z\\\"}}\" for pod \"openshift-multus\"/\"multus-8plkg\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T18:17:34Z is after 2025-08-24T17:21:41Z" Oct 08 18:17:34 crc kubenswrapper[4859]: I1008 18:17:34.914458 4859 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-08T18:17:33Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T18:17:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T18:17:33Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5d99e071bd083914418d9c128f38955f26295c08b1520f2dc774c99b1c843693\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T18:17:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T18:17:34Z is after 2025-08-24T17:21:41Z" Oct 08 18:17:34 crc kubenswrapper[4859]: I1008 18:17:34.955185 4859 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-08T18:17:31Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T18:17:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T18:17:31Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c6d15905694b6bbd4999d49dc8f7a53eb6bfaa7adb59efb0d7e4973c21d1c7be\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T18:17:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T18:17:34Z is after 2025-08-24T17:21:41Z" Oct 08 18:17:34 crc kubenswrapper[4859]: I1008 18:17:34.993917 4859 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"63a623de-5608-4228-b1c2-77188c276986\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:17:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:17:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:17:10Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:17:10Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:17:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d4786a8fb8abd96fb22e60733c974f599017407d589ed815fb8541fd3b64f696\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T18:17:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://183d481502ed5108d5e19e483b714c289f355ea3f8b0891883e5335b802c83b6\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T18:17:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ec848c46fd5dde0469b00d3cab14ba4a652533a265490fadbee7a8b8843ef249\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T18:17:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4fe1828b4d63ff8106bfcd15e1dc1c1bd1ba7f77777cdb23ff1b291d0418774c\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2e700b2a20c81b414617f5b5c018a738278a3fa44e6c6d14b2170051d0725431\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-08T18:17:24Z\\\",\\\"message\\\":\\\"W1008 18:17:13.670625 1 cmd.go:257] Using insecure, self-signed certificates\\\\nI1008 18:17:13.671202 1 crypto.go:601] Generating new CA for check-endpoints-signer@1759947433 cert, and key in /tmp/serving-cert-3867218267/serving-signer.crt, /tmp/serving-cert-3867218267/serving-signer.key\\\\nI1008 18:17:13.944463 1 observer_polling.go:159] Starting file observer\\\\nW1008 18:17:13.951958 1 builder.go:272] unable to get owner reference (falling back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": dial tcp [::1]:6443: connect: connection refused\\\\nI1008 18:17:13.952088 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1008 18:17:13.956755 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3867218267/tls.crt::/tmp/serving-cert-3867218267/tls.key\\\\\\\"\\\\nF1008 18:17:24.290991 1 cmd.go:182] error initializing delegating authentication: unable to load configmap based request-header-client-ca-file: Get \\\\\\\"https://localhost:6443/api/v1/namespaces/kube-system/configmaps/extension-apiserver-authentication\\\\\\\": net/http: TLS handshake timeout\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-08T18:17:13Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T18:17:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c65b66f62d325ca10ef8d5d31f04529ccf4e691cf3303d231e7a414fb400e4d6\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T18:17:13Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://58c7bba4ace28215b062653a960e561b1dc1f890f2baf450145638a065af038b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://58c7bba4ace28215b062653a960e561b1dc1f890f2baf450145638a065af038b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-08T18:17:11Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-08T18:17:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T18:17:10Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T18:17:34Z is after 2025-08-24T17:21:41Z" Oct 08 18:17:35 crc kubenswrapper[4859]: I1008 18:17:35.005112 4859 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-image-registry/node-ca-btjmq"] Oct 08 18:17:35 crc kubenswrapper[4859]: I1008 18:17:35.005539 4859 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/node-ca-btjmq" Oct 08 18:17:35 crc kubenswrapper[4859]: I1008 18:17:35.023793 4859 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"node-ca-dockercfg-4777p" Oct 08 18:17:35 crc kubenswrapper[4859]: I1008 18:17:35.045275 4859 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-image-registry"/"openshift-service-ca.crt" Oct 08 18:17:35 crc kubenswrapper[4859]: I1008 18:17:35.065083 4859 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-image-registry"/"image-registry-certificates" Oct 08 18:17:35 crc kubenswrapper[4859]: I1008 18:17:35.084142 4859 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-image-registry"/"kube-root-ca.crt" Oct 08 18:17:35 crc kubenswrapper[4859]: I1008 18:17:35.117621 4859 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-08T18:17:31Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T18:17:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T18:17:31Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://99e2d51bd7757498dae60949eea6d732b5f7b481fa00f808dd5308a530a1daf1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T18:17:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e32aaba93699944665660ae41e9b5779178c1354c31e15ad8188bf8c3c8fa500\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T18:17:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T18:17:35Z is after 2025-08-24T17:21:41Z" Oct 08 18:17:35 crc kubenswrapper[4859]: I1008 18:17:35.159656 4859 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/3b8605f7-fed7-493f-9b4f-6a6358adf907-host\") pod \"node-ca-btjmq\" (UID: \"3b8605f7-fed7-493f-9b4f-6a6358adf907\") " pod="openshift-image-registry/node-ca-btjmq" Oct 08 18:17:35 crc kubenswrapper[4859]: I1008 18:17:35.159805 4859 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serviceca\" (UniqueName: \"kubernetes.io/configmap/3b8605f7-fed7-493f-9b4f-6a6358adf907-serviceca\") pod \"node-ca-btjmq\" (UID: \"3b8605f7-fed7-493f-9b4f-6a6358adf907\") " pod="openshift-image-registry/node-ca-btjmq" Oct 08 18:17:35 crc kubenswrapper[4859]: I1008 18:17:35.159830 4859 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-xspg4\" (UniqueName: \"kubernetes.io/projected/3b8605f7-fed7-493f-9b4f-6a6358adf907-kube-api-access-xspg4\") pod \"node-ca-btjmq\" (UID: \"3b8605f7-fed7-493f-9b4f-6a6358adf907\") " pod="openshift-image-registry/node-ca-btjmq" Oct 08 18:17:35 crc kubenswrapper[4859]: I1008 18:17:35.161143 4859 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-08T18:17:29Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T18:17:29Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T18:17:35Z is after 2025-08-24T17:21:41Z" Oct 08 18:17:35 crc kubenswrapper[4859]: I1008 18:17:35.194317 4859 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-82s52" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b23a6a6c-9d92-4e7b-840e-55cfda873a2d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:17:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:17:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:17:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:17:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://71302c766485b9f55743d85f97440ff578cd44ba1e5a6f5b1c024c0e808072d2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T18:17:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6vx22\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c23ea3580998adf90621a0155c20ee4bd590c512629812174e95c28857bcfef9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T18:17:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6vx22\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T18:17:31Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-82s52\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T18:17:35Z is after 2025-08-24T17:21:41Z" Oct 08 18:17:35 crc kubenswrapper[4859]: I1008 18:17:35.237679 4859 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-4g8gf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1dff864e-b75d-4e0b-b182-75f710eac8df\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:17:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:17:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:17:32Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:17:32Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s8jnn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s8jnn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s8jnn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s8jnn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s8jnn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s8jnn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s8jnn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s8jnn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7a211d3da02b173867ff3a912eecd3b379483c7cc77e6b91122edd1ffd30fd53\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7a211d3da02b173867ff3a912eecd3b379483c7cc77e6b91122edd1ffd30fd53\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-08T18:17:32Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-08T18:17:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s8jnn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T18:17:32Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-4g8gf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T18:17:35Z is after 2025-08-24T17:21:41Z" Oct 08 18:17:35 crc kubenswrapper[4859]: I1008 18:17:35.260852 4859 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/3b8605f7-fed7-493f-9b4f-6a6358adf907-host\") pod \"node-ca-btjmq\" (UID: \"3b8605f7-fed7-493f-9b4f-6a6358adf907\") " pod="openshift-image-registry/node-ca-btjmq" Oct 08 18:17:35 crc kubenswrapper[4859]: I1008 18:17:35.260925 4859 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serviceca\" (UniqueName: \"kubernetes.io/configmap/3b8605f7-fed7-493f-9b4f-6a6358adf907-serviceca\") pod \"node-ca-btjmq\" (UID: \"3b8605f7-fed7-493f-9b4f-6a6358adf907\") " pod="openshift-image-registry/node-ca-btjmq" Oct 08 18:17:35 crc kubenswrapper[4859]: I1008 18:17:35.260955 4859 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-xspg4\" (UniqueName: \"kubernetes.io/projected/3b8605f7-fed7-493f-9b4f-6a6358adf907-kube-api-access-xspg4\") pod \"node-ca-btjmq\" (UID: \"3b8605f7-fed7-493f-9b4f-6a6358adf907\") " pod="openshift-image-registry/node-ca-btjmq" Oct 08 18:17:35 crc kubenswrapper[4859]: I1008 18:17:35.260995 4859 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host\" (UniqueName: \"kubernetes.io/host-path/3b8605f7-fed7-493f-9b4f-6a6358adf907-host\") pod \"node-ca-btjmq\" (UID: \"3b8605f7-fed7-493f-9b4f-6a6358adf907\") " pod="openshift-image-registry/node-ca-btjmq" Oct 08 18:17:35 crc kubenswrapper[4859]: I1008 18:17:35.262416 4859 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serviceca\" (UniqueName: \"kubernetes.io/configmap/3b8605f7-fed7-493f-9b4f-6a6358adf907-serviceca\") pod \"node-ca-btjmq\" (UID: \"3b8605f7-fed7-493f-9b4f-6a6358adf907\") " pod="openshift-image-registry/node-ca-btjmq" Oct 08 18:17:35 crc kubenswrapper[4859]: I1008 18:17:35.273498 4859 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-08T18:17:29Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T18:17:29Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T18:17:35Z is after 2025-08-24T17:21:41Z" Oct 08 18:17:35 crc kubenswrapper[4859]: I1008 18:17:35.303529 4859 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-xspg4\" (UniqueName: \"kubernetes.io/projected/3b8605f7-fed7-493f-9b4f-6a6358adf907-kube-api-access-xspg4\") pod \"node-ca-btjmq\" (UID: \"3b8605f7-fed7-493f-9b4f-6a6358adf907\") " pod="openshift-image-registry/node-ca-btjmq" Oct 08 18:17:35 crc kubenswrapper[4859]: I1008 18:17:35.323319 4859 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/node-ca-btjmq" Oct 08 18:17:35 crc kubenswrapper[4859]: I1008 18:17:35.337588 4859 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-84k8x" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1c24078c-c9e8-4160-883a-f483b4b3dd23\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:17:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:17:31Z\\\",\\\"message\\\":\\\"containers with incomplete status: [bond-cni-plugin routeoverride-cni whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:17:31Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:17:31Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hk5xb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://18b291eb3344fb7c00ef5679dca9a6f7ed9d08cea684a803ff7dad2e8a18c585\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://18b291eb3344fb7c00ef5679dca9a6f7ed9d08cea684a803ff7dad2e8a18c585\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-08T18:17:32Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-08T18:17:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hk5xb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://160163f1a91d981583452db3c7a9244182d1c8ced45749a61b721d0d45c1dfe4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://160163f1a91d981583452db3c7a9244182d1c8ced45749a61b721d0d45c1dfe4\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-08T18:17:34Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-08T18:17:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hk5xb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hk5xb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hk5xb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hk5xb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hk5xb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T18:17:31Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-84k8x\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T18:17:35Z is after 2025-08-24T17:21:41Z" Oct 08 18:17:35 crc kubenswrapper[4859]: W1008 18:17:35.345615 4859 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod3b8605f7_fed7_493f_9b4f_6a6358adf907.slice/crio-715700021ed48379cb5bdce45ec111623537b0b37c15f76aab3133b98a35bb8b WatchSource:0}: Error finding container 715700021ed48379cb5bdce45ec111623537b0b37c15f76aab3133b98a35bb8b: Status 404 returned error can't find the container with id 715700021ed48379cb5bdce45ec111623537b0b37c15f76aab3133b98a35bb8b Oct 08 18:17:35 crc kubenswrapper[4859]: I1008 18:17:35.373010 4859 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4efc16fb-e679-464d-8b3c-0f46f082593e\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:17:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:17:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:17:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:17:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:17:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://aa8f8ea895272cd79a08d1c47cd7863a1ec10e21bb60d12aa10862a56c99e4bf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T18:17:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5c4e95d8a712ab219455564539a90286ae48bbc46ddc322a5bfd96c1a8268110\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T18:17:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://bae6ecfd8b446ef04d393d823e5ffe37fe749337c0ca17fde79c5f88d5e31393\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T18:17:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6411d1e7e18d5ceadaaceed5a8a6b20a0cacc55cede793ca9fb4dcad70c64fc6\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T18:17:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T18:17:10Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T18:17:35Z is after 2025-08-24T17:21:41Z" Oct 08 18:17:35 crc kubenswrapper[4859]: I1008 18:17:35.416018 4859 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-08T18:17:29Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T18:17:29Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T18:17:35Z is after 2025-08-24T17:21:41Z" Oct 08 18:17:35 crc kubenswrapper[4859]: I1008 18:17:35.451415 4859 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-btw28" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"97bc5598-3ead-4300-a938-3aebbff1e8c6\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:17:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:17:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:17:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:17:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d6d4d6ba771cb5fde637449bba05660f634df7e833a0c8dad2b2c7f3a422bb2b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T18:17:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qcxsd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T18:17:31Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-btw28\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T18:17:35Z is after 2025-08-24T17:21:41Z" Oct 08 18:17:35 crc kubenswrapper[4859]: I1008 18:17:35.469445 4859 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 08 18:17:35 crc kubenswrapper[4859]: I1008 18:17:35.469493 4859 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 08 18:17:35 crc kubenswrapper[4859]: E1008 18:17:35.469739 4859 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 08 18:17:35 crc kubenswrapper[4859]: E1008 18:17:35.469960 4859 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 08 18:17:35 crc kubenswrapper[4859]: I1008 18:17:35.470277 4859 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 08 18:17:35 crc kubenswrapper[4859]: E1008 18:17:35.470593 4859 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 08 18:17:35 crc kubenswrapper[4859]: I1008 18:17:35.495199 4859 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-8plkg" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"92b2d7de-31cc-4a91-95a5-ed5ea964b028\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:17:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:17:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:17:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:17:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fa00b8345d40b74a917ddc23cc1b2c5c684a9ca61b6119d090c92cc1439b71c4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T18:17:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q9d7m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T18:17:31Z\\\"}}\" for pod \"openshift-multus\"/\"multus-8plkg\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T18:17:35Z is after 2025-08-24T17:21:41Z" Oct 08 18:17:35 crc kubenswrapper[4859]: I1008 18:17:35.544099 4859 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-btjmq" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b8605f7-fed7-493f-9b4f-6a6358adf907\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:17:35Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:17:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:17:35Z\\\",\\\"message\\\":\\\"containers with unready status: [node-ca]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:17:35Z\\\",\\\"message\\\":\\\"containers with unready status: [node-ca]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xspg4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T18:17:35Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-btjmq\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T18:17:35Z is after 2025-08-24T17:21:41Z" Oct 08 18:17:35 crc kubenswrapper[4859]: I1008 18:17:35.580867 4859 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-08T18:17:31Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T18:17:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T18:17:31Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c6d15905694b6bbd4999d49dc8f7a53eb6bfaa7adb59efb0d7e4973c21d1c7be\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T18:17:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T18:17:35Z is after 2025-08-24T17:21:41Z" Oct 08 18:17:35 crc kubenswrapper[4859]: I1008 18:17:35.639988 4859 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-08T18:17:33Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T18:17:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T18:17:33Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5d99e071bd083914418d9c128f38955f26295c08b1520f2dc774c99b1c843693\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T18:17:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T18:17:35Z is after 2025-08-24T17:21:41Z" Oct 08 18:17:35 crc kubenswrapper[4859]: I1008 18:17:35.678014 4859 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/node-ca-btjmq" event={"ID":"3b8605f7-fed7-493f-9b4f-6a6358adf907","Type":"ContainerStarted","Data":"14e9b6c4a4403f36adb78fcf1135a620ca6fc56ed974c9639dcdfc59e6917559"} Oct 08 18:17:35 crc kubenswrapper[4859]: I1008 18:17:35.678079 4859 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/node-ca-btjmq" event={"ID":"3b8605f7-fed7-493f-9b4f-6a6358adf907","Type":"ContainerStarted","Data":"715700021ed48379cb5bdce45ec111623537b0b37c15f76aab3133b98a35bb8b"} Oct 08 18:17:35 crc kubenswrapper[4859]: I1008 18:17:35.680831 4859 generic.go:334] "Generic (PLEG): container finished" podID="1c24078c-c9e8-4160-883a-f483b4b3dd23" containerID="51e9b53f3847ac53f378ad83870debf8e71518d4edbc7fc0dc66644818e2103e" exitCode=0 Oct 08 18:17:35 crc kubenswrapper[4859]: I1008 18:17:35.680890 4859 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-84k8x" event={"ID":"1c24078c-c9e8-4160-883a-f483b4b3dd23","Type":"ContainerDied","Data":"51e9b53f3847ac53f378ad83870debf8e71518d4edbc7fc0dc66644818e2103e"} Oct 08 18:17:35 crc kubenswrapper[4859]: I1008 18:17:35.698248 4859 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-4g8gf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1dff864e-b75d-4e0b-b182-75f710eac8df\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:17:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:17:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:17:32Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:17:32Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s8jnn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s8jnn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s8jnn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s8jnn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s8jnn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s8jnn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s8jnn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s8jnn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7a211d3da02b173867ff3a912eecd3b379483c7cc77e6b91122edd1ffd30fd53\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7a211d3da02b173867ff3a912eecd3b379483c7cc77e6b91122edd1ffd30fd53\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-08T18:17:32Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-08T18:17:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s8jnn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T18:17:32Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-4g8gf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T18:17:35Z is after 2025-08-24T17:21:41Z" Oct 08 18:17:35 crc kubenswrapper[4859]: I1008 18:17:35.724636 4859 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"63a623de-5608-4228-b1c2-77188c276986\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:17:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:17:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:17:10Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:17:10Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:17:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d4786a8fb8abd96fb22e60733c974f599017407d589ed815fb8541fd3b64f696\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T18:17:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://183d481502ed5108d5e19e483b714c289f355ea3f8b0891883e5335b802c83b6\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T18:17:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ec848c46fd5dde0469b00d3cab14ba4a652533a265490fadbee7a8b8843ef249\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T18:17:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4fe1828b4d63ff8106bfcd15e1dc1c1bd1ba7f77777cdb23ff1b291d0418774c\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2e700b2a20c81b414617f5b5c018a738278a3fa44e6c6d14b2170051d0725431\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-08T18:17:24Z\\\",\\\"message\\\":\\\"W1008 18:17:13.670625 1 cmd.go:257] Using insecure, self-signed certificates\\\\nI1008 18:17:13.671202 1 crypto.go:601] Generating new CA for check-endpoints-signer@1759947433 cert, and key in /tmp/serving-cert-3867218267/serving-signer.crt, /tmp/serving-cert-3867218267/serving-signer.key\\\\nI1008 18:17:13.944463 1 observer_polling.go:159] Starting file observer\\\\nW1008 18:17:13.951958 1 builder.go:272] unable to get owner reference (falling back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": dial tcp [::1]:6443: connect: connection refused\\\\nI1008 18:17:13.952088 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1008 18:17:13.956755 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3867218267/tls.crt::/tmp/serving-cert-3867218267/tls.key\\\\\\\"\\\\nF1008 18:17:24.290991 1 cmd.go:182] error initializing delegating authentication: unable to load configmap based request-header-client-ca-file: Get \\\\\\\"https://localhost:6443/api/v1/namespaces/kube-system/configmaps/extension-apiserver-authentication\\\\\\\": net/http: TLS handshake timeout\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-08T18:17:13Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T18:17:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c65b66f62d325ca10ef8d5d31f04529ccf4e691cf3303d231e7a414fb400e4d6\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T18:17:13Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://58c7bba4ace28215b062653a960e561b1dc1f890f2baf450145638a065af038b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://58c7bba4ace28215b062653a960e561b1dc1f890f2baf450145638a065af038b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-08T18:17:11Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-08T18:17:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T18:17:10Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T18:17:35Z is after 2025-08-24T17:21:41Z" Oct 08 18:17:35 crc kubenswrapper[4859]: I1008 18:17:35.740265 4859 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-08T18:17:31Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T18:17:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T18:17:31Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://99e2d51bd7757498dae60949eea6d732b5f7b481fa00f808dd5308a530a1daf1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T18:17:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e32aaba93699944665660ae41e9b5779178c1354c31e15ad8188bf8c3c8fa500\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T18:17:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T18:17:35Z is after 2025-08-24T17:21:41Z" Oct 08 18:17:35 crc kubenswrapper[4859]: I1008 18:17:35.772536 4859 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-08T18:17:29Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T18:17:29Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T18:17:35Z is after 2025-08-24T17:21:41Z" Oct 08 18:17:35 crc kubenswrapper[4859]: I1008 18:17:35.811528 4859 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-82s52" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b23a6a6c-9d92-4e7b-840e-55cfda873a2d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:17:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:17:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:17:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:17:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://71302c766485b9f55743d85f97440ff578cd44ba1e5a6f5b1c024c0e808072d2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T18:17:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6vx22\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c23ea3580998adf90621a0155c20ee4bd590c512629812174e95c28857bcfef9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T18:17:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6vx22\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T18:17:31Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-82s52\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T18:17:35Z is after 2025-08-24T17:21:41Z" Oct 08 18:17:35 crc kubenswrapper[4859]: I1008 18:17:35.851046 4859 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4efc16fb-e679-464d-8b3c-0f46f082593e\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:17:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:17:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:17:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:17:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:17:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://aa8f8ea895272cd79a08d1c47cd7863a1ec10e21bb60d12aa10862a56c99e4bf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T18:17:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5c4e95d8a712ab219455564539a90286ae48bbc46ddc322a5bfd96c1a8268110\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T18:17:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://bae6ecfd8b446ef04d393d823e5ffe37fe749337c0ca17fde79c5f88d5e31393\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T18:17:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6411d1e7e18d5ceadaaceed5a8a6b20a0cacc55cede793ca9fb4dcad70c64fc6\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T18:17:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T18:17:10Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T18:17:35Z is after 2025-08-24T17:21:41Z" Oct 08 18:17:35 crc kubenswrapper[4859]: I1008 18:17:35.893950 4859 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-08T18:17:29Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T18:17:29Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T18:17:35Z is after 2025-08-24T17:21:41Z" Oct 08 18:17:35 crc kubenswrapper[4859]: I1008 18:17:35.934709 4859 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-84k8x" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1c24078c-c9e8-4160-883a-f483b4b3dd23\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:17:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:17:31Z\\\",\\\"message\\\":\\\"containers with incomplete status: [bond-cni-plugin routeoverride-cni whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:17:31Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:17:31Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hk5xb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://18b291eb3344fb7c00ef5679dca9a6f7ed9d08cea684a803ff7dad2e8a18c585\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://18b291eb3344fb7c00ef5679dca9a6f7ed9d08cea684a803ff7dad2e8a18c585\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-08T18:17:32Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-08T18:17:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hk5xb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://160163f1a91d981583452db3c7a9244182d1c8ced45749a61b721d0d45c1dfe4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://160163f1a91d981583452db3c7a9244182d1c8ced45749a61b721d0d45c1dfe4\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-08T18:17:34Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-08T18:17:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hk5xb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hk5xb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hk5xb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hk5xb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hk5xb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T18:17:31Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-84k8x\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T18:17:35Z is after 2025-08-24T17:21:41Z" Oct 08 18:17:35 crc kubenswrapper[4859]: I1008 18:17:35.972833 4859 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4efc16fb-e679-464d-8b3c-0f46f082593e\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:17:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:17:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:17:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:17:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:17:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://aa8f8ea895272cd79a08d1c47cd7863a1ec10e21bb60d12aa10862a56c99e4bf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T18:17:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5c4e95d8a712ab219455564539a90286ae48bbc46ddc322a5bfd96c1a8268110\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T18:17:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://bae6ecfd8b446ef04d393d823e5ffe37fe749337c0ca17fde79c5f88d5e31393\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T18:17:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6411d1e7e18d5ceadaaceed5a8a6b20a0cacc55cede793ca9fb4dcad70c64fc6\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T18:17:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T18:17:10Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T18:17:35Z is after 2025-08-24T17:21:41Z" Oct 08 18:17:36 crc kubenswrapper[4859]: I1008 18:17:36.012802 4859 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-08T18:17:29Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T18:17:29Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T18:17:36Z is after 2025-08-24T17:21:41Z" Oct 08 18:17:36 crc kubenswrapper[4859]: I1008 18:17:36.055144 4859 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-84k8x" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1c24078c-c9e8-4160-883a-f483b4b3dd23\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:17:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:17:31Z\\\",\\\"message\\\":\\\"containers with incomplete status: [routeoverride-cni whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:17:31Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:17:31Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hk5xb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://18b291eb3344fb7c00ef5679dca9a6f7ed9d08cea684a803ff7dad2e8a18c585\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://18b291eb3344fb7c00ef5679dca9a6f7ed9d08cea684a803ff7dad2e8a18c585\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-08T18:17:32Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-08T18:17:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hk5xb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://160163f1a91d981583452db3c7a9244182d1c8ced45749a61b721d0d45c1dfe4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://160163f1a91d981583452db3c7a9244182d1c8ced45749a61b721d0d45c1dfe4\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-08T18:17:34Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-08T18:17:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hk5xb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://51e9b53f3847ac53f378ad83870debf8e71518d4edbc7fc0dc66644818e2103e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://51e9b53f3847ac53f378ad83870debf8e71518d4edbc7fc0dc66644818e2103e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-08T18:17:34Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-08T18:17:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hk5xb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hk5xb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hk5xb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hk5xb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T18:17:31Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-84k8x\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T18:17:36Z is after 2025-08-24T17:21:41Z" Oct 08 18:17:36 crc kubenswrapper[4859]: I1008 18:17:36.092263 4859 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-btjmq" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b8605f7-fed7-493f-9b4f-6a6358adf907\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:17:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:17:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:17:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:17:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://14e9b6c4a4403f36adb78fcf1135a620ca6fc56ed974c9639dcdfc59e6917559\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T18:17:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xspg4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T18:17:35Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-btjmq\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T18:17:36Z is after 2025-08-24T17:21:41Z" Oct 08 18:17:36 crc kubenswrapper[4859]: I1008 18:17:36.133195 4859 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-08T18:17:29Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T18:17:29Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T18:17:36Z is after 2025-08-24T17:21:41Z" Oct 08 18:17:36 crc kubenswrapper[4859]: I1008 18:17:36.172213 4859 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-btw28" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"97bc5598-3ead-4300-a938-3aebbff1e8c6\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:17:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:17:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:17:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:17:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d6d4d6ba771cb5fde637449bba05660f634df7e833a0c8dad2b2c7f3a422bb2b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T18:17:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qcxsd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T18:17:31Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-btw28\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T18:17:36Z is after 2025-08-24T17:21:41Z" Oct 08 18:17:36 crc kubenswrapper[4859]: I1008 18:17:36.215971 4859 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-8plkg" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"92b2d7de-31cc-4a91-95a5-ed5ea964b028\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:17:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:17:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:17:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:17:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fa00b8345d40b74a917ddc23cc1b2c5c684a9ca61b6119d090c92cc1439b71c4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T18:17:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q9d7m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T18:17:31Z\\\"}}\" for pod \"openshift-multus\"/\"multus-8plkg\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T18:17:36Z is after 2025-08-24T17:21:41Z" Oct 08 18:17:36 crc kubenswrapper[4859]: I1008 18:17:36.256431 4859 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-08T18:17:31Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T18:17:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T18:17:31Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c6d15905694b6bbd4999d49dc8f7a53eb6bfaa7adb59efb0d7e4973c21d1c7be\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T18:17:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T18:17:36Z is after 2025-08-24T17:21:41Z" Oct 08 18:17:36 crc kubenswrapper[4859]: I1008 18:17:36.299628 4859 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-08T18:17:33Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T18:17:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T18:17:33Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5d99e071bd083914418d9c128f38955f26295c08b1520f2dc774c99b1c843693\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T18:17:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T18:17:36Z is after 2025-08-24T17:21:41Z" Oct 08 18:17:36 crc kubenswrapper[4859]: I1008 18:17:36.336877 4859 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-08T18:17:29Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T18:17:29Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T18:17:36Z is after 2025-08-24T17:21:41Z" Oct 08 18:17:36 crc kubenswrapper[4859]: I1008 18:17:36.375771 4859 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-82s52" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b23a6a6c-9d92-4e7b-840e-55cfda873a2d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:17:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:17:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:17:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:17:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://71302c766485b9f55743d85f97440ff578cd44ba1e5a6f5b1c024c0e808072d2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T18:17:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6vx22\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c23ea3580998adf90621a0155c20ee4bd590c512629812174e95c28857bcfef9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T18:17:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6vx22\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T18:17:31Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-82s52\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T18:17:36Z is after 2025-08-24T17:21:41Z" Oct 08 18:17:36 crc kubenswrapper[4859]: I1008 18:17:36.426231 4859 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-4g8gf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1dff864e-b75d-4e0b-b182-75f710eac8df\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:17:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:17:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:17:32Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:17:32Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s8jnn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s8jnn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s8jnn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s8jnn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s8jnn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s8jnn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s8jnn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s8jnn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7a211d3da02b173867ff3a912eecd3b379483c7cc77e6b91122edd1ffd30fd53\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7a211d3da02b173867ff3a912eecd3b379483c7cc77e6b91122edd1ffd30fd53\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-08T18:17:32Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-08T18:17:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s8jnn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T18:17:32Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-4g8gf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T18:17:36Z is after 2025-08-24T17:21:41Z" Oct 08 18:17:36 crc kubenswrapper[4859]: I1008 18:17:36.457393 4859 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"63a623de-5608-4228-b1c2-77188c276986\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:17:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:17:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:17:10Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:17:10Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:17:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d4786a8fb8abd96fb22e60733c974f599017407d589ed815fb8541fd3b64f696\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T18:17:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://183d481502ed5108d5e19e483b714c289f355ea3f8b0891883e5335b802c83b6\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T18:17:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ec848c46fd5dde0469b00d3cab14ba4a652533a265490fadbee7a8b8843ef249\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T18:17:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4fe1828b4d63ff8106bfcd15e1dc1c1bd1ba7f77777cdb23ff1b291d0418774c\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2e700b2a20c81b414617f5b5c018a738278a3fa44e6c6d14b2170051d0725431\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-08T18:17:24Z\\\",\\\"message\\\":\\\"W1008 18:17:13.670625 1 cmd.go:257] Using insecure, self-signed certificates\\\\nI1008 18:17:13.671202 1 crypto.go:601] Generating new CA for check-endpoints-signer@1759947433 cert, and key in /tmp/serving-cert-3867218267/serving-signer.crt, /tmp/serving-cert-3867218267/serving-signer.key\\\\nI1008 18:17:13.944463 1 observer_polling.go:159] Starting file observer\\\\nW1008 18:17:13.951958 1 builder.go:272] unable to get owner reference (falling back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": dial tcp [::1]:6443: connect: connection refused\\\\nI1008 18:17:13.952088 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1008 18:17:13.956755 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3867218267/tls.crt::/tmp/serving-cert-3867218267/tls.key\\\\\\\"\\\\nF1008 18:17:24.290991 1 cmd.go:182] error initializing delegating authentication: unable to load configmap based request-header-client-ca-file: Get \\\\\\\"https://localhost:6443/api/v1/namespaces/kube-system/configmaps/extension-apiserver-authentication\\\\\\\": net/http: TLS handshake timeout\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-08T18:17:13Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T18:17:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c65b66f62d325ca10ef8d5d31f04529ccf4e691cf3303d231e7a414fb400e4d6\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T18:17:13Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://58c7bba4ace28215b062653a960e561b1dc1f890f2baf450145638a065af038b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://58c7bba4ace28215b062653a960e561b1dc1f890f2baf450145638a065af038b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-08T18:17:11Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-08T18:17:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T18:17:10Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T18:17:36Z is after 2025-08-24T17:21:41Z" Oct 08 18:17:36 crc kubenswrapper[4859]: I1008 18:17:36.497367 4859 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-08T18:17:31Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T18:17:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T18:17:31Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://99e2d51bd7757498dae60949eea6d732b5f7b481fa00f808dd5308a530a1daf1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T18:17:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e32aaba93699944665660ae41e9b5779178c1354c31e15ad8188bf8c3c8fa500\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T18:17:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T18:17:36Z is after 2025-08-24T17:21:41Z" Oct 08 18:17:36 crc kubenswrapper[4859]: I1008 18:17:36.688530 4859 generic.go:334] "Generic (PLEG): container finished" podID="1c24078c-c9e8-4160-883a-f483b4b3dd23" containerID="1ff118919739759ac01b5cbe3240e6673db54eaa16353c609d670071ed7964cb" exitCode=0 Oct 08 18:17:36 crc kubenswrapper[4859]: I1008 18:17:36.688593 4859 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-84k8x" event={"ID":"1c24078c-c9e8-4160-883a-f483b4b3dd23","Type":"ContainerDied","Data":"1ff118919739759ac01b5cbe3240e6673db54eaa16353c609d670071ed7964cb"} Oct 08 18:17:36 crc kubenswrapper[4859]: I1008 18:17:36.709274 4859 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-82s52" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b23a6a6c-9d92-4e7b-840e-55cfda873a2d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:17:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:17:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:17:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:17:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://71302c766485b9f55743d85f97440ff578cd44ba1e5a6f5b1c024c0e808072d2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T18:17:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6vx22\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c23ea3580998adf90621a0155c20ee4bd590c512629812174e95c28857bcfef9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T18:17:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6vx22\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T18:17:31Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-82s52\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T18:17:36Z is after 2025-08-24T17:21:41Z" Oct 08 18:17:36 crc kubenswrapper[4859]: I1008 18:17:36.740443 4859 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-4g8gf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1dff864e-b75d-4e0b-b182-75f710eac8df\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:17:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:17:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:17:32Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:17:32Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s8jnn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s8jnn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s8jnn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s8jnn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s8jnn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s8jnn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s8jnn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s8jnn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7a211d3da02b173867ff3a912eecd3b379483c7cc77e6b91122edd1ffd30fd53\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7a211d3da02b173867ff3a912eecd3b379483c7cc77e6b91122edd1ffd30fd53\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-08T18:17:32Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-08T18:17:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s8jnn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T18:17:32Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-4g8gf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T18:17:36Z is after 2025-08-24T17:21:41Z" Oct 08 18:17:36 crc kubenswrapper[4859]: I1008 18:17:36.764331 4859 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"63a623de-5608-4228-b1c2-77188c276986\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:17:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:17:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:17:10Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:17:10Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:17:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d4786a8fb8abd96fb22e60733c974f599017407d589ed815fb8541fd3b64f696\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T18:17:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://183d481502ed5108d5e19e483b714c289f355ea3f8b0891883e5335b802c83b6\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T18:17:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ec848c46fd5dde0469b00d3cab14ba4a652533a265490fadbee7a8b8843ef249\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T18:17:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4fe1828b4d63ff8106bfcd15e1dc1c1bd1ba7f77777cdb23ff1b291d0418774c\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2e700b2a20c81b414617f5b5c018a738278a3fa44e6c6d14b2170051d0725431\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-08T18:17:24Z\\\",\\\"message\\\":\\\"W1008 18:17:13.670625 1 cmd.go:257] Using insecure, self-signed certificates\\\\nI1008 18:17:13.671202 1 crypto.go:601] Generating new CA for check-endpoints-signer@1759947433 cert, and key in /tmp/serving-cert-3867218267/serving-signer.crt, /tmp/serving-cert-3867218267/serving-signer.key\\\\nI1008 18:17:13.944463 1 observer_polling.go:159] Starting file observer\\\\nW1008 18:17:13.951958 1 builder.go:272] unable to get owner reference (falling back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": dial tcp [::1]:6443: connect: connection refused\\\\nI1008 18:17:13.952088 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1008 18:17:13.956755 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3867218267/tls.crt::/tmp/serving-cert-3867218267/tls.key\\\\\\\"\\\\nF1008 18:17:24.290991 1 cmd.go:182] error initializing delegating authentication: unable to load configmap based request-header-client-ca-file: Get \\\\\\\"https://localhost:6443/api/v1/namespaces/kube-system/configmaps/extension-apiserver-authentication\\\\\\\": net/http: TLS handshake timeout\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-08T18:17:13Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T18:17:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c65b66f62d325ca10ef8d5d31f04529ccf4e691cf3303d231e7a414fb400e4d6\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T18:17:13Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://58c7bba4ace28215b062653a960e561b1dc1f890f2baf450145638a065af038b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://58c7bba4ace28215b062653a960e561b1dc1f890f2baf450145638a065af038b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-08T18:17:11Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-08T18:17:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T18:17:10Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T18:17:36Z is after 2025-08-24T17:21:41Z" Oct 08 18:17:36 crc kubenswrapper[4859]: I1008 18:17:36.784538 4859 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-08T18:17:31Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T18:17:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T18:17:31Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://99e2d51bd7757498dae60949eea6d732b5f7b481fa00f808dd5308a530a1daf1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T18:17:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e32aaba93699944665660ae41e9b5779178c1354c31e15ad8188bf8c3c8fa500\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T18:17:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T18:17:36Z is after 2025-08-24T17:21:41Z" Oct 08 18:17:36 crc kubenswrapper[4859]: I1008 18:17:36.800039 4859 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-08T18:17:29Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T18:17:29Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T18:17:36Z is after 2025-08-24T17:21:41Z" Oct 08 18:17:36 crc kubenswrapper[4859]: I1008 18:17:36.819655 4859 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4efc16fb-e679-464d-8b3c-0f46f082593e\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:17:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:17:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:17:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:17:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:17:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://aa8f8ea895272cd79a08d1c47cd7863a1ec10e21bb60d12aa10862a56c99e4bf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T18:17:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5c4e95d8a712ab219455564539a90286ae48bbc46ddc322a5bfd96c1a8268110\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T18:17:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://bae6ecfd8b446ef04d393d823e5ffe37fe749337c0ca17fde79c5f88d5e31393\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T18:17:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6411d1e7e18d5ceadaaceed5a8a6b20a0cacc55cede793ca9fb4dcad70c64fc6\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T18:17:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T18:17:10Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T18:17:36Z is after 2025-08-24T17:21:41Z" Oct 08 18:17:36 crc kubenswrapper[4859]: I1008 18:17:36.841123 4859 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-08T18:17:29Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T18:17:29Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T18:17:36Z is after 2025-08-24T17:21:41Z" Oct 08 18:17:36 crc kubenswrapper[4859]: I1008 18:17:36.861545 4859 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-84k8x" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1c24078c-c9e8-4160-883a-f483b4b3dd23\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:17:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:17:31Z\\\",\\\"message\\\":\\\"containers with incomplete status: [whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:17:31Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:17:31Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hk5xb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://18b291eb3344fb7c00ef5679dca9a6f7ed9d08cea684a803ff7dad2e8a18c585\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://18b291eb3344fb7c00ef5679dca9a6f7ed9d08cea684a803ff7dad2e8a18c585\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-08T18:17:32Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-08T18:17:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hk5xb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://160163f1a91d981583452db3c7a9244182d1c8ced45749a61b721d0d45c1dfe4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://160163f1a91d981583452db3c7a9244182d1c8ced45749a61b721d0d45c1dfe4\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-08T18:17:34Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-08T18:17:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hk5xb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://51e9b53f3847ac53f378ad83870debf8e71518d4edbc7fc0dc66644818e2103e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://51e9b53f3847ac53f378ad83870debf8e71518d4edbc7fc0dc66644818e2103e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-08T18:17:34Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-08T18:17:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hk5xb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1ff118919739759ac01b5cbe3240e6673db54eaa16353c609d670071ed7964cb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1ff118919739759ac01b5cbe3240e6673db54eaa16353c609d670071ed7964cb\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-08T18:17:35Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-08T18:17:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hk5xb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hk5xb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hk5xb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T18:17:31Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-84k8x\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T18:17:36Z is after 2025-08-24T17:21:41Z" Oct 08 18:17:36 crc kubenswrapper[4859]: I1008 18:17:36.869839 4859 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 08 18:17:36 crc kubenswrapper[4859]: I1008 18:17:36.871854 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 18:17:36 crc kubenswrapper[4859]: I1008 18:17:36.871890 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 18:17:36 crc kubenswrapper[4859]: I1008 18:17:36.871917 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 18:17:36 crc kubenswrapper[4859]: I1008 18:17:36.872025 4859 kubelet_node_status.go:76] "Attempting to register node" node="crc" Oct 08 18:17:36 crc kubenswrapper[4859]: I1008 18:17:36.876463 4859 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-08T18:17:29Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T18:17:29Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T18:17:36Z is after 2025-08-24T17:21:41Z" Oct 08 18:17:36 crc kubenswrapper[4859]: I1008 18:17:36.905425 4859 kubelet_node_status.go:115] "Node was previously registered" node="crc" Oct 08 18:17:36 crc kubenswrapper[4859]: I1008 18:17:36.905792 4859 kubelet_node_status.go:79] "Successfully registered node" node="crc" Oct 08 18:17:36 crc kubenswrapper[4859]: I1008 18:17:36.907024 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 18:17:36 crc kubenswrapper[4859]: I1008 18:17:36.907067 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 18:17:36 crc kubenswrapper[4859]: I1008 18:17:36.907081 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 18:17:36 crc kubenswrapper[4859]: I1008 18:17:36.907102 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 18:17:36 crc kubenswrapper[4859]: I1008 18:17:36.907117 4859 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T18:17:36Z","lastTransitionTime":"2025-10-08T18:17:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 18:17:36 crc kubenswrapper[4859]: E1008 18:17:36.924275 4859 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404552Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865352Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-10-08T18:17:36Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-08T18:17:36Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-08T18:17:36Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-08T18:17:36Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-08T18:17:36Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-08T18:17:36Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-08T18:17:36Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-08T18:17:36Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"edb3ee6d-476b-407c-bb7d-4480786e8777\\\",\\\"systemUUID\\\":\\\"8f73c7ed-44aa-4d16-bae1-d4e684803cfc\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T18:17:36Z is after 2025-08-24T17:21:41Z" Oct 08 18:17:36 crc kubenswrapper[4859]: I1008 18:17:36.929129 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 18:17:36 crc kubenswrapper[4859]: I1008 18:17:36.929164 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 18:17:36 crc kubenswrapper[4859]: I1008 18:17:36.929181 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 18:17:36 crc kubenswrapper[4859]: I1008 18:17:36.929199 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 18:17:36 crc kubenswrapper[4859]: I1008 18:17:36.929211 4859 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T18:17:36Z","lastTransitionTime":"2025-10-08T18:17:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 18:17:36 crc kubenswrapper[4859]: I1008 18:17:36.934668 4859 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-btw28" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"97bc5598-3ead-4300-a938-3aebbff1e8c6\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:17:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:17:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:17:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:17:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d6d4d6ba771cb5fde637449bba05660f634df7e833a0c8dad2b2c7f3a422bb2b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T18:17:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qcxsd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T18:17:31Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-btw28\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T18:17:36Z is after 2025-08-24T17:21:41Z" Oct 08 18:17:36 crc kubenswrapper[4859]: E1008 18:17:36.943416 4859 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404552Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865352Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-10-08T18:17:36Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-08T18:17:36Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-08T18:17:36Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-08T18:17:36Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-08T18:17:36Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-08T18:17:36Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-08T18:17:36Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-08T18:17:36Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"edb3ee6d-476b-407c-bb7d-4480786e8777\\\",\\\"systemUUID\\\":\\\"8f73c7ed-44aa-4d16-bae1-d4e684803cfc\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T18:17:36Z is after 2025-08-24T17:21:41Z" Oct 08 18:17:36 crc kubenswrapper[4859]: I1008 18:17:36.948856 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 18:17:36 crc kubenswrapper[4859]: I1008 18:17:36.948897 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 18:17:36 crc kubenswrapper[4859]: I1008 18:17:36.948908 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 18:17:36 crc kubenswrapper[4859]: I1008 18:17:36.948924 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 18:17:36 crc kubenswrapper[4859]: I1008 18:17:36.948935 4859 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T18:17:36Z","lastTransitionTime":"2025-10-08T18:17:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 18:17:36 crc kubenswrapper[4859]: E1008 18:17:36.961342 4859 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404552Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865352Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-10-08T18:17:36Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-08T18:17:36Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-08T18:17:36Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-08T18:17:36Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-08T18:17:36Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-08T18:17:36Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-08T18:17:36Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-08T18:17:36Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"edb3ee6d-476b-407c-bb7d-4480786e8777\\\",\\\"systemUUID\\\":\\\"8f73c7ed-44aa-4d16-bae1-d4e684803cfc\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T18:17:36Z is after 2025-08-24T17:21:41Z" Oct 08 18:17:36 crc kubenswrapper[4859]: I1008 18:17:36.965266 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 18:17:36 crc kubenswrapper[4859]: I1008 18:17:36.965300 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 18:17:36 crc kubenswrapper[4859]: I1008 18:17:36.965310 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 18:17:36 crc kubenswrapper[4859]: I1008 18:17:36.965328 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 18:17:36 crc kubenswrapper[4859]: I1008 18:17:36.965341 4859 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T18:17:36Z","lastTransitionTime":"2025-10-08T18:17:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 18:17:36 crc kubenswrapper[4859]: I1008 18:17:36.973547 4859 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-8plkg" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"92b2d7de-31cc-4a91-95a5-ed5ea964b028\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:17:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:17:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:17:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:17:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fa00b8345d40b74a917ddc23cc1b2c5c684a9ca61b6119d090c92cc1439b71c4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T18:17:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q9d7m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T18:17:31Z\\\"}}\" for pod \"openshift-multus\"/\"multus-8plkg\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T18:17:36Z is after 2025-08-24T17:21:41Z" Oct 08 18:17:36 crc kubenswrapper[4859]: E1008 18:17:36.982035 4859 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404552Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865352Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-10-08T18:17:36Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-08T18:17:36Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-08T18:17:36Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-08T18:17:36Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-08T18:17:36Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-08T18:17:36Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-08T18:17:36Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-08T18:17:36Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"edb3ee6d-476b-407c-bb7d-4480786e8777\\\",\\\"systemUUID\\\":\\\"8f73c7ed-44aa-4d16-bae1-d4e684803cfc\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T18:17:36Z is after 2025-08-24T17:21:41Z" Oct 08 18:17:36 crc kubenswrapper[4859]: I1008 18:17:36.985507 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 18:17:36 crc kubenswrapper[4859]: I1008 18:17:36.985535 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 18:17:36 crc kubenswrapper[4859]: I1008 18:17:36.985544 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 18:17:36 crc kubenswrapper[4859]: I1008 18:17:36.985560 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 18:17:36 crc kubenswrapper[4859]: I1008 18:17:36.985573 4859 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T18:17:36Z","lastTransitionTime":"2025-10-08T18:17:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 18:17:36 crc kubenswrapper[4859]: E1008 18:17:36.998948 4859 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404552Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865352Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-10-08T18:17:36Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-08T18:17:36Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-08T18:17:36Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-08T18:17:36Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-08T18:17:36Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-08T18:17:36Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-08T18:17:36Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-08T18:17:36Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"edb3ee6d-476b-407c-bb7d-4480786e8777\\\",\\\"systemUUID\\\":\\\"8f73c7ed-44aa-4d16-bae1-d4e684803cfc\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T18:17:36Z is after 2025-08-24T17:21:41Z" Oct 08 18:17:36 crc kubenswrapper[4859]: E1008 18:17:36.999115 4859 kubelet_node_status.go:572] "Unable to update node status" err="update node status exceeds retry count" Oct 08 18:17:37 crc kubenswrapper[4859]: I1008 18:17:37.001162 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 18:17:37 crc kubenswrapper[4859]: I1008 18:17:37.001191 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 18:17:37 crc kubenswrapper[4859]: I1008 18:17:37.001207 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 18:17:37 crc kubenswrapper[4859]: I1008 18:17:37.001227 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 18:17:37 crc kubenswrapper[4859]: I1008 18:17:37.001240 4859 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T18:17:37Z","lastTransitionTime":"2025-10-08T18:17:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 18:17:37 crc kubenswrapper[4859]: I1008 18:17:37.011748 4859 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-btjmq" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b8605f7-fed7-493f-9b4f-6a6358adf907\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:17:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:17:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:17:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:17:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://14e9b6c4a4403f36adb78fcf1135a620ca6fc56ed974c9639dcdfc59e6917559\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T18:17:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xspg4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T18:17:35Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-btjmq\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T18:17:37Z is after 2025-08-24T17:21:41Z" Oct 08 18:17:37 crc kubenswrapper[4859]: I1008 18:17:37.055935 4859 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-08T18:17:31Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T18:17:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T18:17:31Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c6d15905694b6bbd4999d49dc8f7a53eb6bfaa7adb59efb0d7e4973c21d1c7be\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T18:17:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T18:17:37Z is after 2025-08-24T17:21:41Z" Oct 08 18:17:37 crc kubenswrapper[4859]: I1008 18:17:37.096928 4859 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-08T18:17:33Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T18:17:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T18:17:33Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5d99e071bd083914418d9c128f38955f26295c08b1520f2dc774c99b1c843693\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T18:17:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T18:17:37Z is after 2025-08-24T17:21:41Z" Oct 08 18:17:37 crc kubenswrapper[4859]: I1008 18:17:37.104160 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 18:17:37 crc kubenswrapper[4859]: I1008 18:17:37.104199 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 18:17:37 crc kubenswrapper[4859]: I1008 18:17:37.104210 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 18:17:37 crc kubenswrapper[4859]: I1008 18:17:37.104228 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 18:17:37 crc kubenswrapper[4859]: I1008 18:17:37.104247 4859 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T18:17:37Z","lastTransitionTime":"2025-10-08T18:17:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 18:17:37 crc kubenswrapper[4859]: I1008 18:17:37.213077 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 18:17:37 crc kubenswrapper[4859]: I1008 18:17:37.213113 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 18:17:37 crc kubenswrapper[4859]: I1008 18:17:37.213122 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 18:17:37 crc kubenswrapper[4859]: I1008 18:17:37.213138 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 18:17:37 crc kubenswrapper[4859]: I1008 18:17:37.213150 4859 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T18:17:37Z","lastTransitionTime":"2025-10-08T18:17:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 18:17:37 crc kubenswrapper[4859]: I1008 18:17:37.316266 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 18:17:37 crc kubenswrapper[4859]: I1008 18:17:37.316315 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 18:17:37 crc kubenswrapper[4859]: I1008 18:17:37.316326 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 18:17:37 crc kubenswrapper[4859]: I1008 18:17:37.316347 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 18:17:37 crc kubenswrapper[4859]: I1008 18:17:37.316359 4859 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T18:17:37Z","lastTransitionTime":"2025-10-08T18:17:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 18:17:37 crc kubenswrapper[4859]: I1008 18:17:37.419264 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 18:17:37 crc kubenswrapper[4859]: I1008 18:17:37.419313 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 18:17:37 crc kubenswrapper[4859]: I1008 18:17:37.419324 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 18:17:37 crc kubenswrapper[4859]: I1008 18:17:37.419342 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 18:17:37 crc kubenswrapper[4859]: I1008 18:17:37.419353 4859 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T18:17:37Z","lastTransitionTime":"2025-10-08T18:17:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 18:17:37 crc kubenswrapper[4859]: I1008 18:17:37.469438 4859 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 08 18:17:37 crc kubenswrapper[4859]: I1008 18:17:37.469506 4859 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 08 18:17:37 crc kubenswrapper[4859]: I1008 18:17:37.469476 4859 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 08 18:17:37 crc kubenswrapper[4859]: E1008 18:17:37.469750 4859 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 08 18:17:37 crc kubenswrapper[4859]: E1008 18:17:37.469904 4859 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 08 18:17:37 crc kubenswrapper[4859]: E1008 18:17:37.469997 4859 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 08 18:17:37 crc kubenswrapper[4859]: I1008 18:17:37.522584 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 18:17:37 crc kubenswrapper[4859]: I1008 18:17:37.522736 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 18:17:37 crc kubenswrapper[4859]: I1008 18:17:37.522762 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 18:17:37 crc kubenswrapper[4859]: I1008 18:17:37.522796 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 18:17:37 crc kubenswrapper[4859]: I1008 18:17:37.522819 4859 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T18:17:37Z","lastTransitionTime":"2025-10-08T18:17:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 18:17:37 crc kubenswrapper[4859]: I1008 18:17:37.625988 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 18:17:37 crc kubenswrapper[4859]: I1008 18:17:37.626027 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 18:17:37 crc kubenswrapper[4859]: I1008 18:17:37.626037 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 18:17:37 crc kubenswrapper[4859]: I1008 18:17:37.626051 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 18:17:37 crc kubenswrapper[4859]: I1008 18:17:37.626061 4859 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T18:17:37Z","lastTransitionTime":"2025-10-08T18:17:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 18:17:37 crc kubenswrapper[4859]: I1008 18:17:37.694992 4859 generic.go:334] "Generic (PLEG): container finished" podID="1c24078c-c9e8-4160-883a-f483b4b3dd23" containerID="8522e803ff56eb57cd4d9fd036e0c3fb16240ea21d4a094c1db7aa7d50bfe2c2" exitCode=0 Oct 08 18:17:37 crc kubenswrapper[4859]: I1008 18:17:37.695073 4859 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-84k8x" event={"ID":"1c24078c-c9e8-4160-883a-f483b4b3dd23","Type":"ContainerDied","Data":"8522e803ff56eb57cd4d9fd036e0c3fb16240ea21d4a094c1db7aa7d50bfe2c2"} Oct 08 18:17:37 crc kubenswrapper[4859]: I1008 18:17:37.699871 4859 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-4g8gf" event={"ID":"1dff864e-b75d-4e0b-b182-75f710eac8df","Type":"ContainerStarted","Data":"2c67d2aed7126d76989e057d8e2d8e188a8902b2f3fe26ed6d1d45d3b013d897"} Oct 08 18:17:37 crc kubenswrapper[4859]: I1008 18:17:37.714149 4859 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-08T18:17:31Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T18:17:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T18:17:31Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c6d15905694b6bbd4999d49dc8f7a53eb6bfaa7adb59efb0d7e4973c21d1c7be\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T18:17:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T18:17:37Z is after 2025-08-24T17:21:41Z" Oct 08 18:17:37 crc kubenswrapper[4859]: I1008 18:17:37.733576 4859 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-08T18:17:33Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T18:17:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T18:17:33Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5d99e071bd083914418d9c128f38955f26295c08b1520f2dc774c99b1c843693\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T18:17:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T18:17:37Z is after 2025-08-24T17:21:41Z" Oct 08 18:17:37 crc kubenswrapper[4859]: I1008 18:17:37.735995 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 18:17:37 crc kubenswrapper[4859]: I1008 18:17:37.736094 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 18:17:37 crc kubenswrapper[4859]: I1008 18:17:37.736121 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 18:17:37 crc kubenswrapper[4859]: I1008 18:17:37.736162 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 18:17:37 crc kubenswrapper[4859]: I1008 18:17:37.736186 4859 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T18:17:37Z","lastTransitionTime":"2025-10-08T18:17:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 18:17:37 crc kubenswrapper[4859]: I1008 18:17:37.750339 4859 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"63a623de-5608-4228-b1c2-77188c276986\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:17:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:17:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:17:10Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:17:10Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:17:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d4786a8fb8abd96fb22e60733c974f599017407d589ed815fb8541fd3b64f696\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T18:17:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://183d481502ed5108d5e19e483b714c289f355ea3f8b0891883e5335b802c83b6\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T18:17:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ec848c46fd5dde0469b00d3cab14ba4a652533a265490fadbee7a8b8843ef249\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T18:17:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4fe1828b4d63ff8106bfcd15e1dc1c1bd1ba7f77777cdb23ff1b291d0418774c\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2e700b2a20c81b414617f5b5c018a738278a3fa44e6c6d14b2170051d0725431\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-08T18:17:24Z\\\",\\\"message\\\":\\\"W1008 18:17:13.670625 1 cmd.go:257] Using insecure, self-signed certificates\\\\nI1008 18:17:13.671202 1 crypto.go:601] Generating new CA for check-endpoints-signer@1759947433 cert, and key in /tmp/serving-cert-3867218267/serving-signer.crt, /tmp/serving-cert-3867218267/serving-signer.key\\\\nI1008 18:17:13.944463 1 observer_polling.go:159] Starting file observer\\\\nW1008 18:17:13.951958 1 builder.go:272] unable to get owner reference (falling back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": dial tcp [::1]:6443: connect: connection refused\\\\nI1008 18:17:13.952088 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1008 18:17:13.956755 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3867218267/tls.crt::/tmp/serving-cert-3867218267/tls.key\\\\\\\"\\\\nF1008 18:17:24.290991 1 cmd.go:182] error initializing delegating authentication: unable to load configmap based request-header-client-ca-file: Get \\\\\\\"https://localhost:6443/api/v1/namespaces/kube-system/configmaps/extension-apiserver-authentication\\\\\\\": net/http: TLS handshake timeout\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-08T18:17:13Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T18:17:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c65b66f62d325ca10ef8d5d31f04529ccf4e691cf3303d231e7a414fb400e4d6\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T18:17:13Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://58c7bba4ace28215b062653a960e561b1dc1f890f2baf450145638a065af038b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://58c7bba4ace28215b062653a960e561b1dc1f890f2baf450145638a065af038b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-08T18:17:11Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-08T18:17:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T18:17:10Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T18:17:37Z is after 2025-08-24T17:21:41Z" Oct 08 18:17:37 crc kubenswrapper[4859]: I1008 18:17:37.769047 4859 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-08T18:17:31Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T18:17:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T18:17:31Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://99e2d51bd7757498dae60949eea6d732b5f7b481fa00f808dd5308a530a1daf1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T18:17:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e32aaba93699944665660ae41e9b5779178c1354c31e15ad8188bf8c3c8fa500\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T18:17:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T18:17:37Z is after 2025-08-24T17:21:41Z" Oct 08 18:17:37 crc kubenswrapper[4859]: I1008 18:17:37.782583 4859 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-08T18:17:29Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T18:17:29Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T18:17:37Z is after 2025-08-24T17:21:41Z" Oct 08 18:17:37 crc kubenswrapper[4859]: I1008 18:17:37.797369 4859 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-82s52" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b23a6a6c-9d92-4e7b-840e-55cfda873a2d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:17:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:17:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:17:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:17:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://71302c766485b9f55743d85f97440ff578cd44ba1e5a6f5b1c024c0e808072d2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T18:17:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6vx22\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c23ea3580998adf90621a0155c20ee4bd590c512629812174e95c28857bcfef9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T18:17:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6vx22\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T18:17:31Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-82s52\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T18:17:37Z is after 2025-08-24T17:21:41Z" Oct 08 18:17:37 crc kubenswrapper[4859]: I1008 18:17:37.827823 4859 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-4g8gf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1dff864e-b75d-4e0b-b182-75f710eac8df\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:17:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:17:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:17:32Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:17:32Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s8jnn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s8jnn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s8jnn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s8jnn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s8jnn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s8jnn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s8jnn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s8jnn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7a211d3da02b173867ff3a912eecd3b379483c7cc77e6b91122edd1ffd30fd53\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7a211d3da02b173867ff3a912eecd3b379483c7cc77e6b91122edd1ffd30fd53\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-08T18:17:32Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-08T18:17:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s8jnn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T18:17:32Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-4g8gf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T18:17:37Z is after 2025-08-24T17:21:41Z" Oct 08 18:17:37 crc kubenswrapper[4859]: I1008 18:17:37.839383 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 18:17:37 crc kubenswrapper[4859]: I1008 18:17:37.839442 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 18:17:37 crc kubenswrapper[4859]: I1008 18:17:37.839465 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 18:17:37 crc kubenswrapper[4859]: I1008 18:17:37.839486 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 18:17:37 crc kubenswrapper[4859]: I1008 18:17:37.839496 4859 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T18:17:37Z","lastTransitionTime":"2025-10-08T18:17:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 18:17:37 crc kubenswrapper[4859]: I1008 18:17:37.845898 4859 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4efc16fb-e679-464d-8b3c-0f46f082593e\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:17:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:17:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:17:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:17:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:17:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://aa8f8ea895272cd79a08d1c47cd7863a1ec10e21bb60d12aa10862a56c99e4bf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T18:17:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5c4e95d8a712ab219455564539a90286ae48bbc46ddc322a5bfd96c1a8268110\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T18:17:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://bae6ecfd8b446ef04d393d823e5ffe37fe749337c0ca17fde79c5f88d5e31393\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T18:17:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6411d1e7e18d5ceadaaceed5a8a6b20a0cacc55cede793ca9fb4dcad70c64fc6\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T18:17:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T18:17:10Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T18:17:37Z is after 2025-08-24T17:21:41Z" Oct 08 18:17:37 crc kubenswrapper[4859]: I1008 18:17:37.864262 4859 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-08T18:17:29Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T18:17:29Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T18:17:37Z is after 2025-08-24T17:21:41Z" Oct 08 18:17:37 crc kubenswrapper[4859]: I1008 18:17:37.887277 4859 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-84k8x" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1c24078c-c9e8-4160-883a-f483b4b3dd23\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:17:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:17:31Z\\\",\\\"message\\\":\\\"containers with incomplete status: [whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:17:31Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:17:31Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hk5xb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://18b291eb3344fb7c00ef5679dca9a6f7ed9d08cea684a803ff7dad2e8a18c585\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://18b291eb3344fb7c00ef5679dca9a6f7ed9d08cea684a803ff7dad2e8a18c585\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-08T18:17:32Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-08T18:17:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hk5xb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://160163f1a91d981583452db3c7a9244182d1c8ced45749a61b721d0d45c1dfe4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://160163f1a91d981583452db3c7a9244182d1c8ced45749a61b721d0d45c1dfe4\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-08T18:17:34Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-08T18:17:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hk5xb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://51e9b53f3847ac53f378ad83870debf8e71518d4edbc7fc0dc66644818e2103e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://51e9b53f3847ac53f378ad83870debf8e71518d4edbc7fc0dc66644818e2103e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-08T18:17:34Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-08T18:17:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hk5xb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1ff118919739759ac01b5cbe3240e6673db54eaa16353c609d670071ed7964cb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1ff118919739759ac01b5cbe3240e6673db54eaa16353c609d670071ed7964cb\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-08T18:17:35Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-08T18:17:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hk5xb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8522e803ff56eb57cd4d9fd036e0c3fb16240ea21d4a094c1db7aa7d50bfe2c2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8522e803ff56eb57cd4d9fd036e0c3fb16240ea21d4a094c1db7aa7d50bfe2c2\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-08T18:17:37Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-08T18:17:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hk5xb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hk5xb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T18:17:31Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-84k8x\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T18:17:37Z is after 2025-08-24T17:21:41Z" Oct 08 18:17:37 crc kubenswrapper[4859]: I1008 18:17:37.904210 4859 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-08T18:17:29Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T18:17:29Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T18:17:37Z is after 2025-08-24T17:21:41Z" Oct 08 18:17:37 crc kubenswrapper[4859]: I1008 18:17:37.918816 4859 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-btw28" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"97bc5598-3ead-4300-a938-3aebbff1e8c6\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:17:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:17:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:17:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:17:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d6d4d6ba771cb5fde637449bba05660f634df7e833a0c8dad2b2c7f3a422bb2b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T18:17:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qcxsd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T18:17:31Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-btw28\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T18:17:37Z is after 2025-08-24T17:21:41Z" Oct 08 18:17:37 crc kubenswrapper[4859]: I1008 18:17:37.935943 4859 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-8plkg" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"92b2d7de-31cc-4a91-95a5-ed5ea964b028\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:17:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:17:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:17:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:17:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fa00b8345d40b74a917ddc23cc1b2c5c684a9ca61b6119d090c92cc1439b71c4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T18:17:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q9d7m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T18:17:31Z\\\"}}\" for pod \"openshift-multus\"/\"multus-8plkg\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T18:17:37Z is after 2025-08-24T17:21:41Z" Oct 08 18:17:37 crc kubenswrapper[4859]: I1008 18:17:37.942525 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 18:17:37 crc kubenswrapper[4859]: I1008 18:17:37.942566 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 18:17:37 crc kubenswrapper[4859]: I1008 18:17:37.942579 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 18:17:37 crc kubenswrapper[4859]: I1008 18:17:37.942603 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 18:17:37 crc kubenswrapper[4859]: I1008 18:17:37.942619 4859 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T18:17:37Z","lastTransitionTime":"2025-10-08T18:17:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 18:17:37 crc kubenswrapper[4859]: I1008 18:17:37.951630 4859 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-btjmq" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b8605f7-fed7-493f-9b4f-6a6358adf907\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:17:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:17:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:17:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:17:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://14e9b6c4a4403f36adb78fcf1135a620ca6fc56ed974c9639dcdfc59e6917559\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T18:17:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xspg4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T18:17:35Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-btjmq\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T18:17:37Z is after 2025-08-24T17:21:41Z" Oct 08 18:17:38 crc kubenswrapper[4859]: I1008 18:17:38.045597 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 18:17:38 crc kubenswrapper[4859]: I1008 18:17:38.045652 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 18:17:38 crc kubenswrapper[4859]: I1008 18:17:38.045665 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 18:17:38 crc kubenswrapper[4859]: I1008 18:17:38.045704 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 18:17:38 crc kubenswrapper[4859]: I1008 18:17:38.045723 4859 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T18:17:38Z","lastTransitionTime":"2025-10-08T18:17:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 18:17:38 crc kubenswrapper[4859]: I1008 18:17:38.149560 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 18:17:38 crc kubenswrapper[4859]: I1008 18:17:38.149616 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 18:17:38 crc kubenswrapper[4859]: I1008 18:17:38.149630 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 18:17:38 crc kubenswrapper[4859]: I1008 18:17:38.149652 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 18:17:38 crc kubenswrapper[4859]: I1008 18:17:38.149665 4859 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T18:17:38Z","lastTransitionTime":"2025-10-08T18:17:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 18:17:38 crc kubenswrapper[4859]: I1008 18:17:38.200048 4859 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 08 18:17:38 crc kubenswrapper[4859]: E1008 18:17:38.200119 4859 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-08 18:17:46.200080704 +0000 UTC m=+36.446920113 (durationBeforeRetry 8s). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 08 18:17:38 crc kubenswrapper[4859]: I1008 18:17:38.252375 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 18:17:38 crc kubenswrapper[4859]: I1008 18:17:38.252455 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 18:17:38 crc kubenswrapper[4859]: I1008 18:17:38.252466 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 18:17:38 crc kubenswrapper[4859]: I1008 18:17:38.252484 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 18:17:38 crc kubenswrapper[4859]: I1008 18:17:38.252496 4859 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T18:17:38Z","lastTransitionTime":"2025-10-08T18:17:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 18:17:38 crc kubenswrapper[4859]: I1008 18:17:38.301816 4859 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 08 18:17:38 crc kubenswrapper[4859]: I1008 18:17:38.302262 4859 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 08 18:17:38 crc kubenswrapper[4859]: E1008 18:17:38.302112 4859 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Oct 08 18:17:38 crc kubenswrapper[4859]: I1008 18:17:38.302308 4859 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 08 18:17:38 crc kubenswrapper[4859]: E1008 18:17:38.302331 4859 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Oct 08 18:17:38 crc kubenswrapper[4859]: E1008 18:17:38.302356 4859 projected.go:194] Error preparing data for projected volume kube-api-access-s2dwl for pod openshift-network-diagnostics/network-check-source-55646444c4-trplf: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Oct 08 18:17:38 crc kubenswrapper[4859]: I1008 18:17:38.302370 4859 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 08 18:17:38 crc kubenswrapper[4859]: E1008 18:17:38.302474 4859 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl podName:9d751cbb-f2e2-430d-9754-c882a5e924a5 nodeName:}" failed. No retries permitted until 2025-10-08 18:17:46.302436595 +0000 UTC m=+36.549276114 (durationBeforeRetry 8s). Error: MountVolume.SetUp failed for volume "kube-api-access-s2dwl" (UniqueName: "kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl") pod "network-check-source-55646444c4-trplf" (UID: "9d751cbb-f2e2-430d-9754-c882a5e924a5") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Oct 08 18:17:38 crc kubenswrapper[4859]: E1008 18:17:38.302493 4859 secret.go:188] Couldn't get secret openshift-network-console/networking-console-plugin-cert: object "openshift-network-console"/"networking-console-plugin-cert" not registered Oct 08 18:17:38 crc kubenswrapper[4859]: E1008 18:17:38.302507 4859 configmap.go:193] Couldn't get configMap openshift-network-console/networking-console-plugin: object "openshift-network-console"/"networking-console-plugin" not registered Oct 08 18:17:38 crc kubenswrapper[4859]: E1008 18:17:38.302570 4859 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-10-08 18:17:46.302550758 +0000 UTC m=+36.549390167 (durationBeforeRetry 8s). Error: MountVolume.SetUp failed for volume "networking-console-plugin-cert" (UniqueName: "kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin-cert" not registered Oct 08 18:17:38 crc kubenswrapper[4859]: E1008 18:17:38.302610 4859 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-10-08 18:17:46.302584439 +0000 UTC m=+36.549423848 (durationBeforeRetry 8s). Error: MountVolume.SetUp failed for volume "nginx-conf" (UniqueName: "kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin" not registered Oct 08 18:17:38 crc kubenswrapper[4859]: E1008 18:17:38.302607 4859 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Oct 08 18:17:38 crc kubenswrapper[4859]: E1008 18:17:38.302655 4859 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Oct 08 18:17:38 crc kubenswrapper[4859]: E1008 18:17:38.302674 4859 projected.go:194] Error preparing data for projected volume kube-api-access-cqllr for pod openshift-network-diagnostics/network-check-target-xd92c: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Oct 08 18:17:38 crc kubenswrapper[4859]: E1008 18:17:38.302802 4859 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr podName:3b6479f0-333b-4a96-9adf-2099afdc2447 nodeName:}" failed. No retries permitted until 2025-10-08 18:17:46.302770044 +0000 UTC m=+36.549609613 (durationBeforeRetry 8s). Error: MountVolume.SetUp failed for volume "kube-api-access-cqllr" (UniqueName: "kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr") pod "network-check-target-xd92c" (UID: "3b6479f0-333b-4a96-9adf-2099afdc2447") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Oct 08 18:17:38 crc kubenswrapper[4859]: I1008 18:17:38.355853 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 18:17:38 crc kubenswrapper[4859]: I1008 18:17:38.355927 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 18:17:38 crc kubenswrapper[4859]: I1008 18:17:38.355948 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 18:17:38 crc kubenswrapper[4859]: I1008 18:17:38.356012 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 18:17:38 crc kubenswrapper[4859]: I1008 18:17:38.356041 4859 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T18:17:38Z","lastTransitionTime":"2025-10-08T18:17:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 18:17:38 crc kubenswrapper[4859]: I1008 18:17:38.459383 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 18:17:38 crc kubenswrapper[4859]: I1008 18:17:38.459443 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 18:17:38 crc kubenswrapper[4859]: I1008 18:17:38.459454 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 18:17:38 crc kubenswrapper[4859]: I1008 18:17:38.459473 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 18:17:38 crc kubenswrapper[4859]: I1008 18:17:38.459485 4859 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T18:17:38Z","lastTransitionTime":"2025-10-08T18:17:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 18:17:38 crc kubenswrapper[4859]: I1008 18:17:38.563470 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 18:17:38 crc kubenswrapper[4859]: I1008 18:17:38.563565 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 18:17:38 crc kubenswrapper[4859]: I1008 18:17:38.563600 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 18:17:38 crc kubenswrapper[4859]: I1008 18:17:38.563633 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 18:17:38 crc kubenswrapper[4859]: I1008 18:17:38.563655 4859 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T18:17:38Z","lastTransitionTime":"2025-10-08T18:17:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 18:17:38 crc kubenswrapper[4859]: I1008 18:17:38.667241 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 18:17:38 crc kubenswrapper[4859]: I1008 18:17:38.667332 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 18:17:38 crc kubenswrapper[4859]: I1008 18:17:38.667355 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 18:17:38 crc kubenswrapper[4859]: I1008 18:17:38.667386 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 18:17:38 crc kubenswrapper[4859]: I1008 18:17:38.667408 4859 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T18:17:38Z","lastTransitionTime":"2025-10-08T18:17:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 18:17:38 crc kubenswrapper[4859]: I1008 18:17:38.716160 4859 generic.go:334] "Generic (PLEG): container finished" podID="1c24078c-c9e8-4160-883a-f483b4b3dd23" containerID="be61056edf515e037403ca2a0bfd5feb9b2537dd42212b58e133b23b294f327b" exitCode=0 Oct 08 18:17:38 crc kubenswrapper[4859]: I1008 18:17:38.716227 4859 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-84k8x" event={"ID":"1c24078c-c9e8-4160-883a-f483b4b3dd23","Type":"ContainerDied","Data":"be61056edf515e037403ca2a0bfd5feb9b2537dd42212b58e133b23b294f327b"} Oct 08 18:17:38 crc kubenswrapper[4859]: I1008 18:17:38.743493 4859 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-08T18:17:29Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T18:17:29Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T18:17:38Z is after 2025-08-24T17:21:41Z" Oct 08 18:17:38 crc kubenswrapper[4859]: I1008 18:17:38.769222 4859 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-btw28" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"97bc5598-3ead-4300-a938-3aebbff1e8c6\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:17:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:17:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:17:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:17:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d6d4d6ba771cb5fde637449bba05660f634df7e833a0c8dad2b2c7f3a422bb2b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T18:17:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qcxsd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T18:17:31Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-btw28\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T18:17:38Z is after 2025-08-24T17:21:41Z" Oct 08 18:17:38 crc kubenswrapper[4859]: I1008 18:17:38.774236 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 18:17:38 crc kubenswrapper[4859]: I1008 18:17:38.774327 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 18:17:38 crc kubenswrapper[4859]: I1008 18:17:38.774353 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 18:17:38 crc kubenswrapper[4859]: I1008 18:17:38.774385 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 18:17:38 crc kubenswrapper[4859]: I1008 18:17:38.774409 4859 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T18:17:38Z","lastTransitionTime":"2025-10-08T18:17:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 18:17:38 crc kubenswrapper[4859]: I1008 18:17:38.791963 4859 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-8plkg" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"92b2d7de-31cc-4a91-95a5-ed5ea964b028\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:17:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:17:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:17:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:17:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fa00b8345d40b74a917ddc23cc1b2c5c684a9ca61b6119d090c92cc1439b71c4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T18:17:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q9d7m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T18:17:31Z\\\"}}\" for pod \"openshift-multus\"/\"multus-8plkg\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T18:17:38Z is after 2025-08-24T17:21:41Z" Oct 08 18:17:38 crc kubenswrapper[4859]: I1008 18:17:38.807091 4859 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-btjmq" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b8605f7-fed7-493f-9b4f-6a6358adf907\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:17:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:17:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:17:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:17:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://14e9b6c4a4403f36adb78fcf1135a620ca6fc56ed974c9639dcdfc59e6917559\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T18:17:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xspg4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T18:17:35Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-btjmq\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T18:17:38Z is after 2025-08-24T17:21:41Z" Oct 08 18:17:38 crc kubenswrapper[4859]: I1008 18:17:38.822581 4859 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-08T18:17:31Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T18:17:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T18:17:31Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c6d15905694b6bbd4999d49dc8f7a53eb6bfaa7adb59efb0d7e4973c21d1c7be\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T18:17:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T18:17:38Z is after 2025-08-24T17:21:41Z" Oct 08 18:17:38 crc kubenswrapper[4859]: I1008 18:17:38.835172 4859 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-08T18:17:33Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T18:17:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T18:17:33Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5d99e071bd083914418d9c128f38955f26295c08b1520f2dc774c99b1c843693\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T18:17:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T18:17:38Z is after 2025-08-24T17:21:41Z" Oct 08 18:17:38 crc kubenswrapper[4859]: I1008 18:17:38.854221 4859 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-82s52" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b23a6a6c-9d92-4e7b-840e-55cfda873a2d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:17:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:17:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:17:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:17:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://71302c766485b9f55743d85f97440ff578cd44ba1e5a6f5b1c024c0e808072d2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T18:17:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6vx22\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c23ea3580998adf90621a0155c20ee4bd590c512629812174e95c28857bcfef9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T18:17:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6vx22\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T18:17:31Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-82s52\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T18:17:38Z is after 2025-08-24T17:21:41Z" Oct 08 18:17:38 crc kubenswrapper[4859]: I1008 18:17:38.874255 4859 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-4g8gf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1dff864e-b75d-4e0b-b182-75f710eac8df\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:17:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:17:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:17:32Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:17:32Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s8jnn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s8jnn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s8jnn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s8jnn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s8jnn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s8jnn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s8jnn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s8jnn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7a211d3da02b173867ff3a912eecd3b379483c7cc77e6b91122edd1ffd30fd53\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7a211d3da02b173867ff3a912eecd3b379483c7cc77e6b91122edd1ffd30fd53\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-08T18:17:32Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-08T18:17:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s8jnn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T18:17:32Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-4g8gf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T18:17:38Z is after 2025-08-24T17:21:41Z" Oct 08 18:17:38 crc kubenswrapper[4859]: I1008 18:17:38.877801 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 18:17:38 crc kubenswrapper[4859]: I1008 18:17:38.877826 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 18:17:38 crc kubenswrapper[4859]: I1008 18:17:38.877835 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 18:17:38 crc kubenswrapper[4859]: I1008 18:17:38.877849 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 18:17:38 crc kubenswrapper[4859]: I1008 18:17:38.877859 4859 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T18:17:38Z","lastTransitionTime":"2025-10-08T18:17:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 18:17:38 crc kubenswrapper[4859]: I1008 18:17:38.892868 4859 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"63a623de-5608-4228-b1c2-77188c276986\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:17:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:17:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:17:10Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:17:10Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:17:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d4786a8fb8abd96fb22e60733c974f599017407d589ed815fb8541fd3b64f696\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T18:17:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://183d481502ed5108d5e19e483b714c289f355ea3f8b0891883e5335b802c83b6\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T18:17:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ec848c46fd5dde0469b00d3cab14ba4a652533a265490fadbee7a8b8843ef249\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T18:17:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4fe1828b4d63ff8106bfcd15e1dc1c1bd1ba7f77777cdb23ff1b291d0418774c\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2e700b2a20c81b414617f5b5c018a738278a3fa44e6c6d14b2170051d0725431\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-08T18:17:24Z\\\",\\\"message\\\":\\\"W1008 18:17:13.670625 1 cmd.go:257] Using insecure, self-signed certificates\\\\nI1008 18:17:13.671202 1 crypto.go:601] Generating new CA for check-endpoints-signer@1759947433 cert, and key in /tmp/serving-cert-3867218267/serving-signer.crt, /tmp/serving-cert-3867218267/serving-signer.key\\\\nI1008 18:17:13.944463 1 observer_polling.go:159] Starting file observer\\\\nW1008 18:17:13.951958 1 builder.go:272] unable to get owner reference (falling back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": dial tcp [::1]:6443: connect: connection refused\\\\nI1008 18:17:13.952088 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1008 18:17:13.956755 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3867218267/tls.crt::/tmp/serving-cert-3867218267/tls.key\\\\\\\"\\\\nF1008 18:17:24.290991 1 cmd.go:182] error initializing delegating authentication: unable to load configmap based request-header-client-ca-file: Get \\\\\\\"https://localhost:6443/api/v1/namespaces/kube-system/configmaps/extension-apiserver-authentication\\\\\\\": net/http: TLS handshake timeout\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-08T18:17:13Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T18:17:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c65b66f62d325ca10ef8d5d31f04529ccf4e691cf3303d231e7a414fb400e4d6\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T18:17:13Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://58c7bba4ace28215b062653a960e561b1dc1f890f2baf450145638a065af038b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://58c7bba4ace28215b062653a960e561b1dc1f890f2baf450145638a065af038b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-08T18:17:11Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-08T18:17:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T18:17:10Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T18:17:38Z is after 2025-08-24T17:21:41Z" Oct 08 18:17:38 crc kubenswrapper[4859]: I1008 18:17:38.918163 4859 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-08T18:17:31Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T18:17:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T18:17:31Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://99e2d51bd7757498dae60949eea6d732b5f7b481fa00f808dd5308a530a1daf1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T18:17:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e32aaba93699944665660ae41e9b5779178c1354c31e15ad8188bf8c3c8fa500\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T18:17:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T18:17:38Z is after 2025-08-24T17:21:41Z" Oct 08 18:17:38 crc kubenswrapper[4859]: I1008 18:17:38.936045 4859 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-08T18:17:29Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T18:17:29Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T18:17:38Z is after 2025-08-24T17:21:41Z" Oct 08 18:17:38 crc kubenswrapper[4859]: I1008 18:17:38.950501 4859 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4efc16fb-e679-464d-8b3c-0f46f082593e\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:17:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:17:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:17:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:17:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:17:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://aa8f8ea895272cd79a08d1c47cd7863a1ec10e21bb60d12aa10862a56c99e4bf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T18:17:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5c4e95d8a712ab219455564539a90286ae48bbc46ddc322a5bfd96c1a8268110\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T18:17:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://bae6ecfd8b446ef04d393d823e5ffe37fe749337c0ca17fde79c5f88d5e31393\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T18:17:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6411d1e7e18d5ceadaaceed5a8a6b20a0cacc55cede793ca9fb4dcad70c64fc6\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T18:17:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T18:17:10Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T18:17:38Z is after 2025-08-24T17:21:41Z" Oct 08 18:17:38 crc kubenswrapper[4859]: I1008 18:17:38.966445 4859 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-08T18:17:29Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T18:17:29Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T18:17:38Z is after 2025-08-24T17:21:41Z" Oct 08 18:17:38 crc kubenswrapper[4859]: I1008 18:17:38.980534 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 18:17:38 crc kubenswrapper[4859]: I1008 18:17:38.980579 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 18:17:38 crc kubenswrapper[4859]: I1008 18:17:38.980593 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 18:17:38 crc kubenswrapper[4859]: I1008 18:17:38.980613 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 18:17:38 crc kubenswrapper[4859]: I1008 18:17:38.980625 4859 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T18:17:38Z","lastTransitionTime":"2025-10-08T18:17:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 18:17:38 crc kubenswrapper[4859]: I1008 18:17:38.984676 4859 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-84k8x" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1c24078c-c9e8-4160-883a-f483b4b3dd23\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:17:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:17:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:17:31Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:17:31Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hk5xb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://18b291eb3344fb7c00ef5679dca9a6f7ed9d08cea684a803ff7dad2e8a18c585\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://18b291eb3344fb7c00ef5679dca9a6f7ed9d08cea684a803ff7dad2e8a18c585\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-08T18:17:32Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-08T18:17:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hk5xb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://160163f1a91d981583452db3c7a9244182d1c8ced45749a61b721d0d45c1dfe4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://160163f1a91d981583452db3c7a9244182d1c8ced45749a61b721d0d45c1dfe4\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-08T18:17:34Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-08T18:17:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hk5xb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://51e9b53f3847ac53f378ad83870debf8e71518d4edbc7fc0dc66644818e2103e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://51e9b53f3847ac53f378ad83870debf8e71518d4edbc7fc0dc66644818e2103e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-08T18:17:34Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-08T18:17:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hk5xb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1ff118919739759ac01b5cbe3240e6673db54eaa16353c609d670071ed7964cb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1ff118919739759ac01b5cbe3240e6673db54eaa16353c609d670071ed7964cb\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-08T18:17:35Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-08T18:17:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hk5xb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8522e803ff56eb57cd4d9fd036e0c3fb16240ea21d4a094c1db7aa7d50bfe2c2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8522e803ff56eb57cd4d9fd036e0c3fb16240ea21d4a094c1db7aa7d50bfe2c2\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-08T18:17:37Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-08T18:17:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hk5xb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://be61056edf515e037403ca2a0bfd5feb9b2537dd42212b58e133b23b294f327b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://be61056edf515e037403ca2a0bfd5feb9b2537dd42212b58e133b23b294f327b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-08T18:17:38Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-08T18:17:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hk5xb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T18:17:31Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-84k8x\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T18:17:38Z is after 2025-08-24T17:21:41Z" Oct 08 18:17:39 crc kubenswrapper[4859]: I1008 18:17:39.083209 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 18:17:39 crc kubenswrapper[4859]: I1008 18:17:39.083251 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 18:17:39 crc kubenswrapper[4859]: I1008 18:17:39.083264 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 18:17:39 crc kubenswrapper[4859]: I1008 18:17:39.083283 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 18:17:39 crc kubenswrapper[4859]: I1008 18:17:39.083296 4859 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T18:17:39Z","lastTransitionTime":"2025-10-08T18:17:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 18:17:39 crc kubenswrapper[4859]: I1008 18:17:39.186370 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 18:17:39 crc kubenswrapper[4859]: I1008 18:17:39.186429 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 18:17:39 crc kubenswrapper[4859]: I1008 18:17:39.186440 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 18:17:39 crc kubenswrapper[4859]: I1008 18:17:39.186458 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 18:17:39 crc kubenswrapper[4859]: I1008 18:17:39.186476 4859 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T18:17:39Z","lastTransitionTime":"2025-10-08T18:17:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 18:17:39 crc kubenswrapper[4859]: I1008 18:17:39.289822 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 18:17:39 crc kubenswrapper[4859]: I1008 18:17:39.289870 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 18:17:39 crc kubenswrapper[4859]: I1008 18:17:39.289883 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 18:17:39 crc kubenswrapper[4859]: I1008 18:17:39.289903 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 18:17:39 crc kubenswrapper[4859]: I1008 18:17:39.289914 4859 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T18:17:39Z","lastTransitionTime":"2025-10-08T18:17:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 18:17:39 crc kubenswrapper[4859]: I1008 18:17:39.393641 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 18:17:39 crc kubenswrapper[4859]: I1008 18:17:39.393680 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 18:17:39 crc kubenswrapper[4859]: I1008 18:17:39.393724 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 18:17:39 crc kubenswrapper[4859]: I1008 18:17:39.393742 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 18:17:39 crc kubenswrapper[4859]: I1008 18:17:39.393754 4859 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T18:17:39Z","lastTransitionTime":"2025-10-08T18:17:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 18:17:39 crc kubenswrapper[4859]: I1008 18:17:39.469784 4859 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 08 18:17:39 crc kubenswrapper[4859]: I1008 18:17:39.469852 4859 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 08 18:17:39 crc kubenswrapper[4859]: I1008 18:17:39.469864 4859 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 08 18:17:39 crc kubenswrapper[4859]: E1008 18:17:39.469993 4859 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 08 18:17:39 crc kubenswrapper[4859]: E1008 18:17:39.470126 4859 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 08 18:17:39 crc kubenswrapper[4859]: E1008 18:17:39.470505 4859 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 08 18:17:39 crc kubenswrapper[4859]: I1008 18:17:39.497569 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 18:17:39 crc kubenswrapper[4859]: I1008 18:17:39.497620 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 18:17:39 crc kubenswrapper[4859]: I1008 18:17:39.497632 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 18:17:39 crc kubenswrapper[4859]: I1008 18:17:39.497715 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 18:17:39 crc kubenswrapper[4859]: I1008 18:17:39.497733 4859 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T18:17:39Z","lastTransitionTime":"2025-10-08T18:17:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 18:17:39 crc kubenswrapper[4859]: I1008 18:17:39.601871 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 18:17:39 crc kubenswrapper[4859]: I1008 18:17:39.601931 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 18:17:39 crc kubenswrapper[4859]: I1008 18:17:39.601943 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 18:17:39 crc kubenswrapper[4859]: I1008 18:17:39.601964 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 18:17:39 crc kubenswrapper[4859]: I1008 18:17:39.601977 4859 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T18:17:39Z","lastTransitionTime":"2025-10-08T18:17:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 18:17:39 crc kubenswrapper[4859]: I1008 18:17:39.706977 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 18:17:39 crc kubenswrapper[4859]: I1008 18:17:39.707040 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 18:17:39 crc kubenswrapper[4859]: I1008 18:17:39.707052 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 18:17:39 crc kubenswrapper[4859]: I1008 18:17:39.707076 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 18:17:39 crc kubenswrapper[4859]: I1008 18:17:39.707089 4859 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T18:17:39Z","lastTransitionTime":"2025-10-08T18:17:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 18:17:39 crc kubenswrapper[4859]: I1008 18:17:39.727682 4859 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-4g8gf" event={"ID":"1dff864e-b75d-4e0b-b182-75f710eac8df","Type":"ContainerStarted","Data":"73a74d9d109262badcdac9ad105c1f34039fa7954910cb148175551bbd53da40"} Oct 08 18:17:39 crc kubenswrapper[4859]: I1008 18:17:39.728294 4859 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ovn-kubernetes/ovnkube-node-4g8gf" Oct 08 18:17:39 crc kubenswrapper[4859]: I1008 18:17:39.734439 4859 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-84k8x" event={"ID":"1c24078c-c9e8-4160-883a-f483b4b3dd23","Type":"ContainerStarted","Data":"8661006834d0f50eaada7394369584ee95fc61ff54437eec35d755526afc5b48"} Oct 08 18:17:39 crc kubenswrapper[4859]: I1008 18:17:39.747948 4859 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-08T18:17:31Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T18:17:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T18:17:31Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c6d15905694b6bbd4999d49dc8f7a53eb6bfaa7adb59efb0d7e4973c21d1c7be\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T18:17:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T18:17:39Z is after 2025-08-24T17:21:41Z" Oct 08 18:17:39 crc kubenswrapper[4859]: I1008 18:17:39.765207 4859 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-ovn-kubernetes/ovnkube-node-4g8gf" Oct 08 18:17:39 crc kubenswrapper[4859]: I1008 18:17:39.765624 4859 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-08T18:17:33Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T18:17:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T18:17:33Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5d99e071bd083914418d9c128f38955f26295c08b1520f2dc774c99b1c843693\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T18:17:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T18:17:39Z is after 2025-08-24T17:21:41Z" Oct 08 18:17:39 crc kubenswrapper[4859]: I1008 18:17:39.789453 4859 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-4g8gf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1dff864e-b75d-4e0b-b182-75f710eac8df\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:17:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:17:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:17:32Z\\\",\\\"message\\\":\\\"containers with unready status: [nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:17:32Z\\\",\\\"message\\\":\\\"containers with unready status: [nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://768ca274dedcc6847fc21d15234d3665fa9ccae371b5b6a882cfcb9b182f527d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T18:17:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s8jnn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://669f5f3c6d3c1ec0c4b4870e090b0a853adc01727c23383e595715b476c32668\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T18:17:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s8jnn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b8009a42cc4c9a816457a0678bdb2f7b7c53d0c8604e007a6fb141f66d06a516\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T18:17:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s8jnn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://676511a11ed4102084b57ebdd5f3fda4969741b2bc8443b415c4bf94f39ae4c9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T18:17:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s8jnn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4a74c1f0d21c90cdd7e776ee14e1f48757a04e2220fbd1f884449181ec6d1c3a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T18:17:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s8jnn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6ddc0e842bb8f81663406070023b0e5ed9cc99c98db0123b1c84f20da59fb5b5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T18:17:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s8jnn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://73a74d9d109262badcdac9ad105c1f34039fa7954910cb148175551bbd53da40\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T18:17:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s8jnn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2c67d2aed7126d76989e057d8e2d8e188a8902b2f3fe26ed6d1d45d3b013d897\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T18:17:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s8jnn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7a211d3da02b173867ff3a912eecd3b379483c7cc77e6b91122edd1ffd30fd53\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7a211d3da02b173867ff3a912eecd3b379483c7cc77e6b91122edd1ffd30fd53\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-08T18:17:32Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-08T18:17:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s8jnn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T18:17:32Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-4g8gf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T18:17:39Z is after 2025-08-24T17:21:41Z" Oct 08 18:17:39 crc kubenswrapper[4859]: I1008 18:17:39.804304 4859 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"63a623de-5608-4228-b1c2-77188c276986\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:17:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:17:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:17:10Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:17:10Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:17:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d4786a8fb8abd96fb22e60733c974f599017407d589ed815fb8541fd3b64f696\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T18:17:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://183d481502ed5108d5e19e483b714c289f355ea3f8b0891883e5335b802c83b6\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T18:17:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ec848c46fd5dde0469b00d3cab14ba4a652533a265490fadbee7a8b8843ef249\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T18:17:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4fe1828b4d63ff8106bfcd15e1dc1c1bd1ba7f77777cdb23ff1b291d0418774c\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2e700b2a20c81b414617f5b5c018a738278a3fa44e6c6d14b2170051d0725431\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-08T18:17:24Z\\\",\\\"message\\\":\\\"W1008 18:17:13.670625 1 cmd.go:257] Using insecure, self-signed certificates\\\\nI1008 18:17:13.671202 1 crypto.go:601] Generating new CA for check-endpoints-signer@1759947433 cert, and key in /tmp/serving-cert-3867218267/serving-signer.crt, /tmp/serving-cert-3867218267/serving-signer.key\\\\nI1008 18:17:13.944463 1 observer_polling.go:159] Starting file observer\\\\nW1008 18:17:13.951958 1 builder.go:272] unable to get owner reference (falling back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": dial tcp [::1]:6443: connect: connection refused\\\\nI1008 18:17:13.952088 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1008 18:17:13.956755 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3867218267/tls.crt::/tmp/serving-cert-3867218267/tls.key\\\\\\\"\\\\nF1008 18:17:24.290991 1 cmd.go:182] error initializing delegating authentication: unable to load configmap based request-header-client-ca-file: Get \\\\\\\"https://localhost:6443/api/v1/namespaces/kube-system/configmaps/extension-apiserver-authentication\\\\\\\": net/http: TLS handshake timeout\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-08T18:17:13Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T18:17:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c65b66f62d325ca10ef8d5d31f04529ccf4e691cf3303d231e7a414fb400e4d6\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T18:17:13Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://58c7bba4ace28215b062653a960e561b1dc1f890f2baf450145638a065af038b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://58c7bba4ace28215b062653a960e561b1dc1f890f2baf450145638a065af038b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-08T18:17:11Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-08T18:17:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T18:17:10Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T18:17:39Z is after 2025-08-24T17:21:41Z" Oct 08 18:17:39 crc kubenswrapper[4859]: I1008 18:17:39.809410 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 18:17:39 crc kubenswrapper[4859]: I1008 18:17:39.809462 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 18:17:39 crc kubenswrapper[4859]: I1008 18:17:39.809476 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 18:17:39 crc kubenswrapper[4859]: I1008 18:17:39.809497 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 18:17:39 crc kubenswrapper[4859]: I1008 18:17:39.809512 4859 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T18:17:39Z","lastTransitionTime":"2025-10-08T18:17:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 18:17:39 crc kubenswrapper[4859]: I1008 18:17:39.817821 4859 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-08T18:17:31Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T18:17:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T18:17:31Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://99e2d51bd7757498dae60949eea6d732b5f7b481fa00f808dd5308a530a1daf1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T18:17:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e32aaba93699944665660ae41e9b5779178c1354c31e15ad8188bf8c3c8fa500\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T18:17:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T18:17:39Z is after 2025-08-24T17:21:41Z" Oct 08 18:17:39 crc kubenswrapper[4859]: I1008 18:17:39.835630 4859 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-08T18:17:29Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T18:17:29Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T18:17:39Z is after 2025-08-24T17:21:41Z" Oct 08 18:17:39 crc kubenswrapper[4859]: I1008 18:17:39.851299 4859 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-82s52" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b23a6a6c-9d92-4e7b-840e-55cfda873a2d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:17:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:17:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:17:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:17:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://71302c766485b9f55743d85f97440ff578cd44ba1e5a6f5b1c024c0e808072d2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T18:17:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6vx22\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c23ea3580998adf90621a0155c20ee4bd590c512629812174e95c28857bcfef9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T18:17:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6vx22\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T18:17:31Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-82s52\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T18:17:39Z is after 2025-08-24T17:21:41Z" Oct 08 18:17:39 crc kubenswrapper[4859]: I1008 18:17:39.864154 4859 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4efc16fb-e679-464d-8b3c-0f46f082593e\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:17:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:17:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:17:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:17:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:17:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://aa8f8ea895272cd79a08d1c47cd7863a1ec10e21bb60d12aa10862a56c99e4bf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T18:17:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5c4e95d8a712ab219455564539a90286ae48bbc46ddc322a5bfd96c1a8268110\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T18:17:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://bae6ecfd8b446ef04d393d823e5ffe37fe749337c0ca17fde79c5f88d5e31393\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T18:17:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6411d1e7e18d5ceadaaceed5a8a6b20a0cacc55cede793ca9fb4dcad70c64fc6\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T18:17:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T18:17:10Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T18:17:39Z is after 2025-08-24T17:21:41Z" Oct 08 18:17:39 crc kubenswrapper[4859]: I1008 18:17:39.877153 4859 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-08T18:17:29Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T18:17:29Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T18:17:39Z is after 2025-08-24T17:21:41Z" Oct 08 18:17:39 crc kubenswrapper[4859]: I1008 18:17:39.899113 4859 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-84k8x" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1c24078c-c9e8-4160-883a-f483b4b3dd23\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:17:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:17:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:17:31Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:17:31Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hk5xb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://18b291eb3344fb7c00ef5679dca9a6f7ed9d08cea684a803ff7dad2e8a18c585\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://18b291eb3344fb7c00ef5679dca9a6f7ed9d08cea684a803ff7dad2e8a18c585\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-08T18:17:32Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-08T18:17:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hk5xb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://160163f1a91d981583452db3c7a9244182d1c8ced45749a61b721d0d45c1dfe4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://160163f1a91d981583452db3c7a9244182d1c8ced45749a61b721d0d45c1dfe4\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-08T18:17:34Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-08T18:17:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hk5xb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://51e9b53f3847ac53f378ad83870debf8e71518d4edbc7fc0dc66644818e2103e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://51e9b53f3847ac53f378ad83870debf8e71518d4edbc7fc0dc66644818e2103e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-08T18:17:34Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-08T18:17:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hk5xb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1ff118919739759ac01b5cbe3240e6673db54eaa16353c609d670071ed7964cb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1ff118919739759ac01b5cbe3240e6673db54eaa16353c609d670071ed7964cb\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-08T18:17:35Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-08T18:17:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hk5xb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8522e803ff56eb57cd4d9fd036e0c3fb16240ea21d4a094c1db7aa7d50bfe2c2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8522e803ff56eb57cd4d9fd036e0c3fb16240ea21d4a094c1db7aa7d50bfe2c2\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-08T18:17:37Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-08T18:17:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hk5xb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://be61056edf515e037403ca2a0bfd5feb9b2537dd42212b58e133b23b294f327b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://be61056edf515e037403ca2a0bfd5feb9b2537dd42212b58e133b23b294f327b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-08T18:17:38Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-08T18:17:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hk5xb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T18:17:31Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-84k8x\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T18:17:39Z is after 2025-08-24T17:21:41Z" Oct 08 18:17:39 crc kubenswrapper[4859]: I1008 18:17:39.913864 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 18:17:39 crc kubenswrapper[4859]: I1008 18:17:39.913903 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 18:17:39 crc kubenswrapper[4859]: I1008 18:17:39.913922 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 18:17:39 crc kubenswrapper[4859]: I1008 18:17:39.913941 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 18:17:39 crc kubenswrapper[4859]: I1008 18:17:39.913956 4859 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T18:17:39Z","lastTransitionTime":"2025-10-08T18:17:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 18:17:39 crc kubenswrapper[4859]: I1008 18:17:39.914749 4859 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-08T18:17:29Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T18:17:29Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T18:17:39Z is after 2025-08-24T17:21:41Z" Oct 08 18:17:39 crc kubenswrapper[4859]: I1008 18:17:39.926779 4859 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-btw28" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"97bc5598-3ead-4300-a938-3aebbff1e8c6\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:17:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:17:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:17:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:17:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d6d4d6ba771cb5fde637449bba05660f634df7e833a0c8dad2b2c7f3a422bb2b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T18:17:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qcxsd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T18:17:31Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-btw28\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T18:17:39Z is after 2025-08-24T17:21:41Z" Oct 08 18:17:39 crc kubenswrapper[4859]: I1008 18:17:39.939874 4859 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-8plkg" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"92b2d7de-31cc-4a91-95a5-ed5ea964b028\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:17:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:17:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:17:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:17:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fa00b8345d40b74a917ddc23cc1b2c5c684a9ca61b6119d090c92cc1439b71c4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T18:17:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q9d7m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T18:17:31Z\\\"}}\" for pod \"openshift-multus\"/\"multus-8plkg\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T18:17:39Z is after 2025-08-24T17:21:41Z" Oct 08 18:17:39 crc kubenswrapper[4859]: I1008 18:17:39.953614 4859 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-btjmq" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b8605f7-fed7-493f-9b4f-6a6358adf907\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:17:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:17:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:17:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:17:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://14e9b6c4a4403f36adb78fcf1135a620ca6fc56ed974c9639dcdfc59e6917559\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T18:17:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xspg4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T18:17:35Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-btjmq\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T18:17:39Z is after 2025-08-24T17:21:41Z" Oct 08 18:17:39 crc kubenswrapper[4859]: I1008 18:17:39.968435 4859 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-08T18:17:29Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T18:17:29Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T18:17:39Z is after 2025-08-24T17:21:41Z" Oct 08 18:17:39 crc kubenswrapper[4859]: I1008 18:17:39.980799 4859 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-btw28" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"97bc5598-3ead-4300-a938-3aebbff1e8c6\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:17:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:17:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:17:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:17:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d6d4d6ba771cb5fde637449bba05660f634df7e833a0c8dad2b2c7f3a422bb2b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T18:17:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qcxsd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T18:17:31Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-btw28\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T18:17:39Z is after 2025-08-24T17:21:41Z" Oct 08 18:17:39 crc kubenswrapper[4859]: I1008 18:17:39.994897 4859 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-8plkg" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"92b2d7de-31cc-4a91-95a5-ed5ea964b028\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:17:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:17:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:17:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:17:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fa00b8345d40b74a917ddc23cc1b2c5c684a9ca61b6119d090c92cc1439b71c4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T18:17:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q9d7m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T18:17:31Z\\\"}}\" for pod \"openshift-multus\"/\"multus-8plkg\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T18:17:39Z is after 2025-08-24T17:21:41Z" Oct 08 18:17:40 crc kubenswrapper[4859]: I1008 18:17:40.007617 4859 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-btjmq" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b8605f7-fed7-493f-9b4f-6a6358adf907\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:17:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:17:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:17:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:17:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://14e9b6c4a4403f36adb78fcf1135a620ca6fc56ed974c9639dcdfc59e6917559\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T18:17:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xspg4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T18:17:35Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-btjmq\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T18:17:40Z is after 2025-08-24T17:21:41Z" Oct 08 18:17:40 crc kubenswrapper[4859]: I1008 18:17:40.016833 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 18:17:40 crc kubenswrapper[4859]: I1008 18:17:40.016900 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 18:17:40 crc kubenswrapper[4859]: I1008 18:17:40.016912 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 18:17:40 crc kubenswrapper[4859]: I1008 18:17:40.016936 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 18:17:40 crc kubenswrapper[4859]: I1008 18:17:40.016952 4859 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T18:17:40Z","lastTransitionTime":"2025-10-08T18:17:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 18:17:40 crc kubenswrapper[4859]: I1008 18:17:40.024374 4859 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-08T18:17:33Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T18:17:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T18:17:33Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5d99e071bd083914418d9c128f38955f26295c08b1520f2dc774c99b1c843693\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T18:17:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T18:17:40Z is after 2025-08-24T17:21:41Z" Oct 08 18:17:40 crc kubenswrapper[4859]: I1008 18:17:40.041271 4859 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-08T18:17:31Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T18:17:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T18:17:31Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c6d15905694b6bbd4999d49dc8f7a53eb6bfaa7adb59efb0d7e4973c21d1c7be\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T18:17:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T18:17:40Z is after 2025-08-24T17:21:41Z" Oct 08 18:17:40 crc kubenswrapper[4859]: I1008 18:17:40.059874 4859 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"63a623de-5608-4228-b1c2-77188c276986\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:17:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:17:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:17:10Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:17:10Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:17:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d4786a8fb8abd96fb22e60733c974f599017407d589ed815fb8541fd3b64f696\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T18:17:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://183d481502ed5108d5e19e483b714c289f355ea3f8b0891883e5335b802c83b6\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T18:17:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ec848c46fd5dde0469b00d3cab14ba4a652533a265490fadbee7a8b8843ef249\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T18:17:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4fe1828b4d63ff8106bfcd15e1dc1c1bd1ba7f77777cdb23ff1b291d0418774c\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2e700b2a20c81b414617f5b5c018a738278a3fa44e6c6d14b2170051d0725431\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-08T18:17:24Z\\\",\\\"message\\\":\\\"W1008 18:17:13.670625 1 cmd.go:257] Using insecure, self-signed certificates\\\\nI1008 18:17:13.671202 1 crypto.go:601] Generating new CA for check-endpoints-signer@1759947433 cert, and key in /tmp/serving-cert-3867218267/serving-signer.crt, /tmp/serving-cert-3867218267/serving-signer.key\\\\nI1008 18:17:13.944463 1 observer_polling.go:159] Starting file observer\\\\nW1008 18:17:13.951958 1 builder.go:272] unable to get owner reference (falling back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": dial tcp [::1]:6443: connect: connection refused\\\\nI1008 18:17:13.952088 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1008 18:17:13.956755 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3867218267/tls.crt::/tmp/serving-cert-3867218267/tls.key\\\\\\\"\\\\nF1008 18:17:24.290991 1 cmd.go:182] error initializing delegating authentication: unable to load configmap based request-header-client-ca-file: Get \\\\\\\"https://localhost:6443/api/v1/namespaces/kube-system/configmaps/extension-apiserver-authentication\\\\\\\": net/http: TLS handshake timeout\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-08T18:17:13Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T18:17:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c65b66f62d325ca10ef8d5d31f04529ccf4e691cf3303d231e7a414fb400e4d6\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T18:17:13Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://58c7bba4ace28215b062653a960e561b1dc1f890f2baf450145638a065af038b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://58c7bba4ace28215b062653a960e561b1dc1f890f2baf450145638a065af038b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-08T18:17:11Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-08T18:17:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T18:17:10Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T18:17:40Z is after 2025-08-24T17:21:41Z" Oct 08 18:17:40 crc kubenswrapper[4859]: I1008 18:17:40.078957 4859 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-08T18:17:31Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T18:17:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T18:17:31Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://99e2d51bd7757498dae60949eea6d732b5f7b481fa00f808dd5308a530a1daf1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T18:17:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e32aaba93699944665660ae41e9b5779178c1354c31e15ad8188bf8c3c8fa500\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T18:17:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T18:17:40Z is after 2025-08-24T17:21:41Z" Oct 08 18:17:40 crc kubenswrapper[4859]: I1008 18:17:40.093149 4859 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-08T18:17:29Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T18:17:29Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T18:17:40Z is after 2025-08-24T17:21:41Z" Oct 08 18:17:40 crc kubenswrapper[4859]: I1008 18:17:40.107482 4859 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-82s52" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b23a6a6c-9d92-4e7b-840e-55cfda873a2d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:17:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:17:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:17:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:17:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://71302c766485b9f55743d85f97440ff578cd44ba1e5a6f5b1c024c0e808072d2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T18:17:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6vx22\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c23ea3580998adf90621a0155c20ee4bd590c512629812174e95c28857bcfef9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T18:17:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6vx22\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T18:17:31Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-82s52\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T18:17:40Z is after 2025-08-24T17:21:41Z" Oct 08 18:17:40 crc kubenswrapper[4859]: I1008 18:17:40.119735 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 18:17:40 crc kubenswrapper[4859]: I1008 18:17:40.119764 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 18:17:40 crc kubenswrapper[4859]: I1008 18:17:40.119772 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 18:17:40 crc kubenswrapper[4859]: I1008 18:17:40.119787 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 18:17:40 crc kubenswrapper[4859]: I1008 18:17:40.119797 4859 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T18:17:40Z","lastTransitionTime":"2025-10-08T18:17:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 18:17:40 crc kubenswrapper[4859]: I1008 18:17:40.129844 4859 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-4g8gf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1dff864e-b75d-4e0b-b182-75f710eac8df\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:17:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:17:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:17:32Z\\\",\\\"message\\\":\\\"containers with unready status: [nbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:17:32Z\\\",\\\"message\\\":\\\"containers with unready status: [nbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://768ca274dedcc6847fc21d15234d3665fa9ccae371b5b6a882cfcb9b182f527d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T18:17:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s8jnn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://669f5f3c6d3c1ec0c4b4870e090b0a853adc01727c23383e595715b476c32668\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T18:17:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s8jnn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b8009a42cc4c9a816457a0678bdb2f7b7c53d0c8604e007a6fb141f66d06a516\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T18:17:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s8jnn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://676511a11ed4102084b57ebdd5f3fda4969741b2bc8443b415c4bf94f39ae4c9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T18:17:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s8jnn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4a74c1f0d21c90cdd7e776ee14e1f48757a04e2220fbd1f884449181ec6d1c3a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T18:17:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s8jnn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6ddc0e842bb8f81663406070023b0e5ed9cc99c98db0123b1c84f20da59fb5b5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T18:17:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s8jnn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://73a74d9d109262badcdac9ad105c1f34039fa7954910cb148175551bbd53da40\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T18:17:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s8jnn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2c67d2aed7126d76989e057d8e2d8e188a8902b2f3fe26ed6d1d45d3b013d897\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T18:17:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s8jnn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7a211d3da02b173867ff3a912eecd3b379483c7cc77e6b91122edd1ffd30fd53\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7a211d3da02b173867ff3a912eecd3b379483c7cc77e6b91122edd1ffd30fd53\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-08T18:17:32Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-08T18:17:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s8jnn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T18:17:32Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-4g8gf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T18:17:40Z is after 2025-08-24T17:21:41Z" Oct 08 18:17:40 crc kubenswrapper[4859]: I1008 18:17:40.147583 4859 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-08T18:17:29Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T18:17:29Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T18:17:40Z is after 2025-08-24T17:21:41Z" Oct 08 18:17:40 crc kubenswrapper[4859]: I1008 18:17:40.165406 4859 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-84k8x" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1c24078c-c9e8-4160-883a-f483b4b3dd23\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:17:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:17:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:17:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:17:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8661006834d0f50eaada7394369584ee95fc61ff54437eec35d755526afc5b48\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T18:17:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hk5xb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://18b291eb3344fb7c00ef5679dca9a6f7ed9d08cea684a803ff7dad2e8a18c585\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://18b291eb3344fb7c00ef5679dca9a6f7ed9d08cea684a803ff7dad2e8a18c585\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-08T18:17:32Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-08T18:17:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hk5xb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://160163f1a91d981583452db3c7a9244182d1c8ced45749a61b721d0d45c1dfe4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://160163f1a91d981583452db3c7a9244182d1c8ced45749a61b721d0d45c1dfe4\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-08T18:17:34Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-08T18:17:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hk5xb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://51e9b53f3847ac53f378ad83870debf8e71518d4edbc7fc0dc66644818e2103e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://51e9b53f3847ac53f378ad83870debf8e71518d4edbc7fc0dc66644818e2103e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-08T18:17:34Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-08T18:17:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hk5xb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1ff118919739759ac01b5cbe3240e6673db54eaa16353c609d670071ed7964cb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1ff118919739759ac01b5cbe3240e6673db54eaa16353c609d670071ed7964cb\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-08T18:17:35Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-08T18:17:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hk5xb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8522e803ff56eb57cd4d9fd036e0c3fb16240ea21d4a094c1db7aa7d50bfe2c2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8522e803ff56eb57cd4d9fd036e0c3fb16240ea21d4a094c1db7aa7d50bfe2c2\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-08T18:17:37Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-08T18:17:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hk5xb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://be61056edf515e037403ca2a0bfd5feb9b2537dd42212b58e133b23b294f327b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://be61056edf515e037403ca2a0bfd5feb9b2537dd42212b58e133b23b294f327b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-08T18:17:38Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-08T18:17:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hk5xb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T18:17:31Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-84k8x\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T18:17:40Z is after 2025-08-24T17:21:41Z" Oct 08 18:17:40 crc kubenswrapper[4859]: I1008 18:17:40.181936 4859 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4efc16fb-e679-464d-8b3c-0f46f082593e\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:17:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:17:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:17:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:17:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:17:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://aa8f8ea895272cd79a08d1c47cd7863a1ec10e21bb60d12aa10862a56c99e4bf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T18:17:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5c4e95d8a712ab219455564539a90286ae48bbc46ddc322a5bfd96c1a8268110\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T18:17:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://bae6ecfd8b446ef04d393d823e5ffe37fe749337c0ca17fde79c5f88d5e31393\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T18:17:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6411d1e7e18d5ceadaaceed5a8a6b20a0cacc55cede793ca9fb4dcad70c64fc6\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T18:17:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T18:17:10Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T18:17:40Z is after 2025-08-24T17:21:41Z" Oct 08 18:17:40 crc kubenswrapper[4859]: I1008 18:17:40.222593 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 18:17:40 crc kubenswrapper[4859]: I1008 18:17:40.223001 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 18:17:40 crc kubenswrapper[4859]: I1008 18:17:40.223094 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 18:17:40 crc kubenswrapper[4859]: I1008 18:17:40.223194 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 18:17:40 crc kubenswrapper[4859]: I1008 18:17:40.223286 4859 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T18:17:40Z","lastTransitionTime":"2025-10-08T18:17:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 18:17:40 crc kubenswrapper[4859]: I1008 18:17:40.326314 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 18:17:40 crc kubenswrapper[4859]: I1008 18:17:40.326402 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 18:17:40 crc kubenswrapper[4859]: I1008 18:17:40.326423 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 18:17:40 crc kubenswrapper[4859]: I1008 18:17:40.326454 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 18:17:40 crc kubenswrapper[4859]: I1008 18:17:40.326477 4859 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T18:17:40Z","lastTransitionTime":"2025-10-08T18:17:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 18:17:40 crc kubenswrapper[4859]: I1008 18:17:40.429205 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 18:17:40 crc kubenswrapper[4859]: I1008 18:17:40.429262 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 18:17:40 crc kubenswrapper[4859]: I1008 18:17:40.429275 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 18:17:40 crc kubenswrapper[4859]: I1008 18:17:40.429295 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 18:17:40 crc kubenswrapper[4859]: I1008 18:17:40.429308 4859 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T18:17:40Z","lastTransitionTime":"2025-10-08T18:17:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 18:17:40 crc kubenswrapper[4859]: I1008 18:17:40.512184 4859 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-08T18:17:31Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T18:17:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T18:17:31Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c6d15905694b6bbd4999d49dc8f7a53eb6bfaa7adb59efb0d7e4973c21d1c7be\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T18:17:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T18:17:40Z is after 2025-08-24T17:21:41Z" Oct 08 18:17:40 crc kubenswrapper[4859]: I1008 18:17:40.531766 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 18:17:40 crc kubenswrapper[4859]: I1008 18:17:40.531814 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 18:17:40 crc kubenswrapper[4859]: I1008 18:17:40.531827 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 18:17:40 crc kubenswrapper[4859]: I1008 18:17:40.531846 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 18:17:40 crc kubenswrapper[4859]: I1008 18:17:40.531858 4859 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T18:17:40Z","lastTransitionTime":"2025-10-08T18:17:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 18:17:40 crc kubenswrapper[4859]: I1008 18:17:40.545736 4859 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-08T18:17:33Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T18:17:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T18:17:33Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5d99e071bd083914418d9c128f38955f26295c08b1520f2dc774c99b1c843693\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T18:17:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T18:17:40Z is after 2025-08-24T17:21:41Z" Oct 08 18:17:40 crc kubenswrapper[4859]: I1008 18:17:40.574194 4859 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-82s52" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b23a6a6c-9d92-4e7b-840e-55cfda873a2d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:17:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:17:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:17:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:17:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://71302c766485b9f55743d85f97440ff578cd44ba1e5a6f5b1c024c0e808072d2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T18:17:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6vx22\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c23ea3580998adf90621a0155c20ee4bd590c512629812174e95c28857bcfef9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T18:17:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6vx22\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T18:17:31Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-82s52\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T18:17:40Z is after 2025-08-24T17:21:41Z" Oct 08 18:17:40 crc kubenswrapper[4859]: I1008 18:17:40.593150 4859 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-4g8gf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1dff864e-b75d-4e0b-b182-75f710eac8df\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:17:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:17:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:17:32Z\\\",\\\"message\\\":\\\"containers with unready status: [nbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:17:32Z\\\",\\\"message\\\":\\\"containers with unready status: [nbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://768ca274dedcc6847fc21d15234d3665fa9ccae371b5b6a882cfcb9b182f527d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T18:17:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s8jnn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://669f5f3c6d3c1ec0c4b4870e090b0a853adc01727c23383e595715b476c32668\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T18:17:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s8jnn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b8009a42cc4c9a816457a0678bdb2f7b7c53d0c8604e007a6fb141f66d06a516\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T18:17:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s8jnn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://676511a11ed4102084b57ebdd5f3fda4969741b2bc8443b415c4bf94f39ae4c9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T18:17:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s8jnn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4a74c1f0d21c90cdd7e776ee14e1f48757a04e2220fbd1f884449181ec6d1c3a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T18:17:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s8jnn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6ddc0e842bb8f81663406070023b0e5ed9cc99c98db0123b1c84f20da59fb5b5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T18:17:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s8jnn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://73a74d9d109262badcdac9ad105c1f34039fa7954910cb148175551bbd53da40\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T18:17:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s8jnn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2c67d2aed7126d76989e057d8e2d8e188a8902b2f3fe26ed6d1d45d3b013d897\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T18:17:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s8jnn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7a211d3da02b173867ff3a912eecd3b379483c7cc77e6b91122edd1ffd30fd53\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7a211d3da02b173867ff3a912eecd3b379483c7cc77e6b91122edd1ffd30fd53\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-08T18:17:32Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-08T18:17:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s8jnn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T18:17:32Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-4g8gf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T18:17:40Z is after 2025-08-24T17:21:41Z" Oct 08 18:17:40 crc kubenswrapper[4859]: I1008 18:17:40.606469 4859 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"63a623de-5608-4228-b1c2-77188c276986\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:17:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:17:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:17:10Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:17:10Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:17:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d4786a8fb8abd96fb22e60733c974f599017407d589ed815fb8541fd3b64f696\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T18:17:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://183d481502ed5108d5e19e483b714c289f355ea3f8b0891883e5335b802c83b6\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T18:17:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ec848c46fd5dde0469b00d3cab14ba4a652533a265490fadbee7a8b8843ef249\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T18:17:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4fe1828b4d63ff8106bfcd15e1dc1c1bd1ba7f77777cdb23ff1b291d0418774c\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2e700b2a20c81b414617f5b5c018a738278a3fa44e6c6d14b2170051d0725431\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-08T18:17:24Z\\\",\\\"message\\\":\\\"W1008 18:17:13.670625 1 cmd.go:257] Using insecure, self-signed certificates\\\\nI1008 18:17:13.671202 1 crypto.go:601] Generating new CA for check-endpoints-signer@1759947433 cert, and key in /tmp/serving-cert-3867218267/serving-signer.crt, /tmp/serving-cert-3867218267/serving-signer.key\\\\nI1008 18:17:13.944463 1 observer_polling.go:159] Starting file observer\\\\nW1008 18:17:13.951958 1 builder.go:272] unable to get owner reference (falling back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": dial tcp [::1]:6443: connect: connection refused\\\\nI1008 18:17:13.952088 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1008 18:17:13.956755 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3867218267/tls.crt::/tmp/serving-cert-3867218267/tls.key\\\\\\\"\\\\nF1008 18:17:24.290991 1 cmd.go:182] error initializing delegating authentication: unable to load configmap based request-header-client-ca-file: Get \\\\\\\"https://localhost:6443/api/v1/namespaces/kube-system/configmaps/extension-apiserver-authentication\\\\\\\": net/http: TLS handshake timeout\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-08T18:17:13Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T18:17:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c65b66f62d325ca10ef8d5d31f04529ccf4e691cf3303d231e7a414fb400e4d6\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T18:17:13Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://58c7bba4ace28215b062653a960e561b1dc1f890f2baf450145638a065af038b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://58c7bba4ace28215b062653a960e561b1dc1f890f2baf450145638a065af038b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-08T18:17:11Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-08T18:17:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T18:17:10Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T18:17:40Z is after 2025-08-24T17:21:41Z" Oct 08 18:17:40 crc kubenswrapper[4859]: I1008 18:17:40.625412 4859 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-08T18:17:31Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T18:17:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T18:17:31Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://99e2d51bd7757498dae60949eea6d732b5f7b481fa00f808dd5308a530a1daf1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T18:17:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e32aaba93699944665660ae41e9b5779178c1354c31e15ad8188bf8c3c8fa500\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T18:17:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T18:17:40Z is after 2025-08-24T17:21:41Z" Oct 08 18:17:40 crc kubenswrapper[4859]: I1008 18:17:40.633503 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 18:17:40 crc kubenswrapper[4859]: I1008 18:17:40.633529 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 18:17:40 crc kubenswrapper[4859]: I1008 18:17:40.633537 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 18:17:40 crc kubenswrapper[4859]: I1008 18:17:40.633552 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 18:17:40 crc kubenswrapper[4859]: I1008 18:17:40.633560 4859 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T18:17:40Z","lastTransitionTime":"2025-10-08T18:17:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 18:17:40 crc kubenswrapper[4859]: I1008 18:17:40.636627 4859 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-08T18:17:29Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T18:17:29Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T18:17:40Z is after 2025-08-24T17:21:41Z" Oct 08 18:17:40 crc kubenswrapper[4859]: I1008 18:17:40.649618 4859 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4efc16fb-e679-464d-8b3c-0f46f082593e\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:17:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:17:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:17:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:17:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:17:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://aa8f8ea895272cd79a08d1c47cd7863a1ec10e21bb60d12aa10862a56c99e4bf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T18:17:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5c4e95d8a712ab219455564539a90286ae48bbc46ddc322a5bfd96c1a8268110\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T18:17:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://bae6ecfd8b446ef04d393d823e5ffe37fe749337c0ca17fde79c5f88d5e31393\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T18:17:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6411d1e7e18d5ceadaaceed5a8a6b20a0cacc55cede793ca9fb4dcad70c64fc6\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T18:17:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T18:17:10Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T18:17:40Z is after 2025-08-24T17:21:41Z" Oct 08 18:17:40 crc kubenswrapper[4859]: I1008 18:17:40.663426 4859 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-08T18:17:29Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T18:17:29Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T18:17:40Z is after 2025-08-24T17:21:41Z" Oct 08 18:17:40 crc kubenswrapper[4859]: I1008 18:17:40.678018 4859 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-84k8x" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1c24078c-c9e8-4160-883a-f483b4b3dd23\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:17:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:17:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:17:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:17:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8661006834d0f50eaada7394369584ee95fc61ff54437eec35d755526afc5b48\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T18:17:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hk5xb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://18b291eb3344fb7c00ef5679dca9a6f7ed9d08cea684a803ff7dad2e8a18c585\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://18b291eb3344fb7c00ef5679dca9a6f7ed9d08cea684a803ff7dad2e8a18c585\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-08T18:17:32Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-08T18:17:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hk5xb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://160163f1a91d981583452db3c7a9244182d1c8ced45749a61b721d0d45c1dfe4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://160163f1a91d981583452db3c7a9244182d1c8ced45749a61b721d0d45c1dfe4\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-08T18:17:34Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-08T18:17:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hk5xb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://51e9b53f3847ac53f378ad83870debf8e71518d4edbc7fc0dc66644818e2103e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://51e9b53f3847ac53f378ad83870debf8e71518d4edbc7fc0dc66644818e2103e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-08T18:17:34Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-08T18:17:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hk5xb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1ff118919739759ac01b5cbe3240e6673db54eaa16353c609d670071ed7964cb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1ff118919739759ac01b5cbe3240e6673db54eaa16353c609d670071ed7964cb\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-08T18:17:35Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-08T18:17:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hk5xb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8522e803ff56eb57cd4d9fd036e0c3fb16240ea21d4a094c1db7aa7d50bfe2c2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8522e803ff56eb57cd4d9fd036e0c3fb16240ea21d4a094c1db7aa7d50bfe2c2\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-08T18:17:37Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-08T18:17:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hk5xb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://be61056edf515e037403ca2a0bfd5feb9b2537dd42212b58e133b23b294f327b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://be61056edf515e037403ca2a0bfd5feb9b2537dd42212b58e133b23b294f327b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-08T18:17:38Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-08T18:17:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hk5xb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T18:17:31Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-84k8x\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T18:17:40Z is after 2025-08-24T17:21:41Z" Oct 08 18:17:40 crc kubenswrapper[4859]: I1008 18:17:40.692791 4859 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-08T18:17:29Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T18:17:29Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T18:17:40Z is after 2025-08-24T17:21:41Z" Oct 08 18:17:40 crc kubenswrapper[4859]: I1008 18:17:40.704776 4859 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-btw28" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"97bc5598-3ead-4300-a938-3aebbff1e8c6\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:17:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:17:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:17:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:17:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d6d4d6ba771cb5fde637449bba05660f634df7e833a0c8dad2b2c7f3a422bb2b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T18:17:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qcxsd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T18:17:31Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-btw28\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T18:17:40Z is after 2025-08-24T17:21:41Z" Oct 08 18:17:40 crc kubenswrapper[4859]: I1008 18:17:40.723146 4859 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-8plkg" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"92b2d7de-31cc-4a91-95a5-ed5ea964b028\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:17:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:17:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:17:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:17:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fa00b8345d40b74a917ddc23cc1b2c5c684a9ca61b6119d090c92cc1439b71c4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T18:17:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q9d7m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T18:17:31Z\\\"}}\" for pod \"openshift-multus\"/\"multus-8plkg\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T18:17:40Z is after 2025-08-24T17:21:41Z" Oct 08 18:17:40 crc kubenswrapper[4859]: I1008 18:17:40.736442 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 18:17:40 crc kubenswrapper[4859]: I1008 18:17:40.736527 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 18:17:40 crc kubenswrapper[4859]: I1008 18:17:40.736570 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 18:17:40 crc kubenswrapper[4859]: I1008 18:17:40.736606 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 18:17:40 crc kubenswrapper[4859]: I1008 18:17:40.736633 4859 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T18:17:40Z","lastTransitionTime":"2025-10-08T18:17:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 18:17:40 crc kubenswrapper[4859]: I1008 18:17:40.736678 4859 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-btjmq" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b8605f7-fed7-493f-9b4f-6a6358adf907\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:17:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:17:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:17:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:17:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://14e9b6c4a4403f36adb78fcf1135a620ca6fc56ed974c9639dcdfc59e6917559\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T18:17:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xspg4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T18:17:35Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-btjmq\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T18:17:40Z is after 2025-08-24T17:21:41Z" Oct 08 18:17:40 crc kubenswrapper[4859]: I1008 18:17:40.737102 4859 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" Oct 08 18:17:40 crc kubenswrapper[4859]: I1008 18:17:40.737672 4859 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ovn-kubernetes/ovnkube-node-4g8gf" Oct 08 18:17:40 crc kubenswrapper[4859]: I1008 18:17:40.769662 4859 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-ovn-kubernetes/ovnkube-node-4g8gf" Oct 08 18:17:40 crc kubenswrapper[4859]: I1008 18:17:40.784739 4859 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-08T18:17:31Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T18:17:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T18:17:31Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c6d15905694b6bbd4999d49dc8f7a53eb6bfaa7adb59efb0d7e4973c21d1c7be\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T18:17:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T18:17:40Z is after 2025-08-24T17:21:41Z" Oct 08 18:17:40 crc kubenswrapper[4859]: I1008 18:17:40.798822 4859 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-08T18:17:33Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T18:17:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T18:17:33Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5d99e071bd083914418d9c128f38955f26295c08b1520f2dc774c99b1c843693\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T18:17:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T18:17:40Z is after 2025-08-24T17:21:41Z" Oct 08 18:17:40 crc kubenswrapper[4859]: I1008 18:17:40.811195 4859 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-82s52" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b23a6a6c-9d92-4e7b-840e-55cfda873a2d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:17:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:17:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:17:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:17:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://71302c766485b9f55743d85f97440ff578cd44ba1e5a6f5b1c024c0e808072d2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T18:17:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6vx22\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c23ea3580998adf90621a0155c20ee4bd590c512629812174e95c28857bcfef9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T18:17:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6vx22\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T18:17:31Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-82s52\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T18:17:40Z is after 2025-08-24T17:21:41Z" Oct 08 18:17:40 crc kubenswrapper[4859]: I1008 18:17:40.830967 4859 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-4g8gf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1dff864e-b75d-4e0b-b182-75f710eac8df\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:17:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:17:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:17:32Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:17:32Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://768ca274dedcc6847fc21d15234d3665fa9ccae371b5b6a882cfcb9b182f527d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T18:17:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s8jnn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://669f5f3c6d3c1ec0c4b4870e090b0a853adc01727c23383e595715b476c32668\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T18:17:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s8jnn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b8009a42cc4c9a816457a0678bdb2f7b7c53d0c8604e007a6fb141f66d06a516\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T18:17:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s8jnn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://676511a11ed4102084b57ebdd5f3fda4969741b2bc8443b415c4bf94f39ae4c9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T18:17:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s8jnn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4a74c1f0d21c90cdd7e776ee14e1f48757a04e2220fbd1f884449181ec6d1c3a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T18:17:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s8jnn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6ddc0e842bb8f81663406070023b0e5ed9cc99c98db0123b1c84f20da59fb5b5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T18:17:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s8jnn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://73a74d9d109262badcdac9ad105c1f34039fa7954910cb148175551bbd53da40\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T18:17:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s8jnn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2c67d2aed7126d76989e057d8e2d8e188a8902b2f3fe26ed6d1d45d3b013d897\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T18:17:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s8jnn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7a211d3da02b173867ff3a912eecd3b379483c7cc77e6b91122edd1ffd30fd53\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7a211d3da02b173867ff3a912eecd3b379483c7cc77e6b91122edd1ffd30fd53\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-08T18:17:32Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-08T18:17:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s8jnn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T18:17:32Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-4g8gf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T18:17:40Z is after 2025-08-24T17:21:41Z" Oct 08 18:17:40 crc kubenswrapper[4859]: I1008 18:17:40.840446 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 18:17:40 crc kubenswrapper[4859]: I1008 18:17:40.840501 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 18:17:40 crc kubenswrapper[4859]: I1008 18:17:40.840517 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 18:17:40 crc kubenswrapper[4859]: I1008 18:17:40.840540 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 18:17:40 crc kubenswrapper[4859]: I1008 18:17:40.840554 4859 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T18:17:40Z","lastTransitionTime":"2025-10-08T18:17:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 18:17:40 crc kubenswrapper[4859]: I1008 18:17:40.854238 4859 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"63a623de-5608-4228-b1c2-77188c276986\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:17:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:17:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:17:10Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:17:10Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:17:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d4786a8fb8abd96fb22e60733c974f599017407d589ed815fb8541fd3b64f696\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T18:17:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://183d481502ed5108d5e19e483b714c289f355ea3f8b0891883e5335b802c83b6\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T18:17:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ec848c46fd5dde0469b00d3cab14ba4a652533a265490fadbee7a8b8843ef249\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T18:17:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4fe1828b4d63ff8106bfcd15e1dc1c1bd1ba7f77777cdb23ff1b291d0418774c\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2e700b2a20c81b414617f5b5c018a738278a3fa44e6c6d14b2170051d0725431\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-08T18:17:24Z\\\",\\\"message\\\":\\\"W1008 18:17:13.670625 1 cmd.go:257] Using insecure, self-signed certificates\\\\nI1008 18:17:13.671202 1 crypto.go:601] Generating new CA for check-endpoints-signer@1759947433 cert, and key in /tmp/serving-cert-3867218267/serving-signer.crt, /tmp/serving-cert-3867218267/serving-signer.key\\\\nI1008 18:17:13.944463 1 observer_polling.go:159] Starting file observer\\\\nW1008 18:17:13.951958 1 builder.go:272] unable to get owner reference (falling back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": dial tcp [::1]:6443: connect: connection refused\\\\nI1008 18:17:13.952088 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1008 18:17:13.956755 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3867218267/tls.crt::/tmp/serving-cert-3867218267/tls.key\\\\\\\"\\\\nF1008 18:17:24.290991 1 cmd.go:182] error initializing delegating authentication: unable to load configmap based request-header-client-ca-file: Get \\\\\\\"https://localhost:6443/api/v1/namespaces/kube-system/configmaps/extension-apiserver-authentication\\\\\\\": net/http: TLS handshake timeout\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-08T18:17:13Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T18:17:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c65b66f62d325ca10ef8d5d31f04529ccf4e691cf3303d231e7a414fb400e4d6\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T18:17:13Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://58c7bba4ace28215b062653a960e561b1dc1f890f2baf450145638a065af038b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://58c7bba4ace28215b062653a960e561b1dc1f890f2baf450145638a065af038b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-08T18:17:11Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-08T18:17:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T18:17:10Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T18:17:40Z is after 2025-08-24T17:21:41Z" Oct 08 18:17:40 crc kubenswrapper[4859]: I1008 18:17:40.871368 4859 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-08T18:17:31Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T18:17:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T18:17:31Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://99e2d51bd7757498dae60949eea6d732b5f7b481fa00f808dd5308a530a1daf1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T18:17:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e32aaba93699944665660ae41e9b5779178c1354c31e15ad8188bf8c3c8fa500\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T18:17:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T18:17:40Z is after 2025-08-24T17:21:41Z" Oct 08 18:17:40 crc kubenswrapper[4859]: I1008 18:17:40.885293 4859 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-08T18:17:29Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T18:17:29Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T18:17:40Z is after 2025-08-24T17:21:41Z" Oct 08 18:17:40 crc kubenswrapper[4859]: I1008 18:17:40.900256 4859 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4efc16fb-e679-464d-8b3c-0f46f082593e\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:17:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:17:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:17:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:17:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:17:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://aa8f8ea895272cd79a08d1c47cd7863a1ec10e21bb60d12aa10862a56c99e4bf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T18:17:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5c4e95d8a712ab219455564539a90286ae48bbc46ddc322a5bfd96c1a8268110\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T18:17:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://bae6ecfd8b446ef04d393d823e5ffe37fe749337c0ca17fde79c5f88d5e31393\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T18:17:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6411d1e7e18d5ceadaaceed5a8a6b20a0cacc55cede793ca9fb4dcad70c64fc6\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T18:17:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T18:17:10Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T18:17:40Z is after 2025-08-24T17:21:41Z" Oct 08 18:17:40 crc kubenswrapper[4859]: I1008 18:17:40.916784 4859 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-08T18:17:29Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T18:17:29Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T18:17:40Z is after 2025-08-24T17:21:41Z" Oct 08 18:17:40 crc kubenswrapper[4859]: I1008 18:17:40.935580 4859 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-84k8x" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1c24078c-c9e8-4160-883a-f483b4b3dd23\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:17:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:17:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:17:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:17:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8661006834d0f50eaada7394369584ee95fc61ff54437eec35d755526afc5b48\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T18:17:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hk5xb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://18b291eb3344fb7c00ef5679dca9a6f7ed9d08cea684a803ff7dad2e8a18c585\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://18b291eb3344fb7c00ef5679dca9a6f7ed9d08cea684a803ff7dad2e8a18c585\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-08T18:17:32Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-08T18:17:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hk5xb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://160163f1a91d981583452db3c7a9244182d1c8ced45749a61b721d0d45c1dfe4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://160163f1a91d981583452db3c7a9244182d1c8ced45749a61b721d0d45c1dfe4\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-08T18:17:34Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-08T18:17:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hk5xb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://51e9b53f3847ac53f378ad83870debf8e71518d4edbc7fc0dc66644818e2103e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://51e9b53f3847ac53f378ad83870debf8e71518d4edbc7fc0dc66644818e2103e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-08T18:17:34Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-08T18:17:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hk5xb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1ff118919739759ac01b5cbe3240e6673db54eaa16353c609d670071ed7964cb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1ff118919739759ac01b5cbe3240e6673db54eaa16353c609d670071ed7964cb\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-08T18:17:35Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-08T18:17:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hk5xb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8522e803ff56eb57cd4d9fd036e0c3fb16240ea21d4a094c1db7aa7d50bfe2c2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8522e803ff56eb57cd4d9fd036e0c3fb16240ea21d4a094c1db7aa7d50bfe2c2\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-08T18:17:37Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-08T18:17:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hk5xb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://be61056edf515e037403ca2a0bfd5feb9b2537dd42212b58e133b23b294f327b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://be61056edf515e037403ca2a0bfd5feb9b2537dd42212b58e133b23b294f327b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-08T18:17:38Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-08T18:17:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hk5xb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T18:17:31Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-84k8x\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T18:17:40Z is after 2025-08-24T17:21:41Z" Oct 08 18:17:40 crc kubenswrapper[4859]: I1008 18:17:40.942950 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 18:17:40 crc kubenswrapper[4859]: I1008 18:17:40.942997 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 18:17:40 crc kubenswrapper[4859]: I1008 18:17:40.943057 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 18:17:40 crc kubenswrapper[4859]: I1008 18:17:40.943113 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 18:17:40 crc kubenswrapper[4859]: I1008 18:17:40.943132 4859 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T18:17:40Z","lastTransitionTime":"2025-10-08T18:17:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 18:17:40 crc kubenswrapper[4859]: I1008 18:17:40.949833 4859 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-08T18:17:29Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T18:17:29Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T18:17:40Z is after 2025-08-24T17:21:41Z" Oct 08 18:17:40 crc kubenswrapper[4859]: I1008 18:17:40.963241 4859 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-btw28" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"97bc5598-3ead-4300-a938-3aebbff1e8c6\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:17:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:17:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:17:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:17:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d6d4d6ba771cb5fde637449bba05660f634df7e833a0c8dad2b2c7f3a422bb2b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T18:17:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qcxsd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T18:17:31Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-btw28\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T18:17:40Z is after 2025-08-24T17:21:41Z" Oct 08 18:17:40 crc kubenswrapper[4859]: I1008 18:17:40.977740 4859 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-8plkg" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"92b2d7de-31cc-4a91-95a5-ed5ea964b028\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:17:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:17:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:17:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:17:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fa00b8345d40b74a917ddc23cc1b2c5c684a9ca61b6119d090c92cc1439b71c4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T18:17:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q9d7m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T18:17:31Z\\\"}}\" for pod \"openshift-multus\"/\"multus-8plkg\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T18:17:40Z is after 2025-08-24T17:21:41Z" Oct 08 18:17:40 crc kubenswrapper[4859]: I1008 18:17:40.989032 4859 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-btjmq" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b8605f7-fed7-493f-9b4f-6a6358adf907\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:17:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:17:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:17:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:17:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://14e9b6c4a4403f36adb78fcf1135a620ca6fc56ed974c9639dcdfc59e6917559\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T18:17:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xspg4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T18:17:35Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-btjmq\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T18:17:40Z is after 2025-08-24T17:21:41Z" Oct 08 18:17:41 crc kubenswrapper[4859]: I1008 18:17:41.046233 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 18:17:41 crc kubenswrapper[4859]: I1008 18:17:41.046293 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 18:17:41 crc kubenswrapper[4859]: I1008 18:17:41.046302 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 18:17:41 crc kubenswrapper[4859]: I1008 18:17:41.046320 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 18:17:41 crc kubenswrapper[4859]: I1008 18:17:41.046330 4859 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T18:17:41Z","lastTransitionTime":"2025-10-08T18:17:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 18:17:41 crc kubenswrapper[4859]: I1008 18:17:41.149666 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 18:17:41 crc kubenswrapper[4859]: I1008 18:17:41.149734 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 18:17:41 crc kubenswrapper[4859]: I1008 18:17:41.149747 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 18:17:41 crc kubenswrapper[4859]: I1008 18:17:41.149766 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 18:17:41 crc kubenswrapper[4859]: I1008 18:17:41.149778 4859 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T18:17:41Z","lastTransitionTime":"2025-10-08T18:17:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 18:17:41 crc kubenswrapper[4859]: I1008 18:17:41.253313 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 18:17:41 crc kubenswrapper[4859]: I1008 18:17:41.253374 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 18:17:41 crc kubenswrapper[4859]: I1008 18:17:41.253387 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 18:17:41 crc kubenswrapper[4859]: I1008 18:17:41.253442 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 18:17:41 crc kubenswrapper[4859]: I1008 18:17:41.253458 4859 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T18:17:41Z","lastTransitionTime":"2025-10-08T18:17:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 18:17:41 crc kubenswrapper[4859]: I1008 18:17:41.356354 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 18:17:41 crc kubenswrapper[4859]: I1008 18:17:41.356395 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 18:17:41 crc kubenswrapper[4859]: I1008 18:17:41.356404 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 18:17:41 crc kubenswrapper[4859]: I1008 18:17:41.356420 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 18:17:41 crc kubenswrapper[4859]: I1008 18:17:41.356429 4859 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T18:17:41Z","lastTransitionTime":"2025-10-08T18:17:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 18:17:41 crc kubenswrapper[4859]: I1008 18:17:41.459152 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 18:17:41 crc kubenswrapper[4859]: I1008 18:17:41.459192 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 18:17:41 crc kubenswrapper[4859]: I1008 18:17:41.459202 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 18:17:41 crc kubenswrapper[4859]: I1008 18:17:41.459218 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 18:17:41 crc kubenswrapper[4859]: I1008 18:17:41.459228 4859 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T18:17:41Z","lastTransitionTime":"2025-10-08T18:17:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 18:17:41 crc kubenswrapper[4859]: I1008 18:17:41.467118 4859 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-kube-apiserver/kube-apiserver-crc" Oct 08 18:17:41 crc kubenswrapper[4859]: I1008 18:17:41.468998 4859 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 08 18:17:41 crc kubenswrapper[4859]: I1008 18:17:41.469030 4859 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 08 18:17:41 crc kubenswrapper[4859]: I1008 18:17:41.469006 4859 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 08 18:17:41 crc kubenswrapper[4859]: E1008 18:17:41.469124 4859 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 08 18:17:41 crc kubenswrapper[4859]: E1008 18:17:41.469276 4859 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 08 18:17:41 crc kubenswrapper[4859]: E1008 18:17:41.469411 4859 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 08 18:17:41 crc kubenswrapper[4859]: I1008 18:17:41.483115 4859 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4efc16fb-e679-464d-8b3c-0f46f082593e\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:17:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:17:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:17:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:17:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:17:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://aa8f8ea895272cd79a08d1c47cd7863a1ec10e21bb60d12aa10862a56c99e4bf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T18:17:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5c4e95d8a712ab219455564539a90286ae48bbc46ddc322a5bfd96c1a8268110\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T18:17:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://bae6ecfd8b446ef04d393d823e5ffe37fe749337c0ca17fde79c5f88d5e31393\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T18:17:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6411d1e7e18d5ceadaaceed5a8a6b20a0cacc55cede793ca9fb4dcad70c64fc6\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T18:17:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T18:17:10Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T18:17:41Z is after 2025-08-24T17:21:41Z" Oct 08 18:17:41 crc kubenswrapper[4859]: I1008 18:17:41.498808 4859 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-08T18:17:29Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T18:17:29Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T18:17:41Z is after 2025-08-24T17:21:41Z" Oct 08 18:17:41 crc kubenswrapper[4859]: I1008 18:17:41.514332 4859 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-84k8x" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1c24078c-c9e8-4160-883a-f483b4b3dd23\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:17:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:17:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:17:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:17:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8661006834d0f50eaada7394369584ee95fc61ff54437eec35d755526afc5b48\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T18:17:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hk5xb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://18b291eb3344fb7c00ef5679dca9a6f7ed9d08cea684a803ff7dad2e8a18c585\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://18b291eb3344fb7c00ef5679dca9a6f7ed9d08cea684a803ff7dad2e8a18c585\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-08T18:17:32Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-08T18:17:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hk5xb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://160163f1a91d981583452db3c7a9244182d1c8ced45749a61b721d0d45c1dfe4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://160163f1a91d981583452db3c7a9244182d1c8ced45749a61b721d0d45c1dfe4\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-08T18:17:34Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-08T18:17:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hk5xb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://51e9b53f3847ac53f378ad83870debf8e71518d4edbc7fc0dc66644818e2103e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://51e9b53f3847ac53f378ad83870debf8e71518d4edbc7fc0dc66644818e2103e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-08T18:17:34Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-08T18:17:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hk5xb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1ff118919739759ac01b5cbe3240e6673db54eaa16353c609d670071ed7964cb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1ff118919739759ac01b5cbe3240e6673db54eaa16353c609d670071ed7964cb\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-08T18:17:35Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-08T18:17:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hk5xb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8522e803ff56eb57cd4d9fd036e0c3fb16240ea21d4a094c1db7aa7d50bfe2c2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8522e803ff56eb57cd4d9fd036e0c3fb16240ea21d4a094c1db7aa7d50bfe2c2\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-08T18:17:37Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-08T18:17:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hk5xb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://be61056edf515e037403ca2a0bfd5feb9b2537dd42212b58e133b23b294f327b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://be61056edf515e037403ca2a0bfd5feb9b2537dd42212b58e133b23b294f327b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-08T18:17:38Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-08T18:17:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hk5xb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T18:17:31Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-84k8x\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T18:17:41Z is after 2025-08-24T17:21:41Z" Oct 08 18:17:41 crc kubenswrapper[4859]: I1008 18:17:41.530957 4859 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-08T18:17:29Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T18:17:29Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T18:17:41Z is after 2025-08-24T17:21:41Z" Oct 08 18:17:41 crc kubenswrapper[4859]: I1008 18:17:41.545615 4859 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-btw28" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"97bc5598-3ead-4300-a938-3aebbff1e8c6\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:17:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:17:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:17:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:17:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d6d4d6ba771cb5fde637449bba05660f634df7e833a0c8dad2b2c7f3a422bb2b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T18:17:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qcxsd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T18:17:31Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-btw28\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T18:17:41Z is after 2025-08-24T17:21:41Z" Oct 08 18:17:41 crc kubenswrapper[4859]: I1008 18:17:41.561966 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 18:17:41 crc kubenswrapper[4859]: I1008 18:17:41.562050 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 18:17:41 crc kubenswrapper[4859]: I1008 18:17:41.562067 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 18:17:41 crc kubenswrapper[4859]: I1008 18:17:41.562103 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 18:17:41 crc kubenswrapper[4859]: I1008 18:17:41.562120 4859 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T18:17:41Z","lastTransitionTime":"2025-10-08T18:17:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 18:17:41 crc kubenswrapper[4859]: I1008 18:17:41.565135 4859 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-8plkg" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"92b2d7de-31cc-4a91-95a5-ed5ea964b028\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:17:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:17:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:17:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:17:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fa00b8345d40b74a917ddc23cc1b2c5c684a9ca61b6119d090c92cc1439b71c4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T18:17:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q9d7m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T18:17:31Z\\\"}}\" for pod \"openshift-multus\"/\"multus-8plkg\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T18:17:41Z is after 2025-08-24T17:21:41Z" Oct 08 18:17:41 crc kubenswrapper[4859]: I1008 18:17:41.580958 4859 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-btjmq" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b8605f7-fed7-493f-9b4f-6a6358adf907\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:17:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:17:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:17:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:17:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://14e9b6c4a4403f36adb78fcf1135a620ca6fc56ed974c9639dcdfc59e6917559\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T18:17:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xspg4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T18:17:35Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-btjmq\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T18:17:41Z is after 2025-08-24T17:21:41Z" Oct 08 18:17:41 crc kubenswrapper[4859]: I1008 18:17:41.598056 4859 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-08T18:17:31Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T18:17:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T18:17:31Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c6d15905694b6bbd4999d49dc8f7a53eb6bfaa7adb59efb0d7e4973c21d1c7be\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T18:17:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T18:17:41Z is after 2025-08-24T17:21:41Z" Oct 08 18:17:41 crc kubenswrapper[4859]: I1008 18:17:41.614120 4859 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-08T18:17:33Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T18:17:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T18:17:33Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5d99e071bd083914418d9c128f38955f26295c08b1520f2dc774c99b1c843693\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T18:17:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T18:17:41Z is after 2025-08-24T17:21:41Z" Oct 08 18:17:41 crc kubenswrapper[4859]: I1008 18:17:41.630199 4859 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"63a623de-5608-4228-b1c2-77188c276986\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:17:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:17:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:17:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:17:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:17:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d4786a8fb8abd96fb22e60733c974f599017407d589ed815fb8541fd3b64f696\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T18:17:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://183d481502ed5108d5e19e483b714c289f355ea3f8b0891883e5335b802c83b6\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T18:17:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ec848c46fd5dde0469b00d3cab14ba4a652533a265490fadbee7a8b8843ef249\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T18:17:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4fe1828b4d63ff8106bfcd15e1dc1c1bd1ba7f77777cdb23ff1b291d0418774c\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2e700b2a20c81b414617f5b5c018a738278a3fa44e6c6d14b2170051d0725431\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-08T18:17:24Z\\\",\\\"message\\\":\\\"W1008 18:17:13.670625 1 cmd.go:257] Using insecure, self-signed certificates\\\\nI1008 18:17:13.671202 1 crypto.go:601] Generating new CA for check-endpoints-signer@1759947433 cert, and key in /tmp/serving-cert-3867218267/serving-signer.crt, /tmp/serving-cert-3867218267/serving-signer.key\\\\nI1008 18:17:13.944463 1 observer_polling.go:159] Starting file observer\\\\nW1008 18:17:13.951958 1 builder.go:272] unable to get owner reference (falling back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": dial tcp [::1]:6443: connect: connection refused\\\\nI1008 18:17:13.952088 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1008 18:17:13.956755 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3867218267/tls.crt::/tmp/serving-cert-3867218267/tls.key\\\\\\\"\\\\nF1008 18:17:24.290991 1 cmd.go:182] error initializing delegating authentication: unable to load configmap based request-header-client-ca-file: Get \\\\\\\"https://localhost:6443/api/v1/namespaces/kube-system/configmaps/extension-apiserver-authentication\\\\\\\": net/http: TLS handshake timeout\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-08T18:17:13Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T18:17:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c65b66f62d325ca10ef8d5d31f04529ccf4e691cf3303d231e7a414fb400e4d6\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T18:17:13Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://58c7bba4ace28215b062653a960e561b1dc1f890f2baf450145638a065af038b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://58c7bba4ace28215b062653a960e561b1dc1f890f2baf450145638a065af038b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-08T18:17:11Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-08T18:17:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T18:17:10Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T18:17:41Z is after 2025-08-24T17:21:41Z" Oct 08 18:17:41 crc kubenswrapper[4859]: I1008 18:17:41.644080 4859 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-08T18:17:31Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T18:17:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T18:17:31Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://99e2d51bd7757498dae60949eea6d732b5f7b481fa00f808dd5308a530a1daf1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T18:17:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e32aaba93699944665660ae41e9b5779178c1354c31e15ad8188bf8c3c8fa500\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T18:17:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T18:17:41Z is after 2025-08-24T17:21:41Z" Oct 08 18:17:41 crc kubenswrapper[4859]: I1008 18:17:41.655407 4859 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-08T18:17:29Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T18:17:29Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T18:17:41Z is after 2025-08-24T17:21:41Z" Oct 08 18:17:41 crc kubenswrapper[4859]: I1008 18:17:41.664478 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 18:17:41 crc kubenswrapper[4859]: I1008 18:17:41.664518 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 18:17:41 crc kubenswrapper[4859]: I1008 18:17:41.664531 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 18:17:41 crc kubenswrapper[4859]: I1008 18:17:41.664551 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 18:17:41 crc kubenswrapper[4859]: I1008 18:17:41.664564 4859 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T18:17:41Z","lastTransitionTime":"2025-10-08T18:17:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 18:17:41 crc kubenswrapper[4859]: I1008 18:17:41.667517 4859 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-82s52" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b23a6a6c-9d92-4e7b-840e-55cfda873a2d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:17:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:17:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:17:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:17:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://71302c766485b9f55743d85f97440ff578cd44ba1e5a6f5b1c024c0e808072d2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T18:17:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6vx22\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c23ea3580998adf90621a0155c20ee4bd590c512629812174e95c28857bcfef9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T18:17:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6vx22\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T18:17:31Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-82s52\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T18:17:41Z is after 2025-08-24T17:21:41Z" Oct 08 18:17:41 crc kubenswrapper[4859]: I1008 18:17:41.687379 4859 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-4g8gf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1dff864e-b75d-4e0b-b182-75f710eac8df\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:17:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:17:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:17:32Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:17:32Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://768ca274dedcc6847fc21d15234d3665fa9ccae371b5b6a882cfcb9b182f527d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T18:17:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s8jnn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://669f5f3c6d3c1ec0c4b4870e090b0a853adc01727c23383e595715b476c32668\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T18:17:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s8jnn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b8009a42cc4c9a816457a0678bdb2f7b7c53d0c8604e007a6fb141f66d06a516\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T18:17:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s8jnn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://676511a11ed4102084b57ebdd5f3fda4969741b2bc8443b415c4bf94f39ae4c9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T18:17:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s8jnn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4a74c1f0d21c90cdd7e776ee14e1f48757a04e2220fbd1f884449181ec6d1c3a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T18:17:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s8jnn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6ddc0e842bb8f81663406070023b0e5ed9cc99c98db0123b1c84f20da59fb5b5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T18:17:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s8jnn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://73a74d9d109262badcdac9ad105c1f34039fa7954910cb148175551bbd53da40\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T18:17:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s8jnn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2c67d2aed7126d76989e057d8e2d8e188a8902b2f3fe26ed6d1d45d3b013d897\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T18:17:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s8jnn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7a211d3da02b173867ff3a912eecd3b379483c7cc77e6b91122edd1ffd30fd53\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7a211d3da02b173867ff3a912eecd3b379483c7cc77e6b91122edd1ffd30fd53\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-08T18:17:32Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-08T18:17:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s8jnn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T18:17:32Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-4g8gf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T18:17:41Z is after 2025-08-24T17:21:41Z" Oct 08 18:17:41 crc kubenswrapper[4859]: I1008 18:17:41.740100 4859 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" Oct 08 18:17:41 crc kubenswrapper[4859]: I1008 18:17:41.770882 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 18:17:41 crc kubenswrapper[4859]: I1008 18:17:41.770952 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 18:17:41 crc kubenswrapper[4859]: I1008 18:17:41.770969 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 18:17:41 crc kubenswrapper[4859]: I1008 18:17:41.770995 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 18:17:41 crc kubenswrapper[4859]: I1008 18:17:41.771017 4859 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T18:17:41Z","lastTransitionTime":"2025-10-08T18:17:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 18:17:41 crc kubenswrapper[4859]: I1008 18:17:41.873598 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 18:17:41 crc kubenswrapper[4859]: I1008 18:17:41.873648 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 18:17:41 crc kubenswrapper[4859]: I1008 18:17:41.873662 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 18:17:41 crc kubenswrapper[4859]: I1008 18:17:41.873706 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 18:17:41 crc kubenswrapper[4859]: I1008 18:17:41.873723 4859 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T18:17:41Z","lastTransitionTime":"2025-10-08T18:17:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 18:17:41 crc kubenswrapper[4859]: I1008 18:17:41.977069 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 18:17:41 crc kubenswrapper[4859]: I1008 18:17:41.977113 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 18:17:41 crc kubenswrapper[4859]: I1008 18:17:41.977144 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 18:17:41 crc kubenswrapper[4859]: I1008 18:17:41.977166 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 18:17:41 crc kubenswrapper[4859]: I1008 18:17:41.977178 4859 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T18:17:41Z","lastTransitionTime":"2025-10-08T18:17:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 18:17:42 crc kubenswrapper[4859]: I1008 18:17:42.080022 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 18:17:42 crc kubenswrapper[4859]: I1008 18:17:42.080070 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 18:17:42 crc kubenswrapper[4859]: I1008 18:17:42.080083 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 18:17:42 crc kubenswrapper[4859]: I1008 18:17:42.080106 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 18:17:42 crc kubenswrapper[4859]: I1008 18:17:42.080119 4859 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T18:17:42Z","lastTransitionTime":"2025-10-08T18:17:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 18:17:42 crc kubenswrapper[4859]: I1008 18:17:42.183272 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 18:17:42 crc kubenswrapper[4859]: I1008 18:17:42.183351 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 18:17:42 crc kubenswrapper[4859]: I1008 18:17:42.183375 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 18:17:42 crc kubenswrapper[4859]: I1008 18:17:42.183406 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 18:17:42 crc kubenswrapper[4859]: I1008 18:17:42.183425 4859 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T18:17:42Z","lastTransitionTime":"2025-10-08T18:17:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 18:17:42 crc kubenswrapper[4859]: I1008 18:17:42.287061 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 18:17:42 crc kubenswrapper[4859]: I1008 18:17:42.287103 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 18:17:42 crc kubenswrapper[4859]: I1008 18:17:42.287115 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 18:17:42 crc kubenswrapper[4859]: I1008 18:17:42.287133 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 18:17:42 crc kubenswrapper[4859]: I1008 18:17:42.287145 4859 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T18:17:42Z","lastTransitionTime":"2025-10-08T18:17:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 18:17:42 crc kubenswrapper[4859]: I1008 18:17:42.390397 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 18:17:42 crc kubenswrapper[4859]: I1008 18:17:42.390450 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 18:17:42 crc kubenswrapper[4859]: I1008 18:17:42.390467 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 18:17:42 crc kubenswrapper[4859]: I1008 18:17:42.390491 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 18:17:42 crc kubenswrapper[4859]: I1008 18:17:42.390509 4859 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T18:17:42Z","lastTransitionTime":"2025-10-08T18:17:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 18:17:42 crc kubenswrapper[4859]: I1008 18:17:42.494137 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 18:17:42 crc kubenswrapper[4859]: I1008 18:17:42.494213 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 18:17:42 crc kubenswrapper[4859]: I1008 18:17:42.494240 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 18:17:42 crc kubenswrapper[4859]: I1008 18:17:42.494273 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 18:17:42 crc kubenswrapper[4859]: I1008 18:17:42.494297 4859 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T18:17:42Z","lastTransitionTime":"2025-10-08T18:17:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 18:17:42 crc kubenswrapper[4859]: I1008 18:17:42.597916 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 18:17:42 crc kubenswrapper[4859]: I1008 18:17:42.598001 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 18:17:42 crc kubenswrapper[4859]: I1008 18:17:42.598024 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 18:17:42 crc kubenswrapper[4859]: I1008 18:17:42.598050 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 18:17:42 crc kubenswrapper[4859]: I1008 18:17:42.598067 4859 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T18:17:42Z","lastTransitionTime":"2025-10-08T18:17:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 18:17:42 crc kubenswrapper[4859]: I1008 18:17:42.701290 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 18:17:42 crc kubenswrapper[4859]: I1008 18:17:42.701364 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 18:17:42 crc kubenswrapper[4859]: I1008 18:17:42.701384 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 18:17:42 crc kubenswrapper[4859]: I1008 18:17:42.701415 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 18:17:42 crc kubenswrapper[4859]: I1008 18:17:42.701438 4859 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T18:17:42Z","lastTransitionTime":"2025-10-08T18:17:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 18:17:42 crc kubenswrapper[4859]: I1008 18:17:42.746092 4859 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-4g8gf_1dff864e-b75d-4e0b-b182-75f710eac8df/ovnkube-controller/0.log" Oct 08 18:17:42 crc kubenswrapper[4859]: I1008 18:17:42.749738 4859 generic.go:334] "Generic (PLEG): container finished" podID="1dff864e-b75d-4e0b-b182-75f710eac8df" containerID="73a74d9d109262badcdac9ad105c1f34039fa7954910cb148175551bbd53da40" exitCode=1 Oct 08 18:17:42 crc kubenswrapper[4859]: I1008 18:17:42.749824 4859 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-4g8gf" event={"ID":"1dff864e-b75d-4e0b-b182-75f710eac8df","Type":"ContainerDied","Data":"73a74d9d109262badcdac9ad105c1f34039fa7954910cb148175551bbd53da40"} Oct 08 18:17:42 crc kubenswrapper[4859]: I1008 18:17:42.751514 4859 scope.go:117] "RemoveContainer" containerID="73a74d9d109262badcdac9ad105c1f34039fa7954910cb148175551bbd53da40" Oct 08 18:17:42 crc kubenswrapper[4859]: I1008 18:17:42.770599 4859 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-8plkg" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"92b2d7de-31cc-4a91-95a5-ed5ea964b028\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:17:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:17:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:17:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:17:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fa00b8345d40b74a917ddc23cc1b2c5c684a9ca61b6119d090c92cc1439b71c4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T18:17:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q9d7m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T18:17:31Z\\\"}}\" for pod \"openshift-multus\"/\"multus-8plkg\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T18:17:42Z is after 2025-08-24T17:21:41Z" Oct 08 18:17:42 crc kubenswrapper[4859]: I1008 18:17:42.788275 4859 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-btjmq" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b8605f7-fed7-493f-9b4f-6a6358adf907\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:17:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:17:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:17:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:17:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://14e9b6c4a4403f36adb78fcf1135a620ca6fc56ed974c9639dcdfc59e6917559\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T18:17:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xspg4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T18:17:35Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-btjmq\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T18:17:42Z is after 2025-08-24T17:21:41Z" Oct 08 18:17:42 crc kubenswrapper[4859]: I1008 18:17:42.804913 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 18:17:42 crc kubenswrapper[4859]: I1008 18:17:42.804958 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 18:17:42 crc kubenswrapper[4859]: I1008 18:17:42.804968 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 18:17:42 crc kubenswrapper[4859]: I1008 18:17:42.804990 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 18:17:42 crc kubenswrapper[4859]: I1008 18:17:42.805011 4859 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T18:17:42Z","lastTransitionTime":"2025-10-08T18:17:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 18:17:42 crc kubenswrapper[4859]: I1008 18:17:42.809121 4859 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-08T18:17:29Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T18:17:29Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T18:17:42Z is after 2025-08-24T17:21:41Z" Oct 08 18:17:42 crc kubenswrapper[4859]: I1008 18:17:42.822924 4859 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-btw28" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"97bc5598-3ead-4300-a938-3aebbff1e8c6\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:17:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:17:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:17:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:17:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d6d4d6ba771cb5fde637449bba05660f634df7e833a0c8dad2b2c7f3a422bb2b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T18:17:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qcxsd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T18:17:31Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-btw28\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T18:17:42Z is after 2025-08-24T17:21:41Z" Oct 08 18:17:42 crc kubenswrapper[4859]: I1008 18:17:42.836668 4859 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-08T18:17:31Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T18:17:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T18:17:31Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c6d15905694b6bbd4999d49dc8f7a53eb6bfaa7adb59efb0d7e4973c21d1c7be\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T18:17:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T18:17:42Z is after 2025-08-24T17:21:41Z" Oct 08 18:17:42 crc kubenswrapper[4859]: I1008 18:17:42.848285 4859 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-08T18:17:33Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T18:17:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T18:17:33Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5d99e071bd083914418d9c128f38955f26295c08b1520f2dc774c99b1c843693\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T18:17:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T18:17:42Z is after 2025-08-24T17:21:41Z" Oct 08 18:17:42 crc kubenswrapper[4859]: I1008 18:17:42.863347 4859 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-08T18:17:31Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T18:17:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T18:17:31Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://99e2d51bd7757498dae60949eea6d732b5f7b481fa00f808dd5308a530a1daf1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T18:17:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e32aaba93699944665660ae41e9b5779178c1354c31e15ad8188bf8c3c8fa500\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T18:17:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T18:17:42Z is after 2025-08-24T17:21:41Z" Oct 08 18:17:42 crc kubenswrapper[4859]: I1008 18:17:42.878290 4859 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-08T18:17:29Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T18:17:29Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T18:17:42Z is after 2025-08-24T17:21:41Z" Oct 08 18:17:42 crc kubenswrapper[4859]: I1008 18:17:42.895608 4859 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-82s52" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b23a6a6c-9d92-4e7b-840e-55cfda873a2d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:17:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:17:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:17:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:17:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://71302c766485b9f55743d85f97440ff578cd44ba1e5a6f5b1c024c0e808072d2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T18:17:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6vx22\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c23ea3580998adf90621a0155c20ee4bd590c512629812174e95c28857bcfef9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T18:17:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6vx22\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T18:17:31Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-82s52\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T18:17:42Z is after 2025-08-24T17:21:41Z" Oct 08 18:17:42 crc kubenswrapper[4859]: I1008 18:17:42.913117 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 18:17:42 crc kubenswrapper[4859]: I1008 18:17:42.913169 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 18:17:42 crc kubenswrapper[4859]: I1008 18:17:42.913179 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 18:17:42 crc kubenswrapper[4859]: I1008 18:17:42.913199 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 18:17:42 crc kubenswrapper[4859]: I1008 18:17:42.913212 4859 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T18:17:42Z","lastTransitionTime":"2025-10-08T18:17:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 18:17:42 crc kubenswrapper[4859]: I1008 18:17:42.925372 4859 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-4g8gf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1dff864e-b75d-4e0b-b182-75f710eac8df\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:17:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:17:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:17:32Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:17:32Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://768ca274dedcc6847fc21d15234d3665fa9ccae371b5b6a882cfcb9b182f527d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T18:17:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s8jnn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://669f5f3c6d3c1ec0c4b4870e090b0a853adc01727c23383e595715b476c32668\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T18:17:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s8jnn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b8009a42cc4c9a816457a0678bdb2f7b7c53d0c8604e007a6fb141f66d06a516\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T18:17:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s8jnn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://676511a11ed4102084b57ebdd5f3fda4969741b2bc8443b415c4bf94f39ae4c9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T18:17:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s8jnn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4a74c1f0d21c90cdd7e776ee14e1f48757a04e2220fbd1f884449181ec6d1c3a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T18:17:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s8jnn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6ddc0e842bb8f81663406070023b0e5ed9cc99c98db0123b1c84f20da59fb5b5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T18:17:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s8jnn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://73a74d9d109262badcdac9ad105c1f34039fa7954910cb148175551bbd53da40\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://73a74d9d109262badcdac9ad105c1f34039fa7954910cb148175551bbd53da40\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-10-08T18:17:41Z\\\",\\\"message\\\":\\\"l\\\\nI1008 18:17:41.904298 6177 handler.go:190] Sending *v1.Namespace event handler 5 for removal\\\\nI1008 18:17:41.904331 6177 handler.go:208] Removed *v1.Namespace event handler 5\\\\nI1008 18:17:41.904343 6177 handler.go:208] Removed *v1.Namespace event handler 1\\\\nI1008 18:17:41.905055 6177 handler.go:190] Sending *v1.Pod event handler 3 for removal\\\\nI1008 18:17:41.905088 6177 handler.go:190] Sending *v1.Pod event handler 6 for removal\\\\nI1008 18:17:41.905116 6177 handler.go:208] Removed *v1.Pod event handler 6\\\\nI1008 18:17:41.905126 6177 handler.go:208] Removed *v1.Pod event handler 3\\\\nI1008 18:17:41.907806 6177 handler.go:190] Sending *v1.Node event handler 2 for removal\\\\nI1008 18:17:41.907827 6177 handler.go:190] Sending *v1.Node event handler 7 for removal\\\\nI1008 18:17:41.907850 6177 handler.go:190] Sending *v1.EgressIP event handler 8 for removal\\\\nI1008 18:17:41.907871 6177 handler.go:190] Sending *v1.EgressFirewall event handler 9 for removal\\\\nI1008 18:17:41.907971 6177 factory.go:656] Stopping watch factory\\\\nI1008 18:17:41.907995 6177 ovnkube.go:599] Stopped ovnkube\\\\nI1008 18:17:41.908032 6177 handler.go:208] Removed *v1.Node event handler 2\\\\nI1008 18:17:41.908051 6177 handler.go:208] Removed *v1.Node event handler 7\\\\nI1008 18:17:41.908065 6177 handler.go:208] Removed *v1.EgressIP event handler 8\\\\nI1008 18:17:4\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-08T18:17:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s8jnn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2c67d2aed7126d76989e057d8e2d8e188a8902b2f3fe26ed6d1d45d3b013d897\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T18:17:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s8jnn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7a211d3da02b173867ff3a912eecd3b379483c7cc77e6b91122edd1ffd30fd53\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7a211d3da02b173867ff3a912eecd3b379483c7cc77e6b91122edd1ffd30fd53\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-08T18:17:32Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-08T18:17:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s8jnn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T18:17:32Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-4g8gf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T18:17:42Z is after 2025-08-24T17:21:41Z" Oct 08 18:17:42 crc kubenswrapper[4859]: I1008 18:17:42.945839 4859 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"63a623de-5608-4228-b1c2-77188c276986\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:17:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:17:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:17:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:17:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:17:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d4786a8fb8abd96fb22e60733c974f599017407d589ed815fb8541fd3b64f696\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T18:17:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://183d481502ed5108d5e19e483b714c289f355ea3f8b0891883e5335b802c83b6\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T18:17:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ec848c46fd5dde0469b00d3cab14ba4a652533a265490fadbee7a8b8843ef249\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T18:17:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4fe1828b4d63ff8106bfcd15e1dc1c1bd1ba7f77777cdb23ff1b291d0418774c\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2e700b2a20c81b414617f5b5c018a738278a3fa44e6c6d14b2170051d0725431\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-08T18:17:24Z\\\",\\\"message\\\":\\\"W1008 18:17:13.670625 1 cmd.go:257] Using insecure, self-signed certificates\\\\nI1008 18:17:13.671202 1 crypto.go:601] Generating new CA for check-endpoints-signer@1759947433 cert, and key in /tmp/serving-cert-3867218267/serving-signer.crt, /tmp/serving-cert-3867218267/serving-signer.key\\\\nI1008 18:17:13.944463 1 observer_polling.go:159] Starting file observer\\\\nW1008 18:17:13.951958 1 builder.go:272] unable to get owner reference (falling back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": dial tcp [::1]:6443: connect: connection refused\\\\nI1008 18:17:13.952088 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1008 18:17:13.956755 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3867218267/tls.crt::/tmp/serving-cert-3867218267/tls.key\\\\\\\"\\\\nF1008 18:17:24.290991 1 cmd.go:182] error initializing delegating authentication: unable to load configmap based request-header-client-ca-file: Get \\\\\\\"https://localhost:6443/api/v1/namespaces/kube-system/configmaps/extension-apiserver-authentication\\\\\\\": net/http: TLS handshake timeout\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-08T18:17:13Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T18:17:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c65b66f62d325ca10ef8d5d31f04529ccf4e691cf3303d231e7a414fb400e4d6\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T18:17:13Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://58c7bba4ace28215b062653a960e561b1dc1f890f2baf450145638a065af038b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://58c7bba4ace28215b062653a960e561b1dc1f890f2baf450145638a065af038b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-08T18:17:11Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-08T18:17:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T18:17:10Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T18:17:42Z is after 2025-08-24T17:21:41Z" Oct 08 18:17:42 crc kubenswrapper[4859]: I1008 18:17:42.964470 4859 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4efc16fb-e679-464d-8b3c-0f46f082593e\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:17:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:17:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:17:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:17:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:17:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://aa8f8ea895272cd79a08d1c47cd7863a1ec10e21bb60d12aa10862a56c99e4bf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T18:17:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5c4e95d8a712ab219455564539a90286ae48bbc46ddc322a5bfd96c1a8268110\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T18:17:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://bae6ecfd8b446ef04d393d823e5ffe37fe749337c0ca17fde79c5f88d5e31393\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T18:17:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6411d1e7e18d5ceadaaceed5a8a6b20a0cacc55cede793ca9fb4dcad70c64fc6\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T18:17:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T18:17:10Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T18:17:42Z is after 2025-08-24T17:21:41Z" Oct 08 18:17:42 crc kubenswrapper[4859]: I1008 18:17:42.979246 4859 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-08T18:17:29Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T18:17:29Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T18:17:42Z is after 2025-08-24T17:21:41Z" Oct 08 18:17:43 crc kubenswrapper[4859]: I1008 18:17:43.003959 4859 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-84k8x" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1c24078c-c9e8-4160-883a-f483b4b3dd23\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:17:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:17:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:17:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:17:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8661006834d0f50eaada7394369584ee95fc61ff54437eec35d755526afc5b48\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T18:17:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hk5xb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://18b291eb3344fb7c00ef5679dca9a6f7ed9d08cea684a803ff7dad2e8a18c585\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://18b291eb3344fb7c00ef5679dca9a6f7ed9d08cea684a803ff7dad2e8a18c585\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-08T18:17:32Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-08T18:17:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hk5xb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://160163f1a91d981583452db3c7a9244182d1c8ced45749a61b721d0d45c1dfe4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://160163f1a91d981583452db3c7a9244182d1c8ced45749a61b721d0d45c1dfe4\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-08T18:17:34Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-08T18:17:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hk5xb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://51e9b53f3847ac53f378ad83870debf8e71518d4edbc7fc0dc66644818e2103e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://51e9b53f3847ac53f378ad83870debf8e71518d4edbc7fc0dc66644818e2103e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-08T18:17:34Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-08T18:17:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hk5xb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1ff118919739759ac01b5cbe3240e6673db54eaa16353c609d670071ed7964cb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1ff118919739759ac01b5cbe3240e6673db54eaa16353c609d670071ed7964cb\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-08T18:17:35Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-08T18:17:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hk5xb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8522e803ff56eb57cd4d9fd036e0c3fb16240ea21d4a094c1db7aa7d50bfe2c2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8522e803ff56eb57cd4d9fd036e0c3fb16240ea21d4a094c1db7aa7d50bfe2c2\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-08T18:17:37Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-08T18:17:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hk5xb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://be61056edf515e037403ca2a0bfd5feb9b2537dd42212b58e133b23b294f327b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://be61056edf515e037403ca2a0bfd5feb9b2537dd42212b58e133b23b294f327b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-08T18:17:38Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-08T18:17:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hk5xb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T18:17:31Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-84k8x\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T18:17:43Z is after 2025-08-24T17:21:41Z" Oct 08 18:17:43 crc kubenswrapper[4859]: I1008 18:17:43.015755 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 18:17:43 crc kubenswrapper[4859]: I1008 18:17:43.015813 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 18:17:43 crc kubenswrapper[4859]: I1008 18:17:43.015873 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 18:17:43 crc kubenswrapper[4859]: I1008 18:17:43.015898 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 18:17:43 crc kubenswrapper[4859]: I1008 18:17:43.015910 4859 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T18:17:43Z","lastTransitionTime":"2025-10-08T18:17:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 18:17:43 crc kubenswrapper[4859]: I1008 18:17:43.119836 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 18:17:43 crc kubenswrapper[4859]: I1008 18:17:43.119904 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 18:17:43 crc kubenswrapper[4859]: I1008 18:17:43.119924 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 18:17:43 crc kubenswrapper[4859]: I1008 18:17:43.119951 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 18:17:43 crc kubenswrapper[4859]: I1008 18:17:43.119971 4859 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T18:17:43Z","lastTransitionTime":"2025-10-08T18:17:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 18:17:43 crc kubenswrapper[4859]: I1008 18:17:43.222835 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 18:17:43 crc kubenswrapper[4859]: I1008 18:17:43.222880 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 18:17:43 crc kubenswrapper[4859]: I1008 18:17:43.222894 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 18:17:43 crc kubenswrapper[4859]: I1008 18:17:43.222912 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 18:17:43 crc kubenswrapper[4859]: I1008 18:17:43.222924 4859 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T18:17:43Z","lastTransitionTime":"2025-10-08T18:17:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 18:17:43 crc kubenswrapper[4859]: I1008 18:17:43.326133 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 18:17:43 crc kubenswrapper[4859]: I1008 18:17:43.326193 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 18:17:43 crc kubenswrapper[4859]: I1008 18:17:43.326214 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 18:17:43 crc kubenswrapper[4859]: I1008 18:17:43.326243 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 18:17:43 crc kubenswrapper[4859]: I1008 18:17:43.326261 4859 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T18:17:43Z","lastTransitionTime":"2025-10-08T18:17:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 18:17:43 crc kubenswrapper[4859]: I1008 18:17:43.429395 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 18:17:43 crc kubenswrapper[4859]: I1008 18:17:43.429458 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 18:17:43 crc kubenswrapper[4859]: I1008 18:17:43.429474 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 18:17:43 crc kubenswrapper[4859]: I1008 18:17:43.429497 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 18:17:43 crc kubenswrapper[4859]: I1008 18:17:43.429514 4859 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T18:17:43Z","lastTransitionTime":"2025-10-08T18:17:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 18:17:43 crc kubenswrapper[4859]: I1008 18:17:43.469378 4859 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 08 18:17:43 crc kubenswrapper[4859]: I1008 18:17:43.469386 4859 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 08 18:17:43 crc kubenswrapper[4859]: E1008 18:17:43.469588 4859 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 08 18:17:43 crc kubenswrapper[4859]: I1008 18:17:43.469404 4859 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 08 18:17:43 crc kubenswrapper[4859]: E1008 18:17:43.469671 4859 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 08 18:17:43 crc kubenswrapper[4859]: E1008 18:17:43.469808 4859 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 08 18:17:43 crc kubenswrapper[4859]: I1008 18:17:43.533061 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 18:17:43 crc kubenswrapper[4859]: I1008 18:17:43.533124 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 18:17:43 crc kubenswrapper[4859]: I1008 18:17:43.533143 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 18:17:43 crc kubenswrapper[4859]: I1008 18:17:43.533170 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 18:17:43 crc kubenswrapper[4859]: I1008 18:17:43.533190 4859 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T18:17:43Z","lastTransitionTime":"2025-10-08T18:17:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 18:17:43 crc kubenswrapper[4859]: I1008 18:17:43.637138 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 18:17:43 crc kubenswrapper[4859]: I1008 18:17:43.637186 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 18:17:43 crc kubenswrapper[4859]: I1008 18:17:43.637198 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 18:17:43 crc kubenswrapper[4859]: I1008 18:17:43.637217 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 18:17:43 crc kubenswrapper[4859]: I1008 18:17:43.637230 4859 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T18:17:43Z","lastTransitionTime":"2025-10-08T18:17:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 18:17:43 crc kubenswrapper[4859]: I1008 18:17:43.740249 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 18:17:43 crc kubenswrapper[4859]: I1008 18:17:43.740283 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 18:17:43 crc kubenswrapper[4859]: I1008 18:17:43.740292 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 18:17:43 crc kubenswrapper[4859]: I1008 18:17:43.740307 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 18:17:43 crc kubenswrapper[4859]: I1008 18:17:43.740317 4859 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T18:17:43Z","lastTransitionTime":"2025-10-08T18:17:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 18:17:43 crc kubenswrapper[4859]: I1008 18:17:43.754969 4859 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-4g8gf_1dff864e-b75d-4e0b-b182-75f710eac8df/ovnkube-controller/0.log" Oct 08 18:17:43 crc kubenswrapper[4859]: I1008 18:17:43.758540 4859 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-4g8gf" event={"ID":"1dff864e-b75d-4e0b-b182-75f710eac8df","Type":"ContainerStarted","Data":"2ce53a49f210c2d6e0001ee4280916611bbd82b4fc21a99c59954ed3d7570b96"} Oct 08 18:17:43 crc kubenswrapper[4859]: I1008 18:17:43.758645 4859 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" Oct 08 18:17:43 crc kubenswrapper[4859]: I1008 18:17:43.775465 4859 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4efc16fb-e679-464d-8b3c-0f46f082593e\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:17:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:17:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:17:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:17:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:17:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://aa8f8ea895272cd79a08d1c47cd7863a1ec10e21bb60d12aa10862a56c99e4bf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T18:17:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5c4e95d8a712ab219455564539a90286ae48bbc46ddc322a5bfd96c1a8268110\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T18:17:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://bae6ecfd8b446ef04d393d823e5ffe37fe749337c0ca17fde79c5f88d5e31393\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T18:17:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6411d1e7e18d5ceadaaceed5a8a6b20a0cacc55cede793ca9fb4dcad70c64fc6\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T18:17:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T18:17:10Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T18:17:43Z is after 2025-08-24T17:21:41Z" Oct 08 18:17:43 crc kubenswrapper[4859]: I1008 18:17:43.790974 4859 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-08T18:17:29Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T18:17:29Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T18:17:43Z is after 2025-08-24T17:21:41Z" Oct 08 18:17:43 crc kubenswrapper[4859]: I1008 18:17:43.807497 4859 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-84k8x" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1c24078c-c9e8-4160-883a-f483b4b3dd23\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:17:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:17:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:17:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:17:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8661006834d0f50eaada7394369584ee95fc61ff54437eec35d755526afc5b48\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T18:17:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hk5xb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://18b291eb3344fb7c00ef5679dca9a6f7ed9d08cea684a803ff7dad2e8a18c585\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://18b291eb3344fb7c00ef5679dca9a6f7ed9d08cea684a803ff7dad2e8a18c585\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-08T18:17:32Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-08T18:17:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hk5xb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://160163f1a91d981583452db3c7a9244182d1c8ced45749a61b721d0d45c1dfe4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://160163f1a91d981583452db3c7a9244182d1c8ced45749a61b721d0d45c1dfe4\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-08T18:17:34Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-08T18:17:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hk5xb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://51e9b53f3847ac53f378ad83870debf8e71518d4edbc7fc0dc66644818e2103e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://51e9b53f3847ac53f378ad83870debf8e71518d4edbc7fc0dc66644818e2103e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-08T18:17:34Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-08T18:17:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hk5xb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1ff118919739759ac01b5cbe3240e6673db54eaa16353c609d670071ed7964cb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1ff118919739759ac01b5cbe3240e6673db54eaa16353c609d670071ed7964cb\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-08T18:17:35Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-08T18:17:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hk5xb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8522e803ff56eb57cd4d9fd036e0c3fb16240ea21d4a094c1db7aa7d50bfe2c2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8522e803ff56eb57cd4d9fd036e0c3fb16240ea21d4a094c1db7aa7d50bfe2c2\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-08T18:17:37Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-08T18:17:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hk5xb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://be61056edf515e037403ca2a0bfd5feb9b2537dd42212b58e133b23b294f327b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://be61056edf515e037403ca2a0bfd5feb9b2537dd42212b58e133b23b294f327b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-08T18:17:38Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-08T18:17:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hk5xb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T18:17:31Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-84k8x\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T18:17:43Z is after 2025-08-24T17:21:41Z" Oct 08 18:17:43 crc kubenswrapper[4859]: I1008 18:17:43.819253 4859 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-08T18:17:29Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T18:17:29Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T18:17:43Z is after 2025-08-24T17:21:41Z" Oct 08 18:17:43 crc kubenswrapper[4859]: I1008 18:17:43.830173 4859 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-btw28" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"97bc5598-3ead-4300-a938-3aebbff1e8c6\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:17:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:17:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:17:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:17:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d6d4d6ba771cb5fde637449bba05660f634df7e833a0c8dad2b2c7f3a422bb2b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T18:17:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qcxsd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T18:17:31Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-btw28\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T18:17:43Z is after 2025-08-24T17:21:41Z" Oct 08 18:17:43 crc kubenswrapper[4859]: I1008 18:17:43.843159 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 18:17:43 crc kubenswrapper[4859]: I1008 18:17:43.843215 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 18:17:43 crc kubenswrapper[4859]: I1008 18:17:43.843228 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 18:17:43 crc kubenswrapper[4859]: I1008 18:17:43.843248 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 18:17:43 crc kubenswrapper[4859]: I1008 18:17:43.843261 4859 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T18:17:43Z","lastTransitionTime":"2025-10-08T18:17:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 18:17:43 crc kubenswrapper[4859]: I1008 18:17:43.843909 4859 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-8plkg" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"92b2d7de-31cc-4a91-95a5-ed5ea964b028\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:17:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:17:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:17:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:17:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fa00b8345d40b74a917ddc23cc1b2c5c684a9ca61b6119d090c92cc1439b71c4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T18:17:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q9d7m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T18:17:31Z\\\"}}\" for pod \"openshift-multus\"/\"multus-8plkg\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T18:17:43Z is after 2025-08-24T17:21:41Z" Oct 08 18:17:43 crc kubenswrapper[4859]: I1008 18:17:43.858488 4859 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-btjmq" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b8605f7-fed7-493f-9b4f-6a6358adf907\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:17:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:17:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:17:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:17:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://14e9b6c4a4403f36adb78fcf1135a620ca6fc56ed974c9639dcdfc59e6917559\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T18:17:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xspg4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T18:17:35Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-btjmq\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T18:17:43Z is after 2025-08-24T17:21:41Z" Oct 08 18:17:43 crc kubenswrapper[4859]: I1008 18:17:43.874366 4859 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-08T18:17:31Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T18:17:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T18:17:31Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c6d15905694b6bbd4999d49dc8f7a53eb6bfaa7adb59efb0d7e4973c21d1c7be\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T18:17:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T18:17:43Z is after 2025-08-24T17:21:41Z" Oct 08 18:17:43 crc kubenswrapper[4859]: I1008 18:17:43.889965 4859 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-08T18:17:33Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T18:17:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T18:17:33Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5d99e071bd083914418d9c128f38955f26295c08b1520f2dc774c99b1c843693\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T18:17:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T18:17:43Z is after 2025-08-24T17:21:41Z" Oct 08 18:17:43 crc kubenswrapper[4859]: I1008 18:17:43.921447 4859 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-4g8gf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1dff864e-b75d-4e0b-b182-75f710eac8df\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:17:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:17:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:17:32Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:17:32Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://768ca274dedcc6847fc21d15234d3665fa9ccae371b5b6a882cfcb9b182f527d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T18:17:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s8jnn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://669f5f3c6d3c1ec0c4b4870e090b0a853adc01727c23383e595715b476c32668\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T18:17:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s8jnn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b8009a42cc4c9a816457a0678bdb2f7b7c53d0c8604e007a6fb141f66d06a516\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T18:17:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s8jnn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://676511a11ed4102084b57ebdd5f3fda4969741b2bc8443b415c4bf94f39ae4c9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T18:17:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s8jnn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4a74c1f0d21c90cdd7e776ee14e1f48757a04e2220fbd1f884449181ec6d1c3a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T18:17:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s8jnn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6ddc0e842bb8f81663406070023b0e5ed9cc99c98db0123b1c84f20da59fb5b5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T18:17:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s8jnn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2ce53a49f210c2d6e0001ee4280916611bbd82b4fc21a99c59954ed3d7570b96\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://73a74d9d109262badcdac9ad105c1f34039fa7954910cb148175551bbd53da40\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-10-08T18:17:41Z\\\",\\\"message\\\":\\\"l\\\\nI1008 18:17:41.904298 6177 handler.go:190] Sending *v1.Namespace event handler 5 for removal\\\\nI1008 18:17:41.904331 6177 handler.go:208] Removed *v1.Namespace event handler 5\\\\nI1008 18:17:41.904343 6177 handler.go:208] Removed *v1.Namespace event handler 1\\\\nI1008 18:17:41.905055 6177 handler.go:190] Sending *v1.Pod event handler 3 for removal\\\\nI1008 18:17:41.905088 6177 handler.go:190] Sending *v1.Pod event handler 6 for removal\\\\nI1008 18:17:41.905116 6177 handler.go:208] Removed *v1.Pod event handler 6\\\\nI1008 18:17:41.905126 6177 handler.go:208] Removed *v1.Pod event handler 3\\\\nI1008 18:17:41.907806 6177 handler.go:190] Sending *v1.Node event handler 2 for removal\\\\nI1008 18:17:41.907827 6177 handler.go:190] Sending *v1.Node event handler 7 for removal\\\\nI1008 18:17:41.907850 6177 handler.go:190] Sending *v1.EgressIP event handler 8 for removal\\\\nI1008 18:17:41.907871 6177 handler.go:190] Sending *v1.EgressFirewall event handler 9 for removal\\\\nI1008 18:17:41.907971 6177 factory.go:656] Stopping watch factory\\\\nI1008 18:17:41.907995 6177 ovnkube.go:599] Stopped ovnkube\\\\nI1008 18:17:41.908032 6177 handler.go:208] Removed *v1.Node event handler 2\\\\nI1008 18:17:41.908051 6177 handler.go:208] Removed *v1.Node event handler 7\\\\nI1008 18:17:41.908065 6177 handler.go:208] Removed *v1.EgressIP event handler 8\\\\nI1008 18:17:4\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-08T18:17:39Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T18:17:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s8jnn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2c67d2aed7126d76989e057d8e2d8e188a8902b2f3fe26ed6d1d45d3b013d897\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T18:17:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s8jnn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7a211d3da02b173867ff3a912eecd3b379483c7cc77e6b91122edd1ffd30fd53\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7a211d3da02b173867ff3a912eecd3b379483c7cc77e6b91122edd1ffd30fd53\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-08T18:17:32Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-08T18:17:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s8jnn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T18:17:32Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-4g8gf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T18:17:43Z is after 2025-08-24T17:21:41Z" Oct 08 18:17:43 crc kubenswrapper[4859]: I1008 18:17:43.943904 4859 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"63a623de-5608-4228-b1c2-77188c276986\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:17:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:17:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:17:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:17:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:17:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d4786a8fb8abd96fb22e60733c974f599017407d589ed815fb8541fd3b64f696\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T18:17:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://183d481502ed5108d5e19e483b714c289f355ea3f8b0891883e5335b802c83b6\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T18:17:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ec848c46fd5dde0469b00d3cab14ba4a652533a265490fadbee7a8b8843ef249\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T18:17:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4fe1828b4d63ff8106bfcd15e1dc1c1bd1ba7f77777cdb23ff1b291d0418774c\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2e700b2a20c81b414617f5b5c018a738278a3fa44e6c6d14b2170051d0725431\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-08T18:17:24Z\\\",\\\"message\\\":\\\"W1008 18:17:13.670625 1 cmd.go:257] Using insecure, self-signed certificates\\\\nI1008 18:17:13.671202 1 crypto.go:601] Generating new CA for check-endpoints-signer@1759947433 cert, and key in /tmp/serving-cert-3867218267/serving-signer.crt, /tmp/serving-cert-3867218267/serving-signer.key\\\\nI1008 18:17:13.944463 1 observer_polling.go:159] Starting file observer\\\\nW1008 18:17:13.951958 1 builder.go:272] unable to get owner reference (falling back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": dial tcp [::1]:6443: connect: connection refused\\\\nI1008 18:17:13.952088 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1008 18:17:13.956755 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3867218267/tls.crt::/tmp/serving-cert-3867218267/tls.key\\\\\\\"\\\\nF1008 18:17:24.290991 1 cmd.go:182] error initializing delegating authentication: unable to load configmap based request-header-client-ca-file: Get \\\\\\\"https://localhost:6443/api/v1/namespaces/kube-system/configmaps/extension-apiserver-authentication\\\\\\\": net/http: TLS handshake timeout\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-08T18:17:13Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T18:17:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c65b66f62d325ca10ef8d5d31f04529ccf4e691cf3303d231e7a414fb400e4d6\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T18:17:13Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://58c7bba4ace28215b062653a960e561b1dc1f890f2baf450145638a065af038b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://58c7bba4ace28215b062653a960e561b1dc1f890f2baf450145638a065af038b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-08T18:17:11Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-08T18:17:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T18:17:10Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T18:17:43Z is after 2025-08-24T17:21:41Z" Oct 08 18:17:43 crc kubenswrapper[4859]: I1008 18:17:43.947027 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 18:17:43 crc kubenswrapper[4859]: I1008 18:17:43.947069 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 18:17:43 crc kubenswrapper[4859]: I1008 18:17:43.947080 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 18:17:43 crc kubenswrapper[4859]: I1008 18:17:43.947102 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 18:17:43 crc kubenswrapper[4859]: I1008 18:17:43.947119 4859 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T18:17:43Z","lastTransitionTime":"2025-10-08T18:17:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 18:17:43 crc kubenswrapper[4859]: I1008 18:17:43.958926 4859 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-08T18:17:31Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T18:17:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T18:17:31Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://99e2d51bd7757498dae60949eea6d732b5f7b481fa00f808dd5308a530a1daf1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T18:17:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e32aaba93699944665660ae41e9b5779178c1354c31e15ad8188bf8c3c8fa500\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T18:17:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T18:17:43Z is after 2025-08-24T17:21:41Z" Oct 08 18:17:43 crc kubenswrapper[4859]: I1008 18:17:43.975899 4859 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-08T18:17:29Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T18:17:29Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T18:17:43Z is after 2025-08-24T17:21:41Z" Oct 08 18:17:43 crc kubenswrapper[4859]: I1008 18:17:43.992582 4859 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-82s52" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b23a6a6c-9d92-4e7b-840e-55cfda873a2d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:17:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:17:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:17:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:17:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://71302c766485b9f55743d85f97440ff578cd44ba1e5a6f5b1c024c0e808072d2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T18:17:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6vx22\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c23ea3580998adf90621a0155c20ee4bd590c512629812174e95c28857bcfef9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T18:17:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6vx22\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T18:17:31Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-82s52\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T18:17:43Z is after 2025-08-24T17:21:41Z" Oct 08 18:17:44 crc kubenswrapper[4859]: I1008 18:17:44.052159 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 18:17:44 crc kubenswrapper[4859]: I1008 18:17:44.052234 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 18:17:44 crc kubenswrapper[4859]: I1008 18:17:44.052298 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 18:17:44 crc kubenswrapper[4859]: I1008 18:17:44.052331 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 18:17:44 crc kubenswrapper[4859]: I1008 18:17:44.052354 4859 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T18:17:44Z","lastTransitionTime":"2025-10-08T18:17:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 18:17:44 crc kubenswrapper[4859]: I1008 18:17:44.156799 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 18:17:44 crc kubenswrapper[4859]: I1008 18:17:44.156875 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 18:17:44 crc kubenswrapper[4859]: I1008 18:17:44.156895 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 18:17:44 crc kubenswrapper[4859]: I1008 18:17:44.156923 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 18:17:44 crc kubenswrapper[4859]: I1008 18:17:44.156943 4859 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T18:17:44Z","lastTransitionTime":"2025-10-08T18:17:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 18:17:44 crc kubenswrapper[4859]: I1008 18:17:44.260665 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 18:17:44 crc kubenswrapper[4859]: I1008 18:17:44.260760 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 18:17:44 crc kubenswrapper[4859]: I1008 18:17:44.260778 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 18:17:44 crc kubenswrapper[4859]: I1008 18:17:44.260804 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 18:17:44 crc kubenswrapper[4859]: I1008 18:17:44.260822 4859 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T18:17:44Z","lastTransitionTime":"2025-10-08T18:17:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 18:17:44 crc kubenswrapper[4859]: I1008 18:17:44.364163 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 18:17:44 crc kubenswrapper[4859]: I1008 18:17:44.364221 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 18:17:44 crc kubenswrapper[4859]: I1008 18:17:44.364236 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 18:17:44 crc kubenswrapper[4859]: I1008 18:17:44.364259 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 18:17:44 crc kubenswrapper[4859]: I1008 18:17:44.364275 4859 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T18:17:44Z","lastTransitionTime":"2025-10-08T18:17:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 18:17:44 crc kubenswrapper[4859]: I1008 18:17:44.467747 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 18:17:44 crc kubenswrapper[4859]: I1008 18:17:44.467804 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 18:17:44 crc kubenswrapper[4859]: I1008 18:17:44.467817 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 18:17:44 crc kubenswrapper[4859]: I1008 18:17:44.467840 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 18:17:44 crc kubenswrapper[4859]: I1008 18:17:44.467854 4859 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T18:17:44Z","lastTransitionTime":"2025-10-08T18:17:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 18:17:44 crc kubenswrapper[4859]: I1008 18:17:44.571040 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 18:17:44 crc kubenswrapper[4859]: I1008 18:17:44.571105 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 18:17:44 crc kubenswrapper[4859]: I1008 18:17:44.571126 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 18:17:44 crc kubenswrapper[4859]: I1008 18:17:44.571154 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 18:17:44 crc kubenswrapper[4859]: I1008 18:17:44.571170 4859 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T18:17:44Z","lastTransitionTime":"2025-10-08T18:17:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 18:17:44 crc kubenswrapper[4859]: I1008 18:17:44.674648 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 18:17:44 crc kubenswrapper[4859]: I1008 18:17:44.674714 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 18:17:44 crc kubenswrapper[4859]: I1008 18:17:44.674730 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 18:17:44 crc kubenswrapper[4859]: I1008 18:17:44.674751 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 18:17:44 crc kubenswrapper[4859]: I1008 18:17:44.674764 4859 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T18:17:44Z","lastTransitionTime":"2025-10-08T18:17:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 18:17:44 crc kubenswrapper[4859]: I1008 18:17:44.765597 4859 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-4g8gf_1dff864e-b75d-4e0b-b182-75f710eac8df/ovnkube-controller/1.log" Oct 08 18:17:44 crc kubenswrapper[4859]: I1008 18:17:44.766725 4859 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-4g8gf_1dff864e-b75d-4e0b-b182-75f710eac8df/ovnkube-controller/0.log" Oct 08 18:17:44 crc kubenswrapper[4859]: I1008 18:17:44.771537 4859 generic.go:334] "Generic (PLEG): container finished" podID="1dff864e-b75d-4e0b-b182-75f710eac8df" containerID="2ce53a49f210c2d6e0001ee4280916611bbd82b4fc21a99c59954ed3d7570b96" exitCode=1 Oct 08 18:17:44 crc kubenswrapper[4859]: I1008 18:17:44.771620 4859 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-4g8gf" event={"ID":"1dff864e-b75d-4e0b-b182-75f710eac8df","Type":"ContainerDied","Data":"2ce53a49f210c2d6e0001ee4280916611bbd82b4fc21a99c59954ed3d7570b96"} Oct 08 18:17:44 crc kubenswrapper[4859]: I1008 18:17:44.771746 4859 scope.go:117] "RemoveContainer" containerID="73a74d9d109262badcdac9ad105c1f34039fa7954910cb148175551bbd53da40" Oct 08 18:17:44 crc kubenswrapper[4859]: I1008 18:17:44.772760 4859 scope.go:117] "RemoveContainer" containerID="2ce53a49f210c2d6e0001ee4280916611bbd82b4fc21a99c59954ed3d7570b96" Oct 08 18:17:44 crc kubenswrapper[4859]: E1008 18:17:44.773111 4859 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ovnkube-controller\" with CrashLoopBackOff: \"back-off 10s restarting failed container=ovnkube-controller pod=ovnkube-node-4g8gf_openshift-ovn-kubernetes(1dff864e-b75d-4e0b-b182-75f710eac8df)\"" pod="openshift-ovn-kubernetes/ovnkube-node-4g8gf" podUID="1dff864e-b75d-4e0b-b182-75f710eac8df" Oct 08 18:17:44 crc kubenswrapper[4859]: I1008 18:17:44.778413 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 18:17:44 crc kubenswrapper[4859]: I1008 18:17:44.778483 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 18:17:44 crc kubenswrapper[4859]: I1008 18:17:44.778509 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 18:17:44 crc kubenswrapper[4859]: I1008 18:17:44.778541 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 18:17:44 crc kubenswrapper[4859]: I1008 18:17:44.778565 4859 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T18:17:44Z","lastTransitionTime":"2025-10-08T18:17:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 18:17:44 crc kubenswrapper[4859]: I1008 18:17:44.798574 4859 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-08T18:17:31Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T18:17:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T18:17:31Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://99e2d51bd7757498dae60949eea6d732b5f7b481fa00f808dd5308a530a1daf1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T18:17:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e32aaba93699944665660ae41e9b5779178c1354c31e15ad8188bf8c3c8fa500\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T18:17:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T18:17:44Z is after 2025-08-24T17:21:41Z" Oct 08 18:17:44 crc kubenswrapper[4859]: I1008 18:17:44.825831 4859 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-08T18:17:29Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T18:17:29Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T18:17:44Z is after 2025-08-24T17:21:41Z" Oct 08 18:17:44 crc kubenswrapper[4859]: I1008 18:17:44.840998 4859 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-82s52" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b23a6a6c-9d92-4e7b-840e-55cfda873a2d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:17:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:17:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:17:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:17:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://71302c766485b9f55743d85f97440ff578cd44ba1e5a6f5b1c024c0e808072d2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T18:17:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6vx22\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c23ea3580998adf90621a0155c20ee4bd590c512629812174e95c28857bcfef9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T18:17:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6vx22\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T18:17:31Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-82s52\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T18:17:44Z is after 2025-08-24T17:21:41Z" Oct 08 18:17:44 crc kubenswrapper[4859]: I1008 18:17:44.870521 4859 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-4g8gf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1dff864e-b75d-4e0b-b182-75f710eac8df\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:17:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:17:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:17:32Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:17:32Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://768ca274dedcc6847fc21d15234d3665fa9ccae371b5b6a882cfcb9b182f527d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T18:17:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s8jnn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://669f5f3c6d3c1ec0c4b4870e090b0a853adc01727c23383e595715b476c32668\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T18:17:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s8jnn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b8009a42cc4c9a816457a0678bdb2f7b7c53d0c8604e007a6fb141f66d06a516\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T18:17:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s8jnn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://676511a11ed4102084b57ebdd5f3fda4969741b2bc8443b415c4bf94f39ae4c9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T18:17:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s8jnn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4a74c1f0d21c90cdd7e776ee14e1f48757a04e2220fbd1f884449181ec6d1c3a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T18:17:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s8jnn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6ddc0e842bb8f81663406070023b0e5ed9cc99c98db0123b1c84f20da59fb5b5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T18:17:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s8jnn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2ce53a49f210c2d6e0001ee4280916611bbd82b4fc21a99c59954ed3d7570b96\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://73a74d9d109262badcdac9ad105c1f34039fa7954910cb148175551bbd53da40\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-10-08T18:17:41Z\\\",\\\"message\\\":\\\"l\\\\nI1008 18:17:41.904298 6177 handler.go:190] Sending *v1.Namespace event handler 5 for removal\\\\nI1008 18:17:41.904331 6177 handler.go:208] Removed *v1.Namespace event handler 5\\\\nI1008 18:17:41.904343 6177 handler.go:208] Removed *v1.Namespace event handler 1\\\\nI1008 18:17:41.905055 6177 handler.go:190] Sending *v1.Pod event handler 3 for removal\\\\nI1008 18:17:41.905088 6177 handler.go:190] Sending *v1.Pod event handler 6 for removal\\\\nI1008 18:17:41.905116 6177 handler.go:208] Removed *v1.Pod event handler 6\\\\nI1008 18:17:41.905126 6177 handler.go:208] Removed *v1.Pod event handler 3\\\\nI1008 18:17:41.907806 6177 handler.go:190] Sending *v1.Node event handler 2 for removal\\\\nI1008 18:17:41.907827 6177 handler.go:190] Sending *v1.Node event handler 7 for removal\\\\nI1008 18:17:41.907850 6177 handler.go:190] Sending *v1.EgressIP event handler 8 for removal\\\\nI1008 18:17:41.907871 6177 handler.go:190] Sending *v1.EgressFirewall event handler 9 for removal\\\\nI1008 18:17:41.907971 6177 factory.go:656] Stopping watch factory\\\\nI1008 18:17:41.907995 6177 ovnkube.go:599] Stopped ovnkube\\\\nI1008 18:17:41.908032 6177 handler.go:208] Removed *v1.Node event handler 2\\\\nI1008 18:17:41.908051 6177 handler.go:208] Removed *v1.Node event handler 7\\\\nI1008 18:17:41.908065 6177 handler.go:208] Removed *v1.EgressIP event handler 8\\\\nI1008 18:17:4\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-08T18:17:39Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2ce53a49f210c2d6e0001ee4280916611bbd82b4fc21a99c59954ed3d7570b96\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-10-08T18:17:43Z\\\",\\\"message\\\":\\\"cert-secret-name:metrics-tls service.beta.openshift.io/serving-cert-signed-by:openshift-service-serving-signer@1740288168] [{config.openshift.io/v1 ClusterVersion version 9101b518-476b-4eea-8fa6-69b0534e5caa 0xc00695d3d7 \\\\u003cnil\\\\u003e}] [] []},Spec:ServiceSpec{Ports:[]ServicePort{ServicePort{Name:metrics,Protocol:TCP,Port:9393,TargetPort:{1 0 metrics},NodePort:0,AppProtocol:nil,},},Selector:map[string]string{name: ingress-operator,},ClusterIP:10.217.5.244,Type:ClusterIP,ExternalIPs:[],SessionAffinity:None,LoadBalancerIP:,LoadBalancerSourceRanges:[],ExternalName:,ExternalTrafficPolicy:,HealthCheckNodePort:0,PublishNotReadyAddresses:false,SessionAffinityConfig:nil,IPFamilyPolicy:*SingleStack,ClusterIPs:[10.217.5.244],IPFamilies:[IPv4],AllocateLoadBalancerNodePorts:nil,LoadBalancerClass:nil,InternalTrafficPolicy:*Cluster,TrafficDistribution:nil,},Status:ServiceStatus{LoadBalancer:LoadBalancerStatus{Ingress:[]LoadBalancerIngress{},},Conditions:[]Condition{},},}\\\\nI1008 18:17:43.764981 6316 base_network_controller_pods.go:477] [default/openshift-network-diagnostics/network-check-target-xd92c] creating logical port openshift-network-diagnostics_network-check-target-xd92c for pod on switch crc\\\\nF1008 18:17:43.764990 6316 ovnkube.go:137] failed to run ovnkube: [failed to start network controller: failed to start default network controller: unable to create\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-08T18:17:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s8jnn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2c67d2aed7126d76989e057d8e2d8e188a8902b2f3fe26ed6d1d45d3b013d897\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T18:17:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s8jnn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7a211d3da02b173867ff3a912eecd3b379483c7cc77e6b91122edd1ffd30fd53\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7a211d3da02b173867ff3a912eecd3b379483c7cc77e6b91122edd1ffd30fd53\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-08T18:17:32Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-08T18:17:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s8jnn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T18:17:32Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-4g8gf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T18:17:44Z is after 2025-08-24T17:21:41Z" Oct 08 18:17:44 crc kubenswrapper[4859]: I1008 18:17:44.880919 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 18:17:44 crc kubenswrapper[4859]: I1008 18:17:44.880964 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 18:17:44 crc kubenswrapper[4859]: I1008 18:17:44.880975 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 18:17:44 crc kubenswrapper[4859]: I1008 18:17:44.880993 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 18:17:44 crc kubenswrapper[4859]: I1008 18:17:44.881005 4859 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T18:17:44Z","lastTransitionTime":"2025-10-08T18:17:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 18:17:44 crc kubenswrapper[4859]: I1008 18:17:44.887557 4859 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"63a623de-5608-4228-b1c2-77188c276986\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:17:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:17:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:17:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:17:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:17:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d4786a8fb8abd96fb22e60733c974f599017407d589ed815fb8541fd3b64f696\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T18:17:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://183d481502ed5108d5e19e483b714c289f355ea3f8b0891883e5335b802c83b6\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T18:17:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ec848c46fd5dde0469b00d3cab14ba4a652533a265490fadbee7a8b8843ef249\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T18:17:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4fe1828b4d63ff8106bfcd15e1dc1c1bd1ba7f77777cdb23ff1b291d0418774c\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2e700b2a20c81b414617f5b5c018a738278a3fa44e6c6d14b2170051d0725431\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-08T18:17:24Z\\\",\\\"message\\\":\\\"W1008 18:17:13.670625 1 cmd.go:257] Using insecure, self-signed certificates\\\\nI1008 18:17:13.671202 1 crypto.go:601] Generating new CA for check-endpoints-signer@1759947433 cert, and key in /tmp/serving-cert-3867218267/serving-signer.crt, /tmp/serving-cert-3867218267/serving-signer.key\\\\nI1008 18:17:13.944463 1 observer_polling.go:159] Starting file observer\\\\nW1008 18:17:13.951958 1 builder.go:272] unable to get owner reference (falling back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": dial tcp [::1]:6443: connect: connection refused\\\\nI1008 18:17:13.952088 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1008 18:17:13.956755 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3867218267/tls.crt::/tmp/serving-cert-3867218267/tls.key\\\\\\\"\\\\nF1008 18:17:24.290991 1 cmd.go:182] error initializing delegating authentication: unable to load configmap based request-header-client-ca-file: Get \\\\\\\"https://localhost:6443/api/v1/namespaces/kube-system/configmaps/extension-apiserver-authentication\\\\\\\": net/http: TLS handshake timeout\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-08T18:17:13Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T18:17:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c65b66f62d325ca10ef8d5d31f04529ccf4e691cf3303d231e7a414fb400e4d6\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T18:17:13Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://58c7bba4ace28215b062653a960e561b1dc1f890f2baf450145638a065af038b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://58c7bba4ace28215b062653a960e561b1dc1f890f2baf450145638a065af038b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-08T18:17:11Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-08T18:17:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T18:17:10Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T18:17:44Z is after 2025-08-24T17:21:41Z" Oct 08 18:17:44 crc kubenswrapper[4859]: I1008 18:17:44.901995 4859 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4efc16fb-e679-464d-8b3c-0f46f082593e\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:17:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:17:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:17:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:17:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:17:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://aa8f8ea895272cd79a08d1c47cd7863a1ec10e21bb60d12aa10862a56c99e4bf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T18:17:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5c4e95d8a712ab219455564539a90286ae48bbc46ddc322a5bfd96c1a8268110\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T18:17:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://bae6ecfd8b446ef04d393d823e5ffe37fe749337c0ca17fde79c5f88d5e31393\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T18:17:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6411d1e7e18d5ceadaaceed5a8a6b20a0cacc55cede793ca9fb4dcad70c64fc6\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T18:17:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T18:17:10Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T18:17:44Z is after 2025-08-24T17:21:41Z" Oct 08 18:17:44 crc kubenswrapper[4859]: I1008 18:17:44.921215 4859 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-08T18:17:29Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T18:17:29Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T18:17:44Z is after 2025-08-24T17:21:41Z" Oct 08 18:17:44 crc kubenswrapper[4859]: I1008 18:17:44.942996 4859 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-84k8x" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1c24078c-c9e8-4160-883a-f483b4b3dd23\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:17:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:17:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:17:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:17:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8661006834d0f50eaada7394369584ee95fc61ff54437eec35d755526afc5b48\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T18:17:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hk5xb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://18b291eb3344fb7c00ef5679dca9a6f7ed9d08cea684a803ff7dad2e8a18c585\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://18b291eb3344fb7c00ef5679dca9a6f7ed9d08cea684a803ff7dad2e8a18c585\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-08T18:17:32Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-08T18:17:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hk5xb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://160163f1a91d981583452db3c7a9244182d1c8ced45749a61b721d0d45c1dfe4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://160163f1a91d981583452db3c7a9244182d1c8ced45749a61b721d0d45c1dfe4\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-08T18:17:34Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-08T18:17:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hk5xb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://51e9b53f3847ac53f378ad83870debf8e71518d4edbc7fc0dc66644818e2103e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://51e9b53f3847ac53f378ad83870debf8e71518d4edbc7fc0dc66644818e2103e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-08T18:17:34Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-08T18:17:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hk5xb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1ff118919739759ac01b5cbe3240e6673db54eaa16353c609d670071ed7964cb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1ff118919739759ac01b5cbe3240e6673db54eaa16353c609d670071ed7964cb\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-08T18:17:35Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-08T18:17:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hk5xb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8522e803ff56eb57cd4d9fd036e0c3fb16240ea21d4a094c1db7aa7d50bfe2c2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8522e803ff56eb57cd4d9fd036e0c3fb16240ea21d4a094c1db7aa7d50bfe2c2\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-08T18:17:37Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-08T18:17:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hk5xb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://be61056edf515e037403ca2a0bfd5feb9b2537dd42212b58e133b23b294f327b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://be61056edf515e037403ca2a0bfd5feb9b2537dd42212b58e133b23b294f327b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-08T18:17:38Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-08T18:17:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hk5xb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T18:17:31Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-84k8x\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T18:17:44Z is after 2025-08-24T17:21:41Z" Oct 08 18:17:44 crc kubenswrapper[4859]: I1008 18:17:44.958142 4859 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-8plkg" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"92b2d7de-31cc-4a91-95a5-ed5ea964b028\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:17:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:17:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:17:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:17:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fa00b8345d40b74a917ddc23cc1b2c5c684a9ca61b6119d090c92cc1439b71c4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T18:17:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q9d7m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T18:17:31Z\\\"}}\" for pod \"openshift-multus\"/\"multus-8plkg\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T18:17:44Z is after 2025-08-24T17:21:41Z" Oct 08 18:17:44 crc kubenswrapper[4859]: I1008 18:17:44.970596 4859 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-btjmq" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b8605f7-fed7-493f-9b4f-6a6358adf907\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:17:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:17:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:17:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:17:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://14e9b6c4a4403f36adb78fcf1135a620ca6fc56ed974c9639dcdfc59e6917559\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T18:17:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xspg4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T18:17:35Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-btjmq\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T18:17:44Z is after 2025-08-24T17:21:41Z" Oct 08 18:17:44 crc kubenswrapper[4859]: I1008 18:17:44.984590 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 18:17:44 crc kubenswrapper[4859]: I1008 18:17:44.984645 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 18:17:44 crc kubenswrapper[4859]: I1008 18:17:44.984665 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 18:17:44 crc kubenswrapper[4859]: I1008 18:17:44.984722 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 18:17:44 crc kubenswrapper[4859]: I1008 18:17:44.984737 4859 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T18:17:44Z","lastTransitionTime":"2025-10-08T18:17:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 18:17:44 crc kubenswrapper[4859]: I1008 18:17:44.989056 4859 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-08T18:17:29Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T18:17:29Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T18:17:44Z is after 2025-08-24T17:21:41Z" Oct 08 18:17:45 crc kubenswrapper[4859]: I1008 18:17:45.000764 4859 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-btw28" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"97bc5598-3ead-4300-a938-3aebbff1e8c6\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:17:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:17:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:17:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:17:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d6d4d6ba771cb5fde637449bba05660f634df7e833a0c8dad2b2c7f3a422bb2b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T18:17:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qcxsd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T18:17:31Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-btw28\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T18:17:44Z is after 2025-08-24T17:21:41Z" Oct 08 18:17:45 crc kubenswrapper[4859]: I1008 18:17:45.017410 4859 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-08T18:17:31Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T18:17:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T18:17:31Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c6d15905694b6bbd4999d49dc8f7a53eb6bfaa7adb59efb0d7e4973c21d1c7be\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T18:17:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T18:17:45Z is after 2025-08-24T17:21:41Z" Oct 08 18:17:45 crc kubenswrapper[4859]: I1008 18:17:45.028641 4859 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-08T18:17:33Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T18:17:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T18:17:33Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5d99e071bd083914418d9c128f38955f26295c08b1520f2dc774c99b1c843693\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T18:17:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T18:17:45Z is after 2025-08-24T17:21:41Z" Oct 08 18:17:45 crc kubenswrapper[4859]: I1008 18:17:45.087481 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 18:17:45 crc kubenswrapper[4859]: I1008 18:17:45.087575 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 18:17:45 crc kubenswrapper[4859]: I1008 18:17:45.087875 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 18:17:45 crc kubenswrapper[4859]: I1008 18:17:45.087904 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 18:17:45 crc kubenswrapper[4859]: I1008 18:17:45.088177 4859 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T18:17:45Z","lastTransitionTime":"2025-10-08T18:17:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 18:17:45 crc kubenswrapper[4859]: I1008 18:17:45.191136 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 18:17:45 crc kubenswrapper[4859]: I1008 18:17:45.191229 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 18:17:45 crc kubenswrapper[4859]: I1008 18:17:45.191266 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 18:17:45 crc kubenswrapper[4859]: I1008 18:17:45.191302 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 18:17:45 crc kubenswrapper[4859]: I1008 18:17:45.191325 4859 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T18:17:45Z","lastTransitionTime":"2025-10-08T18:17:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 18:17:45 crc kubenswrapper[4859]: I1008 18:17:45.219735 4859 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-g4t79"] Oct 08 18:17:45 crc kubenswrapper[4859]: I1008 18:17:45.220532 4859 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-g4t79" Oct 08 18:17:45 crc kubenswrapper[4859]: I1008 18:17:45.225967 4859 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ovn-kubernetes"/"ovn-kubernetes-control-plane-dockercfg-gs7dd" Oct 08 18:17:45 crc kubenswrapper[4859]: I1008 18:17:45.226226 4859 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ovn-kubernetes"/"ovn-control-plane-metrics-cert" Oct 08 18:17:45 crc kubenswrapper[4859]: I1008 18:17:45.239754 4859 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-08T18:17:31Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T18:17:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T18:17:31Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c6d15905694b6bbd4999d49dc8f7a53eb6bfaa7adb59efb0d7e4973c21d1c7be\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T18:17:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T18:17:45Z is after 2025-08-24T17:21:41Z" Oct 08 18:17:45 crc kubenswrapper[4859]: I1008 18:17:45.256150 4859 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-08T18:17:33Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T18:17:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T18:17:33Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5d99e071bd083914418d9c128f38955f26295c08b1520f2dc774c99b1c843693\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T18:17:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T18:17:45Z is after 2025-08-24T17:21:41Z" Oct 08 18:17:45 crc kubenswrapper[4859]: I1008 18:17:45.272307 4859 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"63a623de-5608-4228-b1c2-77188c276986\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:17:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:17:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:17:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:17:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:17:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d4786a8fb8abd96fb22e60733c974f599017407d589ed815fb8541fd3b64f696\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T18:17:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://183d481502ed5108d5e19e483b714c289f355ea3f8b0891883e5335b802c83b6\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T18:17:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ec848c46fd5dde0469b00d3cab14ba4a652533a265490fadbee7a8b8843ef249\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T18:17:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4fe1828b4d63ff8106bfcd15e1dc1c1bd1ba7f77777cdb23ff1b291d0418774c\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2e700b2a20c81b414617f5b5c018a738278a3fa44e6c6d14b2170051d0725431\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-08T18:17:24Z\\\",\\\"message\\\":\\\"W1008 18:17:13.670625 1 cmd.go:257] Using insecure, self-signed certificates\\\\nI1008 18:17:13.671202 1 crypto.go:601] Generating new CA for check-endpoints-signer@1759947433 cert, and key in /tmp/serving-cert-3867218267/serving-signer.crt, /tmp/serving-cert-3867218267/serving-signer.key\\\\nI1008 18:17:13.944463 1 observer_polling.go:159] Starting file observer\\\\nW1008 18:17:13.951958 1 builder.go:272] unable to get owner reference (falling back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": dial tcp [::1]:6443: connect: connection refused\\\\nI1008 18:17:13.952088 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1008 18:17:13.956755 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3867218267/tls.crt::/tmp/serving-cert-3867218267/tls.key\\\\\\\"\\\\nF1008 18:17:24.290991 1 cmd.go:182] error initializing delegating authentication: unable to load configmap based request-header-client-ca-file: Get \\\\\\\"https://localhost:6443/api/v1/namespaces/kube-system/configmaps/extension-apiserver-authentication\\\\\\\": net/http: TLS handshake timeout\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-08T18:17:13Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T18:17:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c65b66f62d325ca10ef8d5d31f04529ccf4e691cf3303d231e7a414fb400e4d6\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T18:17:13Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://58c7bba4ace28215b062653a960e561b1dc1f890f2baf450145638a065af038b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://58c7bba4ace28215b062653a960e561b1dc1f890f2baf450145638a065af038b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-08T18:17:11Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-08T18:17:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T18:17:10Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T18:17:45Z is after 2025-08-24T17:21:41Z" Oct 08 18:17:45 crc kubenswrapper[4859]: I1008 18:17:45.288210 4859 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-08T18:17:31Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T18:17:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T18:17:31Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://99e2d51bd7757498dae60949eea6d732b5f7b481fa00f808dd5308a530a1daf1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T18:17:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e32aaba93699944665660ae41e9b5779178c1354c31e15ad8188bf8c3c8fa500\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T18:17:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T18:17:45Z is after 2025-08-24T17:21:41Z" Oct 08 18:17:45 crc kubenswrapper[4859]: I1008 18:17:45.294449 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 18:17:45 crc kubenswrapper[4859]: I1008 18:17:45.294486 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 18:17:45 crc kubenswrapper[4859]: I1008 18:17:45.294499 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 18:17:45 crc kubenswrapper[4859]: I1008 18:17:45.294525 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 18:17:45 crc kubenswrapper[4859]: I1008 18:17:45.294543 4859 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T18:17:45Z","lastTransitionTime":"2025-10-08T18:17:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 18:17:45 crc kubenswrapper[4859]: I1008 18:17:45.302028 4859 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-08T18:17:29Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T18:17:29Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T18:17:45Z is after 2025-08-24T17:21:41Z" Oct 08 18:17:45 crc kubenswrapper[4859]: I1008 18:17:45.315178 4859 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-82s52" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b23a6a6c-9d92-4e7b-840e-55cfda873a2d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:17:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:17:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:17:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:17:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://71302c766485b9f55743d85f97440ff578cd44ba1e5a6f5b1c024c0e808072d2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T18:17:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6vx22\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c23ea3580998adf90621a0155c20ee4bd590c512629812174e95c28857bcfef9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T18:17:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6vx22\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T18:17:31Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-82s52\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T18:17:45Z is after 2025-08-24T17:21:41Z" Oct 08 18:17:45 crc kubenswrapper[4859]: I1008 18:17:45.335641 4859 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-4g8gf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1dff864e-b75d-4e0b-b182-75f710eac8df\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:17:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:17:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:17:32Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:17:32Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://768ca274dedcc6847fc21d15234d3665fa9ccae371b5b6a882cfcb9b182f527d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T18:17:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s8jnn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://669f5f3c6d3c1ec0c4b4870e090b0a853adc01727c23383e595715b476c32668\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T18:17:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s8jnn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b8009a42cc4c9a816457a0678bdb2f7b7c53d0c8604e007a6fb141f66d06a516\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T18:17:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s8jnn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://676511a11ed4102084b57ebdd5f3fda4969741b2bc8443b415c4bf94f39ae4c9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T18:17:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s8jnn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4a74c1f0d21c90cdd7e776ee14e1f48757a04e2220fbd1f884449181ec6d1c3a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T18:17:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s8jnn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6ddc0e842bb8f81663406070023b0e5ed9cc99c98db0123b1c84f20da59fb5b5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T18:17:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s8jnn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2ce53a49f210c2d6e0001ee4280916611bbd82b4fc21a99c59954ed3d7570b96\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://73a74d9d109262badcdac9ad105c1f34039fa7954910cb148175551bbd53da40\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-10-08T18:17:41Z\\\",\\\"message\\\":\\\"l\\\\nI1008 18:17:41.904298 6177 handler.go:190] Sending *v1.Namespace event handler 5 for removal\\\\nI1008 18:17:41.904331 6177 handler.go:208] Removed *v1.Namespace event handler 5\\\\nI1008 18:17:41.904343 6177 handler.go:208] Removed *v1.Namespace event handler 1\\\\nI1008 18:17:41.905055 6177 handler.go:190] Sending *v1.Pod event handler 3 for removal\\\\nI1008 18:17:41.905088 6177 handler.go:190] Sending *v1.Pod event handler 6 for removal\\\\nI1008 18:17:41.905116 6177 handler.go:208] Removed *v1.Pod event handler 6\\\\nI1008 18:17:41.905126 6177 handler.go:208] Removed *v1.Pod event handler 3\\\\nI1008 18:17:41.907806 6177 handler.go:190] Sending *v1.Node event handler 2 for removal\\\\nI1008 18:17:41.907827 6177 handler.go:190] Sending *v1.Node event handler 7 for removal\\\\nI1008 18:17:41.907850 6177 handler.go:190] Sending *v1.EgressIP event handler 8 for removal\\\\nI1008 18:17:41.907871 6177 handler.go:190] Sending *v1.EgressFirewall event handler 9 for removal\\\\nI1008 18:17:41.907971 6177 factory.go:656] Stopping watch factory\\\\nI1008 18:17:41.907995 6177 ovnkube.go:599] Stopped ovnkube\\\\nI1008 18:17:41.908032 6177 handler.go:208] Removed *v1.Node event handler 2\\\\nI1008 18:17:41.908051 6177 handler.go:208] Removed *v1.Node event handler 7\\\\nI1008 18:17:41.908065 6177 handler.go:208] Removed *v1.EgressIP event handler 8\\\\nI1008 18:17:4\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-08T18:17:39Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2ce53a49f210c2d6e0001ee4280916611bbd82b4fc21a99c59954ed3d7570b96\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-10-08T18:17:43Z\\\",\\\"message\\\":\\\"cert-secret-name:metrics-tls service.beta.openshift.io/serving-cert-signed-by:openshift-service-serving-signer@1740288168] [{config.openshift.io/v1 ClusterVersion version 9101b518-476b-4eea-8fa6-69b0534e5caa 0xc00695d3d7 \\\\u003cnil\\\\u003e}] [] []},Spec:ServiceSpec{Ports:[]ServicePort{ServicePort{Name:metrics,Protocol:TCP,Port:9393,TargetPort:{1 0 metrics},NodePort:0,AppProtocol:nil,},},Selector:map[string]string{name: ingress-operator,},ClusterIP:10.217.5.244,Type:ClusterIP,ExternalIPs:[],SessionAffinity:None,LoadBalancerIP:,LoadBalancerSourceRanges:[],ExternalName:,ExternalTrafficPolicy:,HealthCheckNodePort:0,PublishNotReadyAddresses:false,SessionAffinityConfig:nil,IPFamilyPolicy:*SingleStack,ClusterIPs:[10.217.5.244],IPFamilies:[IPv4],AllocateLoadBalancerNodePorts:nil,LoadBalancerClass:nil,InternalTrafficPolicy:*Cluster,TrafficDistribution:nil,},Status:ServiceStatus{LoadBalancer:LoadBalancerStatus{Ingress:[]LoadBalancerIngress{},},Conditions:[]Condition{},},}\\\\nI1008 18:17:43.764981 6316 base_network_controller_pods.go:477] [default/openshift-network-diagnostics/network-check-target-xd92c] creating logical port openshift-network-diagnostics_network-check-target-xd92c for pod on switch crc\\\\nF1008 18:17:43.764990 6316 ovnkube.go:137] failed to run ovnkube: [failed to start network controller: failed to start default network controller: unable to create\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-08T18:17:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s8jnn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2c67d2aed7126d76989e057d8e2d8e188a8902b2f3fe26ed6d1d45d3b013d897\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T18:17:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s8jnn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7a211d3da02b173867ff3a912eecd3b379483c7cc77e6b91122edd1ffd30fd53\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7a211d3da02b173867ff3a912eecd3b379483c7cc77e6b91122edd1ffd30fd53\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-08T18:17:32Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-08T18:17:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s8jnn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T18:17:32Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-4g8gf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T18:17:45Z is after 2025-08-24T17:21:41Z" Oct 08 18:17:45 crc kubenswrapper[4859]: I1008 18:17:45.351756 4859 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-84k8x" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1c24078c-c9e8-4160-883a-f483b4b3dd23\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:17:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:17:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:17:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:17:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8661006834d0f50eaada7394369584ee95fc61ff54437eec35d755526afc5b48\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T18:17:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hk5xb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://18b291eb3344fb7c00ef5679dca9a6f7ed9d08cea684a803ff7dad2e8a18c585\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://18b291eb3344fb7c00ef5679dca9a6f7ed9d08cea684a803ff7dad2e8a18c585\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-08T18:17:32Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-08T18:17:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hk5xb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://160163f1a91d981583452db3c7a9244182d1c8ced45749a61b721d0d45c1dfe4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://160163f1a91d981583452db3c7a9244182d1c8ced45749a61b721d0d45c1dfe4\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-08T18:17:34Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-08T18:17:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hk5xb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://51e9b53f3847ac53f378ad83870debf8e71518d4edbc7fc0dc66644818e2103e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://51e9b53f3847ac53f378ad83870debf8e71518d4edbc7fc0dc66644818e2103e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-08T18:17:34Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-08T18:17:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hk5xb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1ff118919739759ac01b5cbe3240e6673db54eaa16353c609d670071ed7964cb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1ff118919739759ac01b5cbe3240e6673db54eaa16353c609d670071ed7964cb\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-08T18:17:35Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-08T18:17:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hk5xb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8522e803ff56eb57cd4d9fd036e0c3fb16240ea21d4a094c1db7aa7d50bfe2c2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8522e803ff56eb57cd4d9fd036e0c3fb16240ea21d4a094c1db7aa7d50bfe2c2\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-08T18:17:37Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-08T18:17:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hk5xb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://be61056edf515e037403ca2a0bfd5feb9b2537dd42212b58e133b23b294f327b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://be61056edf515e037403ca2a0bfd5feb9b2537dd42212b58e133b23b294f327b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-08T18:17:38Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-08T18:17:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hk5xb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T18:17:31Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-84k8x\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T18:17:45Z is after 2025-08-24T17:21:41Z" Oct 08 18:17:45 crc kubenswrapper[4859]: I1008 18:17:45.364077 4859 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4efc16fb-e679-464d-8b3c-0f46f082593e\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:17:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:17:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:17:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:17:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:17:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://aa8f8ea895272cd79a08d1c47cd7863a1ec10e21bb60d12aa10862a56c99e4bf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T18:17:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5c4e95d8a712ab219455564539a90286ae48bbc46ddc322a5bfd96c1a8268110\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T18:17:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://bae6ecfd8b446ef04d393d823e5ffe37fe749337c0ca17fde79c5f88d5e31393\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T18:17:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6411d1e7e18d5ceadaaceed5a8a6b20a0cacc55cede793ca9fb4dcad70c64fc6\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T18:17:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T18:17:10Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T18:17:45Z is after 2025-08-24T17:21:41Z" Oct 08 18:17:45 crc kubenswrapper[4859]: I1008 18:17:45.376433 4859 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-08T18:17:29Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T18:17:29Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T18:17:45Z is after 2025-08-24T17:21:41Z" Oct 08 18:17:45 crc kubenswrapper[4859]: I1008 18:17:45.385772 4859 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/3a92dbf1-21a8-4095-96bb-2d701d7f3d56-env-overrides\") pod \"ovnkube-control-plane-749d76644c-g4t79\" (UID: \"3a92dbf1-21a8-4095-96bb-2d701d7f3d56\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-g4t79" Oct 08 18:17:45 crc kubenswrapper[4859]: I1008 18:17:45.385857 4859 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/3a92dbf1-21a8-4095-96bb-2d701d7f3d56-ovnkube-config\") pod \"ovnkube-control-plane-749d76644c-g4t79\" (UID: \"3a92dbf1-21a8-4095-96bb-2d701d7f3d56\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-g4t79" Oct 08 18:17:45 crc kubenswrapper[4859]: I1008 18:17:45.385889 4859 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovn-control-plane-metrics-cert\" (UniqueName: \"kubernetes.io/secret/3a92dbf1-21a8-4095-96bb-2d701d7f3d56-ovn-control-plane-metrics-cert\") pod \"ovnkube-control-plane-749d76644c-g4t79\" (UID: \"3a92dbf1-21a8-4095-96bb-2d701d7f3d56\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-g4t79" Oct 08 18:17:45 crc kubenswrapper[4859]: I1008 18:17:45.385915 4859 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-sv79p\" (UniqueName: \"kubernetes.io/projected/3a92dbf1-21a8-4095-96bb-2d701d7f3d56-kube-api-access-sv79p\") pod \"ovnkube-control-plane-749d76644c-g4t79\" (UID: \"3a92dbf1-21a8-4095-96bb-2d701d7f3d56\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-g4t79" Oct 08 18:17:45 crc kubenswrapper[4859]: I1008 18:17:45.387194 4859 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-btw28" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"97bc5598-3ead-4300-a938-3aebbff1e8c6\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:17:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:17:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:17:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:17:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d6d4d6ba771cb5fde637449bba05660f634df7e833a0c8dad2b2c7f3a422bb2b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T18:17:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qcxsd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T18:17:31Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-btw28\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T18:17:45Z is after 2025-08-24T17:21:41Z" Oct 08 18:17:45 crc kubenswrapper[4859]: I1008 18:17:45.397158 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 18:17:45 crc kubenswrapper[4859]: I1008 18:17:45.397267 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 18:17:45 crc kubenswrapper[4859]: I1008 18:17:45.397299 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 18:17:45 crc kubenswrapper[4859]: I1008 18:17:45.397317 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 18:17:45 crc kubenswrapper[4859]: I1008 18:17:45.397355 4859 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T18:17:45Z","lastTransitionTime":"2025-10-08T18:17:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 18:17:45 crc kubenswrapper[4859]: I1008 18:17:45.402197 4859 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-8plkg" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"92b2d7de-31cc-4a91-95a5-ed5ea964b028\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:17:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:17:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:17:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:17:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fa00b8345d40b74a917ddc23cc1b2c5c684a9ca61b6119d090c92cc1439b71c4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T18:17:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q9d7m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T18:17:31Z\\\"}}\" for pod \"openshift-multus\"/\"multus-8plkg\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T18:17:45Z is after 2025-08-24T17:21:41Z" Oct 08 18:17:45 crc kubenswrapper[4859]: I1008 18:17:45.414652 4859 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-btjmq" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b8605f7-fed7-493f-9b4f-6a6358adf907\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:17:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:17:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:17:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:17:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://14e9b6c4a4403f36adb78fcf1135a620ca6fc56ed974c9639dcdfc59e6917559\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T18:17:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xspg4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T18:17:35Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-btjmq\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T18:17:45Z is after 2025-08-24T17:21:41Z" Oct 08 18:17:45 crc kubenswrapper[4859]: I1008 18:17:45.425755 4859 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-g4t79" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3a92dbf1-21a8-4095-96bb-2d701d7f3d56\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:17:45Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:17:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:17:45Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-rbac-proxy ovnkube-cluster-manager]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:17:45Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-rbac-proxy ovnkube-cluster-manager]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sv79p\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sv79p\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T18:17:45Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-g4t79\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T18:17:45Z is after 2025-08-24T17:21:41Z" Oct 08 18:17:45 crc kubenswrapper[4859]: I1008 18:17:45.441164 4859 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-08T18:17:29Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T18:17:29Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T18:17:45Z is after 2025-08-24T17:21:41Z" Oct 08 18:17:45 crc kubenswrapper[4859]: I1008 18:17:45.469559 4859 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 08 18:17:45 crc kubenswrapper[4859]: I1008 18:17:45.469660 4859 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 08 18:17:45 crc kubenswrapper[4859]: I1008 18:17:45.469743 4859 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 08 18:17:45 crc kubenswrapper[4859]: E1008 18:17:45.469780 4859 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 08 18:17:45 crc kubenswrapper[4859]: E1008 18:17:45.469980 4859 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 08 18:17:45 crc kubenswrapper[4859]: E1008 18:17:45.470128 4859 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 08 18:17:45 crc kubenswrapper[4859]: I1008 18:17:45.486706 4859 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/3a92dbf1-21a8-4095-96bb-2d701d7f3d56-ovnkube-config\") pod \"ovnkube-control-plane-749d76644c-g4t79\" (UID: \"3a92dbf1-21a8-4095-96bb-2d701d7f3d56\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-g4t79" Oct 08 18:17:45 crc kubenswrapper[4859]: I1008 18:17:45.486746 4859 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovn-control-plane-metrics-cert\" (UniqueName: \"kubernetes.io/secret/3a92dbf1-21a8-4095-96bb-2d701d7f3d56-ovn-control-plane-metrics-cert\") pod \"ovnkube-control-plane-749d76644c-g4t79\" (UID: \"3a92dbf1-21a8-4095-96bb-2d701d7f3d56\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-g4t79" Oct 08 18:17:45 crc kubenswrapper[4859]: I1008 18:17:45.486770 4859 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-sv79p\" (UniqueName: \"kubernetes.io/projected/3a92dbf1-21a8-4095-96bb-2d701d7f3d56-kube-api-access-sv79p\") pod \"ovnkube-control-plane-749d76644c-g4t79\" (UID: \"3a92dbf1-21a8-4095-96bb-2d701d7f3d56\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-g4t79" Oct 08 18:17:45 crc kubenswrapper[4859]: I1008 18:17:45.486800 4859 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/3a92dbf1-21a8-4095-96bb-2d701d7f3d56-env-overrides\") pod \"ovnkube-control-plane-749d76644c-g4t79\" (UID: \"3a92dbf1-21a8-4095-96bb-2d701d7f3d56\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-g4t79" Oct 08 18:17:45 crc kubenswrapper[4859]: I1008 18:17:45.487411 4859 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/3a92dbf1-21a8-4095-96bb-2d701d7f3d56-env-overrides\") pod \"ovnkube-control-plane-749d76644c-g4t79\" (UID: \"3a92dbf1-21a8-4095-96bb-2d701d7f3d56\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-g4t79" Oct 08 18:17:45 crc kubenswrapper[4859]: I1008 18:17:45.488112 4859 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/3a92dbf1-21a8-4095-96bb-2d701d7f3d56-ovnkube-config\") pod \"ovnkube-control-plane-749d76644c-g4t79\" (UID: \"3a92dbf1-21a8-4095-96bb-2d701d7f3d56\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-g4t79" Oct 08 18:17:45 crc kubenswrapper[4859]: I1008 18:17:45.496497 4859 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovn-control-plane-metrics-cert\" (UniqueName: \"kubernetes.io/secret/3a92dbf1-21a8-4095-96bb-2d701d7f3d56-ovn-control-plane-metrics-cert\") pod \"ovnkube-control-plane-749d76644c-g4t79\" (UID: \"3a92dbf1-21a8-4095-96bb-2d701d7f3d56\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-g4t79" Oct 08 18:17:45 crc kubenswrapper[4859]: I1008 18:17:45.501548 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 18:17:45 crc kubenswrapper[4859]: I1008 18:17:45.501604 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 18:17:45 crc kubenswrapper[4859]: I1008 18:17:45.501625 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 18:17:45 crc kubenswrapper[4859]: I1008 18:17:45.501653 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 18:17:45 crc kubenswrapper[4859]: I1008 18:17:45.501677 4859 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T18:17:45Z","lastTransitionTime":"2025-10-08T18:17:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 18:17:45 crc kubenswrapper[4859]: I1008 18:17:45.506076 4859 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-sv79p\" (UniqueName: \"kubernetes.io/projected/3a92dbf1-21a8-4095-96bb-2d701d7f3d56-kube-api-access-sv79p\") pod \"ovnkube-control-plane-749d76644c-g4t79\" (UID: \"3a92dbf1-21a8-4095-96bb-2d701d7f3d56\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-g4t79" Oct 08 18:17:45 crc kubenswrapper[4859]: I1008 18:17:45.537731 4859 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-g4t79" Oct 08 18:17:45 crc kubenswrapper[4859]: W1008 18:17:45.555785 4859 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod3a92dbf1_21a8_4095_96bb_2d701d7f3d56.slice/crio-9554738a8cad8c049d035c8c49d0888c0fcff67af795e33155d26efd2b118792 WatchSource:0}: Error finding container 9554738a8cad8c049d035c8c49d0888c0fcff67af795e33155d26efd2b118792: Status 404 returned error can't find the container with id 9554738a8cad8c049d035c8c49d0888c0fcff67af795e33155d26efd2b118792 Oct 08 18:17:45 crc kubenswrapper[4859]: I1008 18:17:45.606139 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 18:17:45 crc kubenswrapper[4859]: I1008 18:17:45.606196 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 18:17:45 crc kubenswrapper[4859]: I1008 18:17:45.606209 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 18:17:45 crc kubenswrapper[4859]: I1008 18:17:45.606227 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 18:17:45 crc kubenswrapper[4859]: I1008 18:17:45.606240 4859 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T18:17:45Z","lastTransitionTime":"2025-10-08T18:17:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 18:17:45 crc kubenswrapper[4859]: I1008 18:17:45.711088 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 18:17:45 crc kubenswrapper[4859]: I1008 18:17:45.711174 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 18:17:45 crc kubenswrapper[4859]: I1008 18:17:45.711195 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 18:17:45 crc kubenswrapper[4859]: I1008 18:17:45.711225 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 18:17:45 crc kubenswrapper[4859]: I1008 18:17:45.711245 4859 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T18:17:45Z","lastTransitionTime":"2025-10-08T18:17:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 18:17:45 crc kubenswrapper[4859]: I1008 18:17:45.786864 4859 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-4g8gf_1dff864e-b75d-4e0b-b182-75f710eac8df/ovnkube-controller/1.log" Oct 08 18:17:45 crc kubenswrapper[4859]: I1008 18:17:45.793647 4859 scope.go:117] "RemoveContainer" containerID="2ce53a49f210c2d6e0001ee4280916611bbd82b4fc21a99c59954ed3d7570b96" Oct 08 18:17:45 crc kubenswrapper[4859]: E1008 18:17:45.794381 4859 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ovnkube-controller\" with CrashLoopBackOff: \"back-off 10s restarting failed container=ovnkube-controller pod=ovnkube-node-4g8gf_openshift-ovn-kubernetes(1dff864e-b75d-4e0b-b182-75f710eac8df)\"" pod="openshift-ovn-kubernetes/ovnkube-node-4g8gf" podUID="1dff864e-b75d-4e0b-b182-75f710eac8df" Oct 08 18:17:45 crc kubenswrapper[4859]: I1008 18:17:45.804091 4859 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-g4t79" event={"ID":"3a92dbf1-21a8-4095-96bb-2d701d7f3d56","Type":"ContainerStarted","Data":"f8772a25b6e605efe9a062dc32dbbc2cfe32b4fe164b4e84fbf3bcffab744179"} Oct 08 18:17:45 crc kubenswrapper[4859]: I1008 18:17:45.804451 4859 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-g4t79" event={"ID":"3a92dbf1-21a8-4095-96bb-2d701d7f3d56","Type":"ContainerStarted","Data":"9554738a8cad8c049d035c8c49d0888c0fcff67af795e33155d26efd2b118792"} Oct 08 18:17:45 crc kubenswrapper[4859]: I1008 18:17:45.813991 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 18:17:45 crc kubenswrapper[4859]: I1008 18:17:45.814208 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 18:17:45 crc kubenswrapper[4859]: I1008 18:17:45.814238 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 18:17:45 crc kubenswrapper[4859]: I1008 18:17:45.814275 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 18:17:45 crc kubenswrapper[4859]: I1008 18:17:45.814294 4859 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T18:17:45Z","lastTransitionTime":"2025-10-08T18:17:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 18:17:45 crc kubenswrapper[4859]: I1008 18:17:45.822449 4859 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-08T18:17:31Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T18:17:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T18:17:31Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c6d15905694b6bbd4999d49dc8f7a53eb6bfaa7adb59efb0d7e4973c21d1c7be\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T18:17:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T18:17:45Z is after 2025-08-24T17:21:41Z" Oct 08 18:17:45 crc kubenswrapper[4859]: I1008 18:17:45.840731 4859 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-08T18:17:33Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T18:17:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T18:17:33Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5d99e071bd083914418d9c128f38955f26295c08b1520f2dc774c99b1c843693\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T18:17:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T18:17:45Z is after 2025-08-24T17:21:41Z" Oct 08 18:17:45 crc kubenswrapper[4859]: I1008 18:17:45.856088 4859 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-08T18:17:29Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T18:17:29Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T18:17:45Z is after 2025-08-24T17:21:41Z" Oct 08 18:17:45 crc kubenswrapper[4859]: I1008 18:17:45.868504 4859 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-82s52" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b23a6a6c-9d92-4e7b-840e-55cfda873a2d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:17:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:17:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:17:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:17:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://71302c766485b9f55743d85f97440ff578cd44ba1e5a6f5b1c024c0e808072d2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T18:17:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6vx22\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c23ea3580998adf90621a0155c20ee4bd590c512629812174e95c28857bcfef9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T18:17:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6vx22\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T18:17:31Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-82s52\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T18:17:45Z is after 2025-08-24T17:21:41Z" Oct 08 18:17:45 crc kubenswrapper[4859]: I1008 18:17:45.889361 4859 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-4g8gf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1dff864e-b75d-4e0b-b182-75f710eac8df\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:17:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:17:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:17:32Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:17:32Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://768ca274dedcc6847fc21d15234d3665fa9ccae371b5b6a882cfcb9b182f527d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T18:17:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s8jnn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://669f5f3c6d3c1ec0c4b4870e090b0a853adc01727c23383e595715b476c32668\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T18:17:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s8jnn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b8009a42cc4c9a816457a0678bdb2f7b7c53d0c8604e007a6fb141f66d06a516\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T18:17:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s8jnn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://676511a11ed4102084b57ebdd5f3fda4969741b2bc8443b415c4bf94f39ae4c9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T18:17:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s8jnn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4a74c1f0d21c90cdd7e776ee14e1f48757a04e2220fbd1f884449181ec6d1c3a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T18:17:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s8jnn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6ddc0e842bb8f81663406070023b0e5ed9cc99c98db0123b1c84f20da59fb5b5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T18:17:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s8jnn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2ce53a49f210c2d6e0001ee4280916611bbd82b4fc21a99c59954ed3d7570b96\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2ce53a49f210c2d6e0001ee4280916611bbd82b4fc21a99c59954ed3d7570b96\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-10-08T18:17:43Z\\\",\\\"message\\\":\\\"cert-secret-name:metrics-tls service.beta.openshift.io/serving-cert-signed-by:openshift-service-serving-signer@1740288168] [{config.openshift.io/v1 ClusterVersion version 9101b518-476b-4eea-8fa6-69b0534e5caa 0xc00695d3d7 \\\\u003cnil\\\\u003e}] [] []},Spec:ServiceSpec{Ports:[]ServicePort{ServicePort{Name:metrics,Protocol:TCP,Port:9393,TargetPort:{1 0 metrics},NodePort:0,AppProtocol:nil,},},Selector:map[string]string{name: ingress-operator,},ClusterIP:10.217.5.244,Type:ClusterIP,ExternalIPs:[],SessionAffinity:None,LoadBalancerIP:,LoadBalancerSourceRanges:[],ExternalName:,ExternalTrafficPolicy:,HealthCheckNodePort:0,PublishNotReadyAddresses:false,SessionAffinityConfig:nil,IPFamilyPolicy:*SingleStack,ClusterIPs:[10.217.5.244],IPFamilies:[IPv4],AllocateLoadBalancerNodePorts:nil,LoadBalancerClass:nil,InternalTrafficPolicy:*Cluster,TrafficDistribution:nil,},Status:ServiceStatus{LoadBalancer:LoadBalancerStatus{Ingress:[]LoadBalancerIngress{},},Conditions:[]Condition{},},}\\\\nI1008 18:17:43.764981 6316 base_network_controller_pods.go:477] [default/openshift-network-diagnostics/network-check-target-xd92c] creating logical port openshift-network-diagnostics_network-check-target-xd92c for pod on switch crc\\\\nF1008 18:17:43.764990 6316 ovnkube.go:137] failed to run ovnkube: [failed to start network controller: failed to start default network controller: unable to create\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-08T18:17:42Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 10s restarting failed container=ovnkube-controller pod=ovnkube-node-4g8gf_openshift-ovn-kubernetes(1dff864e-b75d-4e0b-b182-75f710eac8df)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s8jnn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2c67d2aed7126d76989e057d8e2d8e188a8902b2f3fe26ed6d1d45d3b013d897\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T18:17:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s8jnn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7a211d3da02b173867ff3a912eecd3b379483c7cc77e6b91122edd1ffd30fd53\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7a211d3da02b173867ff3a912eecd3b379483c7cc77e6b91122edd1ffd30fd53\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-08T18:17:32Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-08T18:17:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s8jnn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T18:17:32Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-4g8gf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T18:17:45Z is after 2025-08-24T17:21:41Z" Oct 08 18:17:45 crc kubenswrapper[4859]: I1008 18:17:45.904394 4859 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"63a623de-5608-4228-b1c2-77188c276986\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:17:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:17:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:17:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:17:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:17:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d4786a8fb8abd96fb22e60733c974f599017407d589ed815fb8541fd3b64f696\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T18:17:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://183d481502ed5108d5e19e483b714c289f355ea3f8b0891883e5335b802c83b6\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T18:17:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ec848c46fd5dde0469b00d3cab14ba4a652533a265490fadbee7a8b8843ef249\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T18:17:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4fe1828b4d63ff8106bfcd15e1dc1c1bd1ba7f77777cdb23ff1b291d0418774c\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2e700b2a20c81b414617f5b5c018a738278a3fa44e6c6d14b2170051d0725431\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-08T18:17:24Z\\\",\\\"message\\\":\\\"W1008 18:17:13.670625 1 cmd.go:257] Using insecure, self-signed certificates\\\\nI1008 18:17:13.671202 1 crypto.go:601] Generating new CA for check-endpoints-signer@1759947433 cert, and key in /tmp/serving-cert-3867218267/serving-signer.crt, /tmp/serving-cert-3867218267/serving-signer.key\\\\nI1008 18:17:13.944463 1 observer_polling.go:159] Starting file observer\\\\nW1008 18:17:13.951958 1 builder.go:272] unable to get owner reference (falling back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": dial tcp [::1]:6443: connect: connection refused\\\\nI1008 18:17:13.952088 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1008 18:17:13.956755 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3867218267/tls.crt::/tmp/serving-cert-3867218267/tls.key\\\\\\\"\\\\nF1008 18:17:24.290991 1 cmd.go:182] error initializing delegating authentication: unable to load configmap based request-header-client-ca-file: Get \\\\\\\"https://localhost:6443/api/v1/namespaces/kube-system/configmaps/extension-apiserver-authentication\\\\\\\": net/http: TLS handshake timeout\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-08T18:17:13Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T18:17:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c65b66f62d325ca10ef8d5d31f04529ccf4e691cf3303d231e7a414fb400e4d6\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T18:17:13Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://58c7bba4ace28215b062653a960e561b1dc1f890f2baf450145638a065af038b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://58c7bba4ace28215b062653a960e561b1dc1f890f2baf450145638a065af038b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-08T18:17:11Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-08T18:17:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T18:17:10Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T18:17:45Z is after 2025-08-24T17:21:41Z" Oct 08 18:17:45 crc kubenswrapper[4859]: I1008 18:17:45.916436 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 18:17:45 crc kubenswrapper[4859]: I1008 18:17:45.916469 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 18:17:45 crc kubenswrapper[4859]: I1008 18:17:45.916478 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 18:17:45 crc kubenswrapper[4859]: I1008 18:17:45.916494 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 18:17:45 crc kubenswrapper[4859]: I1008 18:17:45.916504 4859 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T18:17:45Z","lastTransitionTime":"2025-10-08T18:17:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 18:17:45 crc kubenswrapper[4859]: I1008 18:17:45.919257 4859 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-08T18:17:31Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T18:17:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T18:17:31Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://99e2d51bd7757498dae60949eea6d732b5f7b481fa00f808dd5308a530a1daf1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T18:17:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e32aaba93699944665660ae41e9b5779178c1354c31e15ad8188bf8c3c8fa500\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T18:17:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T18:17:45Z is after 2025-08-24T17:21:41Z" Oct 08 18:17:45 crc kubenswrapper[4859]: I1008 18:17:45.934790 4859 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4efc16fb-e679-464d-8b3c-0f46f082593e\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:17:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:17:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:17:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:17:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:17:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://aa8f8ea895272cd79a08d1c47cd7863a1ec10e21bb60d12aa10862a56c99e4bf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T18:17:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5c4e95d8a712ab219455564539a90286ae48bbc46ddc322a5bfd96c1a8268110\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T18:17:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://bae6ecfd8b446ef04d393d823e5ffe37fe749337c0ca17fde79c5f88d5e31393\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T18:17:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6411d1e7e18d5ceadaaceed5a8a6b20a0cacc55cede793ca9fb4dcad70c64fc6\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T18:17:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T18:17:10Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T18:17:45Z is after 2025-08-24T17:21:41Z" Oct 08 18:17:45 crc kubenswrapper[4859]: I1008 18:17:45.950483 4859 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-08T18:17:29Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T18:17:29Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T18:17:45Z is after 2025-08-24T17:21:41Z" Oct 08 18:17:45 crc kubenswrapper[4859]: I1008 18:17:45.969337 4859 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-84k8x" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1c24078c-c9e8-4160-883a-f483b4b3dd23\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:17:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:17:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:17:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:17:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8661006834d0f50eaada7394369584ee95fc61ff54437eec35d755526afc5b48\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T18:17:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hk5xb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://18b291eb3344fb7c00ef5679dca9a6f7ed9d08cea684a803ff7dad2e8a18c585\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://18b291eb3344fb7c00ef5679dca9a6f7ed9d08cea684a803ff7dad2e8a18c585\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-08T18:17:32Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-08T18:17:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hk5xb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://160163f1a91d981583452db3c7a9244182d1c8ced45749a61b721d0d45c1dfe4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://160163f1a91d981583452db3c7a9244182d1c8ced45749a61b721d0d45c1dfe4\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-08T18:17:34Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-08T18:17:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hk5xb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://51e9b53f3847ac53f378ad83870debf8e71518d4edbc7fc0dc66644818e2103e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://51e9b53f3847ac53f378ad83870debf8e71518d4edbc7fc0dc66644818e2103e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-08T18:17:34Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-08T18:17:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hk5xb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1ff118919739759ac01b5cbe3240e6673db54eaa16353c609d670071ed7964cb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1ff118919739759ac01b5cbe3240e6673db54eaa16353c609d670071ed7964cb\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-08T18:17:35Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-08T18:17:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hk5xb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8522e803ff56eb57cd4d9fd036e0c3fb16240ea21d4a094c1db7aa7d50bfe2c2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8522e803ff56eb57cd4d9fd036e0c3fb16240ea21d4a094c1db7aa7d50bfe2c2\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-08T18:17:37Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-08T18:17:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hk5xb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://be61056edf515e037403ca2a0bfd5feb9b2537dd42212b58e133b23b294f327b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://be61056edf515e037403ca2a0bfd5feb9b2537dd42212b58e133b23b294f327b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-08T18:17:38Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-08T18:17:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hk5xb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T18:17:31Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-84k8x\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T18:17:45Z is after 2025-08-24T17:21:41Z" Oct 08 18:17:45 crc kubenswrapper[4859]: I1008 18:17:45.971560 4859 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-multus/network-metrics-daemon-4b4sz"] Oct 08 18:17:45 crc kubenswrapper[4859]: I1008 18:17:45.972453 4859 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-4b4sz" Oct 08 18:17:45 crc kubenswrapper[4859]: E1008 18:17:45.972553 4859 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-4b4sz" podUID="2e53cd9b-64f7-4e07-8b96-fb77847c9ac6" Oct 08 18:17:45 crc kubenswrapper[4859]: I1008 18:17:45.979715 4859 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-btjmq" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b8605f7-fed7-493f-9b4f-6a6358adf907\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:17:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:17:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:17:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:17:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://14e9b6c4a4403f36adb78fcf1135a620ca6fc56ed974c9639dcdfc59e6917559\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T18:17:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xspg4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T18:17:35Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-btjmq\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T18:17:45Z is after 2025-08-24T17:21:41Z" Oct 08 18:17:45 crc kubenswrapper[4859]: I1008 18:17:45.991995 4859 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-g4t79" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3a92dbf1-21a8-4095-96bb-2d701d7f3d56\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:17:45Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:17:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:17:45Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-rbac-proxy ovnkube-cluster-manager]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:17:45Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-rbac-proxy ovnkube-cluster-manager]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sv79p\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sv79p\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T18:17:45Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-g4t79\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T18:17:45Z is after 2025-08-24T17:21:41Z" Oct 08 18:17:46 crc kubenswrapper[4859]: I1008 18:17:46.010808 4859 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-08T18:17:29Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T18:17:29Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T18:17:46Z is after 2025-08-24T17:21:41Z" Oct 08 18:17:46 crc kubenswrapper[4859]: I1008 18:17:46.018837 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 18:17:46 crc kubenswrapper[4859]: I1008 18:17:46.018891 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 18:17:46 crc kubenswrapper[4859]: I1008 18:17:46.018901 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 18:17:46 crc kubenswrapper[4859]: I1008 18:17:46.018920 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 18:17:46 crc kubenswrapper[4859]: I1008 18:17:46.018932 4859 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T18:17:46Z","lastTransitionTime":"2025-10-08T18:17:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 18:17:46 crc kubenswrapper[4859]: I1008 18:17:46.030947 4859 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-btw28" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"97bc5598-3ead-4300-a938-3aebbff1e8c6\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:17:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:17:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:17:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:17:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d6d4d6ba771cb5fde637449bba05660f634df7e833a0c8dad2b2c7f3a422bb2b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T18:17:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qcxsd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T18:17:31Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-btw28\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T18:17:46Z is after 2025-08-24T17:21:41Z" Oct 08 18:17:46 crc kubenswrapper[4859]: I1008 18:17:46.044359 4859 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-8plkg" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"92b2d7de-31cc-4a91-95a5-ed5ea964b028\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:17:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:17:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:17:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:17:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fa00b8345d40b74a917ddc23cc1b2c5c684a9ca61b6119d090c92cc1439b71c4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T18:17:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q9d7m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T18:17:31Z\\\"}}\" for pod \"openshift-multus\"/\"multus-8plkg\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T18:17:46Z is after 2025-08-24T17:21:41Z" Oct 08 18:17:46 crc kubenswrapper[4859]: I1008 18:17:46.057171 4859 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-08T18:17:29Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T18:17:29Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T18:17:46Z is after 2025-08-24T17:21:41Z" Oct 08 18:17:46 crc kubenswrapper[4859]: I1008 18:17:46.068961 4859 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-btw28" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"97bc5598-3ead-4300-a938-3aebbff1e8c6\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:17:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:17:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:17:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:17:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d6d4d6ba771cb5fde637449bba05660f634df7e833a0c8dad2b2c7f3a422bb2b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T18:17:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qcxsd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T18:17:31Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-btw28\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T18:17:46Z is after 2025-08-24T17:21:41Z" Oct 08 18:17:46 crc kubenswrapper[4859]: I1008 18:17:46.080299 4859 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-8plkg" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"92b2d7de-31cc-4a91-95a5-ed5ea964b028\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:17:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:17:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:17:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:17:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fa00b8345d40b74a917ddc23cc1b2c5c684a9ca61b6119d090c92cc1439b71c4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T18:17:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q9d7m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T18:17:31Z\\\"}}\" for pod \"openshift-multus\"/\"multus-8plkg\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T18:17:46Z is after 2025-08-24T17:21:41Z" Oct 08 18:17:46 crc kubenswrapper[4859]: I1008 18:17:46.090034 4859 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-btjmq" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b8605f7-fed7-493f-9b4f-6a6358adf907\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:17:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:17:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:17:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:17:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://14e9b6c4a4403f36adb78fcf1135a620ca6fc56ed974c9639dcdfc59e6917559\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T18:17:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xspg4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T18:17:35Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-btjmq\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T18:17:46Z is after 2025-08-24T17:21:41Z" Oct 08 18:17:46 crc kubenswrapper[4859]: I1008 18:17:46.092880 4859 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-kxgbz\" (UniqueName: \"kubernetes.io/projected/2e53cd9b-64f7-4e07-8b96-fb77847c9ac6-kube-api-access-kxgbz\") pod \"network-metrics-daemon-4b4sz\" (UID: \"2e53cd9b-64f7-4e07-8b96-fb77847c9ac6\") " pod="openshift-multus/network-metrics-daemon-4b4sz" Oct 08 18:17:46 crc kubenswrapper[4859]: I1008 18:17:46.092948 4859 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/2e53cd9b-64f7-4e07-8b96-fb77847c9ac6-metrics-certs\") pod \"network-metrics-daemon-4b4sz\" (UID: \"2e53cd9b-64f7-4e07-8b96-fb77847c9ac6\") " pod="openshift-multus/network-metrics-daemon-4b4sz" Oct 08 18:17:46 crc kubenswrapper[4859]: I1008 18:17:46.100190 4859 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-g4t79" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3a92dbf1-21a8-4095-96bb-2d701d7f3d56\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:17:45Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:17:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:17:45Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-rbac-proxy ovnkube-cluster-manager]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:17:45Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-rbac-proxy ovnkube-cluster-manager]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sv79p\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sv79p\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T18:17:45Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-g4t79\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T18:17:46Z is after 2025-08-24T17:21:41Z" Oct 08 18:17:46 crc kubenswrapper[4859]: I1008 18:17:46.114571 4859 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-08T18:17:31Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T18:17:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T18:17:31Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c6d15905694b6bbd4999d49dc8f7a53eb6bfaa7adb59efb0d7e4973c21d1c7be\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T18:17:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T18:17:46Z is after 2025-08-24T17:21:41Z" Oct 08 18:17:46 crc kubenswrapper[4859]: I1008 18:17:46.120904 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 18:17:46 crc kubenswrapper[4859]: I1008 18:17:46.120939 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 18:17:46 crc kubenswrapper[4859]: I1008 18:17:46.120950 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 18:17:46 crc kubenswrapper[4859]: I1008 18:17:46.120967 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 18:17:46 crc kubenswrapper[4859]: I1008 18:17:46.120978 4859 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T18:17:46Z","lastTransitionTime":"2025-10-08T18:17:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 18:17:46 crc kubenswrapper[4859]: I1008 18:17:46.135932 4859 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-08T18:17:33Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T18:17:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T18:17:33Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5d99e071bd083914418d9c128f38955f26295c08b1520f2dc774c99b1c843693\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T18:17:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T18:17:46Z is after 2025-08-24T17:21:41Z" Oct 08 18:17:46 crc kubenswrapper[4859]: I1008 18:17:46.178963 4859 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-4g8gf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1dff864e-b75d-4e0b-b182-75f710eac8df\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:17:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:17:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:17:32Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:17:32Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://768ca274dedcc6847fc21d15234d3665fa9ccae371b5b6a882cfcb9b182f527d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T18:17:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s8jnn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://669f5f3c6d3c1ec0c4b4870e090b0a853adc01727c23383e595715b476c32668\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T18:17:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s8jnn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b8009a42cc4c9a816457a0678bdb2f7b7c53d0c8604e007a6fb141f66d06a516\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T18:17:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s8jnn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://676511a11ed4102084b57ebdd5f3fda4969741b2bc8443b415c4bf94f39ae4c9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T18:17:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s8jnn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4a74c1f0d21c90cdd7e776ee14e1f48757a04e2220fbd1f884449181ec6d1c3a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T18:17:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s8jnn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6ddc0e842bb8f81663406070023b0e5ed9cc99c98db0123b1c84f20da59fb5b5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T18:17:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s8jnn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2ce53a49f210c2d6e0001ee4280916611bbd82b4fc21a99c59954ed3d7570b96\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2ce53a49f210c2d6e0001ee4280916611bbd82b4fc21a99c59954ed3d7570b96\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-10-08T18:17:43Z\\\",\\\"message\\\":\\\"cert-secret-name:metrics-tls service.beta.openshift.io/serving-cert-signed-by:openshift-service-serving-signer@1740288168] [{config.openshift.io/v1 ClusterVersion version 9101b518-476b-4eea-8fa6-69b0534e5caa 0xc00695d3d7 \\\\u003cnil\\\\u003e}] [] []},Spec:ServiceSpec{Ports:[]ServicePort{ServicePort{Name:metrics,Protocol:TCP,Port:9393,TargetPort:{1 0 metrics},NodePort:0,AppProtocol:nil,},},Selector:map[string]string{name: ingress-operator,},ClusterIP:10.217.5.244,Type:ClusterIP,ExternalIPs:[],SessionAffinity:None,LoadBalancerIP:,LoadBalancerSourceRanges:[],ExternalName:,ExternalTrafficPolicy:,HealthCheckNodePort:0,PublishNotReadyAddresses:false,SessionAffinityConfig:nil,IPFamilyPolicy:*SingleStack,ClusterIPs:[10.217.5.244],IPFamilies:[IPv4],AllocateLoadBalancerNodePorts:nil,LoadBalancerClass:nil,InternalTrafficPolicy:*Cluster,TrafficDistribution:nil,},Status:ServiceStatus{LoadBalancer:LoadBalancerStatus{Ingress:[]LoadBalancerIngress{},},Conditions:[]Condition{},},}\\\\nI1008 18:17:43.764981 6316 base_network_controller_pods.go:477] [default/openshift-network-diagnostics/network-check-target-xd92c] creating logical port openshift-network-diagnostics_network-check-target-xd92c for pod on switch crc\\\\nF1008 18:17:43.764990 6316 ovnkube.go:137] failed to run ovnkube: [failed to start network controller: failed to start default network controller: unable to create\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-08T18:17:42Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 10s restarting failed container=ovnkube-controller pod=ovnkube-node-4g8gf_openshift-ovn-kubernetes(1dff864e-b75d-4e0b-b182-75f710eac8df)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s8jnn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2c67d2aed7126d76989e057d8e2d8e188a8902b2f3fe26ed6d1d45d3b013d897\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T18:17:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s8jnn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7a211d3da02b173867ff3a912eecd3b379483c7cc77e6b91122edd1ffd30fd53\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7a211d3da02b173867ff3a912eecd3b379483c7cc77e6b91122edd1ffd30fd53\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-08T18:17:32Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-08T18:17:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s8jnn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T18:17:32Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-4g8gf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T18:17:46Z is after 2025-08-24T17:21:41Z" Oct 08 18:17:46 crc kubenswrapper[4859]: I1008 18:17:46.193126 4859 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-4b4sz" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2e53cd9b-64f7-4e07-8b96-fb77847c9ac6\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:17:45Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:17:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:17:45Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:17:45Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kxgbz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kxgbz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T18:17:45Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-4b4sz\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T18:17:46Z is after 2025-08-24T17:21:41Z" Oct 08 18:17:46 crc kubenswrapper[4859]: I1008 18:17:46.193422 4859 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-kxgbz\" (UniqueName: \"kubernetes.io/projected/2e53cd9b-64f7-4e07-8b96-fb77847c9ac6-kube-api-access-kxgbz\") pod \"network-metrics-daemon-4b4sz\" (UID: \"2e53cd9b-64f7-4e07-8b96-fb77847c9ac6\") " pod="openshift-multus/network-metrics-daemon-4b4sz" Oct 08 18:17:46 crc kubenswrapper[4859]: I1008 18:17:46.193462 4859 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/2e53cd9b-64f7-4e07-8b96-fb77847c9ac6-metrics-certs\") pod \"network-metrics-daemon-4b4sz\" (UID: \"2e53cd9b-64f7-4e07-8b96-fb77847c9ac6\") " pod="openshift-multus/network-metrics-daemon-4b4sz" Oct 08 18:17:46 crc kubenswrapper[4859]: E1008 18:17:46.193576 4859 secret.go:188] Couldn't get secret openshift-multus/metrics-daemon-secret: object "openshift-multus"/"metrics-daemon-secret" not registered Oct 08 18:17:46 crc kubenswrapper[4859]: E1008 18:17:46.193633 4859 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/2e53cd9b-64f7-4e07-8b96-fb77847c9ac6-metrics-certs podName:2e53cd9b-64f7-4e07-8b96-fb77847c9ac6 nodeName:}" failed. No retries permitted until 2025-10-08 18:17:46.69361862 +0000 UTC m=+36.940457999 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/2e53cd9b-64f7-4e07-8b96-fb77847c9ac6-metrics-certs") pod "network-metrics-daemon-4b4sz" (UID: "2e53cd9b-64f7-4e07-8b96-fb77847c9ac6") : object "openshift-multus"/"metrics-daemon-secret" not registered Oct 08 18:17:46 crc kubenswrapper[4859]: I1008 18:17:46.207491 4859 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"63a623de-5608-4228-b1c2-77188c276986\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:17:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:17:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:17:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:17:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:17:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d4786a8fb8abd96fb22e60733c974f599017407d589ed815fb8541fd3b64f696\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T18:17:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://183d481502ed5108d5e19e483b714c289f355ea3f8b0891883e5335b802c83b6\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T18:17:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ec848c46fd5dde0469b00d3cab14ba4a652533a265490fadbee7a8b8843ef249\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T18:17:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4fe1828b4d63ff8106bfcd15e1dc1c1bd1ba7f77777cdb23ff1b291d0418774c\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2e700b2a20c81b414617f5b5c018a738278a3fa44e6c6d14b2170051d0725431\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-08T18:17:24Z\\\",\\\"message\\\":\\\"W1008 18:17:13.670625 1 cmd.go:257] Using insecure, self-signed certificates\\\\nI1008 18:17:13.671202 1 crypto.go:601] Generating new CA for check-endpoints-signer@1759947433 cert, and key in /tmp/serving-cert-3867218267/serving-signer.crt, /tmp/serving-cert-3867218267/serving-signer.key\\\\nI1008 18:17:13.944463 1 observer_polling.go:159] Starting file observer\\\\nW1008 18:17:13.951958 1 builder.go:272] unable to get owner reference (falling back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": dial tcp [::1]:6443: connect: connection refused\\\\nI1008 18:17:13.952088 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1008 18:17:13.956755 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3867218267/tls.crt::/tmp/serving-cert-3867218267/tls.key\\\\\\\"\\\\nF1008 18:17:24.290991 1 cmd.go:182] error initializing delegating authentication: unable to load configmap based request-header-client-ca-file: Get \\\\\\\"https://localhost:6443/api/v1/namespaces/kube-system/configmaps/extension-apiserver-authentication\\\\\\\": net/http: TLS handshake timeout\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-08T18:17:13Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T18:17:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c65b66f62d325ca10ef8d5d31f04529ccf4e691cf3303d231e7a414fb400e4d6\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T18:17:13Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://58c7bba4ace28215b062653a960e561b1dc1f890f2baf450145638a065af038b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://58c7bba4ace28215b062653a960e561b1dc1f890f2baf450145638a065af038b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-08T18:17:11Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-08T18:17:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T18:17:10Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T18:17:46Z is after 2025-08-24T17:21:41Z" Oct 08 18:17:46 crc kubenswrapper[4859]: I1008 18:17:46.208600 4859 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-kxgbz\" (UniqueName: \"kubernetes.io/projected/2e53cd9b-64f7-4e07-8b96-fb77847c9ac6-kube-api-access-kxgbz\") pod \"network-metrics-daemon-4b4sz\" (UID: \"2e53cd9b-64f7-4e07-8b96-fb77847c9ac6\") " pod="openshift-multus/network-metrics-daemon-4b4sz" Oct 08 18:17:46 crc kubenswrapper[4859]: I1008 18:17:46.220259 4859 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-08T18:17:31Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T18:17:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T18:17:31Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://99e2d51bd7757498dae60949eea6d732b5f7b481fa00f808dd5308a530a1daf1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T18:17:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e32aaba93699944665660ae41e9b5779178c1354c31e15ad8188bf8c3c8fa500\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T18:17:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T18:17:46Z is after 2025-08-24T17:21:41Z" Oct 08 18:17:46 crc kubenswrapper[4859]: I1008 18:17:46.222987 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 18:17:46 crc kubenswrapper[4859]: I1008 18:17:46.223016 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 18:17:46 crc kubenswrapper[4859]: I1008 18:17:46.223024 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 18:17:46 crc kubenswrapper[4859]: I1008 18:17:46.223039 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 18:17:46 crc kubenswrapper[4859]: I1008 18:17:46.223049 4859 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T18:17:46Z","lastTransitionTime":"2025-10-08T18:17:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 18:17:46 crc kubenswrapper[4859]: I1008 18:17:46.232561 4859 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-08T18:17:29Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T18:17:29Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T18:17:46Z is after 2025-08-24T17:21:41Z" Oct 08 18:17:46 crc kubenswrapper[4859]: I1008 18:17:46.246626 4859 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-82s52" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b23a6a6c-9d92-4e7b-840e-55cfda873a2d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:17:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:17:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:17:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:17:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://71302c766485b9f55743d85f97440ff578cd44ba1e5a6f5b1c024c0e808072d2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T18:17:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6vx22\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c23ea3580998adf90621a0155c20ee4bd590c512629812174e95c28857bcfef9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T18:17:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6vx22\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T18:17:31Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-82s52\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T18:17:46Z is after 2025-08-24T17:21:41Z" Oct 08 18:17:46 crc kubenswrapper[4859]: I1008 18:17:46.259810 4859 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4efc16fb-e679-464d-8b3c-0f46f082593e\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:17:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:17:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:17:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:17:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:17:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://aa8f8ea895272cd79a08d1c47cd7863a1ec10e21bb60d12aa10862a56c99e4bf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T18:17:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5c4e95d8a712ab219455564539a90286ae48bbc46ddc322a5bfd96c1a8268110\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T18:17:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://bae6ecfd8b446ef04d393d823e5ffe37fe749337c0ca17fde79c5f88d5e31393\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T18:17:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6411d1e7e18d5ceadaaceed5a8a6b20a0cacc55cede793ca9fb4dcad70c64fc6\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T18:17:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T18:17:10Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T18:17:46Z is after 2025-08-24T17:21:41Z" Oct 08 18:17:46 crc kubenswrapper[4859]: I1008 18:17:46.274540 4859 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-08T18:17:29Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T18:17:29Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T18:17:46Z is after 2025-08-24T17:21:41Z" Oct 08 18:17:46 crc kubenswrapper[4859]: I1008 18:17:46.290840 4859 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-84k8x" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1c24078c-c9e8-4160-883a-f483b4b3dd23\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:17:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:17:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:17:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:17:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8661006834d0f50eaada7394369584ee95fc61ff54437eec35d755526afc5b48\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T18:17:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hk5xb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://18b291eb3344fb7c00ef5679dca9a6f7ed9d08cea684a803ff7dad2e8a18c585\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://18b291eb3344fb7c00ef5679dca9a6f7ed9d08cea684a803ff7dad2e8a18c585\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-08T18:17:32Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-08T18:17:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hk5xb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://160163f1a91d981583452db3c7a9244182d1c8ced45749a61b721d0d45c1dfe4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://160163f1a91d981583452db3c7a9244182d1c8ced45749a61b721d0d45c1dfe4\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-08T18:17:34Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-08T18:17:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hk5xb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://51e9b53f3847ac53f378ad83870debf8e71518d4edbc7fc0dc66644818e2103e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://51e9b53f3847ac53f378ad83870debf8e71518d4edbc7fc0dc66644818e2103e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-08T18:17:34Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-08T18:17:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hk5xb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1ff118919739759ac01b5cbe3240e6673db54eaa16353c609d670071ed7964cb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1ff118919739759ac01b5cbe3240e6673db54eaa16353c609d670071ed7964cb\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-08T18:17:35Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-08T18:17:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hk5xb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8522e803ff56eb57cd4d9fd036e0c3fb16240ea21d4a094c1db7aa7d50bfe2c2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8522e803ff56eb57cd4d9fd036e0c3fb16240ea21d4a094c1db7aa7d50bfe2c2\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-08T18:17:37Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-08T18:17:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hk5xb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://be61056edf515e037403ca2a0bfd5feb9b2537dd42212b58e133b23b294f327b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://be61056edf515e037403ca2a0bfd5feb9b2537dd42212b58e133b23b294f327b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-08T18:17:38Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-08T18:17:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hk5xb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T18:17:31Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-84k8x\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T18:17:46Z is after 2025-08-24T17:21:41Z" Oct 08 18:17:46 crc kubenswrapper[4859]: I1008 18:17:46.294014 4859 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 08 18:17:46 crc kubenswrapper[4859]: E1008 18:17:46.294362 4859 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-08 18:18:02.294317887 +0000 UTC m=+52.541157336 (durationBeforeRetry 16s). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 08 18:17:46 crc kubenswrapper[4859]: I1008 18:17:46.332865 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 18:17:46 crc kubenswrapper[4859]: I1008 18:17:46.332952 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 18:17:46 crc kubenswrapper[4859]: I1008 18:17:46.332966 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 18:17:46 crc kubenswrapper[4859]: I1008 18:17:46.332984 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 18:17:46 crc kubenswrapper[4859]: I1008 18:17:46.333000 4859 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T18:17:46Z","lastTransitionTime":"2025-10-08T18:17:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 18:17:46 crc kubenswrapper[4859]: I1008 18:17:46.394679 4859 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 08 18:17:46 crc kubenswrapper[4859]: I1008 18:17:46.394763 4859 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 08 18:17:46 crc kubenswrapper[4859]: I1008 18:17:46.394787 4859 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 08 18:17:46 crc kubenswrapper[4859]: I1008 18:17:46.394821 4859 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 08 18:17:46 crc kubenswrapper[4859]: E1008 18:17:46.394899 4859 configmap.go:193] Couldn't get configMap openshift-network-console/networking-console-plugin: object "openshift-network-console"/"networking-console-plugin" not registered Oct 08 18:17:46 crc kubenswrapper[4859]: E1008 18:17:46.394953 4859 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-10-08 18:18:02.394939191 +0000 UTC m=+52.641778570 (durationBeforeRetry 16s). Error: MountVolume.SetUp failed for volume "nginx-conf" (UniqueName: "kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin" not registered Oct 08 18:17:46 crc kubenswrapper[4859]: E1008 18:17:46.395039 4859 secret.go:188] Couldn't get secret openshift-network-console/networking-console-plugin-cert: object "openshift-network-console"/"networking-console-plugin-cert" not registered Oct 08 18:17:46 crc kubenswrapper[4859]: E1008 18:17:46.395088 4859 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-10-08 18:18:02.395078865 +0000 UTC m=+52.641918244 (durationBeforeRetry 16s). Error: MountVolume.SetUp failed for volume "networking-console-plugin-cert" (UniqueName: "kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin-cert" not registered Oct 08 18:17:46 crc kubenswrapper[4859]: E1008 18:17:46.395096 4859 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Oct 08 18:17:46 crc kubenswrapper[4859]: E1008 18:17:46.395115 4859 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Oct 08 18:17:46 crc kubenswrapper[4859]: E1008 18:17:46.395191 4859 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Oct 08 18:17:46 crc kubenswrapper[4859]: E1008 18:17:46.395144 4859 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Oct 08 18:17:46 crc kubenswrapper[4859]: E1008 18:17:46.395208 4859 projected.go:194] Error preparing data for projected volume kube-api-access-cqllr for pod openshift-network-diagnostics/network-check-target-xd92c: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Oct 08 18:17:46 crc kubenswrapper[4859]: E1008 18:17:46.395228 4859 projected.go:194] Error preparing data for projected volume kube-api-access-s2dwl for pod openshift-network-diagnostics/network-check-source-55646444c4-trplf: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Oct 08 18:17:46 crc kubenswrapper[4859]: E1008 18:17:46.395300 4859 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr podName:3b6479f0-333b-4a96-9adf-2099afdc2447 nodeName:}" failed. No retries permitted until 2025-10-08 18:18:02.39527237 +0000 UTC m=+52.642111879 (durationBeforeRetry 16s). Error: MountVolume.SetUp failed for volume "kube-api-access-cqllr" (UniqueName: "kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr") pod "network-check-target-xd92c" (UID: "3b6479f0-333b-4a96-9adf-2099afdc2447") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Oct 08 18:17:46 crc kubenswrapper[4859]: E1008 18:17:46.395322 4859 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl podName:9d751cbb-f2e2-430d-9754-c882a5e924a5 nodeName:}" failed. No retries permitted until 2025-10-08 18:18:02.395313381 +0000 UTC m=+52.642152770 (durationBeforeRetry 16s). Error: MountVolume.SetUp failed for volume "kube-api-access-s2dwl" (UniqueName: "kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl") pod "network-check-source-55646444c4-trplf" (UID: "9d751cbb-f2e2-430d-9754-c882a5e924a5") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Oct 08 18:17:46 crc kubenswrapper[4859]: I1008 18:17:46.436302 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 18:17:46 crc kubenswrapper[4859]: I1008 18:17:46.436371 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 18:17:46 crc kubenswrapper[4859]: I1008 18:17:46.436381 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 18:17:46 crc kubenswrapper[4859]: I1008 18:17:46.436401 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 18:17:46 crc kubenswrapper[4859]: I1008 18:17:46.436412 4859 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T18:17:46Z","lastTransitionTime":"2025-10-08T18:17:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 18:17:46 crc kubenswrapper[4859]: I1008 18:17:46.540563 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 18:17:46 crc kubenswrapper[4859]: I1008 18:17:46.540645 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 18:17:46 crc kubenswrapper[4859]: I1008 18:17:46.540670 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 18:17:46 crc kubenswrapper[4859]: I1008 18:17:46.540736 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 18:17:46 crc kubenswrapper[4859]: I1008 18:17:46.540761 4859 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T18:17:46Z","lastTransitionTime":"2025-10-08T18:17:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 18:17:46 crc kubenswrapper[4859]: I1008 18:17:46.645666 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 18:17:46 crc kubenswrapper[4859]: I1008 18:17:46.645811 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 18:17:46 crc kubenswrapper[4859]: I1008 18:17:46.645836 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 18:17:46 crc kubenswrapper[4859]: I1008 18:17:46.645865 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 18:17:46 crc kubenswrapper[4859]: I1008 18:17:46.645885 4859 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T18:17:46Z","lastTransitionTime":"2025-10-08T18:17:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 18:17:46 crc kubenswrapper[4859]: I1008 18:17:46.698682 4859 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/2e53cd9b-64f7-4e07-8b96-fb77847c9ac6-metrics-certs\") pod \"network-metrics-daemon-4b4sz\" (UID: \"2e53cd9b-64f7-4e07-8b96-fb77847c9ac6\") " pod="openshift-multus/network-metrics-daemon-4b4sz" Oct 08 18:17:46 crc kubenswrapper[4859]: E1008 18:17:46.699079 4859 secret.go:188] Couldn't get secret openshift-multus/metrics-daemon-secret: object "openshift-multus"/"metrics-daemon-secret" not registered Oct 08 18:17:46 crc kubenswrapper[4859]: E1008 18:17:46.699231 4859 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/2e53cd9b-64f7-4e07-8b96-fb77847c9ac6-metrics-certs podName:2e53cd9b-64f7-4e07-8b96-fb77847c9ac6 nodeName:}" failed. No retries permitted until 2025-10-08 18:17:47.699195039 +0000 UTC m=+37.946034458 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/2e53cd9b-64f7-4e07-8b96-fb77847c9ac6-metrics-certs") pod "network-metrics-daemon-4b4sz" (UID: "2e53cd9b-64f7-4e07-8b96-fb77847c9ac6") : object "openshift-multus"/"metrics-daemon-secret" not registered Oct 08 18:17:46 crc kubenswrapper[4859]: I1008 18:17:46.749514 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 18:17:46 crc kubenswrapper[4859]: I1008 18:17:46.749593 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 18:17:46 crc kubenswrapper[4859]: I1008 18:17:46.749608 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 18:17:46 crc kubenswrapper[4859]: I1008 18:17:46.749632 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 18:17:46 crc kubenswrapper[4859]: I1008 18:17:46.749649 4859 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T18:17:46Z","lastTransitionTime":"2025-10-08T18:17:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 18:17:46 crc kubenswrapper[4859]: I1008 18:17:46.812865 4859 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-g4t79" event={"ID":"3a92dbf1-21a8-4095-96bb-2d701d7f3d56","Type":"ContainerStarted","Data":"92cb9986c09a7f9e1e552e3e3ac1f74f04128241e00ddee38882d2ba5c005a67"} Oct 08 18:17:46 crc kubenswrapper[4859]: I1008 18:17:46.829647 4859 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-4b4sz" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2e53cd9b-64f7-4e07-8b96-fb77847c9ac6\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:17:45Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:17:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:17:45Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:17:45Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kxgbz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kxgbz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T18:17:45Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-4b4sz\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T18:17:46Z is after 2025-08-24T17:21:41Z" Oct 08 18:17:46 crc kubenswrapper[4859]: I1008 18:17:46.854033 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 18:17:46 crc kubenswrapper[4859]: I1008 18:17:46.854088 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 18:17:46 crc kubenswrapper[4859]: I1008 18:17:46.854106 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 18:17:46 crc kubenswrapper[4859]: I1008 18:17:46.854132 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 18:17:46 crc kubenswrapper[4859]: I1008 18:17:46.854152 4859 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T18:17:46Z","lastTransitionTime":"2025-10-08T18:17:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 18:17:46 crc kubenswrapper[4859]: I1008 18:17:46.857419 4859 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"63a623de-5608-4228-b1c2-77188c276986\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:17:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:17:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:17:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:17:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:17:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d4786a8fb8abd96fb22e60733c974f599017407d589ed815fb8541fd3b64f696\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T18:17:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://183d481502ed5108d5e19e483b714c289f355ea3f8b0891883e5335b802c83b6\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T18:17:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ec848c46fd5dde0469b00d3cab14ba4a652533a265490fadbee7a8b8843ef249\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T18:17:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4fe1828b4d63ff8106bfcd15e1dc1c1bd1ba7f77777cdb23ff1b291d0418774c\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2e700b2a20c81b414617f5b5c018a738278a3fa44e6c6d14b2170051d0725431\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-08T18:17:24Z\\\",\\\"message\\\":\\\"W1008 18:17:13.670625 1 cmd.go:257] Using insecure, self-signed certificates\\\\nI1008 18:17:13.671202 1 crypto.go:601] Generating new CA for check-endpoints-signer@1759947433 cert, and key in /tmp/serving-cert-3867218267/serving-signer.crt, /tmp/serving-cert-3867218267/serving-signer.key\\\\nI1008 18:17:13.944463 1 observer_polling.go:159] Starting file observer\\\\nW1008 18:17:13.951958 1 builder.go:272] unable to get owner reference (falling back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": dial tcp [::1]:6443: connect: connection refused\\\\nI1008 18:17:13.952088 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1008 18:17:13.956755 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3867218267/tls.crt::/tmp/serving-cert-3867218267/tls.key\\\\\\\"\\\\nF1008 18:17:24.290991 1 cmd.go:182] error initializing delegating authentication: unable to load configmap based request-header-client-ca-file: Get \\\\\\\"https://localhost:6443/api/v1/namespaces/kube-system/configmaps/extension-apiserver-authentication\\\\\\\": net/http: TLS handshake timeout\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-08T18:17:13Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T18:17:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c65b66f62d325ca10ef8d5d31f04529ccf4e691cf3303d231e7a414fb400e4d6\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T18:17:13Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://58c7bba4ace28215b062653a960e561b1dc1f890f2baf450145638a065af038b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://58c7bba4ace28215b062653a960e561b1dc1f890f2baf450145638a065af038b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-08T18:17:11Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-08T18:17:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T18:17:10Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T18:17:46Z is after 2025-08-24T17:21:41Z" Oct 08 18:17:46 crc kubenswrapper[4859]: I1008 18:17:46.881757 4859 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-08T18:17:31Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T18:17:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T18:17:31Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://99e2d51bd7757498dae60949eea6d732b5f7b481fa00f808dd5308a530a1daf1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T18:17:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e32aaba93699944665660ae41e9b5779178c1354c31e15ad8188bf8c3c8fa500\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T18:17:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T18:17:46Z is after 2025-08-24T17:21:41Z" Oct 08 18:17:46 crc kubenswrapper[4859]: I1008 18:17:46.900120 4859 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-08T18:17:29Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T18:17:29Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T18:17:46Z is after 2025-08-24T17:21:41Z" Oct 08 18:17:46 crc kubenswrapper[4859]: I1008 18:17:46.920017 4859 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-82s52" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b23a6a6c-9d92-4e7b-840e-55cfda873a2d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:17:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:17:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:17:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:17:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://71302c766485b9f55743d85f97440ff578cd44ba1e5a6f5b1c024c0e808072d2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T18:17:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6vx22\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c23ea3580998adf90621a0155c20ee4bd590c512629812174e95c28857bcfef9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T18:17:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6vx22\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T18:17:31Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-82s52\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T18:17:46Z is after 2025-08-24T17:21:41Z" Oct 08 18:17:46 crc kubenswrapper[4859]: I1008 18:17:46.941061 4859 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-4g8gf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1dff864e-b75d-4e0b-b182-75f710eac8df\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:17:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:17:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:17:32Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:17:32Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://768ca274dedcc6847fc21d15234d3665fa9ccae371b5b6a882cfcb9b182f527d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T18:17:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s8jnn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://669f5f3c6d3c1ec0c4b4870e090b0a853adc01727c23383e595715b476c32668\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T18:17:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s8jnn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b8009a42cc4c9a816457a0678bdb2f7b7c53d0c8604e007a6fb141f66d06a516\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T18:17:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s8jnn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://676511a11ed4102084b57ebdd5f3fda4969741b2bc8443b415c4bf94f39ae4c9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T18:17:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s8jnn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4a74c1f0d21c90cdd7e776ee14e1f48757a04e2220fbd1f884449181ec6d1c3a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T18:17:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s8jnn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6ddc0e842bb8f81663406070023b0e5ed9cc99c98db0123b1c84f20da59fb5b5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T18:17:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s8jnn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2ce53a49f210c2d6e0001ee4280916611bbd82b4fc21a99c59954ed3d7570b96\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2ce53a49f210c2d6e0001ee4280916611bbd82b4fc21a99c59954ed3d7570b96\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-10-08T18:17:43Z\\\",\\\"message\\\":\\\"cert-secret-name:metrics-tls service.beta.openshift.io/serving-cert-signed-by:openshift-service-serving-signer@1740288168] [{config.openshift.io/v1 ClusterVersion version 9101b518-476b-4eea-8fa6-69b0534e5caa 0xc00695d3d7 \\\\u003cnil\\\\u003e}] [] []},Spec:ServiceSpec{Ports:[]ServicePort{ServicePort{Name:metrics,Protocol:TCP,Port:9393,TargetPort:{1 0 metrics},NodePort:0,AppProtocol:nil,},},Selector:map[string]string{name: ingress-operator,},ClusterIP:10.217.5.244,Type:ClusterIP,ExternalIPs:[],SessionAffinity:None,LoadBalancerIP:,LoadBalancerSourceRanges:[],ExternalName:,ExternalTrafficPolicy:,HealthCheckNodePort:0,PublishNotReadyAddresses:false,SessionAffinityConfig:nil,IPFamilyPolicy:*SingleStack,ClusterIPs:[10.217.5.244],IPFamilies:[IPv4],AllocateLoadBalancerNodePorts:nil,LoadBalancerClass:nil,InternalTrafficPolicy:*Cluster,TrafficDistribution:nil,},Status:ServiceStatus{LoadBalancer:LoadBalancerStatus{Ingress:[]LoadBalancerIngress{},},Conditions:[]Condition{},},}\\\\nI1008 18:17:43.764981 6316 base_network_controller_pods.go:477] [default/openshift-network-diagnostics/network-check-target-xd92c] creating logical port openshift-network-diagnostics_network-check-target-xd92c for pod on switch crc\\\\nF1008 18:17:43.764990 6316 ovnkube.go:137] failed to run ovnkube: [failed to start network controller: failed to start default network controller: unable to create\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-08T18:17:42Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 10s restarting failed container=ovnkube-controller pod=ovnkube-node-4g8gf_openshift-ovn-kubernetes(1dff864e-b75d-4e0b-b182-75f710eac8df)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s8jnn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2c67d2aed7126d76989e057d8e2d8e188a8902b2f3fe26ed6d1d45d3b013d897\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T18:17:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s8jnn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7a211d3da02b173867ff3a912eecd3b379483c7cc77e6b91122edd1ffd30fd53\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7a211d3da02b173867ff3a912eecd3b379483c7cc77e6b91122edd1ffd30fd53\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-08T18:17:32Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-08T18:17:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s8jnn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T18:17:32Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-4g8gf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T18:17:46Z is after 2025-08-24T17:21:41Z" Oct 08 18:17:46 crc kubenswrapper[4859]: I1008 18:17:46.954831 4859 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4efc16fb-e679-464d-8b3c-0f46f082593e\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:17:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:17:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:17:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:17:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:17:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://aa8f8ea895272cd79a08d1c47cd7863a1ec10e21bb60d12aa10862a56c99e4bf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T18:17:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5c4e95d8a712ab219455564539a90286ae48bbc46ddc322a5bfd96c1a8268110\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T18:17:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://bae6ecfd8b446ef04d393d823e5ffe37fe749337c0ca17fde79c5f88d5e31393\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T18:17:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6411d1e7e18d5ceadaaceed5a8a6b20a0cacc55cede793ca9fb4dcad70c64fc6\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T18:17:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T18:17:10Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T18:17:46Z is after 2025-08-24T17:21:41Z" Oct 08 18:17:46 crc kubenswrapper[4859]: I1008 18:17:46.956706 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 18:17:46 crc kubenswrapper[4859]: I1008 18:17:46.956753 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 18:17:46 crc kubenswrapper[4859]: I1008 18:17:46.956765 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 18:17:46 crc kubenswrapper[4859]: I1008 18:17:46.956782 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 18:17:46 crc kubenswrapper[4859]: I1008 18:17:46.956794 4859 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T18:17:46Z","lastTransitionTime":"2025-10-08T18:17:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 18:17:46 crc kubenswrapper[4859]: I1008 18:17:46.968289 4859 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-08T18:17:29Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T18:17:29Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T18:17:46Z is after 2025-08-24T17:21:41Z" Oct 08 18:17:46 crc kubenswrapper[4859]: I1008 18:17:46.983923 4859 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-84k8x" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1c24078c-c9e8-4160-883a-f483b4b3dd23\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:17:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:17:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:17:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:17:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8661006834d0f50eaada7394369584ee95fc61ff54437eec35d755526afc5b48\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T18:17:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hk5xb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://18b291eb3344fb7c00ef5679dca9a6f7ed9d08cea684a803ff7dad2e8a18c585\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://18b291eb3344fb7c00ef5679dca9a6f7ed9d08cea684a803ff7dad2e8a18c585\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-08T18:17:32Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-08T18:17:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hk5xb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://160163f1a91d981583452db3c7a9244182d1c8ced45749a61b721d0d45c1dfe4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://160163f1a91d981583452db3c7a9244182d1c8ced45749a61b721d0d45c1dfe4\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-08T18:17:34Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-08T18:17:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hk5xb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://51e9b53f3847ac53f378ad83870debf8e71518d4edbc7fc0dc66644818e2103e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://51e9b53f3847ac53f378ad83870debf8e71518d4edbc7fc0dc66644818e2103e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-08T18:17:34Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-08T18:17:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hk5xb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1ff118919739759ac01b5cbe3240e6673db54eaa16353c609d670071ed7964cb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1ff118919739759ac01b5cbe3240e6673db54eaa16353c609d670071ed7964cb\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-08T18:17:35Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-08T18:17:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hk5xb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8522e803ff56eb57cd4d9fd036e0c3fb16240ea21d4a094c1db7aa7d50bfe2c2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8522e803ff56eb57cd4d9fd036e0c3fb16240ea21d4a094c1db7aa7d50bfe2c2\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-08T18:17:37Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-08T18:17:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hk5xb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://be61056edf515e037403ca2a0bfd5feb9b2537dd42212b58e133b23b294f327b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://be61056edf515e037403ca2a0bfd5feb9b2537dd42212b58e133b23b294f327b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-08T18:17:38Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-08T18:17:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hk5xb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T18:17:31Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-84k8x\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T18:17:46Z is after 2025-08-24T17:21:41Z" Oct 08 18:17:46 crc kubenswrapper[4859]: I1008 18:17:46.997085 4859 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-08T18:17:29Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T18:17:29Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T18:17:46Z is after 2025-08-24T17:21:41Z" Oct 08 18:17:47 crc kubenswrapper[4859]: I1008 18:17:47.008189 4859 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-btw28" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"97bc5598-3ead-4300-a938-3aebbff1e8c6\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:17:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:17:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:17:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:17:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d6d4d6ba771cb5fde637449bba05660f634df7e833a0c8dad2b2c7f3a422bb2b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T18:17:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qcxsd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T18:17:31Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-btw28\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T18:17:47Z is after 2025-08-24T17:21:41Z" Oct 08 18:17:47 crc kubenswrapper[4859]: I1008 18:17:47.022338 4859 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-8plkg" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"92b2d7de-31cc-4a91-95a5-ed5ea964b028\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:17:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:17:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:17:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:17:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fa00b8345d40b74a917ddc23cc1b2c5c684a9ca61b6119d090c92cc1439b71c4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T18:17:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q9d7m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T18:17:31Z\\\"}}\" for pod \"openshift-multus\"/\"multus-8plkg\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T18:17:47Z is after 2025-08-24T17:21:41Z" Oct 08 18:17:47 crc kubenswrapper[4859]: I1008 18:17:47.033604 4859 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-btjmq" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b8605f7-fed7-493f-9b4f-6a6358adf907\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:17:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:17:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:17:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:17:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://14e9b6c4a4403f36adb78fcf1135a620ca6fc56ed974c9639dcdfc59e6917559\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T18:17:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xspg4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T18:17:35Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-btjmq\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T18:17:47Z is after 2025-08-24T17:21:41Z" Oct 08 18:17:47 crc kubenswrapper[4859]: I1008 18:17:47.045511 4859 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-g4t79" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3a92dbf1-21a8-4095-96bb-2d701d7f3d56\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:17:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:17:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:17:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:17:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f8772a25b6e605efe9a062dc32dbbc2cfe32b4fe164b4e84fbf3bcffab744179\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T18:17:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sv79p\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://92cb9986c09a7f9e1e552e3e3ac1f74f04128241e00ddee38882d2ba5c005a67\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T18:17:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sv79p\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T18:17:45Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-g4t79\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T18:17:47Z is after 2025-08-24T17:21:41Z" Oct 08 18:17:47 crc kubenswrapper[4859]: I1008 18:17:47.059816 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 18:17:47 crc kubenswrapper[4859]: I1008 18:17:47.059851 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 18:17:47 crc kubenswrapper[4859]: I1008 18:17:47.059862 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 18:17:47 crc kubenswrapper[4859]: I1008 18:17:47.059881 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 18:17:47 crc kubenswrapper[4859]: I1008 18:17:47.059892 4859 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T18:17:47Z","lastTransitionTime":"2025-10-08T18:17:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 18:17:47 crc kubenswrapper[4859]: I1008 18:17:47.060190 4859 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-08T18:17:31Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T18:17:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T18:17:31Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c6d15905694b6bbd4999d49dc8f7a53eb6bfaa7adb59efb0d7e4973c21d1c7be\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T18:17:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T18:17:47Z is after 2025-08-24T17:21:41Z" Oct 08 18:17:47 crc kubenswrapper[4859]: I1008 18:17:47.073376 4859 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-08T18:17:33Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T18:17:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T18:17:33Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5d99e071bd083914418d9c128f38955f26295c08b1520f2dc774c99b1c843693\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T18:17:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T18:17:47Z is after 2025-08-24T17:21:41Z" Oct 08 18:17:47 crc kubenswrapper[4859]: I1008 18:17:47.162744 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 18:17:47 crc kubenswrapper[4859]: I1008 18:17:47.162807 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 18:17:47 crc kubenswrapper[4859]: I1008 18:17:47.162827 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 18:17:47 crc kubenswrapper[4859]: I1008 18:17:47.162852 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 18:17:47 crc kubenswrapper[4859]: I1008 18:17:47.162870 4859 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T18:17:47Z","lastTransitionTime":"2025-10-08T18:17:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 18:17:47 crc kubenswrapper[4859]: I1008 18:17:47.216384 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 18:17:47 crc kubenswrapper[4859]: I1008 18:17:47.216445 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 18:17:47 crc kubenswrapper[4859]: I1008 18:17:47.216462 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 18:17:47 crc kubenswrapper[4859]: I1008 18:17:47.216485 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 18:17:47 crc kubenswrapper[4859]: I1008 18:17:47.216501 4859 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T18:17:47Z","lastTransitionTime":"2025-10-08T18:17:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 18:17:47 crc kubenswrapper[4859]: E1008 18:17:47.236131 4859 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404552Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865352Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-10-08T18:17:47Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-08T18:17:47Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-08T18:17:47Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-08T18:17:47Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-08T18:17:47Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-08T18:17:47Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-08T18:17:47Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-08T18:17:47Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"edb3ee6d-476b-407c-bb7d-4480786e8777\\\",\\\"systemUUID\\\":\\\"8f73c7ed-44aa-4d16-bae1-d4e684803cfc\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T18:17:47Z is after 2025-08-24T17:21:41Z" Oct 08 18:17:47 crc kubenswrapper[4859]: I1008 18:17:47.241297 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 18:17:47 crc kubenswrapper[4859]: I1008 18:17:47.241352 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 18:17:47 crc kubenswrapper[4859]: I1008 18:17:47.241363 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 18:17:47 crc kubenswrapper[4859]: I1008 18:17:47.241386 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 18:17:47 crc kubenswrapper[4859]: I1008 18:17:47.241398 4859 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T18:17:47Z","lastTransitionTime":"2025-10-08T18:17:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 18:17:47 crc kubenswrapper[4859]: E1008 18:17:47.261765 4859 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404552Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865352Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-10-08T18:17:47Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-08T18:17:47Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-08T18:17:47Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-08T18:17:47Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-08T18:17:47Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-08T18:17:47Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-08T18:17:47Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-08T18:17:47Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"edb3ee6d-476b-407c-bb7d-4480786e8777\\\",\\\"systemUUID\\\":\\\"8f73c7ed-44aa-4d16-bae1-d4e684803cfc\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T18:17:47Z is after 2025-08-24T17:21:41Z" Oct 08 18:17:47 crc kubenswrapper[4859]: I1008 18:17:47.266987 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 18:17:47 crc kubenswrapper[4859]: I1008 18:17:47.267049 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 18:17:47 crc kubenswrapper[4859]: I1008 18:17:47.267069 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 18:17:47 crc kubenswrapper[4859]: I1008 18:17:47.267096 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 18:17:47 crc kubenswrapper[4859]: I1008 18:17:47.267159 4859 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T18:17:47Z","lastTransitionTime":"2025-10-08T18:17:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 18:17:47 crc kubenswrapper[4859]: E1008 18:17:47.282122 4859 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404552Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865352Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-10-08T18:17:47Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-08T18:17:47Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-08T18:17:47Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-08T18:17:47Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-08T18:17:47Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-08T18:17:47Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-08T18:17:47Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-08T18:17:47Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"edb3ee6d-476b-407c-bb7d-4480786e8777\\\",\\\"systemUUID\\\":\\\"8f73c7ed-44aa-4d16-bae1-d4e684803cfc\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T18:17:47Z is after 2025-08-24T17:21:41Z" Oct 08 18:17:47 crc kubenswrapper[4859]: I1008 18:17:47.287099 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 18:17:47 crc kubenswrapper[4859]: I1008 18:17:47.287166 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 18:17:47 crc kubenswrapper[4859]: I1008 18:17:47.287185 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 18:17:47 crc kubenswrapper[4859]: I1008 18:17:47.287214 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 18:17:47 crc kubenswrapper[4859]: I1008 18:17:47.287235 4859 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T18:17:47Z","lastTransitionTime":"2025-10-08T18:17:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 18:17:47 crc kubenswrapper[4859]: E1008 18:17:47.308181 4859 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404552Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865352Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-10-08T18:17:47Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-08T18:17:47Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-08T18:17:47Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-08T18:17:47Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-08T18:17:47Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-08T18:17:47Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-08T18:17:47Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-08T18:17:47Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"edb3ee6d-476b-407c-bb7d-4480786e8777\\\",\\\"systemUUID\\\":\\\"8f73c7ed-44aa-4d16-bae1-d4e684803cfc\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T18:17:47Z is after 2025-08-24T17:21:41Z" Oct 08 18:17:47 crc kubenswrapper[4859]: I1008 18:17:47.313715 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 18:17:47 crc kubenswrapper[4859]: I1008 18:17:47.313789 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 18:17:47 crc kubenswrapper[4859]: I1008 18:17:47.313811 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 18:17:47 crc kubenswrapper[4859]: I1008 18:17:47.313839 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 18:17:47 crc kubenswrapper[4859]: I1008 18:17:47.313863 4859 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T18:17:47Z","lastTransitionTime":"2025-10-08T18:17:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 18:17:47 crc kubenswrapper[4859]: E1008 18:17:47.331417 4859 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404552Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865352Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-10-08T18:17:47Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-08T18:17:47Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-08T18:17:47Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-08T18:17:47Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-08T18:17:47Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-08T18:17:47Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-08T18:17:47Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-08T18:17:47Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"edb3ee6d-476b-407c-bb7d-4480786e8777\\\",\\\"systemUUID\\\":\\\"8f73c7ed-44aa-4d16-bae1-d4e684803cfc\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T18:17:47Z is after 2025-08-24T17:21:41Z" Oct 08 18:17:47 crc kubenswrapper[4859]: E1008 18:17:47.331558 4859 kubelet_node_status.go:572] "Unable to update node status" err="update node status exceeds retry count" Oct 08 18:17:47 crc kubenswrapper[4859]: I1008 18:17:47.333571 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 18:17:47 crc kubenswrapper[4859]: I1008 18:17:47.333633 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 18:17:47 crc kubenswrapper[4859]: I1008 18:17:47.333646 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 18:17:47 crc kubenswrapper[4859]: I1008 18:17:47.333667 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 18:17:47 crc kubenswrapper[4859]: I1008 18:17:47.333680 4859 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T18:17:47Z","lastTransitionTime":"2025-10-08T18:17:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 18:17:47 crc kubenswrapper[4859]: I1008 18:17:47.437140 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 18:17:47 crc kubenswrapper[4859]: I1008 18:17:47.437186 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 18:17:47 crc kubenswrapper[4859]: I1008 18:17:47.437197 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 18:17:47 crc kubenswrapper[4859]: I1008 18:17:47.437217 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 18:17:47 crc kubenswrapper[4859]: I1008 18:17:47.437230 4859 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T18:17:47Z","lastTransitionTime":"2025-10-08T18:17:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 18:17:47 crc kubenswrapper[4859]: I1008 18:17:47.469424 4859 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 08 18:17:47 crc kubenswrapper[4859]: E1008 18:17:47.469732 4859 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 08 18:17:47 crc kubenswrapper[4859]: I1008 18:17:47.470066 4859 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-4b4sz" Oct 08 18:17:47 crc kubenswrapper[4859]: I1008 18:17:47.470180 4859 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 08 18:17:47 crc kubenswrapper[4859]: E1008 18:17:47.470259 4859 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-4b4sz" podUID="2e53cd9b-64f7-4e07-8b96-fb77847c9ac6" Oct 08 18:17:47 crc kubenswrapper[4859]: I1008 18:17:47.470180 4859 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 08 18:17:47 crc kubenswrapper[4859]: E1008 18:17:47.470405 4859 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 08 18:17:47 crc kubenswrapper[4859]: E1008 18:17:47.470506 4859 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 08 18:17:47 crc kubenswrapper[4859]: I1008 18:17:47.540529 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 18:17:47 crc kubenswrapper[4859]: I1008 18:17:47.540580 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 18:17:47 crc kubenswrapper[4859]: I1008 18:17:47.540596 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 18:17:47 crc kubenswrapper[4859]: I1008 18:17:47.540616 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 18:17:47 crc kubenswrapper[4859]: I1008 18:17:47.540629 4859 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T18:17:47Z","lastTransitionTime":"2025-10-08T18:17:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 18:17:47 crc kubenswrapper[4859]: I1008 18:17:47.644519 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 18:17:47 crc kubenswrapper[4859]: I1008 18:17:47.644590 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 18:17:47 crc kubenswrapper[4859]: I1008 18:17:47.644604 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 18:17:47 crc kubenswrapper[4859]: I1008 18:17:47.644629 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 18:17:47 crc kubenswrapper[4859]: I1008 18:17:47.644645 4859 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T18:17:47Z","lastTransitionTime":"2025-10-08T18:17:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 18:17:47 crc kubenswrapper[4859]: I1008 18:17:47.710081 4859 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/2e53cd9b-64f7-4e07-8b96-fb77847c9ac6-metrics-certs\") pod \"network-metrics-daemon-4b4sz\" (UID: \"2e53cd9b-64f7-4e07-8b96-fb77847c9ac6\") " pod="openshift-multus/network-metrics-daemon-4b4sz" Oct 08 18:17:47 crc kubenswrapper[4859]: E1008 18:17:47.710362 4859 secret.go:188] Couldn't get secret openshift-multus/metrics-daemon-secret: object "openshift-multus"/"metrics-daemon-secret" not registered Oct 08 18:17:47 crc kubenswrapper[4859]: E1008 18:17:47.710482 4859 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/2e53cd9b-64f7-4e07-8b96-fb77847c9ac6-metrics-certs podName:2e53cd9b-64f7-4e07-8b96-fb77847c9ac6 nodeName:}" failed. No retries permitted until 2025-10-08 18:17:49.710449771 +0000 UTC m=+39.957289190 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/2e53cd9b-64f7-4e07-8b96-fb77847c9ac6-metrics-certs") pod "network-metrics-daemon-4b4sz" (UID: "2e53cd9b-64f7-4e07-8b96-fb77847c9ac6") : object "openshift-multus"/"metrics-daemon-secret" not registered Oct 08 18:17:47 crc kubenswrapper[4859]: I1008 18:17:47.748058 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 18:17:47 crc kubenswrapper[4859]: I1008 18:17:47.748144 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 18:17:47 crc kubenswrapper[4859]: I1008 18:17:47.748161 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 18:17:47 crc kubenswrapper[4859]: I1008 18:17:47.748188 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 18:17:47 crc kubenswrapper[4859]: I1008 18:17:47.748206 4859 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T18:17:47Z","lastTransitionTime":"2025-10-08T18:17:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 18:17:47 crc kubenswrapper[4859]: I1008 18:17:47.851916 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 18:17:47 crc kubenswrapper[4859]: I1008 18:17:47.851971 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 18:17:47 crc kubenswrapper[4859]: I1008 18:17:47.851983 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 18:17:47 crc kubenswrapper[4859]: I1008 18:17:47.852003 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 18:17:47 crc kubenswrapper[4859]: I1008 18:17:47.852019 4859 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T18:17:47Z","lastTransitionTime":"2025-10-08T18:17:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 18:17:47 crc kubenswrapper[4859]: I1008 18:17:47.954625 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 18:17:47 crc kubenswrapper[4859]: I1008 18:17:47.954705 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 18:17:47 crc kubenswrapper[4859]: I1008 18:17:47.954720 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 18:17:47 crc kubenswrapper[4859]: I1008 18:17:47.954739 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 18:17:47 crc kubenswrapper[4859]: I1008 18:17:47.954754 4859 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T18:17:47Z","lastTransitionTime":"2025-10-08T18:17:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 18:17:48 crc kubenswrapper[4859]: I1008 18:17:48.058209 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 18:17:48 crc kubenswrapper[4859]: I1008 18:17:48.058293 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 18:17:48 crc kubenswrapper[4859]: I1008 18:17:48.058315 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 18:17:48 crc kubenswrapper[4859]: I1008 18:17:48.058356 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 18:17:48 crc kubenswrapper[4859]: I1008 18:17:48.058370 4859 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T18:17:48Z","lastTransitionTime":"2025-10-08T18:17:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 18:17:48 crc kubenswrapper[4859]: I1008 18:17:48.165612 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 18:17:48 crc kubenswrapper[4859]: I1008 18:17:48.165752 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 18:17:48 crc kubenswrapper[4859]: I1008 18:17:48.165782 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 18:17:48 crc kubenswrapper[4859]: I1008 18:17:48.165814 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 18:17:48 crc kubenswrapper[4859]: I1008 18:17:48.165842 4859 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T18:17:48Z","lastTransitionTime":"2025-10-08T18:17:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 18:17:48 crc kubenswrapper[4859]: I1008 18:17:48.269369 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 18:17:48 crc kubenswrapper[4859]: I1008 18:17:48.269445 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 18:17:48 crc kubenswrapper[4859]: I1008 18:17:48.269465 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 18:17:48 crc kubenswrapper[4859]: I1008 18:17:48.269494 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 18:17:48 crc kubenswrapper[4859]: I1008 18:17:48.269511 4859 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T18:17:48Z","lastTransitionTime":"2025-10-08T18:17:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 18:17:48 crc kubenswrapper[4859]: I1008 18:17:48.373517 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 18:17:48 crc kubenswrapper[4859]: I1008 18:17:48.373583 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 18:17:48 crc kubenswrapper[4859]: I1008 18:17:48.373599 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 18:17:48 crc kubenswrapper[4859]: I1008 18:17:48.373620 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 18:17:48 crc kubenswrapper[4859]: I1008 18:17:48.373634 4859 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T18:17:48Z","lastTransitionTime":"2025-10-08T18:17:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 18:17:48 crc kubenswrapper[4859]: I1008 18:17:48.476976 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 18:17:48 crc kubenswrapper[4859]: I1008 18:17:48.477050 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 18:17:48 crc kubenswrapper[4859]: I1008 18:17:48.477076 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 18:17:48 crc kubenswrapper[4859]: I1008 18:17:48.477112 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 18:17:48 crc kubenswrapper[4859]: I1008 18:17:48.477139 4859 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T18:17:48Z","lastTransitionTime":"2025-10-08T18:17:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 18:17:48 crc kubenswrapper[4859]: I1008 18:17:48.580353 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 18:17:48 crc kubenswrapper[4859]: I1008 18:17:48.580425 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 18:17:48 crc kubenswrapper[4859]: I1008 18:17:48.580576 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 18:17:48 crc kubenswrapper[4859]: I1008 18:17:48.580674 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 18:17:48 crc kubenswrapper[4859]: I1008 18:17:48.580741 4859 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T18:17:48Z","lastTransitionTime":"2025-10-08T18:17:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 18:17:48 crc kubenswrapper[4859]: I1008 18:17:48.684001 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 18:17:48 crc kubenswrapper[4859]: I1008 18:17:48.684060 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 18:17:48 crc kubenswrapper[4859]: I1008 18:17:48.684070 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 18:17:48 crc kubenswrapper[4859]: I1008 18:17:48.684092 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 18:17:48 crc kubenswrapper[4859]: I1008 18:17:48.684106 4859 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T18:17:48Z","lastTransitionTime":"2025-10-08T18:17:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 18:17:48 crc kubenswrapper[4859]: I1008 18:17:48.787065 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 18:17:48 crc kubenswrapper[4859]: I1008 18:17:48.787209 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 18:17:48 crc kubenswrapper[4859]: I1008 18:17:48.787236 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 18:17:48 crc kubenswrapper[4859]: I1008 18:17:48.787271 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 18:17:48 crc kubenswrapper[4859]: I1008 18:17:48.787301 4859 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T18:17:48Z","lastTransitionTime":"2025-10-08T18:17:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 18:17:48 crc kubenswrapper[4859]: I1008 18:17:48.891611 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 18:17:48 crc kubenswrapper[4859]: I1008 18:17:48.891667 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 18:17:48 crc kubenswrapper[4859]: I1008 18:17:48.891682 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 18:17:48 crc kubenswrapper[4859]: I1008 18:17:48.891729 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 18:17:48 crc kubenswrapper[4859]: I1008 18:17:48.891740 4859 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T18:17:48Z","lastTransitionTime":"2025-10-08T18:17:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 18:17:48 crc kubenswrapper[4859]: I1008 18:17:48.994972 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 18:17:48 crc kubenswrapper[4859]: I1008 18:17:48.995032 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 18:17:48 crc kubenswrapper[4859]: I1008 18:17:48.995048 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 18:17:48 crc kubenswrapper[4859]: I1008 18:17:48.995069 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 18:17:48 crc kubenswrapper[4859]: I1008 18:17:48.995082 4859 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T18:17:48Z","lastTransitionTime":"2025-10-08T18:17:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 18:17:49 crc kubenswrapper[4859]: I1008 18:17:49.098615 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 18:17:49 crc kubenswrapper[4859]: I1008 18:17:49.098674 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 18:17:49 crc kubenswrapper[4859]: I1008 18:17:49.098722 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 18:17:49 crc kubenswrapper[4859]: I1008 18:17:49.098755 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 18:17:49 crc kubenswrapper[4859]: I1008 18:17:49.098774 4859 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T18:17:49Z","lastTransitionTime":"2025-10-08T18:17:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 18:17:49 crc kubenswrapper[4859]: I1008 18:17:49.201958 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 18:17:49 crc kubenswrapper[4859]: I1008 18:17:49.202015 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 18:17:49 crc kubenswrapper[4859]: I1008 18:17:49.202024 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 18:17:49 crc kubenswrapper[4859]: I1008 18:17:49.202041 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 18:17:49 crc kubenswrapper[4859]: I1008 18:17:49.202055 4859 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T18:17:49Z","lastTransitionTime":"2025-10-08T18:17:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 18:17:49 crc kubenswrapper[4859]: I1008 18:17:49.305242 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 18:17:49 crc kubenswrapper[4859]: I1008 18:17:49.305305 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 18:17:49 crc kubenswrapper[4859]: I1008 18:17:49.305315 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 18:17:49 crc kubenswrapper[4859]: I1008 18:17:49.305337 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 18:17:49 crc kubenswrapper[4859]: I1008 18:17:49.305356 4859 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T18:17:49Z","lastTransitionTime":"2025-10-08T18:17:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 18:17:49 crc kubenswrapper[4859]: I1008 18:17:49.410266 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 18:17:49 crc kubenswrapper[4859]: I1008 18:17:49.410339 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 18:17:49 crc kubenswrapper[4859]: I1008 18:17:49.410361 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 18:17:49 crc kubenswrapper[4859]: I1008 18:17:49.410402 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 18:17:49 crc kubenswrapper[4859]: I1008 18:17:49.410423 4859 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T18:17:49Z","lastTransitionTime":"2025-10-08T18:17:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 18:17:49 crc kubenswrapper[4859]: I1008 18:17:49.469659 4859 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-4b4sz" Oct 08 18:17:49 crc kubenswrapper[4859]: I1008 18:17:49.469796 4859 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 08 18:17:49 crc kubenswrapper[4859]: I1008 18:17:49.469661 4859 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 08 18:17:49 crc kubenswrapper[4859]: E1008 18:17:49.469969 4859 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-4b4sz" podUID="2e53cd9b-64f7-4e07-8b96-fb77847c9ac6" Oct 08 18:17:49 crc kubenswrapper[4859]: I1008 18:17:49.469888 4859 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 08 18:17:49 crc kubenswrapper[4859]: E1008 18:17:49.470086 4859 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 08 18:17:49 crc kubenswrapper[4859]: E1008 18:17:49.470256 4859 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 08 18:17:49 crc kubenswrapper[4859]: E1008 18:17:49.470328 4859 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 08 18:17:49 crc kubenswrapper[4859]: I1008 18:17:49.513636 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 18:17:49 crc kubenswrapper[4859]: I1008 18:17:49.514061 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 18:17:49 crc kubenswrapper[4859]: I1008 18:17:49.514199 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 18:17:49 crc kubenswrapper[4859]: I1008 18:17:49.514379 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 18:17:49 crc kubenswrapper[4859]: I1008 18:17:49.514602 4859 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T18:17:49Z","lastTransitionTime":"2025-10-08T18:17:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 18:17:49 crc kubenswrapper[4859]: I1008 18:17:49.616910 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 18:17:49 crc kubenswrapper[4859]: I1008 18:17:49.616964 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 18:17:49 crc kubenswrapper[4859]: I1008 18:17:49.616979 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 18:17:49 crc kubenswrapper[4859]: I1008 18:17:49.617005 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 18:17:49 crc kubenswrapper[4859]: I1008 18:17:49.617022 4859 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T18:17:49Z","lastTransitionTime":"2025-10-08T18:17:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 18:17:49 crc kubenswrapper[4859]: I1008 18:17:49.719397 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 18:17:49 crc kubenswrapper[4859]: I1008 18:17:49.719474 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 18:17:49 crc kubenswrapper[4859]: I1008 18:17:49.719493 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 18:17:49 crc kubenswrapper[4859]: I1008 18:17:49.719522 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 18:17:49 crc kubenswrapper[4859]: I1008 18:17:49.719542 4859 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T18:17:49Z","lastTransitionTime":"2025-10-08T18:17:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 18:17:49 crc kubenswrapper[4859]: I1008 18:17:49.733406 4859 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/2e53cd9b-64f7-4e07-8b96-fb77847c9ac6-metrics-certs\") pod \"network-metrics-daemon-4b4sz\" (UID: \"2e53cd9b-64f7-4e07-8b96-fb77847c9ac6\") " pod="openshift-multus/network-metrics-daemon-4b4sz" Oct 08 18:17:49 crc kubenswrapper[4859]: E1008 18:17:49.733595 4859 secret.go:188] Couldn't get secret openshift-multus/metrics-daemon-secret: object "openshift-multus"/"metrics-daemon-secret" not registered Oct 08 18:17:49 crc kubenswrapper[4859]: E1008 18:17:49.733706 4859 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/2e53cd9b-64f7-4e07-8b96-fb77847c9ac6-metrics-certs podName:2e53cd9b-64f7-4e07-8b96-fb77847c9ac6 nodeName:}" failed. No retries permitted until 2025-10-08 18:17:53.733656532 +0000 UTC m=+43.980495911 (durationBeforeRetry 4s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/2e53cd9b-64f7-4e07-8b96-fb77847c9ac6-metrics-certs") pod "network-metrics-daemon-4b4sz" (UID: "2e53cd9b-64f7-4e07-8b96-fb77847c9ac6") : object "openshift-multus"/"metrics-daemon-secret" not registered Oct 08 18:17:49 crc kubenswrapper[4859]: I1008 18:17:49.821870 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 18:17:49 crc kubenswrapper[4859]: I1008 18:17:49.821912 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 18:17:49 crc kubenswrapper[4859]: I1008 18:17:49.821922 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 18:17:49 crc kubenswrapper[4859]: I1008 18:17:49.821938 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 18:17:49 crc kubenswrapper[4859]: I1008 18:17:49.821951 4859 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T18:17:49Z","lastTransitionTime":"2025-10-08T18:17:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 18:17:49 crc kubenswrapper[4859]: I1008 18:17:49.925441 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 18:17:49 crc kubenswrapper[4859]: I1008 18:17:49.925789 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 18:17:49 crc kubenswrapper[4859]: I1008 18:17:49.925874 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 18:17:49 crc kubenswrapper[4859]: I1008 18:17:49.925968 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 18:17:49 crc kubenswrapper[4859]: I1008 18:17:49.926049 4859 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T18:17:49Z","lastTransitionTime":"2025-10-08T18:17:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 18:17:50 crc kubenswrapper[4859]: I1008 18:17:50.028928 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 18:17:50 crc kubenswrapper[4859]: I1008 18:17:50.028976 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 18:17:50 crc kubenswrapper[4859]: I1008 18:17:50.028989 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 18:17:50 crc kubenswrapper[4859]: I1008 18:17:50.029010 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 18:17:50 crc kubenswrapper[4859]: I1008 18:17:50.029024 4859 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T18:17:50Z","lastTransitionTime":"2025-10-08T18:17:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 18:17:50 crc kubenswrapper[4859]: I1008 18:17:50.132437 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 18:17:50 crc kubenswrapper[4859]: I1008 18:17:50.132767 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 18:17:50 crc kubenswrapper[4859]: I1008 18:17:50.132901 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 18:17:50 crc kubenswrapper[4859]: I1008 18:17:50.133007 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 18:17:50 crc kubenswrapper[4859]: I1008 18:17:50.133096 4859 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T18:17:50Z","lastTransitionTime":"2025-10-08T18:17:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 18:17:50 crc kubenswrapper[4859]: I1008 18:17:50.236957 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 18:17:50 crc kubenswrapper[4859]: I1008 18:17:50.237015 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 18:17:50 crc kubenswrapper[4859]: I1008 18:17:50.237026 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 18:17:50 crc kubenswrapper[4859]: I1008 18:17:50.237050 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 18:17:50 crc kubenswrapper[4859]: I1008 18:17:50.237062 4859 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T18:17:50Z","lastTransitionTime":"2025-10-08T18:17:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 18:17:50 crc kubenswrapper[4859]: I1008 18:17:50.340547 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 18:17:50 crc kubenswrapper[4859]: I1008 18:17:50.340644 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 18:17:50 crc kubenswrapper[4859]: I1008 18:17:50.340668 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 18:17:50 crc kubenswrapper[4859]: I1008 18:17:50.340740 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 18:17:50 crc kubenswrapper[4859]: I1008 18:17:50.340766 4859 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T18:17:50Z","lastTransitionTime":"2025-10-08T18:17:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 18:17:50 crc kubenswrapper[4859]: I1008 18:17:50.445317 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 18:17:50 crc kubenswrapper[4859]: I1008 18:17:50.445960 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 18:17:50 crc kubenswrapper[4859]: I1008 18:17:50.446269 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 18:17:50 crc kubenswrapper[4859]: I1008 18:17:50.446517 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 18:17:50 crc kubenswrapper[4859]: I1008 18:17:50.446721 4859 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T18:17:50Z","lastTransitionTime":"2025-10-08T18:17:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 18:17:50 crc kubenswrapper[4859]: I1008 18:17:50.491216 4859 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"63a623de-5608-4228-b1c2-77188c276986\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:17:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:17:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:17:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:17:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:17:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d4786a8fb8abd96fb22e60733c974f599017407d589ed815fb8541fd3b64f696\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T18:17:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://183d481502ed5108d5e19e483b714c289f355ea3f8b0891883e5335b802c83b6\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T18:17:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ec848c46fd5dde0469b00d3cab14ba4a652533a265490fadbee7a8b8843ef249\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T18:17:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4fe1828b4d63ff8106bfcd15e1dc1c1bd1ba7f77777cdb23ff1b291d0418774c\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2e700b2a20c81b414617f5b5c018a738278a3fa44e6c6d14b2170051d0725431\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-08T18:17:24Z\\\",\\\"message\\\":\\\"W1008 18:17:13.670625 1 cmd.go:257] Using insecure, self-signed certificates\\\\nI1008 18:17:13.671202 1 crypto.go:601] Generating new CA for check-endpoints-signer@1759947433 cert, and key in /tmp/serving-cert-3867218267/serving-signer.crt, /tmp/serving-cert-3867218267/serving-signer.key\\\\nI1008 18:17:13.944463 1 observer_polling.go:159] Starting file observer\\\\nW1008 18:17:13.951958 1 builder.go:272] unable to get owner reference (falling back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": dial tcp [::1]:6443: connect: connection refused\\\\nI1008 18:17:13.952088 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1008 18:17:13.956755 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3867218267/tls.crt::/tmp/serving-cert-3867218267/tls.key\\\\\\\"\\\\nF1008 18:17:24.290991 1 cmd.go:182] error initializing delegating authentication: unable to load configmap based request-header-client-ca-file: Get \\\\\\\"https://localhost:6443/api/v1/namespaces/kube-system/configmaps/extension-apiserver-authentication\\\\\\\": net/http: TLS handshake timeout\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-08T18:17:13Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T18:17:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c65b66f62d325ca10ef8d5d31f04529ccf4e691cf3303d231e7a414fb400e4d6\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T18:17:13Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://58c7bba4ace28215b062653a960e561b1dc1f890f2baf450145638a065af038b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://58c7bba4ace28215b062653a960e561b1dc1f890f2baf450145638a065af038b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-08T18:17:11Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-08T18:17:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T18:17:10Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T18:17:50Z is after 2025-08-24T17:21:41Z" Oct 08 18:17:50 crc kubenswrapper[4859]: I1008 18:17:50.513161 4859 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-08T18:17:31Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T18:17:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T18:17:31Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://99e2d51bd7757498dae60949eea6d732b5f7b481fa00f808dd5308a530a1daf1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T18:17:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e32aaba93699944665660ae41e9b5779178c1354c31e15ad8188bf8c3c8fa500\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T18:17:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T18:17:50Z is after 2025-08-24T17:21:41Z" Oct 08 18:17:50 crc kubenswrapper[4859]: I1008 18:17:50.535293 4859 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-08T18:17:29Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T18:17:29Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T18:17:50Z is after 2025-08-24T17:21:41Z" Oct 08 18:17:50 crc kubenswrapper[4859]: I1008 18:17:50.550572 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 18:17:50 crc kubenswrapper[4859]: I1008 18:17:50.550631 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 18:17:50 crc kubenswrapper[4859]: I1008 18:17:50.550646 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 18:17:50 crc kubenswrapper[4859]: I1008 18:17:50.550671 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 18:17:50 crc kubenswrapper[4859]: I1008 18:17:50.550723 4859 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T18:17:50Z","lastTransitionTime":"2025-10-08T18:17:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 18:17:50 crc kubenswrapper[4859]: I1008 18:17:50.552607 4859 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-82s52" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b23a6a6c-9d92-4e7b-840e-55cfda873a2d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:17:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:17:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:17:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:17:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://71302c766485b9f55743d85f97440ff578cd44ba1e5a6f5b1c024c0e808072d2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T18:17:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6vx22\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c23ea3580998adf90621a0155c20ee4bd590c512629812174e95c28857bcfef9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T18:17:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6vx22\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T18:17:31Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-82s52\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T18:17:50Z is after 2025-08-24T17:21:41Z" Oct 08 18:17:50 crc kubenswrapper[4859]: I1008 18:17:50.578885 4859 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-4g8gf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1dff864e-b75d-4e0b-b182-75f710eac8df\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:17:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:17:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:17:32Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:17:32Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://768ca274dedcc6847fc21d15234d3665fa9ccae371b5b6a882cfcb9b182f527d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T18:17:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s8jnn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://669f5f3c6d3c1ec0c4b4870e090b0a853adc01727c23383e595715b476c32668\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T18:17:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s8jnn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b8009a42cc4c9a816457a0678bdb2f7b7c53d0c8604e007a6fb141f66d06a516\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T18:17:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s8jnn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://676511a11ed4102084b57ebdd5f3fda4969741b2bc8443b415c4bf94f39ae4c9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T18:17:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s8jnn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4a74c1f0d21c90cdd7e776ee14e1f48757a04e2220fbd1f884449181ec6d1c3a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T18:17:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s8jnn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6ddc0e842bb8f81663406070023b0e5ed9cc99c98db0123b1c84f20da59fb5b5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T18:17:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s8jnn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2ce53a49f210c2d6e0001ee4280916611bbd82b4fc21a99c59954ed3d7570b96\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2ce53a49f210c2d6e0001ee4280916611bbd82b4fc21a99c59954ed3d7570b96\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-10-08T18:17:43Z\\\",\\\"message\\\":\\\"cert-secret-name:metrics-tls service.beta.openshift.io/serving-cert-signed-by:openshift-service-serving-signer@1740288168] [{config.openshift.io/v1 ClusterVersion version 9101b518-476b-4eea-8fa6-69b0534e5caa 0xc00695d3d7 \\\\u003cnil\\\\u003e}] [] []},Spec:ServiceSpec{Ports:[]ServicePort{ServicePort{Name:metrics,Protocol:TCP,Port:9393,TargetPort:{1 0 metrics},NodePort:0,AppProtocol:nil,},},Selector:map[string]string{name: ingress-operator,},ClusterIP:10.217.5.244,Type:ClusterIP,ExternalIPs:[],SessionAffinity:None,LoadBalancerIP:,LoadBalancerSourceRanges:[],ExternalName:,ExternalTrafficPolicy:,HealthCheckNodePort:0,PublishNotReadyAddresses:false,SessionAffinityConfig:nil,IPFamilyPolicy:*SingleStack,ClusterIPs:[10.217.5.244],IPFamilies:[IPv4],AllocateLoadBalancerNodePorts:nil,LoadBalancerClass:nil,InternalTrafficPolicy:*Cluster,TrafficDistribution:nil,},Status:ServiceStatus{LoadBalancer:LoadBalancerStatus{Ingress:[]LoadBalancerIngress{},},Conditions:[]Condition{},},}\\\\nI1008 18:17:43.764981 6316 base_network_controller_pods.go:477] [default/openshift-network-diagnostics/network-check-target-xd92c] creating logical port openshift-network-diagnostics_network-check-target-xd92c for pod on switch crc\\\\nF1008 18:17:43.764990 6316 ovnkube.go:137] failed to run ovnkube: [failed to start network controller: failed to start default network controller: unable to create\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-08T18:17:42Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 10s restarting failed container=ovnkube-controller pod=ovnkube-node-4g8gf_openshift-ovn-kubernetes(1dff864e-b75d-4e0b-b182-75f710eac8df)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s8jnn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2c67d2aed7126d76989e057d8e2d8e188a8902b2f3fe26ed6d1d45d3b013d897\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T18:17:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s8jnn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7a211d3da02b173867ff3a912eecd3b379483c7cc77e6b91122edd1ffd30fd53\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7a211d3da02b173867ff3a912eecd3b379483c7cc77e6b91122edd1ffd30fd53\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-08T18:17:32Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-08T18:17:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s8jnn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T18:17:32Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-4g8gf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T18:17:50Z is after 2025-08-24T17:21:41Z" Oct 08 18:17:50 crc kubenswrapper[4859]: I1008 18:17:50.596871 4859 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-4b4sz" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2e53cd9b-64f7-4e07-8b96-fb77847c9ac6\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:17:45Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:17:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:17:45Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:17:45Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kxgbz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kxgbz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T18:17:45Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-4b4sz\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T18:17:50Z is after 2025-08-24T17:21:41Z" Oct 08 18:17:50 crc kubenswrapper[4859]: I1008 18:17:50.619965 4859 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-08T18:17:29Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T18:17:29Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T18:17:50Z is after 2025-08-24T17:21:41Z" Oct 08 18:17:50 crc kubenswrapper[4859]: I1008 18:17:50.640371 4859 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-84k8x" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1c24078c-c9e8-4160-883a-f483b4b3dd23\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:17:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:17:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:17:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:17:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8661006834d0f50eaada7394369584ee95fc61ff54437eec35d755526afc5b48\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T18:17:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hk5xb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://18b291eb3344fb7c00ef5679dca9a6f7ed9d08cea684a803ff7dad2e8a18c585\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://18b291eb3344fb7c00ef5679dca9a6f7ed9d08cea684a803ff7dad2e8a18c585\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-08T18:17:32Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-08T18:17:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hk5xb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://160163f1a91d981583452db3c7a9244182d1c8ced45749a61b721d0d45c1dfe4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://160163f1a91d981583452db3c7a9244182d1c8ced45749a61b721d0d45c1dfe4\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-08T18:17:34Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-08T18:17:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hk5xb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://51e9b53f3847ac53f378ad83870debf8e71518d4edbc7fc0dc66644818e2103e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://51e9b53f3847ac53f378ad83870debf8e71518d4edbc7fc0dc66644818e2103e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-08T18:17:34Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-08T18:17:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hk5xb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1ff118919739759ac01b5cbe3240e6673db54eaa16353c609d670071ed7964cb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1ff118919739759ac01b5cbe3240e6673db54eaa16353c609d670071ed7964cb\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-08T18:17:35Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-08T18:17:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hk5xb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8522e803ff56eb57cd4d9fd036e0c3fb16240ea21d4a094c1db7aa7d50bfe2c2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8522e803ff56eb57cd4d9fd036e0c3fb16240ea21d4a094c1db7aa7d50bfe2c2\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-08T18:17:37Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-08T18:17:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hk5xb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://be61056edf515e037403ca2a0bfd5feb9b2537dd42212b58e133b23b294f327b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://be61056edf515e037403ca2a0bfd5feb9b2537dd42212b58e133b23b294f327b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-08T18:17:38Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-08T18:17:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hk5xb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T18:17:31Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-84k8x\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T18:17:50Z is after 2025-08-24T17:21:41Z" Oct 08 18:17:50 crc kubenswrapper[4859]: I1008 18:17:50.653801 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 18:17:50 crc kubenswrapper[4859]: I1008 18:17:50.653857 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 18:17:50 crc kubenswrapper[4859]: I1008 18:17:50.653868 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 18:17:50 crc kubenswrapper[4859]: I1008 18:17:50.653916 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 18:17:50 crc kubenswrapper[4859]: I1008 18:17:50.653934 4859 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T18:17:50Z","lastTransitionTime":"2025-10-08T18:17:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 18:17:50 crc kubenswrapper[4859]: I1008 18:17:50.655963 4859 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4efc16fb-e679-464d-8b3c-0f46f082593e\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:17:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:17:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:17:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:17:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:17:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://aa8f8ea895272cd79a08d1c47cd7863a1ec10e21bb60d12aa10862a56c99e4bf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T18:17:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5c4e95d8a712ab219455564539a90286ae48bbc46ddc322a5bfd96c1a8268110\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T18:17:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://bae6ecfd8b446ef04d393d823e5ffe37fe749337c0ca17fde79c5f88d5e31393\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T18:17:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6411d1e7e18d5ceadaaceed5a8a6b20a0cacc55cede793ca9fb4dcad70c64fc6\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T18:17:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T18:17:10Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T18:17:50Z is after 2025-08-24T17:21:41Z" Oct 08 18:17:50 crc kubenswrapper[4859]: I1008 18:17:50.673512 4859 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-08T18:17:29Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T18:17:29Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T18:17:50Z is after 2025-08-24T17:21:41Z" Oct 08 18:17:50 crc kubenswrapper[4859]: I1008 18:17:50.691243 4859 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-btw28" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"97bc5598-3ead-4300-a938-3aebbff1e8c6\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:17:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:17:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:17:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:17:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d6d4d6ba771cb5fde637449bba05660f634df7e833a0c8dad2b2c7f3a422bb2b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T18:17:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qcxsd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T18:17:31Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-btw28\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T18:17:50Z is after 2025-08-24T17:21:41Z" Oct 08 18:17:50 crc kubenswrapper[4859]: I1008 18:17:50.711018 4859 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-8plkg" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"92b2d7de-31cc-4a91-95a5-ed5ea964b028\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:17:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:17:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:17:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:17:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fa00b8345d40b74a917ddc23cc1b2c5c684a9ca61b6119d090c92cc1439b71c4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T18:17:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q9d7m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T18:17:31Z\\\"}}\" for pod \"openshift-multus\"/\"multus-8plkg\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T18:17:50Z is after 2025-08-24T17:21:41Z" Oct 08 18:17:50 crc kubenswrapper[4859]: I1008 18:17:50.732242 4859 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-btjmq" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b8605f7-fed7-493f-9b4f-6a6358adf907\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:17:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:17:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:17:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:17:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://14e9b6c4a4403f36adb78fcf1135a620ca6fc56ed974c9639dcdfc59e6917559\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T18:17:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xspg4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T18:17:35Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-btjmq\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T18:17:50Z is after 2025-08-24T17:21:41Z" Oct 08 18:17:50 crc kubenswrapper[4859]: I1008 18:17:50.744352 4859 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-g4t79" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3a92dbf1-21a8-4095-96bb-2d701d7f3d56\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:17:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:17:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:17:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:17:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f8772a25b6e605efe9a062dc32dbbc2cfe32b4fe164b4e84fbf3bcffab744179\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T18:17:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sv79p\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://92cb9986c09a7f9e1e552e3e3ac1f74f04128241e00ddee38882d2ba5c005a67\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T18:17:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sv79p\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T18:17:45Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-g4t79\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T18:17:50Z is after 2025-08-24T17:21:41Z" Oct 08 18:17:50 crc kubenswrapper[4859]: I1008 18:17:50.757418 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 18:17:50 crc kubenswrapper[4859]: I1008 18:17:50.757479 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 18:17:50 crc kubenswrapper[4859]: I1008 18:17:50.757490 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 18:17:50 crc kubenswrapper[4859]: I1008 18:17:50.757509 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 18:17:50 crc kubenswrapper[4859]: I1008 18:17:50.757524 4859 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T18:17:50Z","lastTransitionTime":"2025-10-08T18:17:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 18:17:50 crc kubenswrapper[4859]: I1008 18:17:50.760001 4859 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-08T18:17:33Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T18:17:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T18:17:33Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5d99e071bd083914418d9c128f38955f26295c08b1520f2dc774c99b1c843693\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T18:17:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T18:17:50Z is after 2025-08-24T17:21:41Z" Oct 08 18:17:50 crc kubenswrapper[4859]: I1008 18:17:50.775098 4859 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-08T18:17:31Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T18:17:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T18:17:31Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c6d15905694b6bbd4999d49dc8f7a53eb6bfaa7adb59efb0d7e4973c21d1c7be\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T18:17:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T18:17:50Z is after 2025-08-24T17:21:41Z" Oct 08 18:17:50 crc kubenswrapper[4859]: I1008 18:17:50.860846 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 18:17:50 crc kubenswrapper[4859]: I1008 18:17:50.860917 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 18:17:50 crc kubenswrapper[4859]: I1008 18:17:50.860941 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 18:17:50 crc kubenswrapper[4859]: I1008 18:17:50.860968 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 18:17:50 crc kubenswrapper[4859]: I1008 18:17:50.860988 4859 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T18:17:50Z","lastTransitionTime":"2025-10-08T18:17:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 18:17:50 crc kubenswrapper[4859]: I1008 18:17:50.964537 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 18:17:50 crc kubenswrapper[4859]: I1008 18:17:50.964645 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 18:17:50 crc kubenswrapper[4859]: I1008 18:17:50.964667 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 18:17:50 crc kubenswrapper[4859]: I1008 18:17:50.964758 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 18:17:50 crc kubenswrapper[4859]: I1008 18:17:50.964785 4859 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T18:17:50Z","lastTransitionTime":"2025-10-08T18:17:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 18:17:51 crc kubenswrapper[4859]: I1008 18:17:51.067100 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 18:17:51 crc kubenswrapper[4859]: I1008 18:17:51.067167 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 18:17:51 crc kubenswrapper[4859]: I1008 18:17:51.067185 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 18:17:51 crc kubenswrapper[4859]: I1008 18:17:51.067213 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 18:17:51 crc kubenswrapper[4859]: I1008 18:17:51.067231 4859 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T18:17:51Z","lastTransitionTime":"2025-10-08T18:17:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 18:17:51 crc kubenswrapper[4859]: I1008 18:17:51.170515 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 18:17:51 crc kubenswrapper[4859]: I1008 18:17:51.170586 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 18:17:51 crc kubenswrapper[4859]: I1008 18:17:51.170604 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 18:17:51 crc kubenswrapper[4859]: I1008 18:17:51.170628 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 18:17:51 crc kubenswrapper[4859]: I1008 18:17:51.170648 4859 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T18:17:51Z","lastTransitionTime":"2025-10-08T18:17:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 18:17:51 crc kubenswrapper[4859]: I1008 18:17:51.273622 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 18:17:51 crc kubenswrapper[4859]: I1008 18:17:51.273717 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 18:17:51 crc kubenswrapper[4859]: I1008 18:17:51.273740 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 18:17:51 crc kubenswrapper[4859]: I1008 18:17:51.273770 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 18:17:51 crc kubenswrapper[4859]: I1008 18:17:51.273789 4859 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T18:17:51Z","lastTransitionTime":"2025-10-08T18:17:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 18:17:51 crc kubenswrapper[4859]: I1008 18:17:51.377294 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 18:17:51 crc kubenswrapper[4859]: I1008 18:17:51.377359 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 18:17:51 crc kubenswrapper[4859]: I1008 18:17:51.377383 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 18:17:51 crc kubenswrapper[4859]: I1008 18:17:51.377417 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 18:17:51 crc kubenswrapper[4859]: I1008 18:17:51.377439 4859 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T18:17:51Z","lastTransitionTime":"2025-10-08T18:17:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 18:17:51 crc kubenswrapper[4859]: I1008 18:17:51.469471 4859 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 08 18:17:51 crc kubenswrapper[4859]: I1008 18:17:51.469521 4859 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 08 18:17:51 crc kubenswrapper[4859]: I1008 18:17:51.469521 4859 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 08 18:17:51 crc kubenswrapper[4859]: I1008 18:17:51.469486 4859 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-4b4sz" Oct 08 18:17:51 crc kubenswrapper[4859]: E1008 18:17:51.469738 4859 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 08 18:17:51 crc kubenswrapper[4859]: E1008 18:17:51.469823 4859 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 08 18:17:51 crc kubenswrapper[4859]: E1008 18:17:51.469983 4859 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 08 18:17:51 crc kubenswrapper[4859]: E1008 18:17:51.470077 4859 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-4b4sz" podUID="2e53cd9b-64f7-4e07-8b96-fb77847c9ac6" Oct 08 18:17:51 crc kubenswrapper[4859]: I1008 18:17:51.480314 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 18:17:51 crc kubenswrapper[4859]: I1008 18:17:51.480379 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 18:17:51 crc kubenswrapper[4859]: I1008 18:17:51.480400 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 18:17:51 crc kubenswrapper[4859]: I1008 18:17:51.480428 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 18:17:51 crc kubenswrapper[4859]: I1008 18:17:51.480451 4859 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T18:17:51Z","lastTransitionTime":"2025-10-08T18:17:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 18:17:51 crc kubenswrapper[4859]: I1008 18:17:51.584465 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 18:17:51 crc kubenswrapper[4859]: I1008 18:17:51.584509 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 18:17:51 crc kubenswrapper[4859]: I1008 18:17:51.584518 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 18:17:51 crc kubenswrapper[4859]: I1008 18:17:51.584534 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 18:17:51 crc kubenswrapper[4859]: I1008 18:17:51.584546 4859 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T18:17:51Z","lastTransitionTime":"2025-10-08T18:17:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 18:17:51 crc kubenswrapper[4859]: I1008 18:17:51.687777 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 18:17:51 crc kubenswrapper[4859]: I1008 18:17:51.687850 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 18:17:51 crc kubenswrapper[4859]: I1008 18:17:51.687864 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 18:17:51 crc kubenswrapper[4859]: I1008 18:17:51.687888 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 18:17:51 crc kubenswrapper[4859]: I1008 18:17:51.687904 4859 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T18:17:51Z","lastTransitionTime":"2025-10-08T18:17:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 18:17:51 crc kubenswrapper[4859]: I1008 18:17:51.791807 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 18:17:51 crc kubenswrapper[4859]: I1008 18:17:51.791870 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 18:17:51 crc kubenswrapper[4859]: I1008 18:17:51.791884 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 18:17:51 crc kubenswrapper[4859]: I1008 18:17:51.791909 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 18:17:51 crc kubenswrapper[4859]: I1008 18:17:51.791926 4859 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T18:17:51Z","lastTransitionTime":"2025-10-08T18:17:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 18:17:51 crc kubenswrapper[4859]: I1008 18:17:51.895291 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 18:17:51 crc kubenswrapper[4859]: I1008 18:17:51.895381 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 18:17:51 crc kubenswrapper[4859]: I1008 18:17:51.895400 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 18:17:51 crc kubenswrapper[4859]: I1008 18:17:51.895427 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 18:17:51 crc kubenswrapper[4859]: I1008 18:17:51.895446 4859 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T18:17:51Z","lastTransitionTime":"2025-10-08T18:17:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 18:17:51 crc kubenswrapper[4859]: I1008 18:17:51.998898 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 18:17:51 crc kubenswrapper[4859]: I1008 18:17:51.998945 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 18:17:51 crc kubenswrapper[4859]: I1008 18:17:51.998954 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 18:17:51 crc kubenswrapper[4859]: I1008 18:17:51.998972 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 18:17:51 crc kubenswrapper[4859]: I1008 18:17:51.998986 4859 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T18:17:51Z","lastTransitionTime":"2025-10-08T18:17:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 18:17:52 crc kubenswrapper[4859]: I1008 18:17:52.101673 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 18:17:52 crc kubenswrapper[4859]: I1008 18:17:52.101808 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 18:17:52 crc kubenswrapper[4859]: I1008 18:17:52.101827 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 18:17:52 crc kubenswrapper[4859]: I1008 18:17:52.101854 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 18:17:52 crc kubenswrapper[4859]: I1008 18:17:52.101874 4859 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T18:17:52Z","lastTransitionTime":"2025-10-08T18:17:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 18:17:52 crc kubenswrapper[4859]: I1008 18:17:52.205488 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 18:17:52 crc kubenswrapper[4859]: I1008 18:17:52.205979 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 18:17:52 crc kubenswrapper[4859]: I1008 18:17:52.206151 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 18:17:52 crc kubenswrapper[4859]: I1008 18:17:52.206354 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 18:17:52 crc kubenswrapper[4859]: I1008 18:17:52.206526 4859 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T18:17:52Z","lastTransitionTime":"2025-10-08T18:17:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 18:17:52 crc kubenswrapper[4859]: I1008 18:17:52.310203 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 18:17:52 crc kubenswrapper[4859]: I1008 18:17:52.310823 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 18:17:52 crc kubenswrapper[4859]: I1008 18:17:52.310869 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 18:17:52 crc kubenswrapper[4859]: I1008 18:17:52.310902 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 18:17:52 crc kubenswrapper[4859]: I1008 18:17:52.310923 4859 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T18:17:52Z","lastTransitionTime":"2025-10-08T18:17:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 18:17:52 crc kubenswrapper[4859]: I1008 18:17:52.414592 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 18:17:52 crc kubenswrapper[4859]: I1008 18:17:52.414680 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 18:17:52 crc kubenswrapper[4859]: I1008 18:17:52.414731 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 18:17:52 crc kubenswrapper[4859]: I1008 18:17:52.414759 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 18:17:52 crc kubenswrapper[4859]: I1008 18:17:52.414778 4859 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T18:17:52Z","lastTransitionTime":"2025-10-08T18:17:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 18:17:52 crc kubenswrapper[4859]: I1008 18:17:52.518754 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 18:17:52 crc kubenswrapper[4859]: I1008 18:17:52.518821 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 18:17:52 crc kubenswrapper[4859]: I1008 18:17:52.518840 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 18:17:52 crc kubenswrapper[4859]: I1008 18:17:52.518873 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 18:17:52 crc kubenswrapper[4859]: I1008 18:17:52.518894 4859 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T18:17:52Z","lastTransitionTime":"2025-10-08T18:17:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 18:17:52 crc kubenswrapper[4859]: I1008 18:17:52.622339 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 18:17:52 crc kubenswrapper[4859]: I1008 18:17:52.622393 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 18:17:52 crc kubenswrapper[4859]: I1008 18:17:52.622412 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 18:17:52 crc kubenswrapper[4859]: I1008 18:17:52.622439 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 18:17:52 crc kubenswrapper[4859]: I1008 18:17:52.622456 4859 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T18:17:52Z","lastTransitionTime":"2025-10-08T18:17:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 18:17:52 crc kubenswrapper[4859]: I1008 18:17:52.726490 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 18:17:52 crc kubenswrapper[4859]: I1008 18:17:52.726594 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 18:17:52 crc kubenswrapper[4859]: I1008 18:17:52.726615 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 18:17:52 crc kubenswrapper[4859]: I1008 18:17:52.726673 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 18:17:52 crc kubenswrapper[4859]: I1008 18:17:52.726752 4859 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T18:17:52Z","lastTransitionTime":"2025-10-08T18:17:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 18:17:52 crc kubenswrapper[4859]: I1008 18:17:52.831514 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 18:17:52 crc kubenswrapper[4859]: I1008 18:17:52.831615 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 18:17:52 crc kubenswrapper[4859]: I1008 18:17:52.831651 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 18:17:52 crc kubenswrapper[4859]: I1008 18:17:52.831728 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 18:17:52 crc kubenswrapper[4859]: I1008 18:17:52.831747 4859 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T18:17:52Z","lastTransitionTime":"2025-10-08T18:17:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 18:17:52 crc kubenswrapper[4859]: I1008 18:17:52.934506 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 18:17:52 crc kubenswrapper[4859]: I1008 18:17:52.934550 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 18:17:52 crc kubenswrapper[4859]: I1008 18:17:52.934561 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 18:17:52 crc kubenswrapper[4859]: I1008 18:17:52.934577 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 18:17:52 crc kubenswrapper[4859]: I1008 18:17:52.934589 4859 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T18:17:52Z","lastTransitionTime":"2025-10-08T18:17:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 18:17:53 crc kubenswrapper[4859]: I1008 18:17:53.038359 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 18:17:53 crc kubenswrapper[4859]: I1008 18:17:53.038425 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 18:17:53 crc kubenswrapper[4859]: I1008 18:17:53.038442 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 18:17:53 crc kubenswrapper[4859]: I1008 18:17:53.038462 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 18:17:53 crc kubenswrapper[4859]: I1008 18:17:53.038474 4859 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T18:17:53Z","lastTransitionTime":"2025-10-08T18:17:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 18:17:53 crc kubenswrapper[4859]: I1008 18:17:53.142221 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 18:17:53 crc kubenswrapper[4859]: I1008 18:17:53.142314 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 18:17:53 crc kubenswrapper[4859]: I1008 18:17:53.142340 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 18:17:53 crc kubenswrapper[4859]: I1008 18:17:53.142368 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 18:17:53 crc kubenswrapper[4859]: I1008 18:17:53.142392 4859 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T18:17:53Z","lastTransitionTime":"2025-10-08T18:17:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 18:17:53 crc kubenswrapper[4859]: I1008 18:17:53.245887 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 18:17:53 crc kubenswrapper[4859]: I1008 18:17:53.245969 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 18:17:53 crc kubenswrapper[4859]: I1008 18:17:53.245987 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 18:17:53 crc kubenswrapper[4859]: I1008 18:17:53.246017 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 18:17:53 crc kubenswrapper[4859]: I1008 18:17:53.246036 4859 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T18:17:53Z","lastTransitionTime":"2025-10-08T18:17:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 18:17:53 crc kubenswrapper[4859]: I1008 18:17:53.349403 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 18:17:53 crc kubenswrapper[4859]: I1008 18:17:53.349469 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 18:17:53 crc kubenswrapper[4859]: I1008 18:17:53.349488 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 18:17:53 crc kubenswrapper[4859]: I1008 18:17:53.349519 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 18:17:53 crc kubenswrapper[4859]: I1008 18:17:53.349539 4859 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T18:17:53Z","lastTransitionTime":"2025-10-08T18:17:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 18:17:53 crc kubenswrapper[4859]: I1008 18:17:53.452209 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 18:17:53 crc kubenswrapper[4859]: I1008 18:17:53.452296 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 18:17:53 crc kubenswrapper[4859]: I1008 18:17:53.452324 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 18:17:53 crc kubenswrapper[4859]: I1008 18:17:53.452355 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 18:17:53 crc kubenswrapper[4859]: I1008 18:17:53.452374 4859 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T18:17:53Z","lastTransitionTime":"2025-10-08T18:17:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 18:17:53 crc kubenswrapper[4859]: I1008 18:17:53.469645 4859 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 08 18:17:53 crc kubenswrapper[4859]: I1008 18:17:53.469662 4859 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 08 18:17:53 crc kubenswrapper[4859]: I1008 18:17:53.469821 4859 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 08 18:17:53 crc kubenswrapper[4859]: I1008 18:17:53.469916 4859 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-4b4sz" Oct 08 18:17:53 crc kubenswrapper[4859]: E1008 18:17:53.470104 4859 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 08 18:17:53 crc kubenswrapper[4859]: E1008 18:17:53.470366 4859 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 08 18:17:53 crc kubenswrapper[4859]: E1008 18:17:53.470513 4859 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 08 18:17:53 crc kubenswrapper[4859]: E1008 18:17:53.470655 4859 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-4b4sz" podUID="2e53cd9b-64f7-4e07-8b96-fb77847c9ac6" Oct 08 18:17:53 crc kubenswrapper[4859]: I1008 18:17:53.555720 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 18:17:53 crc kubenswrapper[4859]: I1008 18:17:53.555795 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 18:17:53 crc kubenswrapper[4859]: I1008 18:17:53.555812 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 18:17:53 crc kubenswrapper[4859]: I1008 18:17:53.555834 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 18:17:53 crc kubenswrapper[4859]: I1008 18:17:53.555850 4859 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T18:17:53Z","lastTransitionTime":"2025-10-08T18:17:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 18:17:53 crc kubenswrapper[4859]: I1008 18:17:53.659983 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 18:17:53 crc kubenswrapper[4859]: I1008 18:17:53.660046 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 18:17:53 crc kubenswrapper[4859]: I1008 18:17:53.660057 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 18:17:53 crc kubenswrapper[4859]: I1008 18:17:53.660078 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 18:17:53 crc kubenswrapper[4859]: I1008 18:17:53.660092 4859 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T18:17:53Z","lastTransitionTime":"2025-10-08T18:17:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 18:17:53 crc kubenswrapper[4859]: I1008 18:17:53.763513 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 18:17:53 crc kubenswrapper[4859]: I1008 18:17:53.763629 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 18:17:53 crc kubenswrapper[4859]: I1008 18:17:53.763648 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 18:17:53 crc kubenswrapper[4859]: I1008 18:17:53.763674 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 18:17:53 crc kubenswrapper[4859]: I1008 18:17:53.763738 4859 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T18:17:53Z","lastTransitionTime":"2025-10-08T18:17:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 18:17:53 crc kubenswrapper[4859]: I1008 18:17:53.779731 4859 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/2e53cd9b-64f7-4e07-8b96-fb77847c9ac6-metrics-certs\") pod \"network-metrics-daemon-4b4sz\" (UID: \"2e53cd9b-64f7-4e07-8b96-fb77847c9ac6\") " pod="openshift-multus/network-metrics-daemon-4b4sz" Oct 08 18:17:53 crc kubenswrapper[4859]: E1008 18:17:53.780046 4859 secret.go:188] Couldn't get secret openshift-multus/metrics-daemon-secret: object "openshift-multus"/"metrics-daemon-secret" not registered Oct 08 18:17:53 crc kubenswrapper[4859]: E1008 18:17:53.780207 4859 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/2e53cd9b-64f7-4e07-8b96-fb77847c9ac6-metrics-certs podName:2e53cd9b-64f7-4e07-8b96-fb77847c9ac6 nodeName:}" failed. No retries permitted until 2025-10-08 18:18:01.780169477 +0000 UTC m=+52.027008896 (durationBeforeRetry 8s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/2e53cd9b-64f7-4e07-8b96-fb77847c9ac6-metrics-certs") pod "network-metrics-daemon-4b4sz" (UID: "2e53cd9b-64f7-4e07-8b96-fb77847c9ac6") : object "openshift-multus"/"metrics-daemon-secret" not registered Oct 08 18:17:53 crc kubenswrapper[4859]: I1008 18:17:53.867717 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 18:17:53 crc kubenswrapper[4859]: I1008 18:17:53.867788 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 18:17:53 crc kubenswrapper[4859]: I1008 18:17:53.867810 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 18:17:53 crc kubenswrapper[4859]: I1008 18:17:53.867838 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 18:17:53 crc kubenswrapper[4859]: I1008 18:17:53.867859 4859 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T18:17:53Z","lastTransitionTime":"2025-10-08T18:17:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 18:17:53 crc kubenswrapper[4859]: I1008 18:17:53.970784 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 18:17:53 crc kubenswrapper[4859]: I1008 18:17:53.970828 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 18:17:53 crc kubenswrapper[4859]: I1008 18:17:53.970840 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 18:17:53 crc kubenswrapper[4859]: I1008 18:17:53.970860 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 18:17:53 crc kubenswrapper[4859]: I1008 18:17:53.970872 4859 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T18:17:53Z","lastTransitionTime":"2025-10-08T18:17:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 18:17:54 crc kubenswrapper[4859]: I1008 18:17:54.074104 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 18:17:54 crc kubenswrapper[4859]: I1008 18:17:54.074175 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 18:17:54 crc kubenswrapper[4859]: I1008 18:17:54.074187 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 18:17:54 crc kubenswrapper[4859]: I1008 18:17:54.074276 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 18:17:54 crc kubenswrapper[4859]: I1008 18:17:54.074294 4859 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T18:17:54Z","lastTransitionTime":"2025-10-08T18:17:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 18:17:54 crc kubenswrapper[4859]: I1008 18:17:54.176641 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 18:17:54 crc kubenswrapper[4859]: I1008 18:17:54.176704 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 18:17:54 crc kubenswrapper[4859]: I1008 18:17:54.176713 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 18:17:54 crc kubenswrapper[4859]: I1008 18:17:54.176729 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 18:17:54 crc kubenswrapper[4859]: I1008 18:17:54.176740 4859 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T18:17:54Z","lastTransitionTime":"2025-10-08T18:17:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 18:17:54 crc kubenswrapper[4859]: I1008 18:17:54.280157 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 18:17:54 crc kubenswrapper[4859]: I1008 18:17:54.280204 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 18:17:54 crc kubenswrapper[4859]: I1008 18:17:54.280225 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 18:17:54 crc kubenswrapper[4859]: I1008 18:17:54.280246 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 18:17:54 crc kubenswrapper[4859]: I1008 18:17:54.280257 4859 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T18:17:54Z","lastTransitionTime":"2025-10-08T18:17:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 18:17:54 crc kubenswrapper[4859]: I1008 18:17:54.384092 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 18:17:54 crc kubenswrapper[4859]: I1008 18:17:54.384154 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 18:17:54 crc kubenswrapper[4859]: I1008 18:17:54.384167 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 18:17:54 crc kubenswrapper[4859]: I1008 18:17:54.384188 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 18:17:54 crc kubenswrapper[4859]: I1008 18:17:54.384201 4859 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T18:17:54Z","lastTransitionTime":"2025-10-08T18:17:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 18:17:54 crc kubenswrapper[4859]: I1008 18:17:54.487096 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 18:17:54 crc kubenswrapper[4859]: I1008 18:17:54.487175 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 18:17:54 crc kubenswrapper[4859]: I1008 18:17:54.487194 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 18:17:54 crc kubenswrapper[4859]: I1008 18:17:54.487227 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 18:17:54 crc kubenswrapper[4859]: I1008 18:17:54.487246 4859 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T18:17:54Z","lastTransitionTime":"2025-10-08T18:17:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 18:17:54 crc kubenswrapper[4859]: I1008 18:17:54.590021 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 18:17:54 crc kubenswrapper[4859]: I1008 18:17:54.590083 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 18:17:54 crc kubenswrapper[4859]: I1008 18:17:54.590097 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 18:17:54 crc kubenswrapper[4859]: I1008 18:17:54.590119 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 18:17:54 crc kubenswrapper[4859]: I1008 18:17:54.590135 4859 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T18:17:54Z","lastTransitionTime":"2025-10-08T18:17:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 18:17:54 crc kubenswrapper[4859]: I1008 18:17:54.692995 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 18:17:54 crc kubenswrapper[4859]: I1008 18:17:54.693071 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 18:17:54 crc kubenswrapper[4859]: I1008 18:17:54.693096 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 18:17:54 crc kubenswrapper[4859]: I1008 18:17:54.693123 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 18:17:54 crc kubenswrapper[4859]: I1008 18:17:54.693144 4859 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T18:17:54Z","lastTransitionTime":"2025-10-08T18:17:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 18:17:54 crc kubenswrapper[4859]: I1008 18:17:54.796398 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 18:17:54 crc kubenswrapper[4859]: I1008 18:17:54.796547 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 18:17:54 crc kubenswrapper[4859]: I1008 18:17:54.796577 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 18:17:54 crc kubenswrapper[4859]: I1008 18:17:54.796610 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 18:17:54 crc kubenswrapper[4859]: I1008 18:17:54.796631 4859 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T18:17:54Z","lastTransitionTime":"2025-10-08T18:17:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 18:17:54 crc kubenswrapper[4859]: I1008 18:17:54.899992 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 18:17:54 crc kubenswrapper[4859]: I1008 18:17:54.900047 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 18:17:54 crc kubenswrapper[4859]: I1008 18:17:54.900058 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 18:17:54 crc kubenswrapper[4859]: I1008 18:17:54.900075 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 18:17:54 crc kubenswrapper[4859]: I1008 18:17:54.900086 4859 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T18:17:54Z","lastTransitionTime":"2025-10-08T18:17:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 18:17:55 crc kubenswrapper[4859]: I1008 18:17:55.003911 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 18:17:55 crc kubenswrapper[4859]: I1008 18:17:55.003991 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 18:17:55 crc kubenswrapper[4859]: I1008 18:17:55.004016 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 18:17:55 crc kubenswrapper[4859]: I1008 18:17:55.004052 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 18:17:55 crc kubenswrapper[4859]: I1008 18:17:55.004076 4859 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T18:17:55Z","lastTransitionTime":"2025-10-08T18:17:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 18:17:55 crc kubenswrapper[4859]: I1008 18:17:55.107362 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 18:17:55 crc kubenswrapper[4859]: I1008 18:17:55.107430 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 18:17:55 crc kubenswrapper[4859]: I1008 18:17:55.107459 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 18:17:55 crc kubenswrapper[4859]: I1008 18:17:55.107491 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 18:17:55 crc kubenswrapper[4859]: I1008 18:17:55.107517 4859 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T18:17:55Z","lastTransitionTime":"2025-10-08T18:17:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 18:17:55 crc kubenswrapper[4859]: I1008 18:17:55.211514 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 18:17:55 crc kubenswrapper[4859]: I1008 18:17:55.211612 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 18:17:55 crc kubenswrapper[4859]: I1008 18:17:55.211642 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 18:17:55 crc kubenswrapper[4859]: I1008 18:17:55.211673 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 18:17:55 crc kubenswrapper[4859]: I1008 18:17:55.211736 4859 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T18:17:55Z","lastTransitionTime":"2025-10-08T18:17:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 18:17:55 crc kubenswrapper[4859]: I1008 18:17:55.315600 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 18:17:55 crc kubenswrapper[4859]: I1008 18:17:55.315729 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 18:17:55 crc kubenswrapper[4859]: I1008 18:17:55.315756 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 18:17:55 crc kubenswrapper[4859]: I1008 18:17:55.315791 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 18:17:55 crc kubenswrapper[4859]: I1008 18:17:55.315820 4859 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T18:17:55Z","lastTransitionTime":"2025-10-08T18:17:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 18:17:55 crc kubenswrapper[4859]: I1008 18:17:55.418440 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 18:17:55 crc kubenswrapper[4859]: I1008 18:17:55.418504 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 18:17:55 crc kubenswrapper[4859]: I1008 18:17:55.418514 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 18:17:55 crc kubenswrapper[4859]: I1008 18:17:55.418530 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 18:17:55 crc kubenswrapper[4859]: I1008 18:17:55.418539 4859 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T18:17:55Z","lastTransitionTime":"2025-10-08T18:17:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 18:17:55 crc kubenswrapper[4859]: I1008 18:17:55.469883 4859 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-4b4sz" Oct 08 18:17:55 crc kubenswrapper[4859]: I1008 18:17:55.469981 4859 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 08 18:17:55 crc kubenswrapper[4859]: E1008 18:17:55.470098 4859 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-4b4sz" podUID="2e53cd9b-64f7-4e07-8b96-fb77847c9ac6" Oct 08 18:17:55 crc kubenswrapper[4859]: I1008 18:17:55.469883 4859 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 08 18:17:55 crc kubenswrapper[4859]: I1008 18:17:55.469905 4859 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 08 18:17:55 crc kubenswrapper[4859]: E1008 18:17:55.470198 4859 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 08 18:17:55 crc kubenswrapper[4859]: E1008 18:17:55.470421 4859 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 08 18:17:55 crc kubenswrapper[4859]: E1008 18:17:55.470601 4859 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 08 18:17:55 crc kubenswrapper[4859]: I1008 18:17:55.522127 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 18:17:55 crc kubenswrapper[4859]: I1008 18:17:55.522190 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 18:17:55 crc kubenswrapper[4859]: I1008 18:17:55.522207 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 18:17:55 crc kubenswrapper[4859]: I1008 18:17:55.522237 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 18:17:55 crc kubenswrapper[4859]: I1008 18:17:55.522255 4859 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T18:17:55Z","lastTransitionTime":"2025-10-08T18:17:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 18:17:55 crc kubenswrapper[4859]: I1008 18:17:55.625832 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 18:17:55 crc kubenswrapper[4859]: I1008 18:17:55.625890 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 18:17:55 crc kubenswrapper[4859]: I1008 18:17:55.625906 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 18:17:55 crc kubenswrapper[4859]: I1008 18:17:55.625934 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 18:17:55 crc kubenswrapper[4859]: I1008 18:17:55.625949 4859 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T18:17:55Z","lastTransitionTime":"2025-10-08T18:17:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 18:17:55 crc kubenswrapper[4859]: I1008 18:17:55.730325 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 18:17:55 crc kubenswrapper[4859]: I1008 18:17:55.730389 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 18:17:55 crc kubenswrapper[4859]: I1008 18:17:55.730412 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 18:17:55 crc kubenswrapper[4859]: I1008 18:17:55.730446 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 18:17:55 crc kubenswrapper[4859]: I1008 18:17:55.730472 4859 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T18:17:55Z","lastTransitionTime":"2025-10-08T18:17:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 18:17:55 crc kubenswrapper[4859]: I1008 18:17:55.833943 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 18:17:55 crc kubenswrapper[4859]: I1008 18:17:55.834000 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 18:17:55 crc kubenswrapper[4859]: I1008 18:17:55.834009 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 18:17:55 crc kubenswrapper[4859]: I1008 18:17:55.834032 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 18:17:55 crc kubenswrapper[4859]: I1008 18:17:55.834048 4859 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T18:17:55Z","lastTransitionTime":"2025-10-08T18:17:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 18:17:55 crc kubenswrapper[4859]: I1008 18:17:55.937168 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 18:17:55 crc kubenswrapper[4859]: I1008 18:17:55.937267 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 18:17:55 crc kubenswrapper[4859]: I1008 18:17:55.937292 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 18:17:55 crc kubenswrapper[4859]: I1008 18:17:55.937330 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 18:17:55 crc kubenswrapper[4859]: I1008 18:17:55.937355 4859 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T18:17:55Z","lastTransitionTime":"2025-10-08T18:17:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 18:17:56 crc kubenswrapper[4859]: I1008 18:17:56.041342 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 18:17:56 crc kubenswrapper[4859]: I1008 18:17:56.041392 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 18:17:56 crc kubenswrapper[4859]: I1008 18:17:56.041404 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 18:17:56 crc kubenswrapper[4859]: I1008 18:17:56.041423 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 18:17:56 crc kubenswrapper[4859]: I1008 18:17:56.041436 4859 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T18:17:56Z","lastTransitionTime":"2025-10-08T18:17:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 18:17:56 crc kubenswrapper[4859]: I1008 18:17:56.145635 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 18:17:56 crc kubenswrapper[4859]: I1008 18:17:56.145730 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 18:17:56 crc kubenswrapper[4859]: I1008 18:17:56.145741 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 18:17:56 crc kubenswrapper[4859]: I1008 18:17:56.145760 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 18:17:56 crc kubenswrapper[4859]: I1008 18:17:56.145771 4859 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T18:17:56Z","lastTransitionTime":"2025-10-08T18:17:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 18:17:56 crc kubenswrapper[4859]: I1008 18:17:56.248740 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 18:17:56 crc kubenswrapper[4859]: I1008 18:17:56.248802 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 18:17:56 crc kubenswrapper[4859]: I1008 18:17:56.248823 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 18:17:56 crc kubenswrapper[4859]: I1008 18:17:56.248848 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 18:17:56 crc kubenswrapper[4859]: I1008 18:17:56.248863 4859 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T18:17:56Z","lastTransitionTime":"2025-10-08T18:17:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 18:17:56 crc kubenswrapper[4859]: I1008 18:17:56.352231 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 18:17:56 crc kubenswrapper[4859]: I1008 18:17:56.352290 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 18:17:56 crc kubenswrapper[4859]: I1008 18:17:56.352301 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 18:17:56 crc kubenswrapper[4859]: I1008 18:17:56.352319 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 18:17:56 crc kubenswrapper[4859]: I1008 18:17:56.352336 4859 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T18:17:56Z","lastTransitionTime":"2025-10-08T18:17:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 18:17:56 crc kubenswrapper[4859]: I1008 18:17:56.455428 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 18:17:56 crc kubenswrapper[4859]: I1008 18:17:56.455518 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 18:17:56 crc kubenswrapper[4859]: I1008 18:17:56.455543 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 18:17:56 crc kubenswrapper[4859]: I1008 18:17:56.455578 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 18:17:56 crc kubenswrapper[4859]: I1008 18:17:56.455599 4859 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T18:17:56Z","lastTransitionTime":"2025-10-08T18:17:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 18:17:56 crc kubenswrapper[4859]: I1008 18:17:56.559259 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 18:17:56 crc kubenswrapper[4859]: I1008 18:17:56.559344 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 18:17:56 crc kubenswrapper[4859]: I1008 18:17:56.559365 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 18:17:56 crc kubenswrapper[4859]: I1008 18:17:56.559396 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 18:17:56 crc kubenswrapper[4859]: I1008 18:17:56.559413 4859 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T18:17:56Z","lastTransitionTime":"2025-10-08T18:17:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 18:17:56 crc kubenswrapper[4859]: I1008 18:17:56.661818 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 18:17:56 crc kubenswrapper[4859]: I1008 18:17:56.661865 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 18:17:56 crc kubenswrapper[4859]: I1008 18:17:56.661877 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 18:17:56 crc kubenswrapper[4859]: I1008 18:17:56.661895 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 18:17:56 crc kubenswrapper[4859]: I1008 18:17:56.661906 4859 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T18:17:56Z","lastTransitionTime":"2025-10-08T18:17:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 18:17:56 crc kubenswrapper[4859]: I1008 18:17:56.765227 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 18:17:56 crc kubenswrapper[4859]: I1008 18:17:56.765269 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 18:17:56 crc kubenswrapper[4859]: I1008 18:17:56.765283 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 18:17:56 crc kubenswrapper[4859]: I1008 18:17:56.765305 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 18:17:56 crc kubenswrapper[4859]: I1008 18:17:56.765318 4859 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T18:17:56Z","lastTransitionTime":"2025-10-08T18:17:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 18:17:56 crc kubenswrapper[4859]: I1008 18:17:56.868604 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 18:17:56 crc kubenswrapper[4859]: I1008 18:17:56.868672 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 18:17:56 crc kubenswrapper[4859]: I1008 18:17:56.868710 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 18:17:56 crc kubenswrapper[4859]: I1008 18:17:56.868734 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 18:17:56 crc kubenswrapper[4859]: I1008 18:17:56.868750 4859 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T18:17:56Z","lastTransitionTime":"2025-10-08T18:17:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 18:17:56 crc kubenswrapper[4859]: I1008 18:17:56.971667 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 18:17:56 crc kubenswrapper[4859]: I1008 18:17:56.971786 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 18:17:56 crc kubenswrapper[4859]: I1008 18:17:56.971814 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 18:17:56 crc kubenswrapper[4859]: I1008 18:17:56.971848 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 18:17:56 crc kubenswrapper[4859]: I1008 18:17:56.971872 4859 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T18:17:56Z","lastTransitionTime":"2025-10-08T18:17:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 18:17:57 crc kubenswrapper[4859]: I1008 18:17:57.075570 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 18:17:57 crc kubenswrapper[4859]: I1008 18:17:57.075644 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 18:17:57 crc kubenswrapper[4859]: I1008 18:17:57.075659 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 18:17:57 crc kubenswrapper[4859]: I1008 18:17:57.075706 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 18:17:57 crc kubenswrapper[4859]: I1008 18:17:57.075729 4859 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T18:17:57Z","lastTransitionTime":"2025-10-08T18:17:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 18:17:57 crc kubenswrapper[4859]: I1008 18:17:57.179460 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 18:17:57 crc kubenswrapper[4859]: I1008 18:17:57.179981 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 18:17:57 crc kubenswrapper[4859]: I1008 18:17:57.180184 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 18:17:57 crc kubenswrapper[4859]: I1008 18:17:57.180343 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 18:17:57 crc kubenswrapper[4859]: I1008 18:17:57.180514 4859 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T18:17:57Z","lastTransitionTime":"2025-10-08T18:17:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 18:17:57 crc kubenswrapper[4859]: I1008 18:17:57.284253 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 18:17:57 crc kubenswrapper[4859]: I1008 18:17:57.284635 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 18:17:57 crc kubenswrapper[4859]: I1008 18:17:57.284767 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 18:17:57 crc kubenswrapper[4859]: I1008 18:17:57.284859 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 18:17:57 crc kubenswrapper[4859]: I1008 18:17:57.284955 4859 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T18:17:57Z","lastTransitionTime":"2025-10-08T18:17:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 18:17:57 crc kubenswrapper[4859]: I1008 18:17:57.388233 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 18:17:57 crc kubenswrapper[4859]: I1008 18:17:57.388353 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 18:17:57 crc kubenswrapper[4859]: I1008 18:17:57.388378 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 18:17:57 crc kubenswrapper[4859]: I1008 18:17:57.388409 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 18:17:57 crc kubenswrapper[4859]: I1008 18:17:57.388435 4859 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T18:17:57Z","lastTransitionTime":"2025-10-08T18:17:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 18:17:57 crc kubenswrapper[4859]: I1008 18:17:57.438639 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 18:17:57 crc kubenswrapper[4859]: I1008 18:17:57.438748 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 18:17:57 crc kubenswrapper[4859]: I1008 18:17:57.438771 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 18:17:57 crc kubenswrapper[4859]: I1008 18:17:57.438797 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 18:17:57 crc kubenswrapper[4859]: I1008 18:17:57.438816 4859 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T18:17:57Z","lastTransitionTime":"2025-10-08T18:17:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 18:17:57 crc kubenswrapper[4859]: E1008 18:17:57.461492 4859 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404552Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865352Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-10-08T18:17:57Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-08T18:17:57Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-08T18:17:57Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-08T18:17:57Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-08T18:17:57Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-08T18:17:57Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-08T18:17:57Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-08T18:17:57Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"edb3ee6d-476b-407c-bb7d-4480786e8777\\\",\\\"systemUUID\\\":\\\"8f73c7ed-44aa-4d16-bae1-d4e684803cfc\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T18:17:57Z is after 2025-08-24T17:21:41Z" Oct 08 18:17:57 crc kubenswrapper[4859]: I1008 18:17:57.467023 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 18:17:57 crc kubenswrapper[4859]: I1008 18:17:57.467110 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 18:17:57 crc kubenswrapper[4859]: I1008 18:17:57.467126 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 18:17:57 crc kubenswrapper[4859]: I1008 18:17:57.467151 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 18:17:57 crc kubenswrapper[4859]: I1008 18:17:57.467166 4859 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T18:17:57Z","lastTransitionTime":"2025-10-08T18:17:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 18:17:57 crc kubenswrapper[4859]: I1008 18:17:57.469398 4859 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 08 18:17:57 crc kubenswrapper[4859]: I1008 18:17:57.469632 4859 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-4b4sz" Oct 08 18:17:57 crc kubenswrapper[4859]: I1008 18:17:57.469659 4859 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 08 18:17:57 crc kubenswrapper[4859]: E1008 18:17:57.469606 4859 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 08 18:17:57 crc kubenswrapper[4859]: I1008 18:17:57.469850 4859 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 08 18:17:57 crc kubenswrapper[4859]: E1008 18:17:57.474827 4859 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 08 18:17:57 crc kubenswrapper[4859]: E1008 18:17:57.476748 4859 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-4b4sz" podUID="2e53cd9b-64f7-4e07-8b96-fb77847c9ac6" Oct 08 18:17:57 crc kubenswrapper[4859]: E1008 18:17:57.477890 4859 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 08 18:17:57 crc kubenswrapper[4859]: I1008 18:17:57.477570 4859 scope.go:117] "RemoveContainer" containerID="2ce53a49f210c2d6e0001ee4280916611bbd82b4fc21a99c59954ed3d7570b96" Oct 08 18:17:57 crc kubenswrapper[4859]: E1008 18:17:57.487010 4859 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404552Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865352Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-10-08T18:17:57Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-08T18:17:57Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-08T18:17:57Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-08T18:17:57Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-08T18:17:57Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-08T18:17:57Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-08T18:17:57Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-08T18:17:57Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"edb3ee6d-476b-407c-bb7d-4480786e8777\\\",\\\"systemUUID\\\":\\\"8f73c7ed-44aa-4d16-bae1-d4e684803cfc\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T18:17:57Z is after 2025-08-24T17:21:41Z" Oct 08 18:17:57 crc kubenswrapper[4859]: I1008 18:17:57.495619 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 18:17:57 crc kubenswrapper[4859]: I1008 18:17:57.495678 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 18:17:57 crc kubenswrapper[4859]: I1008 18:17:57.495776 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 18:17:57 crc kubenswrapper[4859]: I1008 18:17:57.495806 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 18:17:57 crc kubenswrapper[4859]: I1008 18:17:57.495840 4859 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T18:17:57Z","lastTransitionTime":"2025-10-08T18:17:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 18:17:57 crc kubenswrapper[4859]: E1008 18:17:57.517896 4859 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404552Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865352Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-10-08T18:17:57Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-08T18:17:57Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-08T18:17:57Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-08T18:17:57Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-08T18:17:57Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-08T18:17:57Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-08T18:17:57Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-08T18:17:57Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"edb3ee6d-476b-407c-bb7d-4480786e8777\\\",\\\"systemUUID\\\":\\\"8f73c7ed-44aa-4d16-bae1-d4e684803cfc\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T18:17:57Z is after 2025-08-24T17:21:41Z" Oct 08 18:17:57 crc kubenswrapper[4859]: I1008 18:17:57.524828 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 18:17:57 crc kubenswrapper[4859]: I1008 18:17:57.524890 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 18:17:57 crc kubenswrapper[4859]: I1008 18:17:57.524915 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 18:17:57 crc kubenswrapper[4859]: I1008 18:17:57.524949 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 18:17:57 crc kubenswrapper[4859]: I1008 18:17:57.524979 4859 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T18:17:57Z","lastTransitionTime":"2025-10-08T18:17:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 18:17:57 crc kubenswrapper[4859]: E1008 18:17:57.546433 4859 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404552Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865352Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-10-08T18:17:57Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-08T18:17:57Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-08T18:17:57Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-08T18:17:57Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-08T18:17:57Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-08T18:17:57Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-08T18:17:57Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-08T18:17:57Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"edb3ee6d-476b-407c-bb7d-4480786e8777\\\",\\\"systemUUID\\\":\\\"8f73c7ed-44aa-4d16-bae1-d4e684803cfc\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T18:17:57Z is after 2025-08-24T17:21:41Z" Oct 08 18:17:57 crc kubenswrapper[4859]: I1008 18:17:57.552001 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 18:17:57 crc kubenswrapper[4859]: I1008 18:17:57.552067 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 18:17:57 crc kubenswrapper[4859]: I1008 18:17:57.552080 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 18:17:57 crc kubenswrapper[4859]: I1008 18:17:57.552106 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 18:17:57 crc kubenswrapper[4859]: I1008 18:17:57.552122 4859 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T18:17:57Z","lastTransitionTime":"2025-10-08T18:17:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 18:17:57 crc kubenswrapper[4859]: E1008 18:17:57.570019 4859 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404552Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865352Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-10-08T18:17:57Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-08T18:17:57Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-08T18:17:57Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-08T18:17:57Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-08T18:17:57Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-08T18:17:57Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-08T18:17:57Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-08T18:17:57Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"edb3ee6d-476b-407c-bb7d-4480786e8777\\\",\\\"systemUUID\\\":\\\"8f73c7ed-44aa-4d16-bae1-d4e684803cfc\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T18:17:57Z is after 2025-08-24T17:21:41Z" Oct 08 18:17:57 crc kubenswrapper[4859]: E1008 18:17:57.570252 4859 kubelet_node_status.go:572] "Unable to update node status" err="update node status exceeds retry count" Oct 08 18:17:57 crc kubenswrapper[4859]: I1008 18:17:57.572233 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 18:17:57 crc kubenswrapper[4859]: I1008 18:17:57.572272 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 18:17:57 crc kubenswrapper[4859]: I1008 18:17:57.572285 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 18:17:57 crc kubenswrapper[4859]: I1008 18:17:57.572305 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 18:17:57 crc kubenswrapper[4859]: I1008 18:17:57.572318 4859 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T18:17:57Z","lastTransitionTime":"2025-10-08T18:17:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 18:17:57 crc kubenswrapper[4859]: I1008 18:17:57.676488 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 18:17:57 crc kubenswrapper[4859]: I1008 18:17:57.676553 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 18:17:57 crc kubenswrapper[4859]: I1008 18:17:57.676572 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 18:17:57 crc kubenswrapper[4859]: I1008 18:17:57.676599 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 18:17:57 crc kubenswrapper[4859]: I1008 18:17:57.676617 4859 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T18:17:57Z","lastTransitionTime":"2025-10-08T18:17:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 18:17:57 crc kubenswrapper[4859]: I1008 18:17:57.781438 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 18:17:57 crc kubenswrapper[4859]: I1008 18:17:57.781988 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 18:17:57 crc kubenswrapper[4859]: I1008 18:17:57.782011 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 18:17:57 crc kubenswrapper[4859]: I1008 18:17:57.782040 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 18:17:57 crc kubenswrapper[4859]: I1008 18:17:57.782060 4859 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T18:17:57Z","lastTransitionTime":"2025-10-08T18:17:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 18:17:57 crc kubenswrapper[4859]: I1008 18:17:57.866530 4859 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-4g8gf_1dff864e-b75d-4e0b-b182-75f710eac8df/ovnkube-controller/1.log" Oct 08 18:17:57 crc kubenswrapper[4859]: I1008 18:17:57.870573 4859 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-4g8gf" event={"ID":"1dff864e-b75d-4e0b-b182-75f710eac8df","Type":"ContainerStarted","Data":"c646f0e0dd8e7f74125b687f0cb4114b1f34667e5a6325f394f0208d7fb89df6"} Oct 08 18:17:57 crc kubenswrapper[4859]: I1008 18:17:57.870769 4859 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" Oct 08 18:17:57 crc kubenswrapper[4859]: I1008 18:17:57.884974 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 18:17:57 crc kubenswrapper[4859]: I1008 18:17:57.885033 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 18:17:57 crc kubenswrapper[4859]: I1008 18:17:57.885045 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 18:17:57 crc kubenswrapper[4859]: I1008 18:17:57.885065 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 18:17:57 crc kubenswrapper[4859]: I1008 18:17:57.885079 4859 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T18:17:57Z","lastTransitionTime":"2025-10-08T18:17:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 18:17:57 crc kubenswrapper[4859]: I1008 18:17:57.885234 4859 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-g4t79" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3a92dbf1-21a8-4095-96bb-2d701d7f3d56\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:17:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:17:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:17:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:17:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f8772a25b6e605efe9a062dc32dbbc2cfe32b4fe164b4e84fbf3bcffab744179\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T18:17:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sv79p\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://92cb9986c09a7f9e1e552e3e3ac1f74f04128241e00ddee38882d2ba5c005a67\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T18:17:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sv79p\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T18:17:45Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-g4t79\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T18:17:57Z is after 2025-08-24T17:21:41Z" Oct 08 18:17:57 crc kubenswrapper[4859]: I1008 18:17:57.901064 4859 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-08T18:17:29Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T18:17:29Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T18:17:57Z is after 2025-08-24T17:21:41Z" Oct 08 18:17:57 crc kubenswrapper[4859]: I1008 18:17:57.913675 4859 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-btw28" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"97bc5598-3ead-4300-a938-3aebbff1e8c6\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:17:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:17:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:17:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:17:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d6d4d6ba771cb5fde637449bba05660f634df7e833a0c8dad2b2c7f3a422bb2b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T18:17:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qcxsd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T18:17:31Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-btw28\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T18:17:57Z is after 2025-08-24T17:21:41Z" Oct 08 18:17:57 crc kubenswrapper[4859]: I1008 18:17:57.927924 4859 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-8plkg" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"92b2d7de-31cc-4a91-95a5-ed5ea964b028\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:17:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:17:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:17:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:17:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fa00b8345d40b74a917ddc23cc1b2c5c684a9ca61b6119d090c92cc1439b71c4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T18:17:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q9d7m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T18:17:31Z\\\"}}\" for pod \"openshift-multus\"/\"multus-8plkg\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T18:17:57Z is after 2025-08-24T17:21:41Z" Oct 08 18:17:57 crc kubenswrapper[4859]: I1008 18:17:57.936748 4859 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-btjmq" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b8605f7-fed7-493f-9b4f-6a6358adf907\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:17:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:17:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:17:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:17:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://14e9b6c4a4403f36adb78fcf1135a620ca6fc56ed974c9639dcdfc59e6917559\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T18:17:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xspg4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T18:17:35Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-btjmq\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T18:17:57Z is after 2025-08-24T17:21:41Z" Oct 08 18:17:57 crc kubenswrapper[4859]: I1008 18:17:57.949806 4859 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-08T18:17:31Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T18:17:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T18:17:31Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c6d15905694b6bbd4999d49dc8f7a53eb6bfaa7adb59efb0d7e4973c21d1c7be\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T18:17:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T18:17:57Z is after 2025-08-24T17:21:41Z" Oct 08 18:17:57 crc kubenswrapper[4859]: I1008 18:17:57.963251 4859 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-08T18:17:33Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T18:17:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T18:17:33Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5d99e071bd083914418d9c128f38955f26295c08b1520f2dc774c99b1c843693\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T18:17:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T18:17:57Z is after 2025-08-24T17:21:41Z" Oct 08 18:17:57 crc kubenswrapper[4859]: I1008 18:17:57.975524 4859 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-82s52" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b23a6a6c-9d92-4e7b-840e-55cfda873a2d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:17:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:17:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:17:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:17:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://71302c766485b9f55743d85f97440ff578cd44ba1e5a6f5b1c024c0e808072d2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T18:17:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6vx22\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c23ea3580998adf90621a0155c20ee4bd590c512629812174e95c28857bcfef9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T18:17:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6vx22\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T18:17:31Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-82s52\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T18:17:57Z is after 2025-08-24T17:21:41Z" Oct 08 18:17:57 crc kubenswrapper[4859]: I1008 18:17:57.988310 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 18:17:57 crc kubenswrapper[4859]: I1008 18:17:57.988367 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 18:17:57 crc kubenswrapper[4859]: I1008 18:17:57.988378 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 18:17:57 crc kubenswrapper[4859]: I1008 18:17:57.988397 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 18:17:57 crc kubenswrapper[4859]: I1008 18:17:57.988411 4859 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T18:17:57Z","lastTransitionTime":"2025-10-08T18:17:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 18:17:58 crc kubenswrapper[4859]: I1008 18:17:58.000335 4859 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-4g8gf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1dff864e-b75d-4e0b-b182-75f710eac8df\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:17:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:17:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:17:32Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:17:32Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://768ca274dedcc6847fc21d15234d3665fa9ccae371b5b6a882cfcb9b182f527d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T18:17:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s8jnn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://669f5f3c6d3c1ec0c4b4870e090b0a853adc01727c23383e595715b476c32668\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T18:17:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s8jnn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b8009a42cc4c9a816457a0678bdb2f7b7c53d0c8604e007a6fb141f66d06a516\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T18:17:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s8jnn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://676511a11ed4102084b57ebdd5f3fda4969741b2bc8443b415c4bf94f39ae4c9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T18:17:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s8jnn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4a74c1f0d21c90cdd7e776ee14e1f48757a04e2220fbd1f884449181ec6d1c3a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T18:17:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s8jnn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6ddc0e842bb8f81663406070023b0e5ed9cc99c98db0123b1c84f20da59fb5b5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T18:17:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s8jnn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c646f0e0dd8e7f74125b687f0cb4114b1f34667e5a6325f394f0208d7fb89df6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2ce53a49f210c2d6e0001ee4280916611bbd82b4fc21a99c59954ed3d7570b96\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-10-08T18:17:43Z\\\",\\\"message\\\":\\\"cert-secret-name:metrics-tls service.beta.openshift.io/serving-cert-signed-by:openshift-service-serving-signer@1740288168] [{config.openshift.io/v1 ClusterVersion version 9101b518-476b-4eea-8fa6-69b0534e5caa 0xc00695d3d7 \\\\u003cnil\\\\u003e}] [] []},Spec:ServiceSpec{Ports:[]ServicePort{ServicePort{Name:metrics,Protocol:TCP,Port:9393,TargetPort:{1 0 metrics},NodePort:0,AppProtocol:nil,},},Selector:map[string]string{name: ingress-operator,},ClusterIP:10.217.5.244,Type:ClusterIP,ExternalIPs:[],SessionAffinity:None,LoadBalancerIP:,LoadBalancerSourceRanges:[],ExternalName:,ExternalTrafficPolicy:,HealthCheckNodePort:0,PublishNotReadyAddresses:false,SessionAffinityConfig:nil,IPFamilyPolicy:*SingleStack,ClusterIPs:[10.217.5.244],IPFamilies:[IPv4],AllocateLoadBalancerNodePorts:nil,LoadBalancerClass:nil,InternalTrafficPolicy:*Cluster,TrafficDistribution:nil,},Status:ServiceStatus{LoadBalancer:LoadBalancerStatus{Ingress:[]LoadBalancerIngress{},},Conditions:[]Condition{},},}\\\\nI1008 18:17:43.764981 6316 base_network_controller_pods.go:477] [default/openshift-network-diagnostics/network-check-target-xd92c] creating logical port openshift-network-diagnostics_network-check-target-xd92c for pod on switch crc\\\\nF1008 18:17:43.764990 6316 ovnkube.go:137] failed to run ovnkube: [failed to start network controller: failed to start default network controller: unable to create\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-08T18:17:42Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T18:17:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s8jnn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2c67d2aed7126d76989e057d8e2d8e188a8902b2f3fe26ed6d1d45d3b013d897\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T18:17:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s8jnn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7a211d3da02b173867ff3a912eecd3b379483c7cc77e6b91122edd1ffd30fd53\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7a211d3da02b173867ff3a912eecd3b379483c7cc77e6b91122edd1ffd30fd53\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-08T18:17:32Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-08T18:17:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s8jnn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T18:17:32Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-4g8gf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T18:17:57Z is after 2025-08-24T17:21:41Z" Oct 08 18:17:58 crc kubenswrapper[4859]: I1008 18:17:58.018825 4859 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-4b4sz" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2e53cd9b-64f7-4e07-8b96-fb77847c9ac6\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:17:45Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:17:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:17:45Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:17:45Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kxgbz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kxgbz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T18:17:45Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-4b4sz\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T18:17:58Z is after 2025-08-24T17:21:41Z" Oct 08 18:17:58 crc kubenswrapper[4859]: I1008 18:17:58.035639 4859 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"63a623de-5608-4228-b1c2-77188c276986\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:17:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:17:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:17:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:17:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:17:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d4786a8fb8abd96fb22e60733c974f599017407d589ed815fb8541fd3b64f696\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T18:17:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://183d481502ed5108d5e19e483b714c289f355ea3f8b0891883e5335b802c83b6\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T18:17:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ec848c46fd5dde0469b00d3cab14ba4a652533a265490fadbee7a8b8843ef249\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T18:17:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4fe1828b4d63ff8106bfcd15e1dc1c1bd1ba7f77777cdb23ff1b291d0418774c\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2e700b2a20c81b414617f5b5c018a738278a3fa44e6c6d14b2170051d0725431\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-08T18:17:24Z\\\",\\\"message\\\":\\\"W1008 18:17:13.670625 1 cmd.go:257] Using insecure, self-signed certificates\\\\nI1008 18:17:13.671202 1 crypto.go:601] Generating new CA for check-endpoints-signer@1759947433 cert, and key in /tmp/serving-cert-3867218267/serving-signer.crt, /tmp/serving-cert-3867218267/serving-signer.key\\\\nI1008 18:17:13.944463 1 observer_polling.go:159] Starting file observer\\\\nW1008 18:17:13.951958 1 builder.go:272] unable to get owner reference (falling back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": dial tcp [::1]:6443: connect: connection refused\\\\nI1008 18:17:13.952088 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1008 18:17:13.956755 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3867218267/tls.crt::/tmp/serving-cert-3867218267/tls.key\\\\\\\"\\\\nF1008 18:17:24.290991 1 cmd.go:182] error initializing delegating authentication: unable to load configmap based request-header-client-ca-file: Get \\\\\\\"https://localhost:6443/api/v1/namespaces/kube-system/configmaps/extension-apiserver-authentication\\\\\\\": net/http: TLS handshake timeout\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-08T18:17:13Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T18:17:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c65b66f62d325ca10ef8d5d31f04529ccf4e691cf3303d231e7a414fb400e4d6\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T18:17:13Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://58c7bba4ace28215b062653a960e561b1dc1f890f2baf450145638a065af038b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://58c7bba4ace28215b062653a960e561b1dc1f890f2baf450145638a065af038b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-08T18:17:11Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-08T18:17:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T18:17:10Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T18:17:58Z is after 2025-08-24T17:21:41Z" Oct 08 18:17:58 crc kubenswrapper[4859]: I1008 18:17:58.054362 4859 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-08T18:17:31Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T18:17:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T18:17:31Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://99e2d51bd7757498dae60949eea6d732b5f7b481fa00f808dd5308a530a1daf1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T18:17:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e32aaba93699944665660ae41e9b5779178c1354c31e15ad8188bf8c3c8fa500\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T18:17:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T18:17:58Z is after 2025-08-24T17:21:41Z" Oct 08 18:17:58 crc kubenswrapper[4859]: I1008 18:17:58.077479 4859 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-08T18:17:29Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T18:17:29Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T18:17:58Z is after 2025-08-24T17:21:41Z" Oct 08 18:17:58 crc kubenswrapper[4859]: I1008 18:17:58.091108 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 18:17:58 crc kubenswrapper[4859]: I1008 18:17:58.091152 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 18:17:58 crc kubenswrapper[4859]: I1008 18:17:58.091164 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 18:17:58 crc kubenswrapper[4859]: I1008 18:17:58.091185 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 18:17:58 crc kubenswrapper[4859]: I1008 18:17:58.091201 4859 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T18:17:58Z","lastTransitionTime":"2025-10-08T18:17:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 18:17:58 crc kubenswrapper[4859]: I1008 18:17:58.094455 4859 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4efc16fb-e679-464d-8b3c-0f46f082593e\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:17:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:17:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:17:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:17:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:17:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://aa8f8ea895272cd79a08d1c47cd7863a1ec10e21bb60d12aa10862a56c99e4bf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T18:17:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5c4e95d8a712ab219455564539a90286ae48bbc46ddc322a5bfd96c1a8268110\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T18:17:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://bae6ecfd8b446ef04d393d823e5ffe37fe749337c0ca17fde79c5f88d5e31393\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T18:17:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6411d1e7e18d5ceadaaceed5a8a6b20a0cacc55cede793ca9fb4dcad70c64fc6\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T18:17:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T18:17:10Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T18:17:58Z is after 2025-08-24T17:21:41Z" Oct 08 18:17:58 crc kubenswrapper[4859]: I1008 18:17:58.107250 4859 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-08T18:17:29Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T18:17:29Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T18:17:58Z is after 2025-08-24T17:21:41Z" Oct 08 18:17:58 crc kubenswrapper[4859]: I1008 18:17:58.121244 4859 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-84k8x" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1c24078c-c9e8-4160-883a-f483b4b3dd23\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:17:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:17:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:17:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:17:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8661006834d0f50eaada7394369584ee95fc61ff54437eec35d755526afc5b48\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T18:17:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hk5xb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://18b291eb3344fb7c00ef5679dca9a6f7ed9d08cea684a803ff7dad2e8a18c585\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://18b291eb3344fb7c00ef5679dca9a6f7ed9d08cea684a803ff7dad2e8a18c585\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-08T18:17:32Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-08T18:17:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hk5xb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://160163f1a91d981583452db3c7a9244182d1c8ced45749a61b721d0d45c1dfe4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://160163f1a91d981583452db3c7a9244182d1c8ced45749a61b721d0d45c1dfe4\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-08T18:17:34Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-08T18:17:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hk5xb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://51e9b53f3847ac53f378ad83870debf8e71518d4edbc7fc0dc66644818e2103e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://51e9b53f3847ac53f378ad83870debf8e71518d4edbc7fc0dc66644818e2103e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-08T18:17:34Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-08T18:17:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hk5xb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1ff118919739759ac01b5cbe3240e6673db54eaa16353c609d670071ed7964cb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1ff118919739759ac01b5cbe3240e6673db54eaa16353c609d670071ed7964cb\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-08T18:17:35Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-08T18:17:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hk5xb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8522e803ff56eb57cd4d9fd036e0c3fb16240ea21d4a094c1db7aa7d50bfe2c2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8522e803ff56eb57cd4d9fd036e0c3fb16240ea21d4a094c1db7aa7d50bfe2c2\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-08T18:17:37Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-08T18:17:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hk5xb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://be61056edf515e037403ca2a0bfd5feb9b2537dd42212b58e133b23b294f327b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://be61056edf515e037403ca2a0bfd5feb9b2537dd42212b58e133b23b294f327b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-08T18:17:38Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-08T18:17:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hk5xb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T18:17:31Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-84k8x\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T18:17:58Z is after 2025-08-24T17:21:41Z" Oct 08 18:17:58 crc kubenswrapper[4859]: I1008 18:17:58.193324 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 18:17:58 crc kubenswrapper[4859]: I1008 18:17:58.193390 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 18:17:58 crc kubenswrapper[4859]: I1008 18:17:58.193408 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 18:17:58 crc kubenswrapper[4859]: I1008 18:17:58.193436 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 18:17:58 crc kubenswrapper[4859]: I1008 18:17:58.193460 4859 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T18:17:58Z","lastTransitionTime":"2025-10-08T18:17:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 18:17:58 crc kubenswrapper[4859]: I1008 18:17:58.295280 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 18:17:58 crc kubenswrapper[4859]: I1008 18:17:58.295312 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 18:17:58 crc kubenswrapper[4859]: I1008 18:17:58.295320 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 18:17:58 crc kubenswrapper[4859]: I1008 18:17:58.295334 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 18:17:58 crc kubenswrapper[4859]: I1008 18:17:58.295346 4859 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T18:17:58Z","lastTransitionTime":"2025-10-08T18:17:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 18:17:58 crc kubenswrapper[4859]: I1008 18:17:58.397563 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 18:17:58 crc kubenswrapper[4859]: I1008 18:17:58.397611 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 18:17:58 crc kubenswrapper[4859]: I1008 18:17:58.397621 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 18:17:58 crc kubenswrapper[4859]: I1008 18:17:58.397637 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 18:17:58 crc kubenswrapper[4859]: I1008 18:17:58.397648 4859 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T18:17:58Z","lastTransitionTime":"2025-10-08T18:17:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 18:17:58 crc kubenswrapper[4859]: I1008 18:17:58.500898 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 18:17:58 crc kubenswrapper[4859]: I1008 18:17:58.500946 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 18:17:58 crc kubenswrapper[4859]: I1008 18:17:58.500958 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 18:17:58 crc kubenswrapper[4859]: I1008 18:17:58.500980 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 18:17:58 crc kubenswrapper[4859]: I1008 18:17:58.500995 4859 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T18:17:58Z","lastTransitionTime":"2025-10-08T18:17:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 18:17:58 crc kubenswrapper[4859]: I1008 18:17:58.603492 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 18:17:58 crc kubenswrapper[4859]: I1008 18:17:58.603554 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 18:17:58 crc kubenswrapper[4859]: I1008 18:17:58.603569 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 18:17:58 crc kubenswrapper[4859]: I1008 18:17:58.603587 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 18:17:58 crc kubenswrapper[4859]: I1008 18:17:58.603598 4859 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T18:17:58Z","lastTransitionTime":"2025-10-08T18:17:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 18:17:58 crc kubenswrapper[4859]: I1008 18:17:58.707661 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 18:17:58 crc kubenswrapper[4859]: I1008 18:17:58.707772 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 18:17:58 crc kubenswrapper[4859]: I1008 18:17:58.707794 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 18:17:58 crc kubenswrapper[4859]: I1008 18:17:58.707821 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 18:17:58 crc kubenswrapper[4859]: I1008 18:17:58.707839 4859 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T18:17:58Z","lastTransitionTime":"2025-10-08T18:17:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 18:17:58 crc kubenswrapper[4859]: I1008 18:17:58.811538 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 18:17:58 crc kubenswrapper[4859]: I1008 18:17:58.811612 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 18:17:58 crc kubenswrapper[4859]: I1008 18:17:58.811633 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 18:17:58 crc kubenswrapper[4859]: I1008 18:17:58.811659 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 18:17:58 crc kubenswrapper[4859]: I1008 18:17:58.811679 4859 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T18:17:58Z","lastTransitionTime":"2025-10-08T18:17:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 18:17:58 crc kubenswrapper[4859]: I1008 18:17:58.878563 4859 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-4g8gf_1dff864e-b75d-4e0b-b182-75f710eac8df/ovnkube-controller/2.log" Oct 08 18:17:58 crc kubenswrapper[4859]: I1008 18:17:58.879897 4859 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-4g8gf_1dff864e-b75d-4e0b-b182-75f710eac8df/ovnkube-controller/1.log" Oct 08 18:17:58 crc kubenswrapper[4859]: I1008 18:17:58.885359 4859 generic.go:334] "Generic (PLEG): container finished" podID="1dff864e-b75d-4e0b-b182-75f710eac8df" containerID="c646f0e0dd8e7f74125b687f0cb4114b1f34667e5a6325f394f0208d7fb89df6" exitCode=1 Oct 08 18:17:58 crc kubenswrapper[4859]: I1008 18:17:58.885433 4859 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-4g8gf" event={"ID":"1dff864e-b75d-4e0b-b182-75f710eac8df","Type":"ContainerDied","Data":"c646f0e0dd8e7f74125b687f0cb4114b1f34667e5a6325f394f0208d7fb89df6"} Oct 08 18:17:58 crc kubenswrapper[4859]: I1008 18:17:58.885493 4859 scope.go:117] "RemoveContainer" containerID="2ce53a49f210c2d6e0001ee4280916611bbd82b4fc21a99c59954ed3d7570b96" Oct 08 18:17:58 crc kubenswrapper[4859]: I1008 18:17:58.886783 4859 scope.go:117] "RemoveContainer" containerID="c646f0e0dd8e7f74125b687f0cb4114b1f34667e5a6325f394f0208d7fb89df6" Oct 08 18:17:58 crc kubenswrapper[4859]: E1008 18:17:58.887123 4859 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ovnkube-controller\" with CrashLoopBackOff: \"back-off 20s restarting failed container=ovnkube-controller pod=ovnkube-node-4g8gf_openshift-ovn-kubernetes(1dff864e-b75d-4e0b-b182-75f710eac8df)\"" pod="openshift-ovn-kubernetes/ovnkube-node-4g8gf" podUID="1dff864e-b75d-4e0b-b182-75f710eac8df" Oct 08 18:17:58 crc kubenswrapper[4859]: I1008 18:17:58.915296 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 18:17:58 crc kubenswrapper[4859]: I1008 18:17:58.915362 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 18:17:58 crc kubenswrapper[4859]: I1008 18:17:58.915384 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 18:17:58 crc kubenswrapper[4859]: I1008 18:17:58.915408 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 18:17:58 crc kubenswrapper[4859]: I1008 18:17:58.915426 4859 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T18:17:58Z","lastTransitionTime":"2025-10-08T18:17:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 18:17:58 crc kubenswrapper[4859]: I1008 18:17:58.918168 4859 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-08T18:17:29Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T18:17:29Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T18:17:58Z is after 2025-08-24T17:21:41Z" Oct 08 18:17:58 crc kubenswrapper[4859]: I1008 18:17:58.931747 4859 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-btw28" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"97bc5598-3ead-4300-a938-3aebbff1e8c6\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:17:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:17:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:17:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:17:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d6d4d6ba771cb5fde637449bba05660f634df7e833a0c8dad2b2c7f3a422bb2b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T18:17:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qcxsd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T18:17:31Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-btw28\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T18:17:58Z is after 2025-08-24T17:21:41Z" Oct 08 18:17:58 crc kubenswrapper[4859]: I1008 18:17:58.946470 4859 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-8plkg" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"92b2d7de-31cc-4a91-95a5-ed5ea964b028\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:17:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:17:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:17:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:17:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fa00b8345d40b74a917ddc23cc1b2c5c684a9ca61b6119d090c92cc1439b71c4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T18:17:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q9d7m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T18:17:31Z\\\"}}\" for pod \"openshift-multus\"/\"multus-8plkg\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T18:17:58Z is after 2025-08-24T17:21:41Z" Oct 08 18:17:58 crc kubenswrapper[4859]: I1008 18:17:58.962035 4859 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-btjmq" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b8605f7-fed7-493f-9b4f-6a6358adf907\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:17:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:17:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:17:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:17:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://14e9b6c4a4403f36adb78fcf1135a620ca6fc56ed974c9639dcdfc59e6917559\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T18:17:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xspg4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T18:17:35Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-btjmq\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T18:17:58Z is after 2025-08-24T17:21:41Z" Oct 08 18:17:58 crc kubenswrapper[4859]: I1008 18:17:58.980022 4859 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-g4t79" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3a92dbf1-21a8-4095-96bb-2d701d7f3d56\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:17:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:17:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:17:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:17:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f8772a25b6e605efe9a062dc32dbbc2cfe32b4fe164b4e84fbf3bcffab744179\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T18:17:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sv79p\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://92cb9986c09a7f9e1e552e3e3ac1f74f04128241e00ddee38882d2ba5c005a67\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T18:17:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sv79p\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T18:17:45Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-g4t79\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T18:17:58Z is after 2025-08-24T17:21:41Z" Oct 08 18:17:59 crc kubenswrapper[4859]: I1008 18:17:59.002066 4859 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-08T18:17:31Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T18:17:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T18:17:31Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c6d15905694b6bbd4999d49dc8f7a53eb6bfaa7adb59efb0d7e4973c21d1c7be\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T18:17:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T18:17:58Z is after 2025-08-24T17:21:41Z" Oct 08 18:17:59 crc kubenswrapper[4859]: I1008 18:17:59.016986 4859 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-08T18:17:33Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T18:17:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T18:17:33Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5d99e071bd083914418d9c128f38955f26295c08b1520f2dc774c99b1c843693\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T18:17:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T18:17:59Z is after 2025-08-24T17:21:41Z" Oct 08 18:17:59 crc kubenswrapper[4859]: I1008 18:17:59.019877 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 18:17:59 crc kubenswrapper[4859]: I1008 18:17:59.019995 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 18:17:59 crc kubenswrapper[4859]: I1008 18:17:59.020073 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 18:17:59 crc kubenswrapper[4859]: I1008 18:17:59.020175 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 18:17:59 crc kubenswrapper[4859]: I1008 18:17:59.020258 4859 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T18:17:59Z","lastTransitionTime":"2025-10-08T18:17:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 18:17:59 crc kubenswrapper[4859]: I1008 18:17:59.050630 4859 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-4g8gf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1dff864e-b75d-4e0b-b182-75f710eac8df\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:17:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:17:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:17:32Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:17:32Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://768ca274dedcc6847fc21d15234d3665fa9ccae371b5b6a882cfcb9b182f527d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T18:17:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s8jnn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://669f5f3c6d3c1ec0c4b4870e090b0a853adc01727c23383e595715b476c32668\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T18:17:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s8jnn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b8009a42cc4c9a816457a0678bdb2f7b7c53d0c8604e007a6fb141f66d06a516\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T18:17:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s8jnn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://676511a11ed4102084b57ebdd5f3fda4969741b2bc8443b415c4bf94f39ae4c9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T18:17:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s8jnn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4a74c1f0d21c90cdd7e776ee14e1f48757a04e2220fbd1f884449181ec6d1c3a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T18:17:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s8jnn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6ddc0e842bb8f81663406070023b0e5ed9cc99c98db0123b1c84f20da59fb5b5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T18:17:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s8jnn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c646f0e0dd8e7f74125b687f0cb4114b1f34667e5a6325f394f0208d7fb89df6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2ce53a49f210c2d6e0001ee4280916611bbd82b4fc21a99c59954ed3d7570b96\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-10-08T18:17:43Z\\\",\\\"message\\\":\\\"cert-secret-name:metrics-tls service.beta.openshift.io/serving-cert-signed-by:openshift-service-serving-signer@1740288168] [{config.openshift.io/v1 ClusterVersion version 9101b518-476b-4eea-8fa6-69b0534e5caa 0xc00695d3d7 \\\\u003cnil\\\\u003e}] [] []},Spec:ServiceSpec{Ports:[]ServicePort{ServicePort{Name:metrics,Protocol:TCP,Port:9393,TargetPort:{1 0 metrics},NodePort:0,AppProtocol:nil,},},Selector:map[string]string{name: ingress-operator,},ClusterIP:10.217.5.244,Type:ClusterIP,ExternalIPs:[],SessionAffinity:None,LoadBalancerIP:,LoadBalancerSourceRanges:[],ExternalName:,ExternalTrafficPolicy:,HealthCheckNodePort:0,PublishNotReadyAddresses:false,SessionAffinityConfig:nil,IPFamilyPolicy:*SingleStack,ClusterIPs:[10.217.5.244],IPFamilies:[IPv4],AllocateLoadBalancerNodePorts:nil,LoadBalancerClass:nil,InternalTrafficPolicy:*Cluster,TrafficDistribution:nil,},Status:ServiceStatus{LoadBalancer:LoadBalancerStatus{Ingress:[]LoadBalancerIngress{},},Conditions:[]Condition{},},}\\\\nI1008 18:17:43.764981 6316 base_network_controller_pods.go:477] [default/openshift-network-diagnostics/network-check-target-xd92c] creating logical port openshift-network-diagnostics_network-check-target-xd92c for pod on switch crc\\\\nF1008 18:17:43.764990 6316 ovnkube.go:137] failed to run ovnkube: [failed to start network controller: failed to start default network controller: unable to create\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-08T18:17:42Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c646f0e0dd8e7f74125b687f0cb4114b1f34667e5a6325f394f0208d7fb89df6\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-10-08T18:17:58Z\\\",\\\"message\\\":\\\"e *v1.Pod retry setup to complete in iterateRetryResources\\\\nI1008 18:17:58.424504 6520 obj_retry.go:365] Adding new object: *v1.Pod openshift-machine-config-operator/machine-config-daemon-82s52\\\\nI1008 18:17:58.424529 6520 ovn.go:134] Ensuring zone local for Pod openshift-machine-config-operator/machine-config-daemon-82s52 in node crc\\\\nI1008 18:17:58.424539 6520 obj_retry.go:303] Retry object setup: *v1.Pod openshift-network-diagnostics/network-check-source-55646444c4-trplf\\\\nI1008 18:17:58.424549 6520 obj_retry.go:365] Adding new object: *v1.Pod openshift-network-diagnostics/network-check-source-55646444c4-trplf\\\\nI1008 18:17:58.424549 6520 obj_retry.go:386] Retry successful for *v1.Pod openshift-machine-config-operator/machine-config-daemon-82s52 after 0 failed attempt(s)\\\\nI1008 18:17:58.424559 6520 ovn.go:134] Ensuring zone local for Pod openshift-network-diagnostics/network-check-source-55646444c4-trplf in node crc\\\\nF1008 18:17:58.424552 6520 ovnkube.go:137] failed to run ovnkube: [failed to start network controller: failed to start default network controller: unable to create admin network policy controller, err: could not add Event Handler for anpInformer during admin network policy controller initialization, handler {0x1fcc6e0 0x1fcc3c0 0x1fcc360} was not added to shared informer because it has stopped already, failed to star\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-08T18:17:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s8jnn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2c67d2aed7126d76989e057d8e2d8e188a8902b2f3fe26ed6d1d45d3b013d897\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T18:17:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s8jnn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7a211d3da02b173867ff3a912eecd3b379483c7cc77e6b91122edd1ffd30fd53\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7a211d3da02b173867ff3a912eecd3b379483c7cc77e6b91122edd1ffd30fd53\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-08T18:17:32Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-08T18:17:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s8jnn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T18:17:32Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-4g8gf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T18:17:59Z is after 2025-08-24T17:21:41Z" Oct 08 18:17:59 crc kubenswrapper[4859]: I1008 18:17:59.064365 4859 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-4b4sz" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2e53cd9b-64f7-4e07-8b96-fb77847c9ac6\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:17:45Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:17:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:17:45Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:17:45Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kxgbz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kxgbz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T18:17:45Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-4b4sz\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T18:17:59Z is after 2025-08-24T17:21:41Z" Oct 08 18:17:59 crc kubenswrapper[4859]: I1008 18:17:59.083512 4859 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"63a623de-5608-4228-b1c2-77188c276986\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:17:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:17:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:17:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:17:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:17:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d4786a8fb8abd96fb22e60733c974f599017407d589ed815fb8541fd3b64f696\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T18:17:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://183d481502ed5108d5e19e483b714c289f355ea3f8b0891883e5335b802c83b6\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T18:17:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ec848c46fd5dde0469b00d3cab14ba4a652533a265490fadbee7a8b8843ef249\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T18:17:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4fe1828b4d63ff8106bfcd15e1dc1c1bd1ba7f77777cdb23ff1b291d0418774c\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2e700b2a20c81b414617f5b5c018a738278a3fa44e6c6d14b2170051d0725431\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-08T18:17:24Z\\\",\\\"message\\\":\\\"W1008 18:17:13.670625 1 cmd.go:257] Using insecure, self-signed certificates\\\\nI1008 18:17:13.671202 1 crypto.go:601] Generating new CA for check-endpoints-signer@1759947433 cert, and key in /tmp/serving-cert-3867218267/serving-signer.crt, /tmp/serving-cert-3867218267/serving-signer.key\\\\nI1008 18:17:13.944463 1 observer_polling.go:159] Starting file observer\\\\nW1008 18:17:13.951958 1 builder.go:272] unable to get owner reference (falling back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": dial tcp [::1]:6443: connect: connection refused\\\\nI1008 18:17:13.952088 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1008 18:17:13.956755 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3867218267/tls.crt::/tmp/serving-cert-3867218267/tls.key\\\\\\\"\\\\nF1008 18:17:24.290991 1 cmd.go:182] error initializing delegating authentication: unable to load configmap based request-header-client-ca-file: Get \\\\\\\"https://localhost:6443/api/v1/namespaces/kube-system/configmaps/extension-apiserver-authentication\\\\\\\": net/http: TLS handshake timeout\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-08T18:17:13Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T18:17:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c65b66f62d325ca10ef8d5d31f04529ccf4e691cf3303d231e7a414fb400e4d6\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T18:17:13Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://58c7bba4ace28215b062653a960e561b1dc1f890f2baf450145638a065af038b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://58c7bba4ace28215b062653a960e561b1dc1f890f2baf450145638a065af038b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-08T18:17:11Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-08T18:17:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T18:17:10Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T18:17:59Z is after 2025-08-24T17:21:41Z" Oct 08 18:17:59 crc kubenswrapper[4859]: I1008 18:17:59.100743 4859 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-08T18:17:31Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T18:17:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T18:17:31Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://99e2d51bd7757498dae60949eea6d732b5f7b481fa00f808dd5308a530a1daf1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T18:17:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e32aaba93699944665660ae41e9b5779178c1354c31e15ad8188bf8c3c8fa500\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T18:17:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T18:17:59Z is after 2025-08-24T17:21:41Z" Oct 08 18:17:59 crc kubenswrapper[4859]: I1008 18:17:59.116810 4859 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-08T18:17:29Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T18:17:29Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T18:17:59Z is after 2025-08-24T17:21:41Z" Oct 08 18:17:59 crc kubenswrapper[4859]: I1008 18:17:59.123480 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 18:17:59 crc kubenswrapper[4859]: I1008 18:17:59.123520 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 18:17:59 crc kubenswrapper[4859]: I1008 18:17:59.123528 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 18:17:59 crc kubenswrapper[4859]: I1008 18:17:59.123548 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 18:17:59 crc kubenswrapper[4859]: I1008 18:17:59.123560 4859 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T18:17:59Z","lastTransitionTime":"2025-10-08T18:17:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 18:17:59 crc kubenswrapper[4859]: I1008 18:17:59.133582 4859 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-82s52" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b23a6a6c-9d92-4e7b-840e-55cfda873a2d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:17:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:17:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:17:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:17:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://71302c766485b9f55743d85f97440ff578cd44ba1e5a6f5b1c024c0e808072d2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T18:17:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6vx22\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c23ea3580998adf90621a0155c20ee4bd590c512629812174e95c28857bcfef9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T18:17:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6vx22\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T18:17:31Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-82s52\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T18:17:59Z is after 2025-08-24T17:21:41Z" Oct 08 18:17:59 crc kubenswrapper[4859]: I1008 18:17:59.150532 4859 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4efc16fb-e679-464d-8b3c-0f46f082593e\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:17:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:17:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:17:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:17:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:17:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://aa8f8ea895272cd79a08d1c47cd7863a1ec10e21bb60d12aa10862a56c99e4bf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T18:17:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5c4e95d8a712ab219455564539a90286ae48bbc46ddc322a5bfd96c1a8268110\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T18:17:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://bae6ecfd8b446ef04d393d823e5ffe37fe749337c0ca17fde79c5f88d5e31393\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T18:17:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6411d1e7e18d5ceadaaceed5a8a6b20a0cacc55cede793ca9fb4dcad70c64fc6\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T18:17:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T18:17:10Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T18:17:59Z is after 2025-08-24T17:21:41Z" Oct 08 18:17:59 crc kubenswrapper[4859]: I1008 18:17:59.166829 4859 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-08T18:17:29Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T18:17:29Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T18:17:59Z is after 2025-08-24T17:21:41Z" Oct 08 18:17:59 crc kubenswrapper[4859]: I1008 18:17:59.183538 4859 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-84k8x" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1c24078c-c9e8-4160-883a-f483b4b3dd23\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:17:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:17:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:17:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:17:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8661006834d0f50eaada7394369584ee95fc61ff54437eec35d755526afc5b48\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T18:17:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hk5xb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://18b291eb3344fb7c00ef5679dca9a6f7ed9d08cea684a803ff7dad2e8a18c585\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://18b291eb3344fb7c00ef5679dca9a6f7ed9d08cea684a803ff7dad2e8a18c585\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-08T18:17:32Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-08T18:17:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hk5xb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://160163f1a91d981583452db3c7a9244182d1c8ced45749a61b721d0d45c1dfe4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://160163f1a91d981583452db3c7a9244182d1c8ced45749a61b721d0d45c1dfe4\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-08T18:17:34Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-08T18:17:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hk5xb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://51e9b53f3847ac53f378ad83870debf8e71518d4edbc7fc0dc66644818e2103e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://51e9b53f3847ac53f378ad83870debf8e71518d4edbc7fc0dc66644818e2103e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-08T18:17:34Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-08T18:17:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hk5xb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1ff118919739759ac01b5cbe3240e6673db54eaa16353c609d670071ed7964cb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1ff118919739759ac01b5cbe3240e6673db54eaa16353c609d670071ed7964cb\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-08T18:17:35Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-08T18:17:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hk5xb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8522e803ff56eb57cd4d9fd036e0c3fb16240ea21d4a094c1db7aa7d50bfe2c2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8522e803ff56eb57cd4d9fd036e0c3fb16240ea21d4a094c1db7aa7d50bfe2c2\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-08T18:17:37Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-08T18:17:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hk5xb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://be61056edf515e037403ca2a0bfd5feb9b2537dd42212b58e133b23b294f327b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://be61056edf515e037403ca2a0bfd5feb9b2537dd42212b58e133b23b294f327b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-08T18:17:38Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-08T18:17:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hk5xb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T18:17:31Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-84k8x\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T18:17:59Z is after 2025-08-24T17:21:41Z" Oct 08 18:17:59 crc kubenswrapper[4859]: I1008 18:17:59.226946 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 18:17:59 crc kubenswrapper[4859]: I1008 18:17:59.227025 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 18:17:59 crc kubenswrapper[4859]: I1008 18:17:59.227041 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 18:17:59 crc kubenswrapper[4859]: I1008 18:17:59.227070 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 18:17:59 crc kubenswrapper[4859]: I1008 18:17:59.227093 4859 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T18:17:59Z","lastTransitionTime":"2025-10-08T18:17:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 18:17:59 crc kubenswrapper[4859]: I1008 18:17:59.330222 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 18:17:59 crc kubenswrapper[4859]: I1008 18:17:59.330363 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 18:17:59 crc kubenswrapper[4859]: I1008 18:17:59.330389 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 18:17:59 crc kubenswrapper[4859]: I1008 18:17:59.330416 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 18:17:59 crc kubenswrapper[4859]: I1008 18:17:59.330434 4859 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T18:17:59Z","lastTransitionTime":"2025-10-08T18:17:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 18:17:59 crc kubenswrapper[4859]: I1008 18:17:59.434470 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 18:17:59 crc kubenswrapper[4859]: I1008 18:17:59.434546 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 18:17:59 crc kubenswrapper[4859]: I1008 18:17:59.434584 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 18:17:59 crc kubenswrapper[4859]: I1008 18:17:59.434622 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 18:17:59 crc kubenswrapper[4859]: I1008 18:17:59.434648 4859 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T18:17:59Z","lastTransitionTime":"2025-10-08T18:17:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 18:17:59 crc kubenswrapper[4859]: I1008 18:17:59.469509 4859 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 08 18:17:59 crc kubenswrapper[4859]: I1008 18:17:59.469559 4859 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 08 18:17:59 crc kubenswrapper[4859]: I1008 18:17:59.469509 4859 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 08 18:17:59 crc kubenswrapper[4859]: E1008 18:17:59.469803 4859 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 08 18:17:59 crc kubenswrapper[4859]: I1008 18:17:59.469873 4859 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-4b4sz" Oct 08 18:17:59 crc kubenswrapper[4859]: E1008 18:17:59.469988 4859 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 08 18:17:59 crc kubenswrapper[4859]: E1008 18:17:59.470124 4859 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-4b4sz" podUID="2e53cd9b-64f7-4e07-8b96-fb77847c9ac6" Oct 08 18:17:59 crc kubenswrapper[4859]: E1008 18:17:59.470231 4859 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 08 18:17:59 crc kubenswrapper[4859]: I1008 18:17:59.538001 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 18:17:59 crc kubenswrapper[4859]: I1008 18:17:59.538046 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 18:17:59 crc kubenswrapper[4859]: I1008 18:17:59.538060 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 18:17:59 crc kubenswrapper[4859]: I1008 18:17:59.538084 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 18:17:59 crc kubenswrapper[4859]: I1008 18:17:59.538094 4859 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T18:17:59Z","lastTransitionTime":"2025-10-08T18:17:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 18:17:59 crc kubenswrapper[4859]: I1008 18:17:59.640653 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 18:17:59 crc kubenswrapper[4859]: I1008 18:17:59.640734 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 18:17:59 crc kubenswrapper[4859]: I1008 18:17:59.640744 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 18:17:59 crc kubenswrapper[4859]: I1008 18:17:59.640761 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 18:17:59 crc kubenswrapper[4859]: I1008 18:17:59.640772 4859 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T18:17:59Z","lastTransitionTime":"2025-10-08T18:17:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 18:17:59 crc kubenswrapper[4859]: I1008 18:17:59.744253 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 18:17:59 crc kubenswrapper[4859]: I1008 18:17:59.744318 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 18:17:59 crc kubenswrapper[4859]: I1008 18:17:59.744331 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 18:17:59 crc kubenswrapper[4859]: I1008 18:17:59.744350 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 18:17:59 crc kubenswrapper[4859]: I1008 18:17:59.744362 4859 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T18:17:59Z","lastTransitionTime":"2025-10-08T18:17:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 18:17:59 crc kubenswrapper[4859]: I1008 18:17:59.847760 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 18:17:59 crc kubenswrapper[4859]: I1008 18:17:59.847835 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 18:17:59 crc kubenswrapper[4859]: I1008 18:17:59.847853 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 18:17:59 crc kubenswrapper[4859]: I1008 18:17:59.847879 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 18:17:59 crc kubenswrapper[4859]: I1008 18:17:59.847898 4859 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T18:17:59Z","lastTransitionTime":"2025-10-08T18:17:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 18:17:59 crc kubenswrapper[4859]: I1008 18:17:59.892009 4859 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-4g8gf_1dff864e-b75d-4e0b-b182-75f710eac8df/ovnkube-controller/2.log" Oct 08 18:17:59 crc kubenswrapper[4859]: I1008 18:17:59.950601 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 18:17:59 crc kubenswrapper[4859]: I1008 18:17:59.950712 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 18:17:59 crc kubenswrapper[4859]: I1008 18:17:59.950738 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 18:17:59 crc kubenswrapper[4859]: I1008 18:17:59.950775 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 18:17:59 crc kubenswrapper[4859]: I1008 18:17:59.950796 4859 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T18:17:59Z","lastTransitionTime":"2025-10-08T18:17:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 18:18:00 crc kubenswrapper[4859]: I1008 18:18:00.053095 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 18:18:00 crc kubenswrapper[4859]: I1008 18:18:00.053206 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 18:18:00 crc kubenswrapper[4859]: I1008 18:18:00.053232 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 18:18:00 crc kubenswrapper[4859]: I1008 18:18:00.053266 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 18:18:00 crc kubenswrapper[4859]: I1008 18:18:00.053289 4859 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T18:18:00Z","lastTransitionTime":"2025-10-08T18:18:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 18:18:00 crc kubenswrapper[4859]: I1008 18:18:00.156782 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 18:18:00 crc kubenswrapper[4859]: I1008 18:18:00.156895 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 18:18:00 crc kubenswrapper[4859]: I1008 18:18:00.156917 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 18:18:00 crc kubenswrapper[4859]: I1008 18:18:00.157368 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 18:18:00 crc kubenswrapper[4859]: I1008 18:18:00.157424 4859 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T18:18:00Z","lastTransitionTime":"2025-10-08T18:18:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 18:18:00 crc kubenswrapper[4859]: I1008 18:18:00.261483 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 18:18:00 crc kubenswrapper[4859]: I1008 18:18:00.261541 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 18:18:00 crc kubenswrapper[4859]: I1008 18:18:00.261560 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 18:18:00 crc kubenswrapper[4859]: I1008 18:18:00.261586 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 18:18:00 crc kubenswrapper[4859]: I1008 18:18:00.261603 4859 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T18:18:00Z","lastTransitionTime":"2025-10-08T18:18:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 18:18:00 crc kubenswrapper[4859]: I1008 18:18:00.365172 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 18:18:00 crc kubenswrapper[4859]: I1008 18:18:00.365227 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 18:18:00 crc kubenswrapper[4859]: I1008 18:18:00.365244 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 18:18:00 crc kubenswrapper[4859]: I1008 18:18:00.365268 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 18:18:00 crc kubenswrapper[4859]: I1008 18:18:00.365285 4859 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T18:18:00Z","lastTransitionTime":"2025-10-08T18:18:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 18:18:00 crc kubenswrapper[4859]: I1008 18:18:00.468307 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 18:18:00 crc kubenswrapper[4859]: I1008 18:18:00.468380 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 18:18:00 crc kubenswrapper[4859]: I1008 18:18:00.468393 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 18:18:00 crc kubenswrapper[4859]: I1008 18:18:00.468415 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 18:18:00 crc kubenswrapper[4859]: I1008 18:18:00.468428 4859 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T18:18:00Z","lastTransitionTime":"2025-10-08T18:18:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 18:18:00 crc kubenswrapper[4859]: I1008 18:18:00.487648 4859 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-08T18:17:31Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T18:17:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T18:17:31Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c6d15905694b6bbd4999d49dc8f7a53eb6bfaa7adb59efb0d7e4973c21d1c7be\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T18:17:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T18:18:00Z is after 2025-08-24T17:21:41Z" Oct 08 18:18:00 crc kubenswrapper[4859]: I1008 18:18:00.504419 4859 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-08T18:17:33Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T18:17:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T18:17:33Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5d99e071bd083914418d9c128f38955f26295c08b1520f2dc774c99b1c843693\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T18:17:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T18:18:00Z is after 2025-08-24T17:21:41Z" Oct 08 18:18:00 crc kubenswrapper[4859]: I1008 18:18:00.519401 4859 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-08T18:17:31Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T18:17:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T18:17:31Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://99e2d51bd7757498dae60949eea6d732b5f7b481fa00f808dd5308a530a1daf1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T18:17:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e32aaba93699944665660ae41e9b5779178c1354c31e15ad8188bf8c3c8fa500\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T18:17:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T18:18:00Z is after 2025-08-24T17:21:41Z" Oct 08 18:18:00 crc kubenswrapper[4859]: I1008 18:18:00.535852 4859 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-08T18:17:29Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T18:17:29Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T18:18:00Z is after 2025-08-24T17:21:41Z" Oct 08 18:18:00 crc kubenswrapper[4859]: I1008 18:18:00.560764 4859 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-82s52" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b23a6a6c-9d92-4e7b-840e-55cfda873a2d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:17:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:17:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:17:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:17:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://71302c766485b9f55743d85f97440ff578cd44ba1e5a6f5b1c024c0e808072d2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T18:17:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6vx22\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c23ea3580998adf90621a0155c20ee4bd590c512629812174e95c28857bcfef9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T18:17:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6vx22\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T18:17:31Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-82s52\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T18:18:00Z is after 2025-08-24T17:21:41Z" Oct 08 18:18:00 crc kubenswrapper[4859]: I1008 18:18:00.570371 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 18:18:00 crc kubenswrapper[4859]: I1008 18:18:00.570414 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 18:18:00 crc kubenswrapper[4859]: I1008 18:18:00.570427 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 18:18:00 crc kubenswrapper[4859]: I1008 18:18:00.570447 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 18:18:00 crc kubenswrapper[4859]: I1008 18:18:00.570462 4859 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T18:18:00Z","lastTransitionTime":"2025-10-08T18:18:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 18:18:00 crc kubenswrapper[4859]: I1008 18:18:00.586162 4859 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-4g8gf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1dff864e-b75d-4e0b-b182-75f710eac8df\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:17:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:17:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:17:32Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:17:32Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://768ca274dedcc6847fc21d15234d3665fa9ccae371b5b6a882cfcb9b182f527d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T18:17:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s8jnn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://669f5f3c6d3c1ec0c4b4870e090b0a853adc01727c23383e595715b476c32668\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T18:17:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s8jnn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b8009a42cc4c9a816457a0678bdb2f7b7c53d0c8604e007a6fb141f66d06a516\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T18:17:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s8jnn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://676511a11ed4102084b57ebdd5f3fda4969741b2bc8443b415c4bf94f39ae4c9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T18:17:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s8jnn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4a74c1f0d21c90cdd7e776ee14e1f48757a04e2220fbd1f884449181ec6d1c3a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T18:17:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s8jnn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6ddc0e842bb8f81663406070023b0e5ed9cc99c98db0123b1c84f20da59fb5b5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T18:17:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s8jnn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c646f0e0dd8e7f74125b687f0cb4114b1f34667e5a6325f394f0208d7fb89df6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2ce53a49f210c2d6e0001ee4280916611bbd82b4fc21a99c59954ed3d7570b96\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-10-08T18:17:43Z\\\",\\\"message\\\":\\\"cert-secret-name:metrics-tls service.beta.openshift.io/serving-cert-signed-by:openshift-service-serving-signer@1740288168] [{config.openshift.io/v1 ClusterVersion version 9101b518-476b-4eea-8fa6-69b0534e5caa 0xc00695d3d7 \\\\u003cnil\\\\u003e}] [] []},Spec:ServiceSpec{Ports:[]ServicePort{ServicePort{Name:metrics,Protocol:TCP,Port:9393,TargetPort:{1 0 metrics},NodePort:0,AppProtocol:nil,},},Selector:map[string]string{name: ingress-operator,},ClusterIP:10.217.5.244,Type:ClusterIP,ExternalIPs:[],SessionAffinity:None,LoadBalancerIP:,LoadBalancerSourceRanges:[],ExternalName:,ExternalTrafficPolicy:,HealthCheckNodePort:0,PublishNotReadyAddresses:false,SessionAffinityConfig:nil,IPFamilyPolicy:*SingleStack,ClusterIPs:[10.217.5.244],IPFamilies:[IPv4],AllocateLoadBalancerNodePorts:nil,LoadBalancerClass:nil,InternalTrafficPolicy:*Cluster,TrafficDistribution:nil,},Status:ServiceStatus{LoadBalancer:LoadBalancerStatus{Ingress:[]LoadBalancerIngress{},},Conditions:[]Condition{},},}\\\\nI1008 18:17:43.764981 6316 base_network_controller_pods.go:477] [default/openshift-network-diagnostics/network-check-target-xd92c] creating logical port openshift-network-diagnostics_network-check-target-xd92c for pod on switch crc\\\\nF1008 18:17:43.764990 6316 ovnkube.go:137] failed to run ovnkube: [failed to start network controller: failed to start default network controller: unable to create\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-08T18:17:42Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c646f0e0dd8e7f74125b687f0cb4114b1f34667e5a6325f394f0208d7fb89df6\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-10-08T18:17:58Z\\\",\\\"message\\\":\\\"e *v1.Pod retry setup to complete in iterateRetryResources\\\\nI1008 18:17:58.424504 6520 obj_retry.go:365] Adding new object: *v1.Pod openshift-machine-config-operator/machine-config-daemon-82s52\\\\nI1008 18:17:58.424529 6520 ovn.go:134] Ensuring zone local for Pod openshift-machine-config-operator/machine-config-daemon-82s52 in node crc\\\\nI1008 18:17:58.424539 6520 obj_retry.go:303] Retry object setup: *v1.Pod openshift-network-diagnostics/network-check-source-55646444c4-trplf\\\\nI1008 18:17:58.424549 6520 obj_retry.go:365] Adding new object: *v1.Pod openshift-network-diagnostics/network-check-source-55646444c4-trplf\\\\nI1008 18:17:58.424549 6520 obj_retry.go:386] Retry successful for *v1.Pod openshift-machine-config-operator/machine-config-daemon-82s52 after 0 failed attempt(s)\\\\nI1008 18:17:58.424559 6520 ovn.go:134] Ensuring zone local for Pod openshift-network-diagnostics/network-check-source-55646444c4-trplf in node crc\\\\nF1008 18:17:58.424552 6520 ovnkube.go:137] failed to run ovnkube: [failed to start network controller: failed to start default network controller: unable to create admin network policy controller, err: could not add Event Handler for anpInformer during admin network policy controller initialization, handler {0x1fcc6e0 0x1fcc3c0 0x1fcc360} was not added to shared informer because it has stopped already, failed to star\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-08T18:17:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s8jnn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2c67d2aed7126d76989e057d8e2d8e188a8902b2f3fe26ed6d1d45d3b013d897\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T18:17:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s8jnn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7a211d3da02b173867ff3a912eecd3b379483c7cc77e6b91122edd1ffd30fd53\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7a211d3da02b173867ff3a912eecd3b379483c7cc77e6b91122edd1ffd30fd53\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-08T18:17:32Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-08T18:17:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s8jnn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T18:17:32Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-4g8gf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T18:18:00Z is after 2025-08-24T17:21:41Z" Oct 08 18:18:00 crc kubenswrapper[4859]: I1008 18:18:00.598120 4859 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-4b4sz" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2e53cd9b-64f7-4e07-8b96-fb77847c9ac6\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:17:45Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:17:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:17:45Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:17:45Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kxgbz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kxgbz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T18:17:45Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-4b4sz\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T18:18:00Z is after 2025-08-24T17:21:41Z" Oct 08 18:18:00 crc kubenswrapper[4859]: I1008 18:18:00.613437 4859 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"63a623de-5608-4228-b1c2-77188c276986\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:17:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:17:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:17:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:17:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:17:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d4786a8fb8abd96fb22e60733c974f599017407d589ed815fb8541fd3b64f696\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T18:17:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://183d481502ed5108d5e19e483b714c289f355ea3f8b0891883e5335b802c83b6\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T18:17:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ec848c46fd5dde0469b00d3cab14ba4a652533a265490fadbee7a8b8843ef249\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T18:17:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4fe1828b4d63ff8106bfcd15e1dc1c1bd1ba7f77777cdb23ff1b291d0418774c\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2e700b2a20c81b414617f5b5c018a738278a3fa44e6c6d14b2170051d0725431\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-08T18:17:24Z\\\",\\\"message\\\":\\\"W1008 18:17:13.670625 1 cmd.go:257] Using insecure, self-signed certificates\\\\nI1008 18:17:13.671202 1 crypto.go:601] Generating new CA for check-endpoints-signer@1759947433 cert, and key in /tmp/serving-cert-3867218267/serving-signer.crt, /tmp/serving-cert-3867218267/serving-signer.key\\\\nI1008 18:17:13.944463 1 observer_polling.go:159] Starting file observer\\\\nW1008 18:17:13.951958 1 builder.go:272] unable to get owner reference (falling back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": dial tcp [::1]:6443: connect: connection refused\\\\nI1008 18:17:13.952088 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1008 18:17:13.956755 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3867218267/tls.crt::/tmp/serving-cert-3867218267/tls.key\\\\\\\"\\\\nF1008 18:17:24.290991 1 cmd.go:182] error initializing delegating authentication: unable to load configmap based request-header-client-ca-file: Get \\\\\\\"https://localhost:6443/api/v1/namespaces/kube-system/configmaps/extension-apiserver-authentication\\\\\\\": net/http: TLS handshake timeout\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-08T18:17:13Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T18:17:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c65b66f62d325ca10ef8d5d31f04529ccf4e691cf3303d231e7a414fb400e4d6\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T18:17:13Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://58c7bba4ace28215b062653a960e561b1dc1f890f2baf450145638a065af038b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://58c7bba4ace28215b062653a960e561b1dc1f890f2baf450145638a065af038b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-08T18:17:11Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-08T18:17:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T18:17:10Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T18:18:00Z is after 2025-08-24T17:21:41Z" Oct 08 18:18:00 crc kubenswrapper[4859]: I1008 18:18:00.627655 4859 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4efc16fb-e679-464d-8b3c-0f46f082593e\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:17:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:17:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:17:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:17:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:17:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://aa8f8ea895272cd79a08d1c47cd7863a1ec10e21bb60d12aa10862a56c99e4bf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T18:17:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5c4e95d8a712ab219455564539a90286ae48bbc46ddc322a5bfd96c1a8268110\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T18:17:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://bae6ecfd8b446ef04d393d823e5ffe37fe749337c0ca17fde79c5f88d5e31393\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T18:17:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6411d1e7e18d5ceadaaceed5a8a6b20a0cacc55cede793ca9fb4dcad70c64fc6\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T18:17:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T18:17:10Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T18:18:00Z is after 2025-08-24T17:21:41Z" Oct 08 18:18:00 crc kubenswrapper[4859]: I1008 18:18:00.641908 4859 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-08T18:17:29Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T18:17:29Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T18:18:00Z is after 2025-08-24T17:21:41Z" Oct 08 18:18:00 crc kubenswrapper[4859]: I1008 18:18:00.655143 4859 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-84k8x" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1c24078c-c9e8-4160-883a-f483b4b3dd23\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:17:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:17:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:17:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:17:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8661006834d0f50eaada7394369584ee95fc61ff54437eec35d755526afc5b48\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T18:17:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hk5xb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://18b291eb3344fb7c00ef5679dca9a6f7ed9d08cea684a803ff7dad2e8a18c585\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://18b291eb3344fb7c00ef5679dca9a6f7ed9d08cea684a803ff7dad2e8a18c585\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-08T18:17:32Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-08T18:17:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hk5xb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://160163f1a91d981583452db3c7a9244182d1c8ced45749a61b721d0d45c1dfe4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://160163f1a91d981583452db3c7a9244182d1c8ced45749a61b721d0d45c1dfe4\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-08T18:17:34Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-08T18:17:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hk5xb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://51e9b53f3847ac53f378ad83870debf8e71518d4edbc7fc0dc66644818e2103e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://51e9b53f3847ac53f378ad83870debf8e71518d4edbc7fc0dc66644818e2103e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-08T18:17:34Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-08T18:17:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hk5xb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1ff118919739759ac01b5cbe3240e6673db54eaa16353c609d670071ed7964cb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1ff118919739759ac01b5cbe3240e6673db54eaa16353c609d670071ed7964cb\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-08T18:17:35Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-08T18:17:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hk5xb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8522e803ff56eb57cd4d9fd036e0c3fb16240ea21d4a094c1db7aa7d50bfe2c2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8522e803ff56eb57cd4d9fd036e0c3fb16240ea21d4a094c1db7aa7d50bfe2c2\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-08T18:17:37Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-08T18:17:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hk5xb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://be61056edf515e037403ca2a0bfd5feb9b2537dd42212b58e133b23b294f327b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://be61056edf515e037403ca2a0bfd5feb9b2537dd42212b58e133b23b294f327b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-08T18:17:38Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-08T18:17:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hk5xb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T18:17:31Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-84k8x\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T18:18:00Z is after 2025-08-24T17:21:41Z" Oct 08 18:18:00 crc kubenswrapper[4859]: I1008 18:18:00.667895 4859 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-8plkg" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"92b2d7de-31cc-4a91-95a5-ed5ea964b028\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:17:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:17:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:17:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:17:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fa00b8345d40b74a917ddc23cc1b2c5c684a9ca61b6119d090c92cc1439b71c4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T18:17:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q9d7m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T18:17:31Z\\\"}}\" for pod \"openshift-multus\"/\"multus-8plkg\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T18:18:00Z is after 2025-08-24T17:21:41Z" Oct 08 18:18:00 crc kubenswrapper[4859]: I1008 18:18:00.672824 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 18:18:00 crc kubenswrapper[4859]: I1008 18:18:00.672886 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 18:18:00 crc kubenswrapper[4859]: I1008 18:18:00.672897 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 18:18:00 crc kubenswrapper[4859]: I1008 18:18:00.672913 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 18:18:00 crc kubenswrapper[4859]: I1008 18:18:00.672924 4859 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T18:18:00Z","lastTransitionTime":"2025-10-08T18:18:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 18:18:00 crc kubenswrapper[4859]: I1008 18:18:00.680721 4859 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-btjmq" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b8605f7-fed7-493f-9b4f-6a6358adf907\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:17:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:17:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:17:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:17:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://14e9b6c4a4403f36adb78fcf1135a620ca6fc56ed974c9639dcdfc59e6917559\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T18:17:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xspg4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T18:17:35Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-btjmq\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T18:18:00Z is after 2025-08-24T17:21:41Z" Oct 08 18:18:00 crc kubenswrapper[4859]: I1008 18:18:00.691813 4859 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-g4t79" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3a92dbf1-21a8-4095-96bb-2d701d7f3d56\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:17:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:17:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:17:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:17:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f8772a25b6e605efe9a062dc32dbbc2cfe32b4fe164b4e84fbf3bcffab744179\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T18:17:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sv79p\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://92cb9986c09a7f9e1e552e3e3ac1f74f04128241e00ddee38882d2ba5c005a67\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T18:17:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sv79p\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T18:17:45Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-g4t79\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T18:18:00Z is after 2025-08-24T17:21:41Z" Oct 08 18:18:00 crc kubenswrapper[4859]: I1008 18:18:00.705811 4859 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-08T18:17:29Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T18:17:29Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T18:18:00Z is after 2025-08-24T17:21:41Z" Oct 08 18:18:00 crc kubenswrapper[4859]: I1008 18:18:00.715458 4859 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-btw28" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"97bc5598-3ead-4300-a938-3aebbff1e8c6\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:17:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:17:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:17:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:17:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d6d4d6ba771cb5fde637449bba05660f634df7e833a0c8dad2b2c7f3a422bb2b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T18:17:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qcxsd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T18:17:31Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-btw28\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T18:18:00Z is after 2025-08-24T17:21:41Z" Oct 08 18:18:00 crc kubenswrapper[4859]: I1008 18:18:00.775881 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 18:18:00 crc kubenswrapper[4859]: I1008 18:18:00.775923 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 18:18:00 crc kubenswrapper[4859]: I1008 18:18:00.775931 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 18:18:00 crc kubenswrapper[4859]: I1008 18:18:00.775947 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 18:18:00 crc kubenswrapper[4859]: I1008 18:18:00.775957 4859 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T18:18:00Z","lastTransitionTime":"2025-10-08T18:18:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 18:18:00 crc kubenswrapper[4859]: I1008 18:18:00.878824 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 18:18:00 crc kubenswrapper[4859]: I1008 18:18:00.878882 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 18:18:00 crc kubenswrapper[4859]: I1008 18:18:00.878896 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 18:18:00 crc kubenswrapper[4859]: I1008 18:18:00.878917 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 18:18:00 crc kubenswrapper[4859]: I1008 18:18:00.878931 4859 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T18:18:00Z","lastTransitionTime":"2025-10-08T18:18:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 18:18:00 crc kubenswrapper[4859]: I1008 18:18:00.957717 4859 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Oct 08 18:18:00 crc kubenswrapper[4859]: I1008 18:18:00.971659 4859 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-scheduler/openshift-kube-scheduler-crc"] Oct 08 18:18:00 crc kubenswrapper[4859]: I1008 18:18:00.982561 4859 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-84k8x" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1c24078c-c9e8-4160-883a-f483b4b3dd23\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:17:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:17:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:17:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:17:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8661006834d0f50eaada7394369584ee95fc61ff54437eec35d755526afc5b48\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T18:17:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hk5xb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://18b291eb3344fb7c00ef5679dca9a6f7ed9d08cea684a803ff7dad2e8a18c585\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://18b291eb3344fb7c00ef5679dca9a6f7ed9d08cea684a803ff7dad2e8a18c585\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-08T18:17:32Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-08T18:17:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hk5xb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://160163f1a91d981583452db3c7a9244182d1c8ced45749a61b721d0d45c1dfe4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://160163f1a91d981583452db3c7a9244182d1c8ced45749a61b721d0d45c1dfe4\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-08T18:17:34Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-08T18:17:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hk5xb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://51e9b53f3847ac53f378ad83870debf8e71518d4edbc7fc0dc66644818e2103e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://51e9b53f3847ac53f378ad83870debf8e71518d4edbc7fc0dc66644818e2103e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-08T18:17:34Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-08T18:17:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hk5xb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1ff118919739759ac01b5cbe3240e6673db54eaa16353c609d670071ed7964cb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1ff118919739759ac01b5cbe3240e6673db54eaa16353c609d670071ed7964cb\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-08T18:17:35Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-08T18:17:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hk5xb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8522e803ff56eb57cd4d9fd036e0c3fb16240ea21d4a094c1db7aa7d50bfe2c2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8522e803ff56eb57cd4d9fd036e0c3fb16240ea21d4a094c1db7aa7d50bfe2c2\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-08T18:17:37Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-08T18:17:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hk5xb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://be61056edf515e037403ca2a0bfd5feb9b2537dd42212b58e133b23b294f327b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://be61056edf515e037403ca2a0bfd5feb9b2537dd42212b58e133b23b294f327b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-08T18:17:38Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-08T18:17:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hk5xb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T18:17:31Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-84k8x\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T18:18:00Z is after 2025-08-24T17:21:41Z" Oct 08 18:18:00 crc kubenswrapper[4859]: I1008 18:18:00.983295 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 18:18:00 crc kubenswrapper[4859]: I1008 18:18:00.983331 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 18:18:00 crc kubenswrapper[4859]: I1008 18:18:00.983345 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 18:18:00 crc kubenswrapper[4859]: I1008 18:18:00.983364 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 18:18:00 crc kubenswrapper[4859]: I1008 18:18:00.983380 4859 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T18:18:00Z","lastTransitionTime":"2025-10-08T18:18:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 18:18:01 crc kubenswrapper[4859]: I1008 18:18:01.005650 4859 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4efc16fb-e679-464d-8b3c-0f46f082593e\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:17:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:17:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:17:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:17:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:17:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://aa8f8ea895272cd79a08d1c47cd7863a1ec10e21bb60d12aa10862a56c99e4bf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T18:17:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5c4e95d8a712ab219455564539a90286ae48bbc46ddc322a5bfd96c1a8268110\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T18:17:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://bae6ecfd8b446ef04d393d823e5ffe37fe749337c0ca17fde79c5f88d5e31393\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T18:17:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6411d1e7e18d5ceadaaceed5a8a6b20a0cacc55cede793ca9fb4dcad70c64fc6\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T18:17:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T18:17:10Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T18:18:01Z is after 2025-08-24T17:21:41Z" Oct 08 18:18:01 crc kubenswrapper[4859]: I1008 18:18:01.021324 4859 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-08T18:17:29Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T18:17:29Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T18:18:01Z is after 2025-08-24T17:21:41Z" Oct 08 18:18:01 crc kubenswrapper[4859]: I1008 18:18:01.034841 4859 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-btw28" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"97bc5598-3ead-4300-a938-3aebbff1e8c6\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:17:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:17:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:17:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:17:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d6d4d6ba771cb5fde637449bba05660f634df7e833a0c8dad2b2c7f3a422bb2b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T18:17:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qcxsd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T18:17:31Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-btw28\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T18:18:01Z is after 2025-08-24T17:21:41Z" Oct 08 18:18:01 crc kubenswrapper[4859]: I1008 18:18:01.051585 4859 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-8plkg" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"92b2d7de-31cc-4a91-95a5-ed5ea964b028\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:17:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:17:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:17:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:17:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fa00b8345d40b74a917ddc23cc1b2c5c684a9ca61b6119d090c92cc1439b71c4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T18:17:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q9d7m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T18:17:31Z\\\"}}\" for pod \"openshift-multus\"/\"multus-8plkg\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T18:18:01Z is after 2025-08-24T17:21:41Z" Oct 08 18:18:01 crc kubenswrapper[4859]: I1008 18:18:01.066417 4859 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-btjmq" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b8605f7-fed7-493f-9b4f-6a6358adf907\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:17:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:17:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:17:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:17:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://14e9b6c4a4403f36adb78fcf1135a620ca6fc56ed974c9639dcdfc59e6917559\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T18:17:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xspg4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T18:17:35Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-btjmq\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T18:18:01Z is after 2025-08-24T17:21:41Z" Oct 08 18:18:01 crc kubenswrapper[4859]: I1008 18:18:01.081559 4859 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-g4t79" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3a92dbf1-21a8-4095-96bb-2d701d7f3d56\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:17:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:17:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:17:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:17:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f8772a25b6e605efe9a062dc32dbbc2cfe32b4fe164b4e84fbf3bcffab744179\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T18:17:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sv79p\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://92cb9986c09a7f9e1e552e3e3ac1f74f04128241e00ddee38882d2ba5c005a67\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T18:17:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sv79p\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T18:17:45Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-g4t79\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T18:18:01Z is after 2025-08-24T17:21:41Z" Oct 08 18:18:01 crc kubenswrapper[4859]: I1008 18:18:01.085899 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 18:18:01 crc kubenswrapper[4859]: I1008 18:18:01.085968 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 18:18:01 crc kubenswrapper[4859]: I1008 18:18:01.085986 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 18:18:01 crc kubenswrapper[4859]: I1008 18:18:01.086012 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 18:18:01 crc kubenswrapper[4859]: I1008 18:18:01.086031 4859 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T18:18:01Z","lastTransitionTime":"2025-10-08T18:18:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 18:18:01 crc kubenswrapper[4859]: I1008 18:18:01.096898 4859 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-08T18:17:29Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T18:17:29Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T18:18:01Z is after 2025-08-24T17:21:41Z" Oct 08 18:18:01 crc kubenswrapper[4859]: I1008 18:18:01.113299 4859 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-08T18:17:31Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T18:17:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T18:17:31Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c6d15905694b6bbd4999d49dc8f7a53eb6bfaa7adb59efb0d7e4973c21d1c7be\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T18:17:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T18:18:01Z is after 2025-08-24T17:21:41Z" Oct 08 18:18:01 crc kubenswrapper[4859]: I1008 18:18:01.123390 4859 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-08T18:17:33Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T18:17:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T18:17:33Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5d99e071bd083914418d9c128f38955f26295c08b1520f2dc774c99b1c843693\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T18:17:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T18:18:01Z is after 2025-08-24T17:21:41Z" Oct 08 18:18:01 crc kubenswrapper[4859]: I1008 18:18:01.138342 4859 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"63a623de-5608-4228-b1c2-77188c276986\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:17:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:17:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:17:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:17:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:17:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d4786a8fb8abd96fb22e60733c974f599017407d589ed815fb8541fd3b64f696\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T18:17:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://183d481502ed5108d5e19e483b714c289f355ea3f8b0891883e5335b802c83b6\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T18:17:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ec848c46fd5dde0469b00d3cab14ba4a652533a265490fadbee7a8b8843ef249\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T18:17:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4fe1828b4d63ff8106bfcd15e1dc1c1bd1ba7f77777cdb23ff1b291d0418774c\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2e700b2a20c81b414617f5b5c018a738278a3fa44e6c6d14b2170051d0725431\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-08T18:17:24Z\\\",\\\"message\\\":\\\"W1008 18:17:13.670625 1 cmd.go:257] Using insecure, self-signed certificates\\\\nI1008 18:17:13.671202 1 crypto.go:601] Generating new CA for check-endpoints-signer@1759947433 cert, and key in /tmp/serving-cert-3867218267/serving-signer.crt, /tmp/serving-cert-3867218267/serving-signer.key\\\\nI1008 18:17:13.944463 1 observer_polling.go:159] Starting file observer\\\\nW1008 18:17:13.951958 1 builder.go:272] unable to get owner reference (falling back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": dial tcp [::1]:6443: connect: connection refused\\\\nI1008 18:17:13.952088 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1008 18:17:13.956755 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3867218267/tls.crt::/tmp/serving-cert-3867218267/tls.key\\\\\\\"\\\\nF1008 18:17:24.290991 1 cmd.go:182] error initializing delegating authentication: unable to load configmap based request-header-client-ca-file: Get \\\\\\\"https://localhost:6443/api/v1/namespaces/kube-system/configmaps/extension-apiserver-authentication\\\\\\\": net/http: TLS handshake timeout\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-08T18:17:13Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T18:17:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c65b66f62d325ca10ef8d5d31f04529ccf4e691cf3303d231e7a414fb400e4d6\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T18:17:13Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://58c7bba4ace28215b062653a960e561b1dc1f890f2baf450145638a065af038b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://58c7bba4ace28215b062653a960e561b1dc1f890f2baf450145638a065af038b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-08T18:17:11Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-08T18:17:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T18:17:10Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T18:18:01Z is after 2025-08-24T17:21:41Z" Oct 08 18:18:01 crc kubenswrapper[4859]: I1008 18:18:01.157186 4859 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-08T18:17:31Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T18:17:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T18:17:31Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://99e2d51bd7757498dae60949eea6d732b5f7b481fa00f808dd5308a530a1daf1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T18:17:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e32aaba93699944665660ae41e9b5779178c1354c31e15ad8188bf8c3c8fa500\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T18:17:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T18:18:01Z is after 2025-08-24T17:21:41Z" Oct 08 18:18:01 crc kubenswrapper[4859]: I1008 18:18:01.173257 4859 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-08T18:17:29Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T18:17:29Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T18:18:01Z is after 2025-08-24T17:21:41Z" Oct 08 18:18:01 crc kubenswrapper[4859]: I1008 18:18:01.185957 4859 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-82s52" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b23a6a6c-9d92-4e7b-840e-55cfda873a2d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:17:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:17:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:17:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:17:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://71302c766485b9f55743d85f97440ff578cd44ba1e5a6f5b1c024c0e808072d2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T18:17:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6vx22\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c23ea3580998adf90621a0155c20ee4bd590c512629812174e95c28857bcfef9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T18:17:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6vx22\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T18:17:31Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-82s52\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T18:18:01Z is after 2025-08-24T17:21:41Z" Oct 08 18:18:01 crc kubenswrapper[4859]: I1008 18:18:01.188697 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 18:18:01 crc kubenswrapper[4859]: I1008 18:18:01.188726 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 18:18:01 crc kubenswrapper[4859]: I1008 18:18:01.188734 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 18:18:01 crc kubenswrapper[4859]: I1008 18:18:01.188750 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 18:18:01 crc kubenswrapper[4859]: I1008 18:18:01.188760 4859 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T18:18:01Z","lastTransitionTime":"2025-10-08T18:18:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 18:18:01 crc kubenswrapper[4859]: I1008 18:18:01.205073 4859 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-4g8gf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1dff864e-b75d-4e0b-b182-75f710eac8df\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:17:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:17:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:17:32Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:17:32Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://768ca274dedcc6847fc21d15234d3665fa9ccae371b5b6a882cfcb9b182f527d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T18:17:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s8jnn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://669f5f3c6d3c1ec0c4b4870e090b0a853adc01727c23383e595715b476c32668\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T18:17:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s8jnn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b8009a42cc4c9a816457a0678bdb2f7b7c53d0c8604e007a6fb141f66d06a516\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T18:17:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s8jnn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://676511a11ed4102084b57ebdd5f3fda4969741b2bc8443b415c4bf94f39ae4c9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T18:17:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s8jnn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4a74c1f0d21c90cdd7e776ee14e1f48757a04e2220fbd1f884449181ec6d1c3a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T18:17:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s8jnn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6ddc0e842bb8f81663406070023b0e5ed9cc99c98db0123b1c84f20da59fb5b5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T18:17:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s8jnn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c646f0e0dd8e7f74125b687f0cb4114b1f34667e5a6325f394f0208d7fb89df6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2ce53a49f210c2d6e0001ee4280916611bbd82b4fc21a99c59954ed3d7570b96\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-10-08T18:17:43Z\\\",\\\"message\\\":\\\"cert-secret-name:metrics-tls service.beta.openshift.io/serving-cert-signed-by:openshift-service-serving-signer@1740288168] [{config.openshift.io/v1 ClusterVersion version 9101b518-476b-4eea-8fa6-69b0534e5caa 0xc00695d3d7 \\\\u003cnil\\\\u003e}] [] []},Spec:ServiceSpec{Ports:[]ServicePort{ServicePort{Name:metrics,Protocol:TCP,Port:9393,TargetPort:{1 0 metrics},NodePort:0,AppProtocol:nil,},},Selector:map[string]string{name: ingress-operator,},ClusterIP:10.217.5.244,Type:ClusterIP,ExternalIPs:[],SessionAffinity:None,LoadBalancerIP:,LoadBalancerSourceRanges:[],ExternalName:,ExternalTrafficPolicy:,HealthCheckNodePort:0,PublishNotReadyAddresses:false,SessionAffinityConfig:nil,IPFamilyPolicy:*SingleStack,ClusterIPs:[10.217.5.244],IPFamilies:[IPv4],AllocateLoadBalancerNodePorts:nil,LoadBalancerClass:nil,InternalTrafficPolicy:*Cluster,TrafficDistribution:nil,},Status:ServiceStatus{LoadBalancer:LoadBalancerStatus{Ingress:[]LoadBalancerIngress{},},Conditions:[]Condition{},},}\\\\nI1008 18:17:43.764981 6316 base_network_controller_pods.go:477] [default/openshift-network-diagnostics/network-check-target-xd92c] creating logical port openshift-network-diagnostics_network-check-target-xd92c for pod on switch crc\\\\nF1008 18:17:43.764990 6316 ovnkube.go:137] failed to run ovnkube: [failed to start network controller: failed to start default network controller: unable to create\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-08T18:17:42Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c646f0e0dd8e7f74125b687f0cb4114b1f34667e5a6325f394f0208d7fb89df6\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-10-08T18:17:58Z\\\",\\\"message\\\":\\\"e *v1.Pod retry setup to complete in iterateRetryResources\\\\nI1008 18:17:58.424504 6520 obj_retry.go:365] Adding new object: *v1.Pod openshift-machine-config-operator/machine-config-daemon-82s52\\\\nI1008 18:17:58.424529 6520 ovn.go:134] Ensuring zone local for Pod openshift-machine-config-operator/machine-config-daemon-82s52 in node crc\\\\nI1008 18:17:58.424539 6520 obj_retry.go:303] Retry object setup: *v1.Pod openshift-network-diagnostics/network-check-source-55646444c4-trplf\\\\nI1008 18:17:58.424549 6520 obj_retry.go:365] Adding new object: *v1.Pod openshift-network-diagnostics/network-check-source-55646444c4-trplf\\\\nI1008 18:17:58.424549 6520 obj_retry.go:386] Retry successful for *v1.Pod openshift-machine-config-operator/machine-config-daemon-82s52 after 0 failed attempt(s)\\\\nI1008 18:17:58.424559 6520 ovn.go:134] Ensuring zone local for Pod openshift-network-diagnostics/network-check-source-55646444c4-trplf in node crc\\\\nF1008 18:17:58.424552 6520 ovnkube.go:137] failed to run ovnkube: [failed to start network controller: failed to start default network controller: unable to create admin network policy controller, err: could not add Event Handler for anpInformer during admin network policy controller initialization, handler {0x1fcc6e0 0x1fcc3c0 0x1fcc360} was not added to shared informer because it has stopped already, failed to star\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-08T18:17:57Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s8jnn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2c67d2aed7126d76989e057d8e2d8e188a8902b2f3fe26ed6d1d45d3b013d897\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T18:17:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s8jnn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7a211d3da02b173867ff3a912eecd3b379483c7cc77e6b91122edd1ffd30fd53\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7a211d3da02b173867ff3a912eecd3b379483c7cc77e6b91122edd1ffd30fd53\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-08T18:17:32Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-08T18:17:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s8jnn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T18:17:32Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-4g8gf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T18:18:01Z is after 2025-08-24T17:21:41Z" Oct 08 18:18:01 crc kubenswrapper[4859]: I1008 18:18:01.217554 4859 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-4b4sz" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2e53cd9b-64f7-4e07-8b96-fb77847c9ac6\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:17:45Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:17:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:17:45Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:17:45Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kxgbz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kxgbz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T18:17:45Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-4b4sz\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T18:18:01Z is after 2025-08-24T17:21:41Z" Oct 08 18:18:01 crc kubenswrapper[4859]: I1008 18:18:01.291611 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 18:18:01 crc kubenswrapper[4859]: I1008 18:18:01.291657 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 18:18:01 crc kubenswrapper[4859]: I1008 18:18:01.291670 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 18:18:01 crc kubenswrapper[4859]: I1008 18:18:01.291709 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 18:18:01 crc kubenswrapper[4859]: I1008 18:18:01.291723 4859 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T18:18:01Z","lastTransitionTime":"2025-10-08T18:18:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 18:18:01 crc kubenswrapper[4859]: I1008 18:18:01.395578 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 18:18:01 crc kubenswrapper[4859]: I1008 18:18:01.395653 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 18:18:01 crc kubenswrapper[4859]: I1008 18:18:01.395671 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 18:18:01 crc kubenswrapper[4859]: I1008 18:18:01.395739 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 18:18:01 crc kubenswrapper[4859]: I1008 18:18:01.395761 4859 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T18:18:01Z","lastTransitionTime":"2025-10-08T18:18:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 18:18:01 crc kubenswrapper[4859]: I1008 18:18:01.470306 4859 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-4b4sz" Oct 08 18:18:01 crc kubenswrapper[4859]: E1008 18:18:01.470558 4859 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-4b4sz" podUID="2e53cd9b-64f7-4e07-8b96-fb77847c9ac6" Oct 08 18:18:01 crc kubenswrapper[4859]: I1008 18:18:01.471398 4859 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 08 18:18:01 crc kubenswrapper[4859]: E1008 18:18:01.471779 4859 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 08 18:18:01 crc kubenswrapper[4859]: I1008 18:18:01.471910 4859 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 08 18:18:01 crc kubenswrapper[4859]: I1008 18:18:01.471986 4859 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 08 18:18:01 crc kubenswrapper[4859]: E1008 18:18:01.472318 4859 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 08 18:18:01 crc kubenswrapper[4859]: E1008 18:18:01.472456 4859 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 08 18:18:01 crc kubenswrapper[4859]: I1008 18:18:01.498677 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 18:18:01 crc kubenswrapper[4859]: I1008 18:18:01.498770 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 18:18:01 crc kubenswrapper[4859]: I1008 18:18:01.498795 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 18:18:01 crc kubenswrapper[4859]: I1008 18:18:01.498823 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 18:18:01 crc kubenswrapper[4859]: I1008 18:18:01.498842 4859 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T18:18:01Z","lastTransitionTime":"2025-10-08T18:18:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 18:18:01 crc kubenswrapper[4859]: I1008 18:18:01.602170 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 18:18:01 crc kubenswrapper[4859]: I1008 18:18:01.602595 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 18:18:01 crc kubenswrapper[4859]: I1008 18:18:01.602772 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 18:18:01 crc kubenswrapper[4859]: I1008 18:18:01.602997 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 18:18:01 crc kubenswrapper[4859]: I1008 18:18:01.603211 4859 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T18:18:01Z","lastTransitionTime":"2025-10-08T18:18:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 18:18:01 crc kubenswrapper[4859]: I1008 18:18:01.706965 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 18:18:01 crc kubenswrapper[4859]: I1008 18:18:01.707011 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 18:18:01 crc kubenswrapper[4859]: I1008 18:18:01.707025 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 18:18:01 crc kubenswrapper[4859]: I1008 18:18:01.707045 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 18:18:01 crc kubenswrapper[4859]: I1008 18:18:01.707058 4859 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T18:18:01Z","lastTransitionTime":"2025-10-08T18:18:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 18:18:01 crc kubenswrapper[4859]: I1008 18:18:01.809592 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 18:18:01 crc kubenswrapper[4859]: I1008 18:18:01.809633 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 18:18:01 crc kubenswrapper[4859]: I1008 18:18:01.809643 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 18:18:01 crc kubenswrapper[4859]: I1008 18:18:01.809678 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 18:18:01 crc kubenswrapper[4859]: I1008 18:18:01.809710 4859 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T18:18:01Z","lastTransitionTime":"2025-10-08T18:18:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 18:18:01 crc kubenswrapper[4859]: I1008 18:18:01.862677 4859 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/2e53cd9b-64f7-4e07-8b96-fb77847c9ac6-metrics-certs\") pod \"network-metrics-daemon-4b4sz\" (UID: \"2e53cd9b-64f7-4e07-8b96-fb77847c9ac6\") " pod="openshift-multus/network-metrics-daemon-4b4sz" Oct 08 18:18:01 crc kubenswrapper[4859]: E1008 18:18:01.862857 4859 secret.go:188] Couldn't get secret openshift-multus/metrics-daemon-secret: object "openshift-multus"/"metrics-daemon-secret" not registered Oct 08 18:18:01 crc kubenswrapper[4859]: E1008 18:18:01.862929 4859 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/2e53cd9b-64f7-4e07-8b96-fb77847c9ac6-metrics-certs podName:2e53cd9b-64f7-4e07-8b96-fb77847c9ac6 nodeName:}" failed. No retries permitted until 2025-10-08 18:18:17.862909859 +0000 UTC m=+68.109749238 (durationBeforeRetry 16s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/2e53cd9b-64f7-4e07-8b96-fb77847c9ac6-metrics-certs") pod "network-metrics-daemon-4b4sz" (UID: "2e53cd9b-64f7-4e07-8b96-fb77847c9ac6") : object "openshift-multus"/"metrics-daemon-secret" not registered Oct 08 18:18:01 crc kubenswrapper[4859]: I1008 18:18:01.912830 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 18:18:01 crc kubenswrapper[4859]: I1008 18:18:01.912879 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 18:18:01 crc kubenswrapper[4859]: I1008 18:18:01.912890 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 18:18:01 crc kubenswrapper[4859]: I1008 18:18:01.912908 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 18:18:01 crc kubenswrapper[4859]: I1008 18:18:01.912920 4859 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T18:18:01Z","lastTransitionTime":"2025-10-08T18:18:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 18:18:02 crc kubenswrapper[4859]: I1008 18:18:02.015473 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 18:18:02 crc kubenswrapper[4859]: I1008 18:18:02.015532 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 18:18:02 crc kubenswrapper[4859]: I1008 18:18:02.015545 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 18:18:02 crc kubenswrapper[4859]: I1008 18:18:02.015563 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 18:18:02 crc kubenswrapper[4859]: I1008 18:18:02.015573 4859 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T18:18:02Z","lastTransitionTime":"2025-10-08T18:18:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 18:18:02 crc kubenswrapper[4859]: I1008 18:18:02.118598 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 18:18:02 crc kubenswrapper[4859]: I1008 18:18:02.118636 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 18:18:02 crc kubenswrapper[4859]: I1008 18:18:02.118646 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 18:18:02 crc kubenswrapper[4859]: I1008 18:18:02.118662 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 18:18:02 crc kubenswrapper[4859]: I1008 18:18:02.118675 4859 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T18:18:02Z","lastTransitionTime":"2025-10-08T18:18:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 18:18:02 crc kubenswrapper[4859]: I1008 18:18:02.221395 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 18:18:02 crc kubenswrapper[4859]: I1008 18:18:02.221449 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 18:18:02 crc kubenswrapper[4859]: I1008 18:18:02.221462 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 18:18:02 crc kubenswrapper[4859]: I1008 18:18:02.221480 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 18:18:02 crc kubenswrapper[4859]: I1008 18:18:02.221490 4859 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T18:18:02Z","lastTransitionTime":"2025-10-08T18:18:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 18:18:02 crc kubenswrapper[4859]: I1008 18:18:02.329750 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 18:18:02 crc kubenswrapper[4859]: I1008 18:18:02.329803 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 18:18:02 crc kubenswrapper[4859]: I1008 18:18:02.329815 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 18:18:02 crc kubenswrapper[4859]: I1008 18:18:02.329833 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 18:18:02 crc kubenswrapper[4859]: I1008 18:18:02.329846 4859 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T18:18:02Z","lastTransitionTime":"2025-10-08T18:18:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 18:18:02 crc kubenswrapper[4859]: I1008 18:18:02.367750 4859 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 08 18:18:02 crc kubenswrapper[4859]: E1008 18:18:02.368032 4859 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-08 18:18:34.367991385 +0000 UTC m=+84.614830764 (durationBeforeRetry 32s). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 08 18:18:02 crc kubenswrapper[4859]: I1008 18:18:02.432940 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 18:18:02 crc kubenswrapper[4859]: I1008 18:18:02.432991 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 18:18:02 crc kubenswrapper[4859]: I1008 18:18:02.433001 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 18:18:02 crc kubenswrapper[4859]: I1008 18:18:02.433021 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 18:18:02 crc kubenswrapper[4859]: I1008 18:18:02.433033 4859 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T18:18:02Z","lastTransitionTime":"2025-10-08T18:18:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 18:18:02 crc kubenswrapper[4859]: I1008 18:18:02.469092 4859 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 08 18:18:02 crc kubenswrapper[4859]: I1008 18:18:02.469150 4859 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 08 18:18:02 crc kubenswrapper[4859]: I1008 18:18:02.469171 4859 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 08 18:18:02 crc kubenswrapper[4859]: I1008 18:18:02.469191 4859 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 08 18:18:02 crc kubenswrapper[4859]: E1008 18:18:02.469259 4859 configmap.go:193] Couldn't get configMap openshift-network-console/networking-console-plugin: object "openshift-network-console"/"networking-console-plugin" not registered Oct 08 18:18:02 crc kubenswrapper[4859]: E1008 18:18:02.469312 4859 secret.go:188] Couldn't get secret openshift-network-console/networking-console-plugin-cert: object "openshift-network-console"/"networking-console-plugin-cert" not registered Oct 08 18:18:02 crc kubenswrapper[4859]: E1008 18:18:02.469385 4859 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-10-08 18:18:34.469353944 +0000 UTC m=+84.716193403 (durationBeforeRetry 32s). Error: MountVolume.SetUp failed for volume "nginx-conf" (UniqueName: "kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin" not registered Oct 08 18:18:02 crc kubenswrapper[4859]: E1008 18:18:02.469415 4859 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-10-08 18:18:34.469404135 +0000 UTC m=+84.716243624 (durationBeforeRetry 32s). Error: MountVolume.SetUp failed for volume "networking-console-plugin-cert" (UniqueName: "kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin-cert" not registered Oct 08 18:18:02 crc kubenswrapper[4859]: E1008 18:18:02.469426 4859 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Oct 08 18:18:02 crc kubenswrapper[4859]: E1008 18:18:02.469454 4859 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Oct 08 18:18:02 crc kubenswrapper[4859]: E1008 18:18:02.469471 4859 projected.go:194] Error preparing data for projected volume kube-api-access-cqllr for pod openshift-network-diagnostics/network-check-target-xd92c: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Oct 08 18:18:02 crc kubenswrapper[4859]: E1008 18:18:02.469517 4859 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr podName:3b6479f0-333b-4a96-9adf-2099afdc2447 nodeName:}" failed. No retries permitted until 2025-10-08 18:18:34.469507978 +0000 UTC m=+84.716347357 (durationBeforeRetry 32s). Error: MountVolume.SetUp failed for volume "kube-api-access-cqllr" (UniqueName: "kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr") pod "network-check-target-xd92c" (UID: "3b6479f0-333b-4a96-9adf-2099afdc2447") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Oct 08 18:18:02 crc kubenswrapper[4859]: E1008 18:18:02.469522 4859 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Oct 08 18:18:02 crc kubenswrapper[4859]: E1008 18:18:02.469603 4859 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Oct 08 18:18:02 crc kubenswrapper[4859]: E1008 18:18:02.469622 4859 projected.go:194] Error preparing data for projected volume kube-api-access-s2dwl for pod openshift-network-diagnostics/network-check-source-55646444c4-trplf: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Oct 08 18:18:02 crc kubenswrapper[4859]: E1008 18:18:02.469777 4859 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl podName:9d751cbb-f2e2-430d-9754-c882a5e924a5 nodeName:}" failed. No retries permitted until 2025-10-08 18:18:34.469750675 +0000 UTC m=+84.716590234 (durationBeforeRetry 32s). Error: MountVolume.SetUp failed for volume "kube-api-access-s2dwl" (UniqueName: "kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl") pod "network-check-source-55646444c4-trplf" (UID: "9d751cbb-f2e2-430d-9754-c882a5e924a5") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Oct 08 18:18:02 crc kubenswrapper[4859]: I1008 18:18:02.535666 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 18:18:02 crc kubenswrapper[4859]: I1008 18:18:02.535748 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 18:18:02 crc kubenswrapper[4859]: I1008 18:18:02.535765 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 18:18:02 crc kubenswrapper[4859]: I1008 18:18:02.535790 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 18:18:02 crc kubenswrapper[4859]: I1008 18:18:02.535805 4859 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T18:18:02Z","lastTransitionTime":"2025-10-08T18:18:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 18:18:02 crc kubenswrapper[4859]: I1008 18:18:02.577824 4859 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ovn-kubernetes/ovnkube-node-4g8gf" Oct 08 18:18:02 crc kubenswrapper[4859]: I1008 18:18:02.578870 4859 scope.go:117] "RemoveContainer" containerID="c646f0e0dd8e7f74125b687f0cb4114b1f34667e5a6325f394f0208d7fb89df6" Oct 08 18:18:02 crc kubenswrapper[4859]: E1008 18:18:02.579258 4859 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ovnkube-controller\" with CrashLoopBackOff: \"back-off 20s restarting failed container=ovnkube-controller pod=ovnkube-node-4g8gf_openshift-ovn-kubernetes(1dff864e-b75d-4e0b-b182-75f710eac8df)\"" pod="openshift-ovn-kubernetes/ovnkube-node-4g8gf" podUID="1dff864e-b75d-4e0b-b182-75f710eac8df" Oct 08 18:18:02 crc kubenswrapper[4859]: I1008 18:18:02.594211 4859 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-4b4sz" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2e53cd9b-64f7-4e07-8b96-fb77847c9ac6\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:17:45Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:17:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:17:45Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:17:45Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kxgbz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kxgbz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T18:17:45Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-4b4sz\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T18:18:02Z is after 2025-08-24T17:21:41Z" Oct 08 18:18:02 crc kubenswrapper[4859]: I1008 18:18:02.608307 4859 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"63a623de-5608-4228-b1c2-77188c276986\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:17:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:17:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:17:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:17:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:17:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d4786a8fb8abd96fb22e60733c974f599017407d589ed815fb8541fd3b64f696\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T18:17:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://183d481502ed5108d5e19e483b714c289f355ea3f8b0891883e5335b802c83b6\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T18:17:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ec848c46fd5dde0469b00d3cab14ba4a652533a265490fadbee7a8b8843ef249\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T18:17:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4fe1828b4d63ff8106bfcd15e1dc1c1bd1ba7f77777cdb23ff1b291d0418774c\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2e700b2a20c81b414617f5b5c018a738278a3fa44e6c6d14b2170051d0725431\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-08T18:17:24Z\\\",\\\"message\\\":\\\"W1008 18:17:13.670625 1 cmd.go:257] Using insecure, self-signed certificates\\\\nI1008 18:17:13.671202 1 crypto.go:601] Generating new CA for check-endpoints-signer@1759947433 cert, and key in /tmp/serving-cert-3867218267/serving-signer.crt, /tmp/serving-cert-3867218267/serving-signer.key\\\\nI1008 18:17:13.944463 1 observer_polling.go:159] Starting file observer\\\\nW1008 18:17:13.951958 1 builder.go:272] unable to get owner reference (falling back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": dial tcp [::1]:6443: connect: connection refused\\\\nI1008 18:17:13.952088 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1008 18:17:13.956755 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3867218267/tls.crt::/tmp/serving-cert-3867218267/tls.key\\\\\\\"\\\\nF1008 18:17:24.290991 1 cmd.go:182] error initializing delegating authentication: unable to load configmap based request-header-client-ca-file: Get \\\\\\\"https://localhost:6443/api/v1/namespaces/kube-system/configmaps/extension-apiserver-authentication\\\\\\\": net/http: TLS handshake timeout\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-08T18:17:13Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T18:17:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c65b66f62d325ca10ef8d5d31f04529ccf4e691cf3303d231e7a414fb400e4d6\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T18:17:13Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://58c7bba4ace28215b062653a960e561b1dc1f890f2baf450145638a065af038b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://58c7bba4ace28215b062653a960e561b1dc1f890f2baf450145638a065af038b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-08T18:17:11Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-08T18:17:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T18:17:10Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T18:18:02Z is after 2025-08-24T17:21:41Z" Oct 08 18:18:02 crc kubenswrapper[4859]: I1008 18:18:02.621644 4859 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-08T18:17:31Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T18:17:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T18:17:31Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://99e2d51bd7757498dae60949eea6d732b5f7b481fa00f808dd5308a530a1daf1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T18:17:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e32aaba93699944665660ae41e9b5779178c1354c31e15ad8188bf8c3c8fa500\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T18:17:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T18:18:02Z is after 2025-08-24T17:21:41Z" Oct 08 18:18:02 crc kubenswrapper[4859]: I1008 18:18:02.637727 4859 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-08T18:17:29Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T18:17:29Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T18:18:02Z is after 2025-08-24T17:21:41Z" Oct 08 18:18:02 crc kubenswrapper[4859]: I1008 18:18:02.639211 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 18:18:02 crc kubenswrapper[4859]: I1008 18:18:02.639310 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 18:18:02 crc kubenswrapper[4859]: I1008 18:18:02.639338 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 18:18:02 crc kubenswrapper[4859]: I1008 18:18:02.639377 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 18:18:02 crc kubenswrapper[4859]: I1008 18:18:02.639596 4859 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T18:18:02Z","lastTransitionTime":"2025-10-08T18:18:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 18:18:02 crc kubenswrapper[4859]: I1008 18:18:02.652921 4859 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-82s52" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b23a6a6c-9d92-4e7b-840e-55cfda873a2d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:17:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:17:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:17:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:17:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://71302c766485b9f55743d85f97440ff578cd44ba1e5a6f5b1c024c0e808072d2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T18:17:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6vx22\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c23ea3580998adf90621a0155c20ee4bd590c512629812174e95c28857bcfef9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T18:17:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6vx22\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T18:17:31Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-82s52\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T18:18:02Z is after 2025-08-24T17:21:41Z" Oct 08 18:18:02 crc kubenswrapper[4859]: I1008 18:18:02.741061 4859 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-4g8gf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1dff864e-b75d-4e0b-b182-75f710eac8df\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:17:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:17:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:17:32Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:17:32Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://768ca274dedcc6847fc21d15234d3665fa9ccae371b5b6a882cfcb9b182f527d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T18:17:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s8jnn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://669f5f3c6d3c1ec0c4b4870e090b0a853adc01727c23383e595715b476c32668\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T18:17:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s8jnn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b8009a42cc4c9a816457a0678bdb2f7b7c53d0c8604e007a6fb141f66d06a516\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T18:17:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s8jnn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://676511a11ed4102084b57ebdd5f3fda4969741b2bc8443b415c4bf94f39ae4c9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T18:17:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s8jnn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4a74c1f0d21c90cdd7e776ee14e1f48757a04e2220fbd1f884449181ec6d1c3a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T18:17:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s8jnn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6ddc0e842bb8f81663406070023b0e5ed9cc99c98db0123b1c84f20da59fb5b5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T18:17:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s8jnn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c646f0e0dd8e7f74125b687f0cb4114b1f34667e5a6325f394f0208d7fb89df6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c646f0e0dd8e7f74125b687f0cb4114b1f34667e5a6325f394f0208d7fb89df6\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-10-08T18:17:58Z\\\",\\\"message\\\":\\\"e *v1.Pod retry setup to complete in iterateRetryResources\\\\nI1008 18:17:58.424504 6520 obj_retry.go:365] Adding new object: *v1.Pod openshift-machine-config-operator/machine-config-daemon-82s52\\\\nI1008 18:17:58.424529 6520 ovn.go:134] Ensuring zone local for Pod openshift-machine-config-operator/machine-config-daemon-82s52 in node crc\\\\nI1008 18:17:58.424539 6520 obj_retry.go:303] Retry object setup: *v1.Pod openshift-network-diagnostics/network-check-source-55646444c4-trplf\\\\nI1008 18:17:58.424549 6520 obj_retry.go:365] Adding new object: *v1.Pod openshift-network-diagnostics/network-check-source-55646444c4-trplf\\\\nI1008 18:17:58.424549 6520 obj_retry.go:386] Retry successful for *v1.Pod openshift-machine-config-operator/machine-config-daemon-82s52 after 0 failed attempt(s)\\\\nI1008 18:17:58.424559 6520 ovn.go:134] Ensuring zone local for Pod openshift-network-diagnostics/network-check-source-55646444c4-trplf in node crc\\\\nF1008 18:17:58.424552 6520 ovnkube.go:137] failed to run ovnkube: [failed to start network controller: failed to start default network controller: unable to create admin network policy controller, err: could not add Event Handler for anpInformer during admin network policy controller initialization, handler {0x1fcc6e0 0x1fcc3c0 0x1fcc360} was not added to shared informer because it has stopped already, failed to star\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-08T18:17:57Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 20s restarting failed container=ovnkube-controller pod=ovnkube-node-4g8gf_openshift-ovn-kubernetes(1dff864e-b75d-4e0b-b182-75f710eac8df)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s8jnn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2c67d2aed7126d76989e057d8e2d8e188a8902b2f3fe26ed6d1d45d3b013d897\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T18:17:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s8jnn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7a211d3da02b173867ff3a912eecd3b379483c7cc77e6b91122edd1ffd30fd53\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7a211d3da02b173867ff3a912eecd3b379483c7cc77e6b91122edd1ffd30fd53\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-08T18:17:32Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-08T18:17:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s8jnn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T18:17:32Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-4g8gf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T18:18:02Z is after 2025-08-24T17:21:41Z" Oct 08 18:18:02 crc kubenswrapper[4859]: I1008 18:18:02.742951 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 18:18:02 crc kubenswrapper[4859]: I1008 18:18:02.743021 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 18:18:02 crc kubenswrapper[4859]: I1008 18:18:02.743040 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 18:18:02 crc kubenswrapper[4859]: I1008 18:18:02.743065 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 18:18:02 crc kubenswrapper[4859]: I1008 18:18:02.743099 4859 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T18:18:02Z","lastTransitionTime":"2025-10-08T18:18:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 18:18:02 crc kubenswrapper[4859]: I1008 18:18:02.757869 4859 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4efc16fb-e679-464d-8b3c-0f46f082593e\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:17:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:17:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:17:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:17:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:17:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://aa8f8ea895272cd79a08d1c47cd7863a1ec10e21bb60d12aa10862a56c99e4bf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T18:17:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5c4e95d8a712ab219455564539a90286ae48bbc46ddc322a5bfd96c1a8268110\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T18:17:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://bae6ecfd8b446ef04d393d823e5ffe37fe749337c0ca17fde79c5f88d5e31393\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T18:17:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6411d1e7e18d5ceadaaceed5a8a6b20a0cacc55cede793ca9fb4dcad70c64fc6\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T18:17:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T18:17:10Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T18:18:02Z is after 2025-08-24T17:21:41Z" Oct 08 18:18:02 crc kubenswrapper[4859]: I1008 18:18:02.774938 4859 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-08T18:17:29Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T18:17:29Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T18:18:02Z is after 2025-08-24T17:21:41Z" Oct 08 18:18:02 crc kubenswrapper[4859]: I1008 18:18:02.791933 4859 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-84k8x" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1c24078c-c9e8-4160-883a-f483b4b3dd23\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:17:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:17:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:17:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:17:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8661006834d0f50eaada7394369584ee95fc61ff54437eec35d755526afc5b48\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T18:17:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hk5xb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://18b291eb3344fb7c00ef5679dca9a6f7ed9d08cea684a803ff7dad2e8a18c585\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://18b291eb3344fb7c00ef5679dca9a6f7ed9d08cea684a803ff7dad2e8a18c585\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-08T18:17:32Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-08T18:17:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hk5xb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://160163f1a91d981583452db3c7a9244182d1c8ced45749a61b721d0d45c1dfe4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://160163f1a91d981583452db3c7a9244182d1c8ced45749a61b721d0d45c1dfe4\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-08T18:17:34Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-08T18:17:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hk5xb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://51e9b53f3847ac53f378ad83870debf8e71518d4edbc7fc0dc66644818e2103e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://51e9b53f3847ac53f378ad83870debf8e71518d4edbc7fc0dc66644818e2103e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-08T18:17:34Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-08T18:17:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hk5xb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1ff118919739759ac01b5cbe3240e6673db54eaa16353c609d670071ed7964cb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1ff118919739759ac01b5cbe3240e6673db54eaa16353c609d670071ed7964cb\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-08T18:17:35Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-08T18:17:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hk5xb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8522e803ff56eb57cd4d9fd036e0c3fb16240ea21d4a094c1db7aa7d50bfe2c2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8522e803ff56eb57cd4d9fd036e0c3fb16240ea21d4a094c1db7aa7d50bfe2c2\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-08T18:17:37Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-08T18:17:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hk5xb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://be61056edf515e037403ca2a0bfd5feb9b2537dd42212b58e133b23b294f327b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://be61056edf515e037403ca2a0bfd5feb9b2537dd42212b58e133b23b294f327b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-08T18:17:38Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-08T18:17:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hk5xb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T18:17:31Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-84k8x\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T18:18:02Z is after 2025-08-24T17:21:41Z" Oct 08 18:18:02 crc kubenswrapper[4859]: I1008 18:18:02.805039 4859 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3ca6ce76-77e8-4e46-a2f6-b1ed4aa0fd4b\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:17:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:17:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:18:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:18:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:17:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f587e149d57ea53626ed3beb83e421c3b89ba1fec9e949705e8ccb5b4f7cc79a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T18:17:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://428cbe63daefbfbcb2ec76dad0fe638cdc65934bac80b791cebe0b36b2d6909a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T18:17:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a41399a37b1ad86a063001224232ecb8f7ea56aeb5686d3858bc08102a0c8f49\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T18:17:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://008757fce552591f7c58193f605eb60792761d5943927cbd63184549618d2b43\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://008757fce552591f7c58193f605eb60792761d5943927cbd63184549618d2b43\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-08T18:17:11Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-08T18:17:11Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T18:17:10Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T18:18:02Z is after 2025-08-24T17:21:41Z" Oct 08 18:18:02 crc kubenswrapper[4859]: I1008 18:18:02.820008 4859 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-08T18:17:29Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T18:17:29Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T18:18:02Z is after 2025-08-24T17:21:41Z" Oct 08 18:18:02 crc kubenswrapper[4859]: I1008 18:18:02.831494 4859 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-btw28" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"97bc5598-3ead-4300-a938-3aebbff1e8c6\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:17:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:17:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:17:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:17:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d6d4d6ba771cb5fde637449bba05660f634df7e833a0c8dad2b2c7f3a422bb2b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T18:17:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qcxsd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T18:17:31Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-btw28\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T18:18:02Z is after 2025-08-24T17:21:41Z" Oct 08 18:18:02 crc kubenswrapper[4859]: I1008 18:18:02.845922 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 18:18:02 crc kubenswrapper[4859]: I1008 18:18:02.846003 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 18:18:02 crc kubenswrapper[4859]: I1008 18:18:02.846014 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 18:18:02 crc kubenswrapper[4859]: I1008 18:18:02.846050 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 18:18:02 crc kubenswrapper[4859]: I1008 18:18:02.846064 4859 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T18:18:02Z","lastTransitionTime":"2025-10-08T18:18:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 18:18:02 crc kubenswrapper[4859]: I1008 18:18:02.846915 4859 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-8plkg" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"92b2d7de-31cc-4a91-95a5-ed5ea964b028\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:17:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:17:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:17:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:17:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fa00b8345d40b74a917ddc23cc1b2c5c684a9ca61b6119d090c92cc1439b71c4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T18:17:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q9d7m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T18:17:31Z\\\"}}\" for pod \"openshift-multus\"/\"multus-8plkg\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T18:18:02Z is after 2025-08-24T17:21:41Z" Oct 08 18:18:02 crc kubenswrapper[4859]: I1008 18:18:02.860397 4859 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-btjmq" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b8605f7-fed7-493f-9b4f-6a6358adf907\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:17:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:17:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:17:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:17:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://14e9b6c4a4403f36adb78fcf1135a620ca6fc56ed974c9639dcdfc59e6917559\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T18:17:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xspg4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T18:17:35Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-btjmq\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T18:18:02Z is after 2025-08-24T17:21:41Z" Oct 08 18:18:02 crc kubenswrapper[4859]: I1008 18:18:02.873865 4859 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-g4t79" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3a92dbf1-21a8-4095-96bb-2d701d7f3d56\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:17:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:17:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:17:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:17:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f8772a25b6e605efe9a062dc32dbbc2cfe32b4fe164b4e84fbf3bcffab744179\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T18:17:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sv79p\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://92cb9986c09a7f9e1e552e3e3ac1f74f04128241e00ddee38882d2ba5c005a67\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T18:17:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sv79p\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T18:17:45Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-g4t79\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T18:18:02Z is after 2025-08-24T17:21:41Z" Oct 08 18:18:02 crc kubenswrapper[4859]: I1008 18:18:02.887453 4859 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-08T18:17:31Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T18:17:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T18:17:31Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c6d15905694b6bbd4999d49dc8f7a53eb6bfaa7adb59efb0d7e4973c21d1c7be\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T18:17:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T18:18:02Z is after 2025-08-24T17:21:41Z" Oct 08 18:18:02 crc kubenswrapper[4859]: I1008 18:18:02.901335 4859 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-08T18:17:33Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T18:17:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T18:17:33Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5d99e071bd083914418d9c128f38955f26295c08b1520f2dc774c99b1c843693\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T18:17:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T18:18:02Z is after 2025-08-24T17:21:41Z" Oct 08 18:18:02 crc kubenswrapper[4859]: I1008 18:18:02.948198 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 18:18:02 crc kubenswrapper[4859]: I1008 18:18:02.948264 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 18:18:02 crc kubenswrapper[4859]: I1008 18:18:02.948280 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 18:18:02 crc kubenswrapper[4859]: I1008 18:18:02.948318 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 18:18:02 crc kubenswrapper[4859]: I1008 18:18:02.948329 4859 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T18:18:02Z","lastTransitionTime":"2025-10-08T18:18:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 18:18:03 crc kubenswrapper[4859]: I1008 18:18:03.051615 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 18:18:03 crc kubenswrapper[4859]: I1008 18:18:03.051702 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 18:18:03 crc kubenswrapper[4859]: I1008 18:18:03.051719 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 18:18:03 crc kubenswrapper[4859]: I1008 18:18:03.051745 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 18:18:03 crc kubenswrapper[4859]: I1008 18:18:03.051759 4859 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T18:18:03Z","lastTransitionTime":"2025-10-08T18:18:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 18:18:03 crc kubenswrapper[4859]: I1008 18:18:03.154678 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 18:18:03 crc kubenswrapper[4859]: I1008 18:18:03.154758 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 18:18:03 crc kubenswrapper[4859]: I1008 18:18:03.154773 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 18:18:03 crc kubenswrapper[4859]: I1008 18:18:03.154797 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 18:18:03 crc kubenswrapper[4859]: I1008 18:18:03.154810 4859 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T18:18:03Z","lastTransitionTime":"2025-10-08T18:18:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 18:18:03 crc kubenswrapper[4859]: I1008 18:18:03.257208 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 18:18:03 crc kubenswrapper[4859]: I1008 18:18:03.257301 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 18:18:03 crc kubenswrapper[4859]: I1008 18:18:03.257320 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 18:18:03 crc kubenswrapper[4859]: I1008 18:18:03.257349 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 18:18:03 crc kubenswrapper[4859]: I1008 18:18:03.257368 4859 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T18:18:03Z","lastTransitionTime":"2025-10-08T18:18:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 18:18:03 crc kubenswrapper[4859]: I1008 18:18:03.360616 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 18:18:03 crc kubenswrapper[4859]: I1008 18:18:03.360842 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 18:18:03 crc kubenswrapper[4859]: I1008 18:18:03.360866 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 18:18:03 crc kubenswrapper[4859]: I1008 18:18:03.360893 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 18:18:03 crc kubenswrapper[4859]: I1008 18:18:03.360915 4859 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T18:18:03Z","lastTransitionTime":"2025-10-08T18:18:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 18:18:03 crc kubenswrapper[4859]: I1008 18:18:03.464428 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 18:18:03 crc kubenswrapper[4859]: I1008 18:18:03.464476 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 18:18:03 crc kubenswrapper[4859]: I1008 18:18:03.464487 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 18:18:03 crc kubenswrapper[4859]: I1008 18:18:03.464506 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 18:18:03 crc kubenswrapper[4859]: I1008 18:18:03.464522 4859 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T18:18:03Z","lastTransitionTime":"2025-10-08T18:18:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 18:18:03 crc kubenswrapper[4859]: I1008 18:18:03.469815 4859 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 08 18:18:03 crc kubenswrapper[4859]: I1008 18:18:03.469862 4859 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 08 18:18:03 crc kubenswrapper[4859]: I1008 18:18:03.469835 4859 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-4b4sz" Oct 08 18:18:03 crc kubenswrapper[4859]: I1008 18:18:03.469835 4859 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 08 18:18:03 crc kubenswrapper[4859]: E1008 18:18:03.469958 4859 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 08 18:18:03 crc kubenswrapper[4859]: E1008 18:18:03.470060 4859 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-4b4sz" podUID="2e53cd9b-64f7-4e07-8b96-fb77847c9ac6" Oct 08 18:18:03 crc kubenswrapper[4859]: E1008 18:18:03.470103 4859 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 08 18:18:03 crc kubenswrapper[4859]: E1008 18:18:03.470140 4859 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 08 18:18:03 crc kubenswrapper[4859]: I1008 18:18:03.568653 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 18:18:03 crc kubenswrapper[4859]: I1008 18:18:03.568743 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 18:18:03 crc kubenswrapper[4859]: I1008 18:18:03.568776 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 18:18:03 crc kubenswrapper[4859]: I1008 18:18:03.568808 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 18:18:03 crc kubenswrapper[4859]: I1008 18:18:03.568829 4859 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T18:18:03Z","lastTransitionTime":"2025-10-08T18:18:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 18:18:03 crc kubenswrapper[4859]: I1008 18:18:03.672065 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 18:18:03 crc kubenswrapper[4859]: I1008 18:18:03.672139 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 18:18:03 crc kubenswrapper[4859]: I1008 18:18:03.672151 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 18:18:03 crc kubenswrapper[4859]: I1008 18:18:03.672172 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 18:18:03 crc kubenswrapper[4859]: I1008 18:18:03.672184 4859 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T18:18:03Z","lastTransitionTime":"2025-10-08T18:18:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 18:18:03 crc kubenswrapper[4859]: I1008 18:18:03.776219 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 18:18:03 crc kubenswrapper[4859]: I1008 18:18:03.776362 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 18:18:03 crc kubenswrapper[4859]: I1008 18:18:03.776373 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 18:18:03 crc kubenswrapper[4859]: I1008 18:18:03.776391 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 18:18:03 crc kubenswrapper[4859]: I1008 18:18:03.776403 4859 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T18:18:03Z","lastTransitionTime":"2025-10-08T18:18:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 18:18:03 crc kubenswrapper[4859]: I1008 18:18:03.879152 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 18:18:03 crc kubenswrapper[4859]: I1008 18:18:03.879241 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 18:18:03 crc kubenswrapper[4859]: I1008 18:18:03.879279 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 18:18:03 crc kubenswrapper[4859]: I1008 18:18:03.879315 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 18:18:03 crc kubenswrapper[4859]: I1008 18:18:03.879337 4859 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T18:18:03Z","lastTransitionTime":"2025-10-08T18:18:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 18:18:03 crc kubenswrapper[4859]: I1008 18:18:03.983824 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 18:18:03 crc kubenswrapper[4859]: I1008 18:18:03.983917 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 18:18:03 crc kubenswrapper[4859]: I1008 18:18:03.983942 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 18:18:03 crc kubenswrapper[4859]: I1008 18:18:03.983978 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 18:18:03 crc kubenswrapper[4859]: I1008 18:18:03.984003 4859 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T18:18:03Z","lastTransitionTime":"2025-10-08T18:18:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 18:18:04 crc kubenswrapper[4859]: I1008 18:18:04.086949 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 18:18:04 crc kubenswrapper[4859]: I1008 18:18:04.087008 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 18:18:04 crc kubenswrapper[4859]: I1008 18:18:04.087025 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 18:18:04 crc kubenswrapper[4859]: I1008 18:18:04.087051 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 18:18:04 crc kubenswrapper[4859]: I1008 18:18:04.087070 4859 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T18:18:04Z","lastTransitionTime":"2025-10-08T18:18:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 18:18:04 crc kubenswrapper[4859]: I1008 18:18:04.190625 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 18:18:04 crc kubenswrapper[4859]: I1008 18:18:04.190732 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 18:18:04 crc kubenswrapper[4859]: I1008 18:18:04.190758 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 18:18:04 crc kubenswrapper[4859]: I1008 18:18:04.190788 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 18:18:04 crc kubenswrapper[4859]: I1008 18:18:04.190812 4859 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T18:18:04Z","lastTransitionTime":"2025-10-08T18:18:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 18:18:04 crc kubenswrapper[4859]: I1008 18:18:04.293220 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 18:18:04 crc kubenswrapper[4859]: I1008 18:18:04.293278 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 18:18:04 crc kubenswrapper[4859]: I1008 18:18:04.293292 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 18:18:04 crc kubenswrapper[4859]: I1008 18:18:04.293313 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 18:18:04 crc kubenswrapper[4859]: I1008 18:18:04.293326 4859 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T18:18:04Z","lastTransitionTime":"2025-10-08T18:18:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 18:18:04 crc kubenswrapper[4859]: I1008 18:18:04.397419 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 18:18:04 crc kubenswrapper[4859]: I1008 18:18:04.397481 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 18:18:04 crc kubenswrapper[4859]: I1008 18:18:04.397495 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 18:18:04 crc kubenswrapper[4859]: I1008 18:18:04.397520 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 18:18:04 crc kubenswrapper[4859]: I1008 18:18:04.397534 4859 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T18:18:04Z","lastTransitionTime":"2025-10-08T18:18:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 18:18:04 crc kubenswrapper[4859]: I1008 18:18:04.500925 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 18:18:04 crc kubenswrapper[4859]: I1008 18:18:04.500998 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 18:18:04 crc kubenswrapper[4859]: I1008 18:18:04.501012 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 18:18:04 crc kubenswrapper[4859]: I1008 18:18:04.501036 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 18:18:04 crc kubenswrapper[4859]: I1008 18:18:04.501050 4859 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T18:18:04Z","lastTransitionTime":"2025-10-08T18:18:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 18:18:04 crc kubenswrapper[4859]: I1008 18:18:04.603591 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 18:18:04 crc kubenswrapper[4859]: I1008 18:18:04.603644 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 18:18:04 crc kubenswrapper[4859]: I1008 18:18:04.603656 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 18:18:04 crc kubenswrapper[4859]: I1008 18:18:04.603673 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 18:18:04 crc kubenswrapper[4859]: I1008 18:18:04.603698 4859 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T18:18:04Z","lastTransitionTime":"2025-10-08T18:18:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 18:18:04 crc kubenswrapper[4859]: I1008 18:18:04.706519 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 18:18:04 crc kubenswrapper[4859]: I1008 18:18:04.706583 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 18:18:04 crc kubenswrapper[4859]: I1008 18:18:04.706596 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 18:18:04 crc kubenswrapper[4859]: I1008 18:18:04.706619 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 18:18:04 crc kubenswrapper[4859]: I1008 18:18:04.706631 4859 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T18:18:04Z","lastTransitionTime":"2025-10-08T18:18:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 18:18:04 crc kubenswrapper[4859]: I1008 18:18:04.809968 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 18:18:04 crc kubenswrapper[4859]: I1008 18:18:04.810031 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 18:18:04 crc kubenswrapper[4859]: I1008 18:18:04.810043 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 18:18:04 crc kubenswrapper[4859]: I1008 18:18:04.810067 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 18:18:04 crc kubenswrapper[4859]: I1008 18:18:04.810080 4859 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T18:18:04Z","lastTransitionTime":"2025-10-08T18:18:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 18:18:04 crc kubenswrapper[4859]: I1008 18:18:04.913004 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 18:18:04 crc kubenswrapper[4859]: I1008 18:18:04.913083 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 18:18:04 crc kubenswrapper[4859]: I1008 18:18:04.913095 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 18:18:04 crc kubenswrapper[4859]: I1008 18:18:04.913116 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 18:18:04 crc kubenswrapper[4859]: I1008 18:18:04.913127 4859 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T18:18:04Z","lastTransitionTime":"2025-10-08T18:18:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 18:18:05 crc kubenswrapper[4859]: I1008 18:18:05.016385 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 18:18:05 crc kubenswrapper[4859]: I1008 18:18:05.016440 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 18:18:05 crc kubenswrapper[4859]: I1008 18:18:05.016453 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 18:18:05 crc kubenswrapper[4859]: I1008 18:18:05.016474 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 18:18:05 crc kubenswrapper[4859]: I1008 18:18:05.016486 4859 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T18:18:05Z","lastTransitionTime":"2025-10-08T18:18:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 18:18:05 crc kubenswrapper[4859]: I1008 18:18:05.120145 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 18:18:05 crc kubenswrapper[4859]: I1008 18:18:05.120210 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 18:18:05 crc kubenswrapper[4859]: I1008 18:18:05.120223 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 18:18:05 crc kubenswrapper[4859]: I1008 18:18:05.120248 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 18:18:05 crc kubenswrapper[4859]: I1008 18:18:05.120266 4859 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T18:18:05Z","lastTransitionTime":"2025-10-08T18:18:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 18:18:05 crc kubenswrapper[4859]: I1008 18:18:05.223504 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 18:18:05 crc kubenswrapper[4859]: I1008 18:18:05.223550 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 18:18:05 crc kubenswrapper[4859]: I1008 18:18:05.223564 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 18:18:05 crc kubenswrapper[4859]: I1008 18:18:05.223582 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 18:18:05 crc kubenswrapper[4859]: I1008 18:18:05.223597 4859 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T18:18:05Z","lastTransitionTime":"2025-10-08T18:18:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 18:18:05 crc kubenswrapper[4859]: I1008 18:18:05.326622 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 18:18:05 crc kubenswrapper[4859]: I1008 18:18:05.326817 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 18:18:05 crc kubenswrapper[4859]: I1008 18:18:05.326838 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 18:18:05 crc kubenswrapper[4859]: I1008 18:18:05.326869 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 18:18:05 crc kubenswrapper[4859]: I1008 18:18:05.326888 4859 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T18:18:05Z","lastTransitionTime":"2025-10-08T18:18:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 18:18:05 crc kubenswrapper[4859]: I1008 18:18:05.430324 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 18:18:05 crc kubenswrapper[4859]: I1008 18:18:05.430388 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 18:18:05 crc kubenswrapper[4859]: I1008 18:18:05.430400 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 18:18:05 crc kubenswrapper[4859]: I1008 18:18:05.430426 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 18:18:05 crc kubenswrapper[4859]: I1008 18:18:05.430440 4859 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T18:18:05Z","lastTransitionTime":"2025-10-08T18:18:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 18:18:05 crc kubenswrapper[4859]: I1008 18:18:05.468938 4859 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 08 18:18:05 crc kubenswrapper[4859]: I1008 18:18:05.468938 4859 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 08 18:18:05 crc kubenswrapper[4859]: I1008 18:18:05.468964 4859 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 08 18:18:05 crc kubenswrapper[4859]: I1008 18:18:05.469129 4859 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-4b4sz" Oct 08 18:18:05 crc kubenswrapper[4859]: E1008 18:18:05.469346 4859 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 08 18:18:05 crc kubenswrapper[4859]: E1008 18:18:05.469435 4859 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 08 18:18:05 crc kubenswrapper[4859]: E1008 18:18:05.469494 4859 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 08 18:18:05 crc kubenswrapper[4859]: E1008 18:18:05.469534 4859 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-4b4sz" podUID="2e53cd9b-64f7-4e07-8b96-fb77847c9ac6" Oct 08 18:18:05 crc kubenswrapper[4859]: I1008 18:18:05.532833 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 18:18:05 crc kubenswrapper[4859]: I1008 18:18:05.532877 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 18:18:05 crc kubenswrapper[4859]: I1008 18:18:05.532890 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 18:18:05 crc kubenswrapper[4859]: I1008 18:18:05.532908 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 18:18:05 crc kubenswrapper[4859]: I1008 18:18:05.532921 4859 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T18:18:05Z","lastTransitionTime":"2025-10-08T18:18:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 18:18:05 crc kubenswrapper[4859]: I1008 18:18:05.636653 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 18:18:05 crc kubenswrapper[4859]: I1008 18:18:05.636812 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 18:18:05 crc kubenswrapper[4859]: I1008 18:18:05.636831 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 18:18:05 crc kubenswrapper[4859]: I1008 18:18:05.636858 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 18:18:05 crc kubenswrapper[4859]: I1008 18:18:05.636875 4859 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T18:18:05Z","lastTransitionTime":"2025-10-08T18:18:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 18:18:05 crc kubenswrapper[4859]: I1008 18:18:05.741106 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 18:18:05 crc kubenswrapper[4859]: I1008 18:18:05.741198 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 18:18:05 crc kubenswrapper[4859]: I1008 18:18:05.741224 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 18:18:05 crc kubenswrapper[4859]: I1008 18:18:05.741258 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 18:18:05 crc kubenswrapper[4859]: I1008 18:18:05.741282 4859 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T18:18:05Z","lastTransitionTime":"2025-10-08T18:18:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 18:18:05 crc kubenswrapper[4859]: I1008 18:18:05.844679 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 18:18:05 crc kubenswrapper[4859]: I1008 18:18:05.844751 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 18:18:05 crc kubenswrapper[4859]: I1008 18:18:05.844764 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 18:18:05 crc kubenswrapper[4859]: I1008 18:18:05.844783 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 18:18:05 crc kubenswrapper[4859]: I1008 18:18:05.844798 4859 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T18:18:05Z","lastTransitionTime":"2025-10-08T18:18:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 18:18:05 crc kubenswrapper[4859]: I1008 18:18:05.947711 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 18:18:05 crc kubenswrapper[4859]: I1008 18:18:05.947774 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 18:18:05 crc kubenswrapper[4859]: I1008 18:18:05.947785 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 18:18:05 crc kubenswrapper[4859]: I1008 18:18:05.947804 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 18:18:05 crc kubenswrapper[4859]: I1008 18:18:05.947826 4859 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T18:18:05Z","lastTransitionTime":"2025-10-08T18:18:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 18:18:06 crc kubenswrapper[4859]: I1008 18:18:06.051193 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 18:18:06 crc kubenswrapper[4859]: I1008 18:18:06.051278 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 18:18:06 crc kubenswrapper[4859]: I1008 18:18:06.051298 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 18:18:06 crc kubenswrapper[4859]: I1008 18:18:06.051329 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 18:18:06 crc kubenswrapper[4859]: I1008 18:18:06.051351 4859 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T18:18:06Z","lastTransitionTime":"2025-10-08T18:18:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 18:18:06 crc kubenswrapper[4859]: I1008 18:18:06.155623 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 18:18:06 crc kubenswrapper[4859]: I1008 18:18:06.155674 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 18:18:06 crc kubenswrapper[4859]: I1008 18:18:06.155714 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 18:18:06 crc kubenswrapper[4859]: I1008 18:18:06.155734 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 18:18:06 crc kubenswrapper[4859]: I1008 18:18:06.155746 4859 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T18:18:06Z","lastTransitionTime":"2025-10-08T18:18:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 18:18:06 crc kubenswrapper[4859]: I1008 18:18:06.259507 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 18:18:06 crc kubenswrapper[4859]: I1008 18:18:06.259586 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 18:18:06 crc kubenswrapper[4859]: I1008 18:18:06.259605 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 18:18:06 crc kubenswrapper[4859]: I1008 18:18:06.259630 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 18:18:06 crc kubenswrapper[4859]: I1008 18:18:06.259648 4859 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T18:18:06Z","lastTransitionTime":"2025-10-08T18:18:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 18:18:06 crc kubenswrapper[4859]: I1008 18:18:06.363029 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 18:18:06 crc kubenswrapper[4859]: I1008 18:18:06.363114 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 18:18:06 crc kubenswrapper[4859]: I1008 18:18:06.363141 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 18:18:06 crc kubenswrapper[4859]: I1008 18:18:06.363175 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 18:18:06 crc kubenswrapper[4859]: I1008 18:18:06.363196 4859 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T18:18:06Z","lastTransitionTime":"2025-10-08T18:18:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 18:18:06 crc kubenswrapper[4859]: I1008 18:18:06.467220 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 18:18:06 crc kubenswrapper[4859]: I1008 18:18:06.467288 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 18:18:06 crc kubenswrapper[4859]: I1008 18:18:06.467314 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 18:18:06 crc kubenswrapper[4859]: I1008 18:18:06.467347 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 18:18:06 crc kubenswrapper[4859]: I1008 18:18:06.467370 4859 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T18:18:06Z","lastTransitionTime":"2025-10-08T18:18:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 18:18:06 crc kubenswrapper[4859]: I1008 18:18:06.570443 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 18:18:06 crc kubenswrapper[4859]: I1008 18:18:06.570497 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 18:18:06 crc kubenswrapper[4859]: I1008 18:18:06.570507 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 18:18:06 crc kubenswrapper[4859]: I1008 18:18:06.570527 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 18:18:06 crc kubenswrapper[4859]: I1008 18:18:06.570540 4859 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T18:18:06Z","lastTransitionTime":"2025-10-08T18:18:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 18:18:06 crc kubenswrapper[4859]: I1008 18:18:06.673734 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 18:18:06 crc kubenswrapper[4859]: I1008 18:18:06.673827 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 18:18:06 crc kubenswrapper[4859]: I1008 18:18:06.673869 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 18:18:06 crc kubenswrapper[4859]: I1008 18:18:06.673892 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 18:18:06 crc kubenswrapper[4859]: I1008 18:18:06.673910 4859 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T18:18:06Z","lastTransitionTime":"2025-10-08T18:18:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 18:18:06 crc kubenswrapper[4859]: I1008 18:18:06.777535 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 18:18:06 crc kubenswrapper[4859]: I1008 18:18:06.777590 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 18:18:06 crc kubenswrapper[4859]: I1008 18:18:06.777602 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 18:18:06 crc kubenswrapper[4859]: I1008 18:18:06.777622 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 18:18:06 crc kubenswrapper[4859]: I1008 18:18:06.777637 4859 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T18:18:06Z","lastTransitionTime":"2025-10-08T18:18:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 18:18:06 crc kubenswrapper[4859]: I1008 18:18:06.880853 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 18:18:06 crc kubenswrapper[4859]: I1008 18:18:06.880903 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 18:18:06 crc kubenswrapper[4859]: I1008 18:18:06.880912 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 18:18:06 crc kubenswrapper[4859]: I1008 18:18:06.880930 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 18:18:06 crc kubenswrapper[4859]: I1008 18:18:06.880943 4859 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T18:18:06Z","lastTransitionTime":"2025-10-08T18:18:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 18:18:06 crc kubenswrapper[4859]: I1008 18:18:06.984190 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 18:18:06 crc kubenswrapper[4859]: I1008 18:18:06.984246 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 18:18:06 crc kubenswrapper[4859]: I1008 18:18:06.984260 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 18:18:06 crc kubenswrapper[4859]: I1008 18:18:06.984281 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 18:18:06 crc kubenswrapper[4859]: I1008 18:18:06.984294 4859 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T18:18:06Z","lastTransitionTime":"2025-10-08T18:18:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 18:18:07 crc kubenswrapper[4859]: I1008 18:18:07.087647 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 18:18:07 crc kubenswrapper[4859]: I1008 18:18:07.087786 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 18:18:07 crc kubenswrapper[4859]: I1008 18:18:07.087807 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 18:18:07 crc kubenswrapper[4859]: I1008 18:18:07.087834 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 18:18:07 crc kubenswrapper[4859]: I1008 18:18:07.087852 4859 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T18:18:07Z","lastTransitionTime":"2025-10-08T18:18:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 18:18:07 crc kubenswrapper[4859]: I1008 18:18:07.190307 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 18:18:07 crc kubenswrapper[4859]: I1008 18:18:07.190372 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 18:18:07 crc kubenswrapper[4859]: I1008 18:18:07.190399 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 18:18:07 crc kubenswrapper[4859]: I1008 18:18:07.190448 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 18:18:07 crc kubenswrapper[4859]: I1008 18:18:07.190472 4859 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T18:18:07Z","lastTransitionTime":"2025-10-08T18:18:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 18:18:07 crc kubenswrapper[4859]: I1008 18:18:07.294067 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 18:18:07 crc kubenswrapper[4859]: I1008 18:18:07.294119 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 18:18:07 crc kubenswrapper[4859]: I1008 18:18:07.294131 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 18:18:07 crc kubenswrapper[4859]: I1008 18:18:07.294150 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 18:18:07 crc kubenswrapper[4859]: I1008 18:18:07.294162 4859 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T18:18:07Z","lastTransitionTime":"2025-10-08T18:18:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 18:18:07 crc kubenswrapper[4859]: I1008 18:18:07.398014 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 18:18:07 crc kubenswrapper[4859]: I1008 18:18:07.398079 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 18:18:07 crc kubenswrapper[4859]: I1008 18:18:07.398096 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 18:18:07 crc kubenswrapper[4859]: I1008 18:18:07.398125 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 18:18:07 crc kubenswrapper[4859]: I1008 18:18:07.398144 4859 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T18:18:07Z","lastTransitionTime":"2025-10-08T18:18:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 18:18:07 crc kubenswrapper[4859]: I1008 18:18:07.469379 4859 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 08 18:18:07 crc kubenswrapper[4859]: I1008 18:18:07.469379 4859 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 08 18:18:07 crc kubenswrapper[4859]: E1008 18:18:07.469538 4859 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 08 18:18:07 crc kubenswrapper[4859]: I1008 18:18:07.469407 4859 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-4b4sz" Oct 08 18:18:07 crc kubenswrapper[4859]: E1008 18:18:07.469597 4859 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 08 18:18:07 crc kubenswrapper[4859]: I1008 18:18:07.469389 4859 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 08 18:18:07 crc kubenswrapper[4859]: E1008 18:18:07.469719 4859 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-4b4sz" podUID="2e53cd9b-64f7-4e07-8b96-fb77847c9ac6" Oct 08 18:18:07 crc kubenswrapper[4859]: E1008 18:18:07.469817 4859 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 08 18:18:07 crc kubenswrapper[4859]: I1008 18:18:07.500674 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 18:18:07 crc kubenswrapper[4859]: I1008 18:18:07.500773 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 18:18:07 crc kubenswrapper[4859]: I1008 18:18:07.500793 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 18:18:07 crc kubenswrapper[4859]: I1008 18:18:07.500820 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 18:18:07 crc kubenswrapper[4859]: I1008 18:18:07.500837 4859 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T18:18:07Z","lastTransitionTime":"2025-10-08T18:18:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 18:18:07 crc kubenswrapper[4859]: I1008 18:18:07.604169 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 18:18:07 crc kubenswrapper[4859]: I1008 18:18:07.604269 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 18:18:07 crc kubenswrapper[4859]: I1008 18:18:07.604290 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 18:18:07 crc kubenswrapper[4859]: I1008 18:18:07.604315 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 18:18:07 crc kubenswrapper[4859]: I1008 18:18:07.604333 4859 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T18:18:07Z","lastTransitionTime":"2025-10-08T18:18:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 18:18:07 crc kubenswrapper[4859]: I1008 18:18:07.628000 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 18:18:07 crc kubenswrapper[4859]: I1008 18:18:07.628065 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 18:18:07 crc kubenswrapper[4859]: I1008 18:18:07.628086 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 18:18:07 crc kubenswrapper[4859]: I1008 18:18:07.628114 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 18:18:07 crc kubenswrapper[4859]: I1008 18:18:07.628135 4859 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T18:18:07Z","lastTransitionTime":"2025-10-08T18:18:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 18:18:07 crc kubenswrapper[4859]: E1008 18:18:07.647015 4859 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404552Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865352Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-10-08T18:18:07Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-08T18:18:07Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-08T18:18:07Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-08T18:18:07Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-08T18:18:07Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-08T18:18:07Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-08T18:18:07Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-08T18:18:07Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"edb3ee6d-476b-407c-bb7d-4480786e8777\\\",\\\"systemUUID\\\":\\\"8f73c7ed-44aa-4d16-bae1-d4e684803cfc\\\"},\\\"runtimeHandlers\\\":[{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":false},\\\"name\\\":\\\"runc\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"crun\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"\\\"}]}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T18:18:07Z is after 2025-08-24T17:21:41Z" Oct 08 18:18:07 crc kubenswrapper[4859]: I1008 18:18:07.652006 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 18:18:07 crc kubenswrapper[4859]: I1008 18:18:07.652076 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 18:18:07 crc kubenswrapper[4859]: I1008 18:18:07.652090 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 18:18:07 crc kubenswrapper[4859]: I1008 18:18:07.652110 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 18:18:07 crc kubenswrapper[4859]: I1008 18:18:07.652122 4859 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T18:18:07Z","lastTransitionTime":"2025-10-08T18:18:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 18:18:07 crc kubenswrapper[4859]: E1008 18:18:07.670382 4859 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404552Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865352Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-10-08T18:18:07Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-08T18:18:07Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-08T18:18:07Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-08T18:18:07Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-08T18:18:07Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-08T18:18:07Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-08T18:18:07Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-08T18:18:07Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"edb3ee6d-476b-407c-bb7d-4480786e8777\\\",\\\"systemUUID\\\":\\\"8f73c7ed-44aa-4d16-bae1-d4e684803cfc\\\"},\\\"runtimeHandlers\\\":[{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":false},\\\"name\\\":\\\"runc\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"crun\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"\\\"}]}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T18:18:07Z is after 2025-08-24T17:21:41Z" Oct 08 18:18:07 crc kubenswrapper[4859]: I1008 18:18:07.675370 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 18:18:07 crc kubenswrapper[4859]: I1008 18:18:07.675418 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 18:18:07 crc kubenswrapper[4859]: I1008 18:18:07.675432 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 18:18:07 crc kubenswrapper[4859]: I1008 18:18:07.675455 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 18:18:07 crc kubenswrapper[4859]: I1008 18:18:07.675475 4859 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T18:18:07Z","lastTransitionTime":"2025-10-08T18:18:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 18:18:07 crc kubenswrapper[4859]: E1008 18:18:07.696844 4859 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404552Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865352Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-10-08T18:18:07Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-08T18:18:07Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-08T18:18:07Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-08T18:18:07Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-08T18:18:07Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-08T18:18:07Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-08T18:18:07Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-08T18:18:07Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"edb3ee6d-476b-407c-bb7d-4480786e8777\\\",\\\"systemUUID\\\":\\\"8f73c7ed-44aa-4d16-bae1-d4e684803cfc\\\"},\\\"runtimeHandlers\\\":[{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":false},\\\"name\\\":\\\"runc\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"crun\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"\\\"}]}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T18:18:07Z is after 2025-08-24T17:21:41Z" Oct 08 18:18:07 crc kubenswrapper[4859]: I1008 18:18:07.701917 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 18:18:07 crc kubenswrapper[4859]: I1008 18:18:07.701957 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 18:18:07 crc kubenswrapper[4859]: I1008 18:18:07.701969 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 18:18:07 crc kubenswrapper[4859]: I1008 18:18:07.701986 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 18:18:07 crc kubenswrapper[4859]: I1008 18:18:07.701997 4859 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T18:18:07Z","lastTransitionTime":"2025-10-08T18:18:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 18:18:07 crc kubenswrapper[4859]: E1008 18:18:07.718458 4859 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404552Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865352Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-10-08T18:18:07Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-08T18:18:07Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-08T18:18:07Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-08T18:18:07Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-08T18:18:07Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-08T18:18:07Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-08T18:18:07Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-08T18:18:07Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"edb3ee6d-476b-407c-bb7d-4480786e8777\\\",\\\"systemUUID\\\":\\\"8f73c7ed-44aa-4d16-bae1-d4e684803cfc\\\"},\\\"runtimeHandlers\\\":[{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":false},\\\"name\\\":\\\"runc\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"crun\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"\\\"}]}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T18:18:07Z is after 2025-08-24T17:21:41Z" Oct 08 18:18:07 crc kubenswrapper[4859]: I1008 18:18:07.723114 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 18:18:07 crc kubenswrapper[4859]: I1008 18:18:07.723145 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 18:18:07 crc kubenswrapper[4859]: I1008 18:18:07.723158 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 18:18:07 crc kubenswrapper[4859]: I1008 18:18:07.723178 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 18:18:07 crc kubenswrapper[4859]: I1008 18:18:07.723192 4859 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T18:18:07Z","lastTransitionTime":"2025-10-08T18:18:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 18:18:07 crc kubenswrapper[4859]: E1008 18:18:07.740086 4859 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404552Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865352Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-10-08T18:18:07Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-08T18:18:07Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-08T18:18:07Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-08T18:18:07Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-08T18:18:07Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-08T18:18:07Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-08T18:18:07Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-08T18:18:07Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"edb3ee6d-476b-407c-bb7d-4480786e8777\\\",\\\"systemUUID\\\":\\\"8f73c7ed-44aa-4d16-bae1-d4e684803cfc\\\"},\\\"runtimeHandlers\\\":[{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":false},\\\"name\\\":\\\"runc\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"crun\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"\\\"}]}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T18:18:07Z is after 2025-08-24T17:21:41Z" Oct 08 18:18:07 crc kubenswrapper[4859]: E1008 18:18:07.740245 4859 kubelet_node_status.go:572] "Unable to update node status" err="update node status exceeds retry count" Oct 08 18:18:07 crc kubenswrapper[4859]: I1008 18:18:07.742214 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 18:18:07 crc kubenswrapper[4859]: I1008 18:18:07.742250 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 18:18:07 crc kubenswrapper[4859]: I1008 18:18:07.742265 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 18:18:07 crc kubenswrapper[4859]: I1008 18:18:07.742286 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 18:18:07 crc kubenswrapper[4859]: I1008 18:18:07.742303 4859 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T18:18:07Z","lastTransitionTime":"2025-10-08T18:18:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 18:18:07 crc kubenswrapper[4859]: I1008 18:18:07.845272 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 18:18:07 crc kubenswrapper[4859]: I1008 18:18:07.845354 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 18:18:07 crc kubenswrapper[4859]: I1008 18:18:07.845380 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 18:18:07 crc kubenswrapper[4859]: I1008 18:18:07.845412 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 18:18:07 crc kubenswrapper[4859]: I1008 18:18:07.845430 4859 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T18:18:07Z","lastTransitionTime":"2025-10-08T18:18:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 18:18:07 crc kubenswrapper[4859]: I1008 18:18:07.948918 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 18:18:07 crc kubenswrapper[4859]: I1008 18:18:07.948986 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 18:18:07 crc kubenswrapper[4859]: I1008 18:18:07.948998 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 18:18:07 crc kubenswrapper[4859]: I1008 18:18:07.949022 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 18:18:07 crc kubenswrapper[4859]: I1008 18:18:07.949042 4859 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T18:18:07Z","lastTransitionTime":"2025-10-08T18:18:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 18:18:08 crc kubenswrapper[4859]: I1008 18:18:08.052141 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 18:18:08 crc kubenswrapper[4859]: I1008 18:18:08.052262 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 18:18:08 crc kubenswrapper[4859]: I1008 18:18:08.052280 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 18:18:08 crc kubenswrapper[4859]: I1008 18:18:08.052307 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 18:18:08 crc kubenswrapper[4859]: I1008 18:18:08.052325 4859 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T18:18:08Z","lastTransitionTime":"2025-10-08T18:18:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 18:18:08 crc kubenswrapper[4859]: I1008 18:18:08.160348 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 18:18:08 crc kubenswrapper[4859]: I1008 18:18:08.160462 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 18:18:08 crc kubenswrapper[4859]: I1008 18:18:08.160504 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 18:18:08 crc kubenswrapper[4859]: I1008 18:18:08.160638 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 18:18:08 crc kubenswrapper[4859]: I1008 18:18:08.160665 4859 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T18:18:08Z","lastTransitionTime":"2025-10-08T18:18:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 18:18:08 crc kubenswrapper[4859]: I1008 18:18:08.264117 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 18:18:08 crc kubenswrapper[4859]: I1008 18:18:08.264185 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 18:18:08 crc kubenswrapper[4859]: I1008 18:18:08.264216 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 18:18:08 crc kubenswrapper[4859]: I1008 18:18:08.264235 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 18:18:08 crc kubenswrapper[4859]: I1008 18:18:08.264249 4859 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T18:18:08Z","lastTransitionTime":"2025-10-08T18:18:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 18:18:08 crc kubenswrapper[4859]: I1008 18:18:08.368288 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 18:18:08 crc kubenswrapper[4859]: I1008 18:18:08.368402 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 18:18:08 crc kubenswrapper[4859]: I1008 18:18:08.368444 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 18:18:08 crc kubenswrapper[4859]: I1008 18:18:08.368479 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 18:18:08 crc kubenswrapper[4859]: I1008 18:18:08.368503 4859 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T18:18:08Z","lastTransitionTime":"2025-10-08T18:18:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 18:18:08 crc kubenswrapper[4859]: I1008 18:18:08.472008 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 18:18:08 crc kubenswrapper[4859]: I1008 18:18:08.472414 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 18:18:08 crc kubenswrapper[4859]: I1008 18:18:08.472857 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 18:18:08 crc kubenswrapper[4859]: I1008 18:18:08.473201 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 18:18:08 crc kubenswrapper[4859]: I1008 18:18:08.473607 4859 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T18:18:08Z","lastTransitionTime":"2025-10-08T18:18:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 18:18:08 crc kubenswrapper[4859]: I1008 18:18:08.577874 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 18:18:08 crc kubenswrapper[4859]: I1008 18:18:08.577936 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 18:18:08 crc kubenswrapper[4859]: I1008 18:18:08.577949 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 18:18:08 crc kubenswrapper[4859]: I1008 18:18:08.577969 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 18:18:08 crc kubenswrapper[4859]: I1008 18:18:08.577983 4859 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T18:18:08Z","lastTransitionTime":"2025-10-08T18:18:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 18:18:08 crc kubenswrapper[4859]: I1008 18:18:08.680756 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 18:18:08 crc kubenswrapper[4859]: I1008 18:18:08.681088 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 18:18:08 crc kubenswrapper[4859]: I1008 18:18:08.681168 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 18:18:08 crc kubenswrapper[4859]: I1008 18:18:08.681258 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 18:18:08 crc kubenswrapper[4859]: I1008 18:18:08.681335 4859 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T18:18:08Z","lastTransitionTime":"2025-10-08T18:18:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 18:18:08 crc kubenswrapper[4859]: I1008 18:18:08.785178 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 18:18:08 crc kubenswrapper[4859]: I1008 18:18:08.785236 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 18:18:08 crc kubenswrapper[4859]: I1008 18:18:08.785249 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 18:18:08 crc kubenswrapper[4859]: I1008 18:18:08.785268 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 18:18:08 crc kubenswrapper[4859]: I1008 18:18:08.785280 4859 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T18:18:08Z","lastTransitionTime":"2025-10-08T18:18:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 18:18:08 crc kubenswrapper[4859]: I1008 18:18:08.888452 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 18:18:08 crc kubenswrapper[4859]: I1008 18:18:08.888535 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 18:18:08 crc kubenswrapper[4859]: I1008 18:18:08.888554 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 18:18:08 crc kubenswrapper[4859]: I1008 18:18:08.888581 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 18:18:08 crc kubenswrapper[4859]: I1008 18:18:08.888604 4859 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T18:18:08Z","lastTransitionTime":"2025-10-08T18:18:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 18:18:08 crc kubenswrapper[4859]: I1008 18:18:08.992125 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 18:18:08 crc kubenswrapper[4859]: I1008 18:18:08.992486 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 18:18:08 crc kubenswrapper[4859]: I1008 18:18:08.992606 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 18:18:08 crc kubenswrapper[4859]: I1008 18:18:08.992854 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 18:18:08 crc kubenswrapper[4859]: I1008 18:18:08.993057 4859 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T18:18:08Z","lastTransitionTime":"2025-10-08T18:18:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 18:18:09 crc kubenswrapper[4859]: I1008 18:18:09.096991 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 18:18:09 crc kubenswrapper[4859]: I1008 18:18:09.097056 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 18:18:09 crc kubenswrapper[4859]: I1008 18:18:09.097066 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 18:18:09 crc kubenswrapper[4859]: I1008 18:18:09.097083 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 18:18:09 crc kubenswrapper[4859]: I1008 18:18:09.097094 4859 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T18:18:09Z","lastTransitionTime":"2025-10-08T18:18:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 18:18:09 crc kubenswrapper[4859]: I1008 18:18:09.200269 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 18:18:09 crc kubenswrapper[4859]: I1008 18:18:09.200347 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 18:18:09 crc kubenswrapper[4859]: I1008 18:18:09.200360 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 18:18:09 crc kubenswrapper[4859]: I1008 18:18:09.200390 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 18:18:09 crc kubenswrapper[4859]: I1008 18:18:09.200409 4859 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T18:18:09Z","lastTransitionTime":"2025-10-08T18:18:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 18:18:09 crc kubenswrapper[4859]: I1008 18:18:09.303564 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 18:18:09 crc kubenswrapper[4859]: I1008 18:18:09.303613 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 18:18:09 crc kubenswrapper[4859]: I1008 18:18:09.303623 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 18:18:09 crc kubenswrapper[4859]: I1008 18:18:09.303640 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 18:18:09 crc kubenswrapper[4859]: I1008 18:18:09.303652 4859 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T18:18:09Z","lastTransitionTime":"2025-10-08T18:18:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 18:18:09 crc kubenswrapper[4859]: I1008 18:18:09.406833 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 18:18:09 crc kubenswrapper[4859]: I1008 18:18:09.406907 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 18:18:09 crc kubenswrapper[4859]: I1008 18:18:09.406930 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 18:18:09 crc kubenswrapper[4859]: I1008 18:18:09.406959 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 18:18:09 crc kubenswrapper[4859]: I1008 18:18:09.406980 4859 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T18:18:09Z","lastTransitionTime":"2025-10-08T18:18:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 18:18:09 crc kubenswrapper[4859]: I1008 18:18:09.469057 4859 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 08 18:18:09 crc kubenswrapper[4859]: I1008 18:18:09.469592 4859 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 08 18:18:09 crc kubenswrapper[4859]: I1008 18:18:09.469904 4859 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-4b4sz" Oct 08 18:18:09 crc kubenswrapper[4859]: E1008 18:18:09.469905 4859 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 08 18:18:09 crc kubenswrapper[4859]: I1008 18:18:09.470059 4859 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 08 18:18:09 crc kubenswrapper[4859]: E1008 18:18:09.470155 4859 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 08 18:18:09 crc kubenswrapper[4859]: E1008 18:18:09.470256 4859 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-4b4sz" podUID="2e53cd9b-64f7-4e07-8b96-fb77847c9ac6" Oct 08 18:18:09 crc kubenswrapper[4859]: E1008 18:18:09.470968 4859 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 08 18:18:09 crc kubenswrapper[4859]: I1008 18:18:09.509796 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 18:18:09 crc kubenswrapper[4859]: I1008 18:18:09.510106 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 18:18:09 crc kubenswrapper[4859]: I1008 18:18:09.510174 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 18:18:09 crc kubenswrapper[4859]: I1008 18:18:09.510256 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 18:18:09 crc kubenswrapper[4859]: I1008 18:18:09.510355 4859 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T18:18:09Z","lastTransitionTime":"2025-10-08T18:18:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 18:18:09 crc kubenswrapper[4859]: I1008 18:18:09.612792 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 18:18:09 crc kubenswrapper[4859]: I1008 18:18:09.612869 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 18:18:09 crc kubenswrapper[4859]: I1008 18:18:09.612896 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 18:18:09 crc kubenswrapper[4859]: I1008 18:18:09.612928 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 18:18:09 crc kubenswrapper[4859]: I1008 18:18:09.612951 4859 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T18:18:09Z","lastTransitionTime":"2025-10-08T18:18:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 18:18:09 crc kubenswrapper[4859]: I1008 18:18:09.716233 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 18:18:09 crc kubenswrapper[4859]: I1008 18:18:09.716768 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 18:18:09 crc kubenswrapper[4859]: I1008 18:18:09.717024 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 18:18:09 crc kubenswrapper[4859]: I1008 18:18:09.717238 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 18:18:09 crc kubenswrapper[4859]: I1008 18:18:09.717447 4859 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T18:18:09Z","lastTransitionTime":"2025-10-08T18:18:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 18:18:09 crc kubenswrapper[4859]: I1008 18:18:09.821317 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 18:18:09 crc kubenswrapper[4859]: I1008 18:18:09.821356 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 18:18:09 crc kubenswrapper[4859]: I1008 18:18:09.821368 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 18:18:09 crc kubenswrapper[4859]: I1008 18:18:09.821393 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 18:18:09 crc kubenswrapper[4859]: I1008 18:18:09.821405 4859 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T18:18:09Z","lastTransitionTime":"2025-10-08T18:18:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 18:18:09 crc kubenswrapper[4859]: I1008 18:18:09.924540 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 18:18:09 crc kubenswrapper[4859]: I1008 18:18:09.924586 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 18:18:09 crc kubenswrapper[4859]: I1008 18:18:09.924596 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 18:18:09 crc kubenswrapper[4859]: I1008 18:18:09.924613 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 18:18:09 crc kubenswrapper[4859]: I1008 18:18:09.924624 4859 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T18:18:09Z","lastTransitionTime":"2025-10-08T18:18:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 18:18:10 crc kubenswrapper[4859]: I1008 18:18:10.026832 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 18:18:10 crc kubenswrapper[4859]: I1008 18:18:10.026884 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 18:18:10 crc kubenswrapper[4859]: I1008 18:18:10.026894 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 18:18:10 crc kubenswrapper[4859]: I1008 18:18:10.026912 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 18:18:10 crc kubenswrapper[4859]: I1008 18:18:10.026922 4859 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T18:18:10Z","lastTransitionTime":"2025-10-08T18:18:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 18:18:10 crc kubenswrapper[4859]: I1008 18:18:10.129702 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 18:18:10 crc kubenswrapper[4859]: I1008 18:18:10.129741 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 18:18:10 crc kubenswrapper[4859]: I1008 18:18:10.129754 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 18:18:10 crc kubenswrapper[4859]: I1008 18:18:10.129770 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 18:18:10 crc kubenswrapper[4859]: I1008 18:18:10.129783 4859 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T18:18:10Z","lastTransitionTime":"2025-10-08T18:18:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 18:18:10 crc kubenswrapper[4859]: I1008 18:18:10.232980 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 18:18:10 crc kubenswrapper[4859]: I1008 18:18:10.233374 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 18:18:10 crc kubenswrapper[4859]: I1008 18:18:10.233451 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 18:18:10 crc kubenswrapper[4859]: I1008 18:18:10.233525 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 18:18:10 crc kubenswrapper[4859]: I1008 18:18:10.233594 4859 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T18:18:10Z","lastTransitionTime":"2025-10-08T18:18:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 18:18:10 crc kubenswrapper[4859]: I1008 18:18:10.336739 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 18:18:10 crc kubenswrapper[4859]: I1008 18:18:10.337413 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 18:18:10 crc kubenswrapper[4859]: I1008 18:18:10.337493 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 18:18:10 crc kubenswrapper[4859]: I1008 18:18:10.337601 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 18:18:10 crc kubenswrapper[4859]: I1008 18:18:10.337712 4859 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T18:18:10Z","lastTransitionTime":"2025-10-08T18:18:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 18:18:10 crc kubenswrapper[4859]: I1008 18:18:10.440794 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 18:18:10 crc kubenswrapper[4859]: I1008 18:18:10.440871 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 18:18:10 crc kubenswrapper[4859]: I1008 18:18:10.440886 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 18:18:10 crc kubenswrapper[4859]: I1008 18:18:10.440912 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 18:18:10 crc kubenswrapper[4859]: I1008 18:18:10.440926 4859 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T18:18:10Z","lastTransitionTime":"2025-10-08T18:18:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 18:18:10 crc kubenswrapper[4859]: I1008 18:18:10.486476 4859 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3ca6ce76-77e8-4e46-a2f6-b1ed4aa0fd4b\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:17:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:17:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:18:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:18:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:17:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f587e149d57ea53626ed3beb83e421c3b89ba1fec9e949705e8ccb5b4f7cc79a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T18:17:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://428cbe63daefbfbcb2ec76dad0fe638cdc65934bac80b791cebe0b36b2d6909a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T18:17:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a41399a37b1ad86a063001224232ecb8f7ea56aeb5686d3858bc08102a0c8f49\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T18:17:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://008757fce552591f7c58193f605eb60792761d5943927cbd63184549618d2b43\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://008757fce552591f7c58193f605eb60792761d5943927cbd63184549618d2b43\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-08T18:17:11Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-08T18:17:11Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T18:17:10Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T18:18:10Z is after 2025-08-24T17:21:41Z" Oct 08 18:18:10 crc kubenswrapper[4859]: I1008 18:18:10.504251 4859 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-08T18:17:29Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T18:17:29Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T18:18:10Z is after 2025-08-24T17:21:41Z" Oct 08 18:18:10 crc kubenswrapper[4859]: I1008 18:18:10.517236 4859 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-btw28" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"97bc5598-3ead-4300-a938-3aebbff1e8c6\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:17:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:17:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:17:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:17:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d6d4d6ba771cb5fde637449bba05660f634df7e833a0c8dad2b2c7f3a422bb2b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T18:17:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qcxsd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T18:17:31Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-btw28\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T18:18:10Z is after 2025-08-24T17:21:41Z" Oct 08 18:18:10 crc kubenswrapper[4859]: I1008 18:18:10.533955 4859 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-8plkg" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"92b2d7de-31cc-4a91-95a5-ed5ea964b028\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:17:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:17:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:17:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:17:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fa00b8345d40b74a917ddc23cc1b2c5c684a9ca61b6119d090c92cc1439b71c4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T18:17:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q9d7m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T18:17:31Z\\\"}}\" for pod \"openshift-multus\"/\"multus-8plkg\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T18:18:10Z is after 2025-08-24T17:21:41Z" Oct 08 18:18:10 crc kubenswrapper[4859]: I1008 18:18:10.545148 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 18:18:10 crc kubenswrapper[4859]: I1008 18:18:10.545197 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 18:18:10 crc kubenswrapper[4859]: I1008 18:18:10.545264 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 18:18:10 crc kubenswrapper[4859]: I1008 18:18:10.545295 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 18:18:10 crc kubenswrapper[4859]: I1008 18:18:10.545310 4859 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T18:18:10Z","lastTransitionTime":"2025-10-08T18:18:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 18:18:10 crc kubenswrapper[4859]: I1008 18:18:10.554137 4859 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-btjmq" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b8605f7-fed7-493f-9b4f-6a6358adf907\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:17:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:17:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:17:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:17:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://14e9b6c4a4403f36adb78fcf1135a620ca6fc56ed974c9639dcdfc59e6917559\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T18:17:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xspg4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T18:17:35Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-btjmq\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T18:18:10Z is after 2025-08-24T17:21:41Z" Oct 08 18:18:10 crc kubenswrapper[4859]: I1008 18:18:10.568641 4859 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-g4t79" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3a92dbf1-21a8-4095-96bb-2d701d7f3d56\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:17:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:17:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:17:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:17:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f8772a25b6e605efe9a062dc32dbbc2cfe32b4fe164b4e84fbf3bcffab744179\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T18:17:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sv79p\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://92cb9986c09a7f9e1e552e3e3ac1f74f04128241e00ddee38882d2ba5c005a67\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T18:17:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sv79p\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T18:17:45Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-g4t79\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T18:18:10Z is after 2025-08-24T17:21:41Z" Oct 08 18:18:10 crc kubenswrapper[4859]: I1008 18:18:10.583488 4859 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-08T18:17:31Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T18:17:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T18:17:31Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c6d15905694b6bbd4999d49dc8f7a53eb6bfaa7adb59efb0d7e4973c21d1c7be\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T18:17:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T18:18:10Z is after 2025-08-24T17:21:41Z" Oct 08 18:18:10 crc kubenswrapper[4859]: I1008 18:18:10.598032 4859 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-08T18:17:33Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T18:17:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T18:17:33Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5d99e071bd083914418d9c128f38955f26295c08b1520f2dc774c99b1c843693\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T18:17:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T18:18:10Z is after 2025-08-24T17:21:41Z" Oct 08 18:18:10 crc kubenswrapper[4859]: I1008 18:18:10.611459 4859 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-4b4sz" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2e53cd9b-64f7-4e07-8b96-fb77847c9ac6\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:17:45Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:17:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:17:45Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:17:45Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kxgbz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kxgbz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T18:17:45Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-4b4sz\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T18:18:10Z is after 2025-08-24T17:21:41Z" Oct 08 18:18:10 crc kubenswrapper[4859]: I1008 18:18:10.629099 4859 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"63a623de-5608-4228-b1c2-77188c276986\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:17:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:17:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:17:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:17:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:17:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d4786a8fb8abd96fb22e60733c974f599017407d589ed815fb8541fd3b64f696\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T18:17:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://183d481502ed5108d5e19e483b714c289f355ea3f8b0891883e5335b802c83b6\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T18:17:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ec848c46fd5dde0469b00d3cab14ba4a652533a265490fadbee7a8b8843ef249\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T18:17:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4fe1828b4d63ff8106bfcd15e1dc1c1bd1ba7f77777cdb23ff1b291d0418774c\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2e700b2a20c81b414617f5b5c018a738278a3fa44e6c6d14b2170051d0725431\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-08T18:17:24Z\\\",\\\"message\\\":\\\"W1008 18:17:13.670625 1 cmd.go:257] Using insecure, self-signed certificates\\\\nI1008 18:17:13.671202 1 crypto.go:601] Generating new CA for check-endpoints-signer@1759947433 cert, and key in /tmp/serving-cert-3867218267/serving-signer.crt, /tmp/serving-cert-3867218267/serving-signer.key\\\\nI1008 18:17:13.944463 1 observer_polling.go:159] Starting file observer\\\\nW1008 18:17:13.951958 1 builder.go:272] unable to get owner reference (falling back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": dial tcp [::1]:6443: connect: connection refused\\\\nI1008 18:17:13.952088 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1008 18:17:13.956755 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3867218267/tls.crt::/tmp/serving-cert-3867218267/tls.key\\\\\\\"\\\\nF1008 18:17:24.290991 1 cmd.go:182] error initializing delegating authentication: unable to load configmap based request-header-client-ca-file: Get \\\\\\\"https://localhost:6443/api/v1/namespaces/kube-system/configmaps/extension-apiserver-authentication\\\\\\\": net/http: TLS handshake timeout\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-08T18:17:13Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T18:17:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c65b66f62d325ca10ef8d5d31f04529ccf4e691cf3303d231e7a414fb400e4d6\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T18:17:13Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://58c7bba4ace28215b062653a960e561b1dc1f890f2baf450145638a065af038b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://58c7bba4ace28215b062653a960e561b1dc1f890f2baf450145638a065af038b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-08T18:17:11Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-08T18:17:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T18:17:10Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T18:18:10Z is after 2025-08-24T17:21:41Z" Oct 08 18:18:10 crc kubenswrapper[4859]: I1008 18:18:10.644624 4859 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-08T18:17:31Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T18:17:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T18:17:31Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://99e2d51bd7757498dae60949eea6d732b5f7b481fa00f808dd5308a530a1daf1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T18:17:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e32aaba93699944665660ae41e9b5779178c1354c31e15ad8188bf8c3c8fa500\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T18:17:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T18:18:10Z is after 2025-08-24T17:21:41Z" Oct 08 18:18:10 crc kubenswrapper[4859]: I1008 18:18:10.649555 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 18:18:10 crc kubenswrapper[4859]: I1008 18:18:10.649613 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 18:18:10 crc kubenswrapper[4859]: I1008 18:18:10.649627 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 18:18:10 crc kubenswrapper[4859]: I1008 18:18:10.649647 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 18:18:10 crc kubenswrapper[4859]: I1008 18:18:10.649661 4859 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T18:18:10Z","lastTransitionTime":"2025-10-08T18:18:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 18:18:10 crc kubenswrapper[4859]: I1008 18:18:10.662265 4859 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-08T18:17:29Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T18:17:29Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T18:18:10Z is after 2025-08-24T17:21:41Z" Oct 08 18:18:10 crc kubenswrapper[4859]: I1008 18:18:10.682482 4859 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-82s52" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b23a6a6c-9d92-4e7b-840e-55cfda873a2d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:17:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:17:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:17:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:17:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://71302c766485b9f55743d85f97440ff578cd44ba1e5a6f5b1c024c0e808072d2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T18:17:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6vx22\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c23ea3580998adf90621a0155c20ee4bd590c512629812174e95c28857bcfef9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T18:17:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6vx22\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T18:17:31Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-82s52\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T18:18:10Z is after 2025-08-24T17:21:41Z" Oct 08 18:18:10 crc kubenswrapper[4859]: I1008 18:18:10.706464 4859 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-4g8gf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1dff864e-b75d-4e0b-b182-75f710eac8df\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:17:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:17:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:17:32Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:17:32Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://768ca274dedcc6847fc21d15234d3665fa9ccae371b5b6a882cfcb9b182f527d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T18:17:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s8jnn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://669f5f3c6d3c1ec0c4b4870e090b0a853adc01727c23383e595715b476c32668\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T18:17:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s8jnn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b8009a42cc4c9a816457a0678bdb2f7b7c53d0c8604e007a6fb141f66d06a516\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T18:17:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s8jnn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://676511a11ed4102084b57ebdd5f3fda4969741b2bc8443b415c4bf94f39ae4c9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T18:17:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s8jnn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4a74c1f0d21c90cdd7e776ee14e1f48757a04e2220fbd1f884449181ec6d1c3a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T18:17:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s8jnn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6ddc0e842bb8f81663406070023b0e5ed9cc99c98db0123b1c84f20da59fb5b5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T18:17:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s8jnn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c646f0e0dd8e7f74125b687f0cb4114b1f34667e5a6325f394f0208d7fb89df6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c646f0e0dd8e7f74125b687f0cb4114b1f34667e5a6325f394f0208d7fb89df6\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-10-08T18:17:58Z\\\",\\\"message\\\":\\\"e *v1.Pod retry setup to complete in iterateRetryResources\\\\nI1008 18:17:58.424504 6520 obj_retry.go:365] Adding new object: *v1.Pod openshift-machine-config-operator/machine-config-daemon-82s52\\\\nI1008 18:17:58.424529 6520 ovn.go:134] Ensuring zone local for Pod openshift-machine-config-operator/machine-config-daemon-82s52 in node crc\\\\nI1008 18:17:58.424539 6520 obj_retry.go:303] Retry object setup: *v1.Pod openshift-network-diagnostics/network-check-source-55646444c4-trplf\\\\nI1008 18:17:58.424549 6520 obj_retry.go:365] Adding new object: *v1.Pod openshift-network-diagnostics/network-check-source-55646444c4-trplf\\\\nI1008 18:17:58.424549 6520 obj_retry.go:386] Retry successful for *v1.Pod openshift-machine-config-operator/machine-config-daemon-82s52 after 0 failed attempt(s)\\\\nI1008 18:17:58.424559 6520 ovn.go:134] Ensuring zone local for Pod openshift-network-diagnostics/network-check-source-55646444c4-trplf in node crc\\\\nF1008 18:17:58.424552 6520 ovnkube.go:137] failed to run ovnkube: [failed to start network controller: failed to start default network controller: unable to create admin network policy controller, err: could not add Event Handler for anpInformer during admin network policy controller initialization, handler {0x1fcc6e0 0x1fcc3c0 0x1fcc360} was not added to shared informer because it has stopped already, failed to star\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-08T18:17:57Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 20s restarting failed container=ovnkube-controller pod=ovnkube-node-4g8gf_openshift-ovn-kubernetes(1dff864e-b75d-4e0b-b182-75f710eac8df)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s8jnn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2c67d2aed7126d76989e057d8e2d8e188a8902b2f3fe26ed6d1d45d3b013d897\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T18:17:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s8jnn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7a211d3da02b173867ff3a912eecd3b379483c7cc77e6b91122edd1ffd30fd53\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7a211d3da02b173867ff3a912eecd3b379483c7cc77e6b91122edd1ffd30fd53\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-08T18:17:32Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-08T18:17:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s8jnn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T18:17:32Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-4g8gf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T18:18:10Z is after 2025-08-24T17:21:41Z" Oct 08 18:18:10 crc kubenswrapper[4859]: I1008 18:18:10.723110 4859 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4efc16fb-e679-464d-8b3c-0f46f082593e\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:17:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:17:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:17:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:17:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:17:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://aa8f8ea895272cd79a08d1c47cd7863a1ec10e21bb60d12aa10862a56c99e4bf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T18:17:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5c4e95d8a712ab219455564539a90286ae48bbc46ddc322a5bfd96c1a8268110\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T18:17:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://bae6ecfd8b446ef04d393d823e5ffe37fe749337c0ca17fde79c5f88d5e31393\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T18:17:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6411d1e7e18d5ceadaaceed5a8a6b20a0cacc55cede793ca9fb4dcad70c64fc6\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T18:17:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T18:17:10Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T18:18:10Z is after 2025-08-24T17:21:41Z" Oct 08 18:18:10 crc kubenswrapper[4859]: I1008 18:18:10.738498 4859 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-08T18:17:29Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T18:17:29Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T18:18:10Z is after 2025-08-24T17:21:41Z" Oct 08 18:18:10 crc kubenswrapper[4859]: I1008 18:18:10.752420 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 18:18:10 crc kubenswrapper[4859]: I1008 18:18:10.752493 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 18:18:10 crc kubenswrapper[4859]: I1008 18:18:10.752532 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 18:18:10 crc kubenswrapper[4859]: I1008 18:18:10.752552 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 18:18:10 crc kubenswrapper[4859]: I1008 18:18:10.752564 4859 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T18:18:10Z","lastTransitionTime":"2025-10-08T18:18:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 18:18:10 crc kubenswrapper[4859]: I1008 18:18:10.755704 4859 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-84k8x" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1c24078c-c9e8-4160-883a-f483b4b3dd23\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:17:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:17:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:17:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:17:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8661006834d0f50eaada7394369584ee95fc61ff54437eec35d755526afc5b48\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T18:17:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hk5xb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://18b291eb3344fb7c00ef5679dca9a6f7ed9d08cea684a803ff7dad2e8a18c585\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://18b291eb3344fb7c00ef5679dca9a6f7ed9d08cea684a803ff7dad2e8a18c585\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-08T18:17:32Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-08T18:17:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hk5xb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://160163f1a91d981583452db3c7a9244182d1c8ced45749a61b721d0d45c1dfe4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://160163f1a91d981583452db3c7a9244182d1c8ced45749a61b721d0d45c1dfe4\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-08T18:17:34Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-08T18:17:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hk5xb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://51e9b53f3847ac53f378ad83870debf8e71518d4edbc7fc0dc66644818e2103e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://51e9b53f3847ac53f378ad83870debf8e71518d4edbc7fc0dc66644818e2103e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-08T18:17:34Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-08T18:17:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hk5xb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1ff118919739759ac01b5cbe3240e6673db54eaa16353c609d670071ed7964cb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1ff118919739759ac01b5cbe3240e6673db54eaa16353c609d670071ed7964cb\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-08T18:17:35Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-08T18:17:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hk5xb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8522e803ff56eb57cd4d9fd036e0c3fb16240ea21d4a094c1db7aa7d50bfe2c2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8522e803ff56eb57cd4d9fd036e0c3fb16240ea21d4a094c1db7aa7d50bfe2c2\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-08T18:17:37Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-08T18:17:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hk5xb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://be61056edf515e037403ca2a0bfd5feb9b2537dd42212b58e133b23b294f327b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://be61056edf515e037403ca2a0bfd5feb9b2537dd42212b58e133b23b294f327b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-08T18:17:38Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-08T18:17:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hk5xb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T18:17:31Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-84k8x\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T18:18:10Z is after 2025-08-24T17:21:41Z" Oct 08 18:18:10 crc kubenswrapper[4859]: I1008 18:18:10.856371 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 18:18:10 crc kubenswrapper[4859]: I1008 18:18:10.856410 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 18:18:10 crc kubenswrapper[4859]: I1008 18:18:10.856419 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 18:18:10 crc kubenswrapper[4859]: I1008 18:18:10.856434 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 18:18:10 crc kubenswrapper[4859]: I1008 18:18:10.856444 4859 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T18:18:10Z","lastTransitionTime":"2025-10-08T18:18:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 18:18:10 crc kubenswrapper[4859]: I1008 18:18:10.959641 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 18:18:10 crc kubenswrapper[4859]: I1008 18:18:10.959738 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 18:18:10 crc kubenswrapper[4859]: I1008 18:18:10.959749 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 18:18:10 crc kubenswrapper[4859]: I1008 18:18:10.959769 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 18:18:10 crc kubenswrapper[4859]: I1008 18:18:10.959781 4859 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T18:18:10Z","lastTransitionTime":"2025-10-08T18:18:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 18:18:11 crc kubenswrapper[4859]: I1008 18:18:11.062354 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 18:18:11 crc kubenswrapper[4859]: I1008 18:18:11.062401 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 18:18:11 crc kubenswrapper[4859]: I1008 18:18:11.062412 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 18:18:11 crc kubenswrapper[4859]: I1008 18:18:11.062430 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 18:18:11 crc kubenswrapper[4859]: I1008 18:18:11.062444 4859 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T18:18:11Z","lastTransitionTime":"2025-10-08T18:18:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 18:18:11 crc kubenswrapper[4859]: I1008 18:18:11.166493 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 18:18:11 crc kubenswrapper[4859]: I1008 18:18:11.166547 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 18:18:11 crc kubenswrapper[4859]: I1008 18:18:11.166558 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 18:18:11 crc kubenswrapper[4859]: I1008 18:18:11.166584 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 18:18:11 crc kubenswrapper[4859]: I1008 18:18:11.166596 4859 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T18:18:11Z","lastTransitionTime":"2025-10-08T18:18:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 18:18:11 crc kubenswrapper[4859]: I1008 18:18:11.270314 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 18:18:11 crc kubenswrapper[4859]: I1008 18:18:11.270364 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 18:18:11 crc kubenswrapper[4859]: I1008 18:18:11.270373 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 18:18:11 crc kubenswrapper[4859]: I1008 18:18:11.270390 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 18:18:11 crc kubenswrapper[4859]: I1008 18:18:11.270402 4859 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T18:18:11Z","lastTransitionTime":"2025-10-08T18:18:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 18:18:11 crc kubenswrapper[4859]: I1008 18:18:11.373196 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 18:18:11 crc kubenswrapper[4859]: I1008 18:18:11.373292 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 18:18:11 crc kubenswrapper[4859]: I1008 18:18:11.373307 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 18:18:11 crc kubenswrapper[4859]: I1008 18:18:11.373329 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 18:18:11 crc kubenswrapper[4859]: I1008 18:18:11.373343 4859 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T18:18:11Z","lastTransitionTime":"2025-10-08T18:18:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 18:18:11 crc kubenswrapper[4859]: I1008 18:18:11.469140 4859 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-4b4sz" Oct 08 18:18:11 crc kubenswrapper[4859]: I1008 18:18:11.469185 4859 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 08 18:18:11 crc kubenswrapper[4859]: I1008 18:18:11.469146 4859 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 08 18:18:11 crc kubenswrapper[4859]: E1008 18:18:11.469298 4859 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-4b4sz" podUID="2e53cd9b-64f7-4e07-8b96-fb77847c9ac6" Oct 08 18:18:11 crc kubenswrapper[4859]: I1008 18:18:11.469326 4859 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 08 18:18:11 crc kubenswrapper[4859]: E1008 18:18:11.469394 4859 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 08 18:18:11 crc kubenswrapper[4859]: E1008 18:18:11.469447 4859 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 08 18:18:11 crc kubenswrapper[4859]: E1008 18:18:11.469483 4859 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 08 18:18:11 crc kubenswrapper[4859]: I1008 18:18:11.476401 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 18:18:11 crc kubenswrapper[4859]: I1008 18:18:11.476464 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 18:18:11 crc kubenswrapper[4859]: I1008 18:18:11.476475 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 18:18:11 crc kubenswrapper[4859]: I1008 18:18:11.476489 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 18:18:11 crc kubenswrapper[4859]: I1008 18:18:11.476500 4859 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T18:18:11Z","lastTransitionTime":"2025-10-08T18:18:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 18:18:11 crc kubenswrapper[4859]: I1008 18:18:11.579169 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 18:18:11 crc kubenswrapper[4859]: I1008 18:18:11.579236 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 18:18:11 crc kubenswrapper[4859]: I1008 18:18:11.579249 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 18:18:11 crc kubenswrapper[4859]: I1008 18:18:11.579271 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 18:18:11 crc kubenswrapper[4859]: I1008 18:18:11.579286 4859 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T18:18:11Z","lastTransitionTime":"2025-10-08T18:18:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 18:18:11 crc kubenswrapper[4859]: I1008 18:18:11.682258 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 18:18:11 crc kubenswrapper[4859]: I1008 18:18:11.682306 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 18:18:11 crc kubenswrapper[4859]: I1008 18:18:11.682316 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 18:18:11 crc kubenswrapper[4859]: I1008 18:18:11.682333 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 18:18:11 crc kubenswrapper[4859]: I1008 18:18:11.682344 4859 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T18:18:11Z","lastTransitionTime":"2025-10-08T18:18:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 18:18:11 crc kubenswrapper[4859]: I1008 18:18:11.785860 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 18:18:11 crc kubenswrapper[4859]: I1008 18:18:11.786269 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 18:18:11 crc kubenswrapper[4859]: I1008 18:18:11.786404 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 18:18:11 crc kubenswrapper[4859]: I1008 18:18:11.786508 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 18:18:11 crc kubenswrapper[4859]: I1008 18:18:11.786596 4859 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T18:18:11Z","lastTransitionTime":"2025-10-08T18:18:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 18:18:11 crc kubenswrapper[4859]: I1008 18:18:11.889552 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 18:18:11 crc kubenswrapper[4859]: I1008 18:18:11.889915 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 18:18:11 crc kubenswrapper[4859]: I1008 18:18:11.890061 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 18:18:11 crc kubenswrapper[4859]: I1008 18:18:11.890400 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 18:18:11 crc kubenswrapper[4859]: I1008 18:18:11.890836 4859 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T18:18:11Z","lastTransitionTime":"2025-10-08T18:18:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 18:18:11 crc kubenswrapper[4859]: I1008 18:18:11.993713 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 18:18:11 crc kubenswrapper[4859]: I1008 18:18:11.993771 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 18:18:11 crc kubenswrapper[4859]: I1008 18:18:11.993784 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 18:18:11 crc kubenswrapper[4859]: I1008 18:18:11.993803 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 18:18:11 crc kubenswrapper[4859]: I1008 18:18:11.993814 4859 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T18:18:11Z","lastTransitionTime":"2025-10-08T18:18:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 18:18:12 crc kubenswrapper[4859]: I1008 18:18:12.096295 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 18:18:12 crc kubenswrapper[4859]: I1008 18:18:12.096366 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 18:18:12 crc kubenswrapper[4859]: I1008 18:18:12.096379 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 18:18:12 crc kubenswrapper[4859]: I1008 18:18:12.096404 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 18:18:12 crc kubenswrapper[4859]: I1008 18:18:12.096420 4859 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T18:18:12Z","lastTransitionTime":"2025-10-08T18:18:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 18:18:12 crc kubenswrapper[4859]: I1008 18:18:12.199775 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 18:18:12 crc kubenswrapper[4859]: I1008 18:18:12.199819 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 18:18:12 crc kubenswrapper[4859]: I1008 18:18:12.199829 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 18:18:12 crc kubenswrapper[4859]: I1008 18:18:12.199846 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 18:18:12 crc kubenswrapper[4859]: I1008 18:18:12.199858 4859 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T18:18:12Z","lastTransitionTime":"2025-10-08T18:18:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 18:18:12 crc kubenswrapper[4859]: I1008 18:18:12.303883 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 18:18:12 crc kubenswrapper[4859]: I1008 18:18:12.303973 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 18:18:12 crc kubenswrapper[4859]: I1008 18:18:12.303990 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 18:18:12 crc kubenswrapper[4859]: I1008 18:18:12.304017 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 18:18:12 crc kubenswrapper[4859]: I1008 18:18:12.304039 4859 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T18:18:12Z","lastTransitionTime":"2025-10-08T18:18:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 18:18:12 crc kubenswrapper[4859]: I1008 18:18:12.406964 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 18:18:12 crc kubenswrapper[4859]: I1008 18:18:12.407036 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 18:18:12 crc kubenswrapper[4859]: I1008 18:18:12.407056 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 18:18:12 crc kubenswrapper[4859]: I1008 18:18:12.407084 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 18:18:12 crc kubenswrapper[4859]: I1008 18:18:12.407101 4859 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T18:18:12Z","lastTransitionTime":"2025-10-08T18:18:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 18:18:12 crc kubenswrapper[4859]: I1008 18:18:12.510113 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 18:18:12 crc kubenswrapper[4859]: I1008 18:18:12.510185 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 18:18:12 crc kubenswrapper[4859]: I1008 18:18:12.510208 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 18:18:12 crc kubenswrapper[4859]: I1008 18:18:12.510240 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 18:18:12 crc kubenswrapper[4859]: I1008 18:18:12.510263 4859 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T18:18:12Z","lastTransitionTime":"2025-10-08T18:18:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 18:18:12 crc kubenswrapper[4859]: I1008 18:18:12.613295 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 18:18:12 crc kubenswrapper[4859]: I1008 18:18:12.613375 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 18:18:12 crc kubenswrapper[4859]: I1008 18:18:12.613398 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 18:18:12 crc kubenswrapper[4859]: I1008 18:18:12.613429 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 18:18:12 crc kubenswrapper[4859]: I1008 18:18:12.613452 4859 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T18:18:12Z","lastTransitionTime":"2025-10-08T18:18:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 18:18:12 crc kubenswrapper[4859]: I1008 18:18:12.718308 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 18:18:12 crc kubenswrapper[4859]: I1008 18:18:12.718360 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 18:18:12 crc kubenswrapper[4859]: I1008 18:18:12.718398 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 18:18:12 crc kubenswrapper[4859]: I1008 18:18:12.718423 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 18:18:12 crc kubenswrapper[4859]: I1008 18:18:12.718436 4859 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T18:18:12Z","lastTransitionTime":"2025-10-08T18:18:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 18:18:12 crc kubenswrapper[4859]: I1008 18:18:12.820869 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 18:18:12 crc kubenswrapper[4859]: I1008 18:18:12.820939 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 18:18:12 crc kubenswrapper[4859]: I1008 18:18:12.820953 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 18:18:12 crc kubenswrapper[4859]: I1008 18:18:12.820975 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 18:18:12 crc kubenswrapper[4859]: I1008 18:18:12.820989 4859 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T18:18:12Z","lastTransitionTime":"2025-10-08T18:18:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 18:18:12 crc kubenswrapper[4859]: I1008 18:18:12.923897 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 18:18:12 crc kubenswrapper[4859]: I1008 18:18:12.923950 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 18:18:12 crc kubenswrapper[4859]: I1008 18:18:12.923962 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 18:18:12 crc kubenswrapper[4859]: I1008 18:18:12.923980 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 18:18:12 crc kubenswrapper[4859]: I1008 18:18:12.923990 4859 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T18:18:12Z","lastTransitionTime":"2025-10-08T18:18:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 18:18:13 crc kubenswrapper[4859]: I1008 18:18:13.026798 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 18:18:13 crc kubenswrapper[4859]: I1008 18:18:13.026864 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 18:18:13 crc kubenswrapper[4859]: I1008 18:18:13.026882 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 18:18:13 crc kubenswrapper[4859]: I1008 18:18:13.026906 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 18:18:13 crc kubenswrapper[4859]: I1008 18:18:13.026926 4859 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T18:18:13Z","lastTransitionTime":"2025-10-08T18:18:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 18:18:13 crc kubenswrapper[4859]: I1008 18:18:13.130064 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 18:18:13 crc kubenswrapper[4859]: I1008 18:18:13.130117 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 18:18:13 crc kubenswrapper[4859]: I1008 18:18:13.130133 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 18:18:13 crc kubenswrapper[4859]: I1008 18:18:13.130159 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 18:18:13 crc kubenswrapper[4859]: I1008 18:18:13.130178 4859 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T18:18:13Z","lastTransitionTime":"2025-10-08T18:18:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 18:18:13 crc kubenswrapper[4859]: I1008 18:18:13.233069 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 18:18:13 crc kubenswrapper[4859]: I1008 18:18:13.233128 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 18:18:13 crc kubenswrapper[4859]: I1008 18:18:13.233137 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 18:18:13 crc kubenswrapper[4859]: I1008 18:18:13.233155 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 18:18:13 crc kubenswrapper[4859]: I1008 18:18:13.233164 4859 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T18:18:13Z","lastTransitionTime":"2025-10-08T18:18:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 18:18:13 crc kubenswrapper[4859]: I1008 18:18:13.337096 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 18:18:13 crc kubenswrapper[4859]: I1008 18:18:13.337166 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 18:18:13 crc kubenswrapper[4859]: I1008 18:18:13.337208 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 18:18:13 crc kubenswrapper[4859]: I1008 18:18:13.337237 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 18:18:13 crc kubenswrapper[4859]: I1008 18:18:13.337255 4859 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T18:18:13Z","lastTransitionTime":"2025-10-08T18:18:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 18:18:13 crc kubenswrapper[4859]: I1008 18:18:13.441911 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 18:18:13 crc kubenswrapper[4859]: I1008 18:18:13.441981 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 18:18:13 crc kubenswrapper[4859]: I1008 18:18:13.442001 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 18:18:13 crc kubenswrapper[4859]: I1008 18:18:13.442028 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 18:18:13 crc kubenswrapper[4859]: I1008 18:18:13.442047 4859 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T18:18:13Z","lastTransitionTime":"2025-10-08T18:18:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 18:18:13 crc kubenswrapper[4859]: I1008 18:18:13.469488 4859 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 08 18:18:13 crc kubenswrapper[4859]: I1008 18:18:13.469644 4859 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-4b4sz" Oct 08 18:18:13 crc kubenswrapper[4859]: I1008 18:18:13.469644 4859 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 08 18:18:13 crc kubenswrapper[4859]: I1008 18:18:13.469578 4859 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 08 18:18:13 crc kubenswrapper[4859]: E1008 18:18:13.469844 4859 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 08 18:18:13 crc kubenswrapper[4859]: E1008 18:18:13.470012 4859 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 08 18:18:13 crc kubenswrapper[4859]: E1008 18:18:13.470231 4859 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-4b4sz" podUID="2e53cd9b-64f7-4e07-8b96-fb77847c9ac6" Oct 08 18:18:13 crc kubenswrapper[4859]: E1008 18:18:13.470398 4859 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 08 18:18:13 crc kubenswrapper[4859]: I1008 18:18:13.545454 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 18:18:13 crc kubenswrapper[4859]: I1008 18:18:13.545514 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 18:18:13 crc kubenswrapper[4859]: I1008 18:18:13.545527 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 18:18:13 crc kubenswrapper[4859]: I1008 18:18:13.545550 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 18:18:13 crc kubenswrapper[4859]: I1008 18:18:13.545565 4859 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T18:18:13Z","lastTransitionTime":"2025-10-08T18:18:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 18:18:13 crc kubenswrapper[4859]: I1008 18:18:13.648980 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 18:18:13 crc kubenswrapper[4859]: I1008 18:18:13.649044 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 18:18:13 crc kubenswrapper[4859]: I1008 18:18:13.649061 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 18:18:13 crc kubenswrapper[4859]: I1008 18:18:13.649087 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 18:18:13 crc kubenswrapper[4859]: I1008 18:18:13.649105 4859 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T18:18:13Z","lastTransitionTime":"2025-10-08T18:18:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 18:18:13 crc kubenswrapper[4859]: I1008 18:18:13.752469 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 18:18:13 crc kubenswrapper[4859]: I1008 18:18:13.752830 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 18:18:13 crc kubenswrapper[4859]: I1008 18:18:13.752857 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 18:18:13 crc kubenswrapper[4859]: I1008 18:18:13.752887 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 18:18:13 crc kubenswrapper[4859]: I1008 18:18:13.752906 4859 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T18:18:13Z","lastTransitionTime":"2025-10-08T18:18:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 18:18:13 crc kubenswrapper[4859]: I1008 18:18:13.856436 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 18:18:13 crc kubenswrapper[4859]: I1008 18:18:13.856530 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 18:18:13 crc kubenswrapper[4859]: I1008 18:18:13.856565 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 18:18:13 crc kubenswrapper[4859]: I1008 18:18:13.856600 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 18:18:13 crc kubenswrapper[4859]: I1008 18:18:13.856627 4859 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T18:18:13Z","lastTransitionTime":"2025-10-08T18:18:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 18:18:13 crc kubenswrapper[4859]: I1008 18:18:13.959095 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 18:18:13 crc kubenswrapper[4859]: I1008 18:18:13.959485 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 18:18:13 crc kubenswrapper[4859]: I1008 18:18:13.959577 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 18:18:13 crc kubenswrapper[4859]: I1008 18:18:13.959696 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 18:18:13 crc kubenswrapper[4859]: I1008 18:18:13.959817 4859 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T18:18:13Z","lastTransitionTime":"2025-10-08T18:18:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 18:18:14 crc kubenswrapper[4859]: I1008 18:18:14.063256 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 18:18:14 crc kubenswrapper[4859]: I1008 18:18:14.063653 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 18:18:14 crc kubenswrapper[4859]: I1008 18:18:14.063748 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 18:18:14 crc kubenswrapper[4859]: I1008 18:18:14.063850 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 18:18:14 crc kubenswrapper[4859]: I1008 18:18:14.063891 4859 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T18:18:14Z","lastTransitionTime":"2025-10-08T18:18:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 18:18:14 crc kubenswrapper[4859]: I1008 18:18:14.167326 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 18:18:14 crc kubenswrapper[4859]: I1008 18:18:14.167378 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 18:18:14 crc kubenswrapper[4859]: I1008 18:18:14.167388 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 18:18:14 crc kubenswrapper[4859]: I1008 18:18:14.167407 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 18:18:14 crc kubenswrapper[4859]: I1008 18:18:14.167418 4859 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T18:18:14Z","lastTransitionTime":"2025-10-08T18:18:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 18:18:14 crc kubenswrapper[4859]: I1008 18:18:14.271195 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 18:18:14 crc kubenswrapper[4859]: I1008 18:18:14.271252 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 18:18:14 crc kubenswrapper[4859]: I1008 18:18:14.271266 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 18:18:14 crc kubenswrapper[4859]: I1008 18:18:14.271288 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 18:18:14 crc kubenswrapper[4859]: I1008 18:18:14.271304 4859 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T18:18:14Z","lastTransitionTime":"2025-10-08T18:18:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 18:18:14 crc kubenswrapper[4859]: I1008 18:18:14.374834 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 18:18:14 crc kubenswrapper[4859]: I1008 18:18:14.374893 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 18:18:14 crc kubenswrapper[4859]: I1008 18:18:14.374907 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 18:18:14 crc kubenswrapper[4859]: I1008 18:18:14.374928 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 18:18:14 crc kubenswrapper[4859]: I1008 18:18:14.374946 4859 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T18:18:14Z","lastTransitionTime":"2025-10-08T18:18:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 18:18:14 crc kubenswrapper[4859]: I1008 18:18:14.478143 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 18:18:14 crc kubenswrapper[4859]: I1008 18:18:14.478208 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 18:18:14 crc kubenswrapper[4859]: I1008 18:18:14.478229 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 18:18:14 crc kubenswrapper[4859]: I1008 18:18:14.478252 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 18:18:14 crc kubenswrapper[4859]: I1008 18:18:14.478267 4859 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T18:18:14Z","lastTransitionTime":"2025-10-08T18:18:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 18:18:14 crc kubenswrapper[4859]: I1008 18:18:14.582242 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 18:18:14 crc kubenswrapper[4859]: I1008 18:18:14.582281 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 18:18:14 crc kubenswrapper[4859]: I1008 18:18:14.582290 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 18:18:14 crc kubenswrapper[4859]: I1008 18:18:14.582308 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 18:18:14 crc kubenswrapper[4859]: I1008 18:18:14.582320 4859 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T18:18:14Z","lastTransitionTime":"2025-10-08T18:18:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 18:18:14 crc kubenswrapper[4859]: I1008 18:18:14.684703 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 18:18:14 crc kubenswrapper[4859]: I1008 18:18:14.684762 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 18:18:14 crc kubenswrapper[4859]: I1008 18:18:14.684772 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 18:18:14 crc kubenswrapper[4859]: I1008 18:18:14.684789 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 18:18:14 crc kubenswrapper[4859]: I1008 18:18:14.684799 4859 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T18:18:14Z","lastTransitionTime":"2025-10-08T18:18:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 18:18:14 crc kubenswrapper[4859]: I1008 18:18:14.787245 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 18:18:14 crc kubenswrapper[4859]: I1008 18:18:14.787622 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 18:18:14 crc kubenswrapper[4859]: I1008 18:18:14.787753 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 18:18:14 crc kubenswrapper[4859]: I1008 18:18:14.787859 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 18:18:14 crc kubenswrapper[4859]: I1008 18:18:14.787958 4859 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T18:18:14Z","lastTransitionTime":"2025-10-08T18:18:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 18:18:14 crc kubenswrapper[4859]: I1008 18:18:14.890533 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 18:18:14 crc kubenswrapper[4859]: I1008 18:18:14.890591 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 18:18:14 crc kubenswrapper[4859]: I1008 18:18:14.890606 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 18:18:14 crc kubenswrapper[4859]: I1008 18:18:14.890627 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 18:18:14 crc kubenswrapper[4859]: I1008 18:18:14.890639 4859 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T18:18:14Z","lastTransitionTime":"2025-10-08T18:18:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 18:18:14 crc kubenswrapper[4859]: I1008 18:18:14.993744 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 18:18:14 crc kubenswrapper[4859]: I1008 18:18:14.993805 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 18:18:14 crc kubenswrapper[4859]: I1008 18:18:14.993819 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 18:18:14 crc kubenswrapper[4859]: I1008 18:18:14.993841 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 18:18:14 crc kubenswrapper[4859]: I1008 18:18:14.993855 4859 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T18:18:14Z","lastTransitionTime":"2025-10-08T18:18:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 18:18:15 crc kubenswrapper[4859]: I1008 18:18:15.097217 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 18:18:15 crc kubenswrapper[4859]: I1008 18:18:15.097625 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 18:18:15 crc kubenswrapper[4859]: I1008 18:18:15.097841 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 18:18:15 crc kubenswrapper[4859]: I1008 18:18:15.097993 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 18:18:15 crc kubenswrapper[4859]: I1008 18:18:15.098149 4859 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T18:18:15Z","lastTransitionTime":"2025-10-08T18:18:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 18:18:15 crc kubenswrapper[4859]: I1008 18:18:15.200669 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 18:18:15 crc kubenswrapper[4859]: I1008 18:18:15.200762 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 18:18:15 crc kubenswrapper[4859]: I1008 18:18:15.200776 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 18:18:15 crc kubenswrapper[4859]: I1008 18:18:15.200801 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 18:18:15 crc kubenswrapper[4859]: I1008 18:18:15.200815 4859 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T18:18:15Z","lastTransitionTime":"2025-10-08T18:18:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 18:18:15 crc kubenswrapper[4859]: I1008 18:18:15.304121 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 18:18:15 crc kubenswrapper[4859]: I1008 18:18:15.304171 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 18:18:15 crc kubenswrapper[4859]: I1008 18:18:15.304183 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 18:18:15 crc kubenswrapper[4859]: I1008 18:18:15.304202 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 18:18:15 crc kubenswrapper[4859]: I1008 18:18:15.304215 4859 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T18:18:15Z","lastTransitionTime":"2025-10-08T18:18:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 18:18:15 crc kubenswrapper[4859]: I1008 18:18:15.407340 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 18:18:15 crc kubenswrapper[4859]: I1008 18:18:15.407414 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 18:18:15 crc kubenswrapper[4859]: I1008 18:18:15.407437 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 18:18:15 crc kubenswrapper[4859]: I1008 18:18:15.407466 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 18:18:15 crc kubenswrapper[4859]: I1008 18:18:15.407489 4859 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T18:18:15Z","lastTransitionTime":"2025-10-08T18:18:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 18:18:15 crc kubenswrapper[4859]: I1008 18:18:15.469810 4859 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 08 18:18:15 crc kubenswrapper[4859]: I1008 18:18:15.469810 4859 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-4b4sz" Oct 08 18:18:15 crc kubenswrapper[4859]: I1008 18:18:15.469846 4859 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 08 18:18:15 crc kubenswrapper[4859]: I1008 18:18:15.469873 4859 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 08 18:18:15 crc kubenswrapper[4859]: E1008 18:18:15.470132 4859 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 08 18:18:15 crc kubenswrapper[4859]: E1008 18:18:15.470317 4859 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 08 18:18:15 crc kubenswrapper[4859]: E1008 18:18:15.470425 4859 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-4b4sz" podUID="2e53cd9b-64f7-4e07-8b96-fb77847c9ac6" Oct 08 18:18:15 crc kubenswrapper[4859]: E1008 18:18:15.470767 4859 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 08 18:18:15 crc kubenswrapper[4859]: I1008 18:18:15.471194 4859 scope.go:117] "RemoveContainer" containerID="c646f0e0dd8e7f74125b687f0cb4114b1f34667e5a6325f394f0208d7fb89df6" Oct 08 18:18:15 crc kubenswrapper[4859]: E1008 18:18:15.471530 4859 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ovnkube-controller\" with CrashLoopBackOff: \"back-off 20s restarting failed container=ovnkube-controller pod=ovnkube-node-4g8gf_openshift-ovn-kubernetes(1dff864e-b75d-4e0b-b182-75f710eac8df)\"" pod="openshift-ovn-kubernetes/ovnkube-node-4g8gf" podUID="1dff864e-b75d-4e0b-b182-75f710eac8df" Oct 08 18:18:15 crc kubenswrapper[4859]: I1008 18:18:15.510989 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 18:18:15 crc kubenswrapper[4859]: I1008 18:18:15.511035 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 18:18:15 crc kubenswrapper[4859]: I1008 18:18:15.511046 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 18:18:15 crc kubenswrapper[4859]: I1008 18:18:15.511067 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 18:18:15 crc kubenswrapper[4859]: I1008 18:18:15.511081 4859 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T18:18:15Z","lastTransitionTime":"2025-10-08T18:18:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 18:18:15 crc kubenswrapper[4859]: I1008 18:18:15.614520 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 18:18:15 crc kubenswrapper[4859]: I1008 18:18:15.614598 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 18:18:15 crc kubenswrapper[4859]: I1008 18:18:15.614610 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 18:18:15 crc kubenswrapper[4859]: I1008 18:18:15.614632 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 18:18:15 crc kubenswrapper[4859]: I1008 18:18:15.614648 4859 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T18:18:15Z","lastTransitionTime":"2025-10-08T18:18:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 18:18:15 crc kubenswrapper[4859]: I1008 18:18:15.717768 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 18:18:15 crc kubenswrapper[4859]: I1008 18:18:15.717834 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 18:18:15 crc kubenswrapper[4859]: I1008 18:18:15.717844 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 18:18:15 crc kubenswrapper[4859]: I1008 18:18:15.717862 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 18:18:15 crc kubenswrapper[4859]: I1008 18:18:15.717909 4859 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T18:18:15Z","lastTransitionTime":"2025-10-08T18:18:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 18:18:15 crc kubenswrapper[4859]: I1008 18:18:15.822020 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 18:18:15 crc kubenswrapper[4859]: I1008 18:18:15.822080 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 18:18:15 crc kubenswrapper[4859]: I1008 18:18:15.822099 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 18:18:15 crc kubenswrapper[4859]: I1008 18:18:15.822136 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 18:18:15 crc kubenswrapper[4859]: I1008 18:18:15.822171 4859 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T18:18:15Z","lastTransitionTime":"2025-10-08T18:18:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 18:18:15 crc kubenswrapper[4859]: I1008 18:18:15.925709 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 18:18:15 crc kubenswrapper[4859]: I1008 18:18:15.925756 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 18:18:15 crc kubenswrapper[4859]: I1008 18:18:15.925767 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 18:18:15 crc kubenswrapper[4859]: I1008 18:18:15.925843 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 18:18:15 crc kubenswrapper[4859]: I1008 18:18:15.925859 4859 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T18:18:15Z","lastTransitionTime":"2025-10-08T18:18:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 18:18:16 crc kubenswrapper[4859]: I1008 18:18:16.028784 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 18:18:16 crc kubenswrapper[4859]: I1008 18:18:16.028839 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 18:18:16 crc kubenswrapper[4859]: I1008 18:18:16.028856 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 18:18:16 crc kubenswrapper[4859]: I1008 18:18:16.028916 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 18:18:16 crc kubenswrapper[4859]: I1008 18:18:16.028937 4859 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T18:18:16Z","lastTransitionTime":"2025-10-08T18:18:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 18:18:16 crc kubenswrapper[4859]: I1008 18:18:16.132174 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 18:18:16 crc kubenswrapper[4859]: I1008 18:18:16.132235 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 18:18:16 crc kubenswrapper[4859]: I1008 18:18:16.132249 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 18:18:16 crc kubenswrapper[4859]: I1008 18:18:16.132266 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 18:18:16 crc kubenswrapper[4859]: I1008 18:18:16.132279 4859 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T18:18:16Z","lastTransitionTime":"2025-10-08T18:18:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 18:18:16 crc kubenswrapper[4859]: I1008 18:18:16.235118 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 18:18:16 crc kubenswrapper[4859]: I1008 18:18:16.235165 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 18:18:16 crc kubenswrapper[4859]: I1008 18:18:16.235175 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 18:18:16 crc kubenswrapper[4859]: I1008 18:18:16.235331 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 18:18:16 crc kubenswrapper[4859]: I1008 18:18:16.235347 4859 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T18:18:16Z","lastTransitionTime":"2025-10-08T18:18:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 18:18:16 crc kubenswrapper[4859]: I1008 18:18:16.337883 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 18:18:16 crc kubenswrapper[4859]: I1008 18:18:16.337937 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 18:18:16 crc kubenswrapper[4859]: I1008 18:18:16.337949 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 18:18:16 crc kubenswrapper[4859]: I1008 18:18:16.337969 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 18:18:16 crc kubenswrapper[4859]: I1008 18:18:16.337982 4859 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T18:18:16Z","lastTransitionTime":"2025-10-08T18:18:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 18:18:16 crc kubenswrapper[4859]: I1008 18:18:16.440483 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 18:18:16 crc kubenswrapper[4859]: I1008 18:18:16.440541 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 18:18:16 crc kubenswrapper[4859]: I1008 18:18:16.440559 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 18:18:16 crc kubenswrapper[4859]: I1008 18:18:16.440581 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 18:18:16 crc kubenswrapper[4859]: I1008 18:18:16.440595 4859 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T18:18:16Z","lastTransitionTime":"2025-10-08T18:18:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 18:18:16 crc kubenswrapper[4859]: I1008 18:18:16.543878 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 18:18:16 crc kubenswrapper[4859]: I1008 18:18:16.543937 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 18:18:16 crc kubenswrapper[4859]: I1008 18:18:16.543952 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 18:18:16 crc kubenswrapper[4859]: I1008 18:18:16.543974 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 18:18:16 crc kubenswrapper[4859]: I1008 18:18:16.543987 4859 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T18:18:16Z","lastTransitionTime":"2025-10-08T18:18:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 18:18:16 crc kubenswrapper[4859]: I1008 18:18:16.646651 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 18:18:16 crc kubenswrapper[4859]: I1008 18:18:16.646730 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 18:18:16 crc kubenswrapper[4859]: I1008 18:18:16.646742 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 18:18:16 crc kubenswrapper[4859]: I1008 18:18:16.646764 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 18:18:16 crc kubenswrapper[4859]: I1008 18:18:16.646775 4859 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T18:18:16Z","lastTransitionTime":"2025-10-08T18:18:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 18:18:16 crc kubenswrapper[4859]: I1008 18:18:16.749881 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 18:18:16 crc kubenswrapper[4859]: I1008 18:18:16.749939 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 18:18:16 crc kubenswrapper[4859]: I1008 18:18:16.749949 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 18:18:16 crc kubenswrapper[4859]: I1008 18:18:16.749967 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 18:18:16 crc kubenswrapper[4859]: I1008 18:18:16.749979 4859 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T18:18:16Z","lastTransitionTime":"2025-10-08T18:18:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 18:18:16 crc kubenswrapper[4859]: I1008 18:18:16.852346 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 18:18:16 crc kubenswrapper[4859]: I1008 18:18:16.852495 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 18:18:16 crc kubenswrapper[4859]: I1008 18:18:16.852538 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 18:18:16 crc kubenswrapper[4859]: I1008 18:18:16.852560 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 18:18:16 crc kubenswrapper[4859]: I1008 18:18:16.852579 4859 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T18:18:16Z","lastTransitionTime":"2025-10-08T18:18:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 18:18:16 crc kubenswrapper[4859]: I1008 18:18:16.954979 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 18:18:16 crc kubenswrapper[4859]: I1008 18:18:16.955031 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 18:18:16 crc kubenswrapper[4859]: I1008 18:18:16.955045 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 18:18:16 crc kubenswrapper[4859]: I1008 18:18:16.955061 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 18:18:16 crc kubenswrapper[4859]: I1008 18:18:16.955072 4859 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T18:18:16Z","lastTransitionTime":"2025-10-08T18:18:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 18:18:17 crc kubenswrapper[4859]: I1008 18:18:17.058344 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 18:18:17 crc kubenswrapper[4859]: I1008 18:18:17.058388 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 18:18:17 crc kubenswrapper[4859]: I1008 18:18:17.058399 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 18:18:17 crc kubenswrapper[4859]: I1008 18:18:17.058416 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 18:18:17 crc kubenswrapper[4859]: I1008 18:18:17.058430 4859 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T18:18:17Z","lastTransitionTime":"2025-10-08T18:18:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 18:18:17 crc kubenswrapper[4859]: I1008 18:18:17.161637 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 18:18:17 crc kubenswrapper[4859]: I1008 18:18:17.161712 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 18:18:17 crc kubenswrapper[4859]: I1008 18:18:17.161727 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 18:18:17 crc kubenswrapper[4859]: I1008 18:18:17.161769 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 18:18:17 crc kubenswrapper[4859]: I1008 18:18:17.161783 4859 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T18:18:17Z","lastTransitionTime":"2025-10-08T18:18:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 18:18:17 crc kubenswrapper[4859]: I1008 18:18:17.265402 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 18:18:17 crc kubenswrapper[4859]: I1008 18:18:17.265465 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 18:18:17 crc kubenswrapper[4859]: I1008 18:18:17.265490 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 18:18:17 crc kubenswrapper[4859]: I1008 18:18:17.265526 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 18:18:17 crc kubenswrapper[4859]: I1008 18:18:17.265550 4859 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T18:18:17Z","lastTransitionTime":"2025-10-08T18:18:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 18:18:17 crc kubenswrapper[4859]: I1008 18:18:17.369733 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 18:18:17 crc kubenswrapper[4859]: I1008 18:18:17.369817 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 18:18:17 crc kubenswrapper[4859]: I1008 18:18:17.369836 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 18:18:17 crc kubenswrapper[4859]: I1008 18:18:17.369861 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 18:18:17 crc kubenswrapper[4859]: I1008 18:18:17.369909 4859 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T18:18:17Z","lastTransitionTime":"2025-10-08T18:18:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 18:18:17 crc kubenswrapper[4859]: I1008 18:18:17.469034 4859 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 08 18:18:17 crc kubenswrapper[4859]: I1008 18:18:17.469183 4859 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 08 18:18:17 crc kubenswrapper[4859]: I1008 18:18:17.469045 4859 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 08 18:18:17 crc kubenswrapper[4859]: E1008 18:18:17.469277 4859 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 08 18:18:17 crc kubenswrapper[4859]: I1008 18:18:17.469071 4859 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-4b4sz" Oct 08 18:18:17 crc kubenswrapper[4859]: E1008 18:18:17.469411 4859 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 08 18:18:17 crc kubenswrapper[4859]: E1008 18:18:17.469543 4859 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-4b4sz" podUID="2e53cd9b-64f7-4e07-8b96-fb77847c9ac6" Oct 08 18:18:17 crc kubenswrapper[4859]: E1008 18:18:17.469680 4859 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 08 18:18:17 crc kubenswrapper[4859]: I1008 18:18:17.473421 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 18:18:17 crc kubenswrapper[4859]: I1008 18:18:17.473463 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 18:18:17 crc kubenswrapper[4859]: I1008 18:18:17.473474 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 18:18:17 crc kubenswrapper[4859]: I1008 18:18:17.473488 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 18:18:17 crc kubenswrapper[4859]: I1008 18:18:17.473499 4859 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T18:18:17Z","lastTransitionTime":"2025-10-08T18:18:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 18:18:17 crc kubenswrapper[4859]: I1008 18:18:17.575775 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 18:18:17 crc kubenswrapper[4859]: I1008 18:18:17.575841 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 18:18:17 crc kubenswrapper[4859]: I1008 18:18:17.575852 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 18:18:17 crc kubenswrapper[4859]: I1008 18:18:17.575875 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 18:18:17 crc kubenswrapper[4859]: I1008 18:18:17.575887 4859 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T18:18:17Z","lastTransitionTime":"2025-10-08T18:18:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 18:18:17 crc kubenswrapper[4859]: I1008 18:18:17.678577 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 18:18:17 crc kubenswrapper[4859]: I1008 18:18:17.678619 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 18:18:17 crc kubenswrapper[4859]: I1008 18:18:17.678630 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 18:18:17 crc kubenswrapper[4859]: I1008 18:18:17.678646 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 18:18:17 crc kubenswrapper[4859]: I1008 18:18:17.678656 4859 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T18:18:17Z","lastTransitionTime":"2025-10-08T18:18:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 18:18:17 crc kubenswrapper[4859]: I1008 18:18:17.782099 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 18:18:17 crc kubenswrapper[4859]: I1008 18:18:17.782140 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 18:18:17 crc kubenswrapper[4859]: I1008 18:18:17.782149 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 18:18:17 crc kubenswrapper[4859]: I1008 18:18:17.782168 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 18:18:17 crc kubenswrapper[4859]: I1008 18:18:17.782177 4859 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T18:18:17Z","lastTransitionTime":"2025-10-08T18:18:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 18:18:17 crc kubenswrapper[4859]: I1008 18:18:17.885241 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 18:18:17 crc kubenswrapper[4859]: I1008 18:18:17.885405 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 18:18:17 crc kubenswrapper[4859]: I1008 18:18:17.885429 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 18:18:17 crc kubenswrapper[4859]: I1008 18:18:17.885472 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 18:18:17 crc kubenswrapper[4859]: I1008 18:18:17.885490 4859 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T18:18:17Z","lastTransitionTime":"2025-10-08T18:18:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 18:18:17 crc kubenswrapper[4859]: I1008 18:18:17.963262 4859 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/2e53cd9b-64f7-4e07-8b96-fb77847c9ac6-metrics-certs\") pod \"network-metrics-daemon-4b4sz\" (UID: \"2e53cd9b-64f7-4e07-8b96-fb77847c9ac6\") " pod="openshift-multus/network-metrics-daemon-4b4sz" Oct 08 18:18:17 crc kubenswrapper[4859]: E1008 18:18:17.963473 4859 secret.go:188] Couldn't get secret openshift-multus/metrics-daemon-secret: object "openshift-multus"/"metrics-daemon-secret" not registered Oct 08 18:18:17 crc kubenswrapper[4859]: E1008 18:18:17.963537 4859 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/2e53cd9b-64f7-4e07-8b96-fb77847c9ac6-metrics-certs podName:2e53cd9b-64f7-4e07-8b96-fb77847c9ac6 nodeName:}" failed. No retries permitted until 2025-10-08 18:18:49.963518803 +0000 UTC m=+100.210358182 (durationBeforeRetry 32s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/2e53cd9b-64f7-4e07-8b96-fb77847c9ac6-metrics-certs") pod "network-metrics-daemon-4b4sz" (UID: "2e53cd9b-64f7-4e07-8b96-fb77847c9ac6") : object "openshift-multus"/"metrics-daemon-secret" not registered Oct 08 18:18:17 crc kubenswrapper[4859]: I1008 18:18:17.988191 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 18:18:17 crc kubenswrapper[4859]: I1008 18:18:17.988252 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 18:18:17 crc kubenswrapper[4859]: I1008 18:18:17.988266 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 18:18:17 crc kubenswrapper[4859]: I1008 18:18:17.988285 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 18:18:17 crc kubenswrapper[4859]: I1008 18:18:17.988299 4859 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T18:18:17Z","lastTransitionTime":"2025-10-08T18:18:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 18:18:18 crc kubenswrapper[4859]: I1008 18:18:18.091396 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 18:18:18 crc kubenswrapper[4859]: I1008 18:18:18.091435 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 18:18:18 crc kubenswrapper[4859]: I1008 18:18:18.091445 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 18:18:18 crc kubenswrapper[4859]: I1008 18:18:18.091462 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 18:18:18 crc kubenswrapper[4859]: I1008 18:18:18.091472 4859 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T18:18:18Z","lastTransitionTime":"2025-10-08T18:18:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 18:18:18 crc kubenswrapper[4859]: I1008 18:18:18.116544 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 18:18:18 crc kubenswrapper[4859]: I1008 18:18:18.116593 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 18:18:18 crc kubenswrapper[4859]: I1008 18:18:18.116607 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 18:18:18 crc kubenswrapper[4859]: I1008 18:18:18.116627 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 18:18:18 crc kubenswrapper[4859]: I1008 18:18:18.116641 4859 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T18:18:18Z","lastTransitionTime":"2025-10-08T18:18:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 18:18:18 crc kubenswrapper[4859]: E1008 18:18:18.129892 4859 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404552Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865352Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-10-08T18:18:18Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-08T18:18:18Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-08T18:18:18Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-08T18:18:18Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-08T18:18:18Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-08T18:18:18Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-08T18:18:18Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-08T18:18:18Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"edb3ee6d-476b-407c-bb7d-4480786e8777\\\",\\\"systemUUID\\\":\\\"8f73c7ed-44aa-4d16-bae1-d4e684803cfc\\\"},\\\"runtimeHandlers\\\":[{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":false},\\\"name\\\":\\\"runc\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"crun\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"\\\"}]}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T18:18:18Z is after 2025-08-24T17:21:41Z" Oct 08 18:18:18 crc kubenswrapper[4859]: I1008 18:18:18.133911 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 18:18:18 crc kubenswrapper[4859]: I1008 18:18:18.133968 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 18:18:18 crc kubenswrapper[4859]: I1008 18:18:18.133984 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 18:18:18 crc kubenswrapper[4859]: I1008 18:18:18.134005 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 18:18:18 crc kubenswrapper[4859]: I1008 18:18:18.134016 4859 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T18:18:18Z","lastTransitionTime":"2025-10-08T18:18:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 18:18:18 crc kubenswrapper[4859]: E1008 18:18:18.154988 4859 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404552Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865352Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-10-08T18:18:18Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-08T18:18:18Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-08T18:18:18Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-08T18:18:18Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-08T18:18:18Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-08T18:18:18Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-08T18:18:18Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-08T18:18:18Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"edb3ee6d-476b-407c-bb7d-4480786e8777\\\",\\\"systemUUID\\\":\\\"8f73c7ed-44aa-4d16-bae1-d4e684803cfc\\\"},\\\"runtimeHandlers\\\":[{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":false},\\\"name\\\":\\\"runc\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"crun\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"\\\"}]}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T18:18:18Z is after 2025-08-24T17:21:41Z" Oct 08 18:18:18 crc kubenswrapper[4859]: I1008 18:18:18.159398 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 18:18:18 crc kubenswrapper[4859]: I1008 18:18:18.159431 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 18:18:18 crc kubenswrapper[4859]: I1008 18:18:18.159445 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 18:18:18 crc kubenswrapper[4859]: I1008 18:18:18.159469 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 18:18:18 crc kubenswrapper[4859]: I1008 18:18:18.159486 4859 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T18:18:18Z","lastTransitionTime":"2025-10-08T18:18:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 18:18:18 crc kubenswrapper[4859]: E1008 18:18:18.175752 4859 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404552Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865352Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-10-08T18:18:18Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-08T18:18:18Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-08T18:18:18Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-08T18:18:18Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-08T18:18:18Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-08T18:18:18Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-08T18:18:18Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-08T18:18:18Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"edb3ee6d-476b-407c-bb7d-4480786e8777\\\",\\\"systemUUID\\\":\\\"8f73c7ed-44aa-4d16-bae1-d4e684803cfc\\\"},\\\"runtimeHandlers\\\":[{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":false},\\\"name\\\":\\\"runc\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"crun\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"\\\"}]}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T18:18:18Z is after 2025-08-24T17:21:41Z" Oct 08 18:18:18 crc kubenswrapper[4859]: I1008 18:18:18.181358 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 18:18:18 crc kubenswrapper[4859]: I1008 18:18:18.181393 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 18:18:18 crc kubenswrapper[4859]: I1008 18:18:18.181405 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 18:18:18 crc kubenswrapper[4859]: I1008 18:18:18.181424 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 18:18:18 crc kubenswrapper[4859]: I1008 18:18:18.181439 4859 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T18:18:18Z","lastTransitionTime":"2025-10-08T18:18:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 18:18:18 crc kubenswrapper[4859]: E1008 18:18:18.196448 4859 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404552Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865352Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-10-08T18:18:18Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-08T18:18:18Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-08T18:18:18Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-08T18:18:18Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-08T18:18:18Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-08T18:18:18Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-08T18:18:18Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-08T18:18:18Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"edb3ee6d-476b-407c-bb7d-4480786e8777\\\",\\\"systemUUID\\\":\\\"8f73c7ed-44aa-4d16-bae1-d4e684803cfc\\\"},\\\"runtimeHandlers\\\":[{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":false},\\\"name\\\":\\\"runc\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"crun\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"\\\"}]}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T18:18:18Z is after 2025-08-24T17:21:41Z" Oct 08 18:18:18 crc kubenswrapper[4859]: I1008 18:18:18.200793 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 18:18:18 crc kubenswrapper[4859]: I1008 18:18:18.200840 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 18:18:18 crc kubenswrapper[4859]: I1008 18:18:18.200855 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 18:18:18 crc kubenswrapper[4859]: I1008 18:18:18.200875 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 18:18:18 crc kubenswrapper[4859]: I1008 18:18:18.200890 4859 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T18:18:18Z","lastTransitionTime":"2025-10-08T18:18:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 18:18:18 crc kubenswrapper[4859]: E1008 18:18:18.213162 4859 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404552Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865352Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-10-08T18:18:18Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-08T18:18:18Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-08T18:18:18Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-08T18:18:18Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-08T18:18:18Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-08T18:18:18Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-08T18:18:18Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-08T18:18:18Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"edb3ee6d-476b-407c-bb7d-4480786e8777\\\",\\\"systemUUID\\\":\\\"8f73c7ed-44aa-4d16-bae1-d4e684803cfc\\\"},\\\"runtimeHandlers\\\":[{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":false},\\\"name\\\":\\\"runc\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"crun\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"\\\"}]}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T18:18:18Z is after 2025-08-24T17:21:41Z" Oct 08 18:18:18 crc kubenswrapper[4859]: E1008 18:18:18.213343 4859 kubelet_node_status.go:572] "Unable to update node status" err="update node status exceeds retry count" Oct 08 18:18:18 crc kubenswrapper[4859]: I1008 18:18:18.215908 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 18:18:18 crc kubenswrapper[4859]: I1008 18:18:18.215965 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 18:18:18 crc kubenswrapper[4859]: I1008 18:18:18.215979 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 18:18:18 crc kubenswrapper[4859]: I1008 18:18:18.215998 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 18:18:18 crc kubenswrapper[4859]: I1008 18:18:18.216009 4859 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T18:18:18Z","lastTransitionTime":"2025-10-08T18:18:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 18:18:18 crc kubenswrapper[4859]: I1008 18:18:18.318512 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 18:18:18 crc kubenswrapper[4859]: I1008 18:18:18.318564 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 18:18:18 crc kubenswrapper[4859]: I1008 18:18:18.318582 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 18:18:18 crc kubenswrapper[4859]: I1008 18:18:18.318601 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 18:18:18 crc kubenswrapper[4859]: I1008 18:18:18.318615 4859 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T18:18:18Z","lastTransitionTime":"2025-10-08T18:18:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 18:18:18 crc kubenswrapper[4859]: I1008 18:18:18.421860 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 18:18:18 crc kubenswrapper[4859]: I1008 18:18:18.421912 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 18:18:18 crc kubenswrapper[4859]: I1008 18:18:18.421920 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 18:18:18 crc kubenswrapper[4859]: I1008 18:18:18.421938 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 18:18:18 crc kubenswrapper[4859]: I1008 18:18:18.421956 4859 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T18:18:18Z","lastTransitionTime":"2025-10-08T18:18:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 18:18:18 crc kubenswrapper[4859]: I1008 18:18:18.524558 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 18:18:18 crc kubenswrapper[4859]: I1008 18:18:18.524621 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 18:18:18 crc kubenswrapper[4859]: I1008 18:18:18.524634 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 18:18:18 crc kubenswrapper[4859]: I1008 18:18:18.524652 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 18:18:18 crc kubenswrapper[4859]: I1008 18:18:18.524666 4859 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T18:18:18Z","lastTransitionTime":"2025-10-08T18:18:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 18:18:18 crc kubenswrapper[4859]: I1008 18:18:18.627732 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 18:18:18 crc kubenswrapper[4859]: I1008 18:18:18.627782 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 18:18:18 crc kubenswrapper[4859]: I1008 18:18:18.627793 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 18:18:18 crc kubenswrapper[4859]: I1008 18:18:18.627811 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 18:18:18 crc kubenswrapper[4859]: I1008 18:18:18.627828 4859 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T18:18:18Z","lastTransitionTime":"2025-10-08T18:18:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 18:18:18 crc kubenswrapper[4859]: I1008 18:18:18.731450 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 18:18:18 crc kubenswrapper[4859]: I1008 18:18:18.731500 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 18:18:18 crc kubenswrapper[4859]: I1008 18:18:18.731520 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 18:18:18 crc kubenswrapper[4859]: I1008 18:18:18.731548 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 18:18:18 crc kubenswrapper[4859]: I1008 18:18:18.731566 4859 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T18:18:18Z","lastTransitionTime":"2025-10-08T18:18:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 18:18:18 crc kubenswrapper[4859]: I1008 18:18:18.834214 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 18:18:18 crc kubenswrapper[4859]: I1008 18:18:18.834253 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 18:18:18 crc kubenswrapper[4859]: I1008 18:18:18.834320 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 18:18:18 crc kubenswrapper[4859]: I1008 18:18:18.834484 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 18:18:18 crc kubenswrapper[4859]: I1008 18:18:18.834504 4859 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T18:18:18Z","lastTransitionTime":"2025-10-08T18:18:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 18:18:18 crc kubenswrapper[4859]: I1008 18:18:18.937017 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 18:18:18 crc kubenswrapper[4859]: I1008 18:18:18.937084 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 18:18:18 crc kubenswrapper[4859]: I1008 18:18:18.937099 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 18:18:18 crc kubenswrapper[4859]: I1008 18:18:18.937122 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 18:18:18 crc kubenswrapper[4859]: I1008 18:18:18.937134 4859 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T18:18:18Z","lastTransitionTime":"2025-10-08T18:18:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 18:18:19 crc kubenswrapper[4859]: I1008 18:18:19.040447 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 18:18:19 crc kubenswrapper[4859]: I1008 18:18:19.040502 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 18:18:19 crc kubenswrapper[4859]: I1008 18:18:19.040512 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 18:18:19 crc kubenswrapper[4859]: I1008 18:18:19.040533 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 18:18:19 crc kubenswrapper[4859]: I1008 18:18:19.040544 4859 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T18:18:19Z","lastTransitionTime":"2025-10-08T18:18:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 18:18:19 crc kubenswrapper[4859]: I1008 18:18:19.144648 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 18:18:19 crc kubenswrapper[4859]: I1008 18:18:19.144740 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 18:18:19 crc kubenswrapper[4859]: I1008 18:18:19.144756 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 18:18:19 crc kubenswrapper[4859]: I1008 18:18:19.144801 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 18:18:19 crc kubenswrapper[4859]: I1008 18:18:19.144814 4859 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T18:18:19Z","lastTransitionTime":"2025-10-08T18:18:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 18:18:19 crc kubenswrapper[4859]: I1008 18:18:19.247678 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 18:18:19 crc kubenswrapper[4859]: I1008 18:18:19.247766 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 18:18:19 crc kubenswrapper[4859]: I1008 18:18:19.247782 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 18:18:19 crc kubenswrapper[4859]: I1008 18:18:19.247804 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 18:18:19 crc kubenswrapper[4859]: I1008 18:18:19.247817 4859 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T18:18:19Z","lastTransitionTime":"2025-10-08T18:18:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 18:18:19 crc kubenswrapper[4859]: I1008 18:18:19.350344 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 18:18:19 crc kubenswrapper[4859]: I1008 18:18:19.350398 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 18:18:19 crc kubenswrapper[4859]: I1008 18:18:19.350411 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 18:18:19 crc kubenswrapper[4859]: I1008 18:18:19.350433 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 18:18:19 crc kubenswrapper[4859]: I1008 18:18:19.350451 4859 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T18:18:19Z","lastTransitionTime":"2025-10-08T18:18:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 18:18:19 crc kubenswrapper[4859]: I1008 18:18:19.453732 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 18:18:19 crc kubenswrapper[4859]: I1008 18:18:19.453786 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 18:18:19 crc kubenswrapper[4859]: I1008 18:18:19.453799 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 18:18:19 crc kubenswrapper[4859]: I1008 18:18:19.453819 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 18:18:19 crc kubenswrapper[4859]: I1008 18:18:19.453833 4859 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T18:18:19Z","lastTransitionTime":"2025-10-08T18:18:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 18:18:19 crc kubenswrapper[4859]: I1008 18:18:19.469238 4859 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 08 18:18:19 crc kubenswrapper[4859]: I1008 18:18:19.469261 4859 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 08 18:18:19 crc kubenswrapper[4859]: I1008 18:18:19.469265 4859 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 08 18:18:19 crc kubenswrapper[4859]: I1008 18:18:19.469355 4859 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-4b4sz" Oct 08 18:18:19 crc kubenswrapper[4859]: E1008 18:18:19.469360 4859 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 08 18:18:19 crc kubenswrapper[4859]: E1008 18:18:19.469428 4859 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 08 18:18:19 crc kubenswrapper[4859]: E1008 18:18:19.469522 4859 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-4b4sz" podUID="2e53cd9b-64f7-4e07-8b96-fb77847c9ac6" Oct 08 18:18:19 crc kubenswrapper[4859]: E1008 18:18:19.469602 4859 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 08 18:18:19 crc kubenswrapper[4859]: I1008 18:18:19.556900 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 18:18:19 crc kubenswrapper[4859]: I1008 18:18:19.556949 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 18:18:19 crc kubenswrapper[4859]: I1008 18:18:19.556962 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 18:18:19 crc kubenswrapper[4859]: I1008 18:18:19.556981 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 18:18:19 crc kubenswrapper[4859]: I1008 18:18:19.556994 4859 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T18:18:19Z","lastTransitionTime":"2025-10-08T18:18:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 18:18:19 crc kubenswrapper[4859]: I1008 18:18:19.660719 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 18:18:19 crc kubenswrapper[4859]: I1008 18:18:19.660773 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 18:18:19 crc kubenswrapper[4859]: I1008 18:18:19.660786 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 18:18:19 crc kubenswrapper[4859]: I1008 18:18:19.660808 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 18:18:19 crc kubenswrapper[4859]: I1008 18:18:19.660827 4859 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T18:18:19Z","lastTransitionTime":"2025-10-08T18:18:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 18:18:19 crc kubenswrapper[4859]: I1008 18:18:19.763539 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 18:18:19 crc kubenswrapper[4859]: I1008 18:18:19.763618 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 18:18:19 crc kubenswrapper[4859]: I1008 18:18:19.763638 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 18:18:19 crc kubenswrapper[4859]: I1008 18:18:19.763663 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 18:18:19 crc kubenswrapper[4859]: I1008 18:18:19.763681 4859 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T18:18:19Z","lastTransitionTime":"2025-10-08T18:18:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 18:18:19 crc kubenswrapper[4859]: I1008 18:18:19.866736 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 18:18:19 crc kubenswrapper[4859]: I1008 18:18:19.866786 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 18:18:19 crc kubenswrapper[4859]: I1008 18:18:19.866800 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 18:18:19 crc kubenswrapper[4859]: I1008 18:18:19.866824 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 18:18:19 crc kubenswrapper[4859]: I1008 18:18:19.866837 4859 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T18:18:19Z","lastTransitionTime":"2025-10-08T18:18:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 18:18:19 crc kubenswrapper[4859]: I1008 18:18:19.969828 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 18:18:19 crc kubenswrapper[4859]: I1008 18:18:19.969889 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 18:18:19 crc kubenswrapper[4859]: I1008 18:18:19.969907 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 18:18:19 crc kubenswrapper[4859]: I1008 18:18:19.969980 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 18:18:19 crc kubenswrapper[4859]: I1008 18:18:19.970004 4859 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T18:18:19Z","lastTransitionTime":"2025-10-08T18:18:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 18:18:19 crc kubenswrapper[4859]: I1008 18:18:19.972090 4859 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-8plkg_92b2d7de-31cc-4a91-95a5-ed5ea964b028/kube-multus/0.log" Oct 08 18:18:19 crc kubenswrapper[4859]: I1008 18:18:19.972167 4859 generic.go:334] "Generic (PLEG): container finished" podID="92b2d7de-31cc-4a91-95a5-ed5ea964b028" containerID="fa00b8345d40b74a917ddc23cc1b2c5c684a9ca61b6119d090c92cc1439b71c4" exitCode=1 Oct 08 18:18:19 crc kubenswrapper[4859]: I1008 18:18:19.972212 4859 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-8plkg" event={"ID":"92b2d7de-31cc-4a91-95a5-ed5ea964b028","Type":"ContainerDied","Data":"fa00b8345d40b74a917ddc23cc1b2c5c684a9ca61b6119d090c92cc1439b71c4"} Oct 08 18:18:19 crc kubenswrapper[4859]: I1008 18:18:19.972924 4859 scope.go:117] "RemoveContainer" containerID="fa00b8345d40b74a917ddc23cc1b2c5c684a9ca61b6119d090c92cc1439b71c4" Oct 08 18:18:19 crc kubenswrapper[4859]: I1008 18:18:19.988124 4859 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3ca6ce76-77e8-4e46-a2f6-b1ed4aa0fd4b\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:17:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:17:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:18:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:18:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:17:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f587e149d57ea53626ed3beb83e421c3b89ba1fec9e949705e8ccb5b4f7cc79a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T18:17:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://428cbe63daefbfbcb2ec76dad0fe638cdc65934bac80b791cebe0b36b2d6909a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T18:17:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a41399a37b1ad86a063001224232ecb8f7ea56aeb5686d3858bc08102a0c8f49\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T18:17:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://008757fce552591f7c58193f605eb60792761d5943927cbd63184549618d2b43\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://008757fce552591f7c58193f605eb60792761d5943927cbd63184549618d2b43\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-08T18:17:11Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-08T18:17:11Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T18:17:10Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T18:18:19Z is after 2025-08-24T17:21:41Z" Oct 08 18:18:20 crc kubenswrapper[4859]: I1008 18:18:20.008397 4859 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-08T18:17:29Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T18:17:29Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T18:18:20Z is after 2025-08-24T17:21:41Z" Oct 08 18:18:20 crc kubenswrapper[4859]: I1008 18:18:20.028487 4859 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-btw28" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"97bc5598-3ead-4300-a938-3aebbff1e8c6\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:17:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:17:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:17:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:17:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d6d4d6ba771cb5fde637449bba05660f634df7e833a0c8dad2b2c7f3a422bb2b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T18:17:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qcxsd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T18:17:31Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-btw28\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T18:18:20Z is after 2025-08-24T17:21:41Z" Oct 08 18:18:20 crc kubenswrapper[4859]: I1008 18:18:20.042882 4859 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-8plkg" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"92b2d7de-31cc-4a91-95a5-ed5ea964b028\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:17:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:17:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:18:19Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:18:19Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fa00b8345d40b74a917ddc23cc1b2c5c684a9ca61b6119d090c92cc1439b71c4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://fa00b8345d40b74a917ddc23cc1b2c5c684a9ca61b6119d090c92cc1439b71c4\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-10-08T18:18:19Z\\\",\\\"message\\\":\\\"2025-10-08T18:17:33+00:00 [cnibincopy] Successfully copied files in /usr/src/multus-cni/rhel9/bin/ to /host/opt/cni/bin/upgrade_888b130f-8754-48c6-b73e-9ff6772793f9\\\\n2025-10-08T18:17:33+00:00 [cnibincopy] Successfully moved files in /host/opt/cni/bin/upgrade_888b130f-8754-48c6-b73e-9ff6772793f9 to /host/opt/cni/bin/\\\\n2025-10-08T18:17:34Z [verbose] multus-daemon started\\\\n2025-10-08T18:17:34Z [verbose] Readiness Indicator file check\\\\n2025-10-08T18:18:19Z [error] have you checked that your default network is ready? still waiting for readinessindicatorfile @ /host/run/multus/cni/net.d/10-ovn-kubernetes.conf. pollimmediate error: timed out waiting for the condition\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-08T18:17:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q9d7m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T18:17:31Z\\\"}}\" for pod \"openshift-multus\"/\"multus-8plkg\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T18:18:20Z is after 2025-08-24T17:21:41Z" Oct 08 18:18:20 crc kubenswrapper[4859]: I1008 18:18:20.054032 4859 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-btjmq" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b8605f7-fed7-493f-9b4f-6a6358adf907\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:17:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:17:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:17:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:17:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://14e9b6c4a4403f36adb78fcf1135a620ca6fc56ed974c9639dcdfc59e6917559\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T18:17:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xspg4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T18:17:35Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-btjmq\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T18:18:20Z is after 2025-08-24T17:21:41Z" Oct 08 18:18:20 crc kubenswrapper[4859]: I1008 18:18:20.065034 4859 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-g4t79" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3a92dbf1-21a8-4095-96bb-2d701d7f3d56\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:17:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:17:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:17:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:17:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f8772a25b6e605efe9a062dc32dbbc2cfe32b4fe164b4e84fbf3bcffab744179\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T18:17:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sv79p\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://92cb9986c09a7f9e1e552e3e3ac1f74f04128241e00ddee38882d2ba5c005a67\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T18:17:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sv79p\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T18:17:45Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-g4t79\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T18:18:20Z is after 2025-08-24T17:21:41Z" Oct 08 18:18:20 crc kubenswrapper[4859]: I1008 18:18:20.073004 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 18:18:20 crc kubenswrapper[4859]: I1008 18:18:20.073036 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 18:18:20 crc kubenswrapper[4859]: I1008 18:18:20.073046 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 18:18:20 crc kubenswrapper[4859]: I1008 18:18:20.073064 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 18:18:20 crc kubenswrapper[4859]: I1008 18:18:20.073076 4859 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T18:18:20Z","lastTransitionTime":"2025-10-08T18:18:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 18:18:20 crc kubenswrapper[4859]: I1008 18:18:20.078623 4859 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-08T18:17:31Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T18:17:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T18:17:31Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c6d15905694b6bbd4999d49dc8f7a53eb6bfaa7adb59efb0d7e4973c21d1c7be\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T18:17:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T18:18:20Z is after 2025-08-24T17:21:41Z" Oct 08 18:18:20 crc kubenswrapper[4859]: I1008 18:18:20.094450 4859 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-08T18:17:33Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T18:17:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T18:17:33Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5d99e071bd083914418d9c128f38955f26295c08b1520f2dc774c99b1c843693\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T18:17:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T18:18:20Z is after 2025-08-24T17:21:41Z" Oct 08 18:18:20 crc kubenswrapper[4859]: I1008 18:18:20.105548 4859 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-4b4sz" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2e53cd9b-64f7-4e07-8b96-fb77847c9ac6\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:17:45Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:17:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:17:45Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:17:45Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kxgbz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kxgbz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T18:17:45Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-4b4sz\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T18:18:20Z is after 2025-08-24T17:21:41Z" Oct 08 18:18:20 crc kubenswrapper[4859]: I1008 18:18:20.119398 4859 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"63a623de-5608-4228-b1c2-77188c276986\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:17:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:17:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:17:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:17:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:17:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d4786a8fb8abd96fb22e60733c974f599017407d589ed815fb8541fd3b64f696\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T18:17:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://183d481502ed5108d5e19e483b714c289f355ea3f8b0891883e5335b802c83b6\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T18:17:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ec848c46fd5dde0469b00d3cab14ba4a652533a265490fadbee7a8b8843ef249\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T18:17:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4fe1828b4d63ff8106bfcd15e1dc1c1bd1ba7f77777cdb23ff1b291d0418774c\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2e700b2a20c81b414617f5b5c018a738278a3fa44e6c6d14b2170051d0725431\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-08T18:17:24Z\\\",\\\"message\\\":\\\"W1008 18:17:13.670625 1 cmd.go:257] Using insecure, self-signed certificates\\\\nI1008 18:17:13.671202 1 crypto.go:601] Generating new CA for check-endpoints-signer@1759947433 cert, and key in /tmp/serving-cert-3867218267/serving-signer.crt, /tmp/serving-cert-3867218267/serving-signer.key\\\\nI1008 18:17:13.944463 1 observer_polling.go:159] Starting file observer\\\\nW1008 18:17:13.951958 1 builder.go:272] unable to get owner reference (falling back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": dial tcp [::1]:6443: connect: connection refused\\\\nI1008 18:17:13.952088 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1008 18:17:13.956755 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3867218267/tls.crt::/tmp/serving-cert-3867218267/tls.key\\\\\\\"\\\\nF1008 18:17:24.290991 1 cmd.go:182] error initializing delegating authentication: unable to load configmap based request-header-client-ca-file: Get \\\\\\\"https://localhost:6443/api/v1/namespaces/kube-system/configmaps/extension-apiserver-authentication\\\\\\\": net/http: TLS handshake timeout\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-08T18:17:13Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T18:17:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c65b66f62d325ca10ef8d5d31f04529ccf4e691cf3303d231e7a414fb400e4d6\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T18:17:13Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://58c7bba4ace28215b062653a960e561b1dc1f890f2baf450145638a065af038b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://58c7bba4ace28215b062653a960e561b1dc1f890f2baf450145638a065af038b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-08T18:17:11Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-08T18:17:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T18:17:10Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T18:18:20Z is after 2025-08-24T17:21:41Z" Oct 08 18:18:20 crc kubenswrapper[4859]: I1008 18:18:20.136062 4859 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-08T18:17:31Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T18:17:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T18:17:31Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://99e2d51bd7757498dae60949eea6d732b5f7b481fa00f808dd5308a530a1daf1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T18:17:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e32aaba93699944665660ae41e9b5779178c1354c31e15ad8188bf8c3c8fa500\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T18:17:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T18:18:20Z is after 2025-08-24T17:21:41Z" Oct 08 18:18:20 crc kubenswrapper[4859]: I1008 18:18:20.150828 4859 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-08T18:17:29Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T18:17:29Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T18:18:20Z is after 2025-08-24T17:21:41Z" Oct 08 18:18:20 crc kubenswrapper[4859]: I1008 18:18:20.165180 4859 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-82s52" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b23a6a6c-9d92-4e7b-840e-55cfda873a2d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:17:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:17:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:17:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:17:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://71302c766485b9f55743d85f97440ff578cd44ba1e5a6f5b1c024c0e808072d2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T18:17:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6vx22\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c23ea3580998adf90621a0155c20ee4bd590c512629812174e95c28857bcfef9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T18:17:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6vx22\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T18:17:31Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-82s52\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T18:18:20Z is after 2025-08-24T17:21:41Z" Oct 08 18:18:20 crc kubenswrapper[4859]: I1008 18:18:20.175726 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 18:18:20 crc kubenswrapper[4859]: I1008 18:18:20.175772 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 18:18:20 crc kubenswrapper[4859]: I1008 18:18:20.175782 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 18:18:20 crc kubenswrapper[4859]: I1008 18:18:20.175801 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 18:18:20 crc kubenswrapper[4859]: I1008 18:18:20.175813 4859 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T18:18:20Z","lastTransitionTime":"2025-10-08T18:18:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 18:18:20 crc kubenswrapper[4859]: I1008 18:18:20.186308 4859 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-4g8gf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1dff864e-b75d-4e0b-b182-75f710eac8df\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:17:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:17:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:17:32Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:17:32Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://768ca274dedcc6847fc21d15234d3665fa9ccae371b5b6a882cfcb9b182f527d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T18:17:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s8jnn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://669f5f3c6d3c1ec0c4b4870e090b0a853adc01727c23383e595715b476c32668\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T18:17:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s8jnn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b8009a42cc4c9a816457a0678bdb2f7b7c53d0c8604e007a6fb141f66d06a516\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T18:17:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s8jnn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://676511a11ed4102084b57ebdd5f3fda4969741b2bc8443b415c4bf94f39ae4c9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T18:17:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s8jnn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4a74c1f0d21c90cdd7e776ee14e1f48757a04e2220fbd1f884449181ec6d1c3a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T18:17:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s8jnn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6ddc0e842bb8f81663406070023b0e5ed9cc99c98db0123b1c84f20da59fb5b5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T18:17:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s8jnn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c646f0e0dd8e7f74125b687f0cb4114b1f34667e5a6325f394f0208d7fb89df6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c646f0e0dd8e7f74125b687f0cb4114b1f34667e5a6325f394f0208d7fb89df6\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-10-08T18:17:58Z\\\",\\\"message\\\":\\\"e *v1.Pod retry setup to complete in iterateRetryResources\\\\nI1008 18:17:58.424504 6520 obj_retry.go:365] Adding new object: *v1.Pod openshift-machine-config-operator/machine-config-daemon-82s52\\\\nI1008 18:17:58.424529 6520 ovn.go:134] Ensuring zone local for Pod openshift-machine-config-operator/machine-config-daemon-82s52 in node crc\\\\nI1008 18:17:58.424539 6520 obj_retry.go:303] Retry object setup: *v1.Pod openshift-network-diagnostics/network-check-source-55646444c4-trplf\\\\nI1008 18:17:58.424549 6520 obj_retry.go:365] Adding new object: *v1.Pod openshift-network-diagnostics/network-check-source-55646444c4-trplf\\\\nI1008 18:17:58.424549 6520 obj_retry.go:386] Retry successful for *v1.Pod openshift-machine-config-operator/machine-config-daemon-82s52 after 0 failed attempt(s)\\\\nI1008 18:17:58.424559 6520 ovn.go:134] Ensuring zone local for Pod openshift-network-diagnostics/network-check-source-55646444c4-trplf in node crc\\\\nF1008 18:17:58.424552 6520 ovnkube.go:137] failed to run ovnkube: [failed to start network controller: failed to start default network controller: unable to create admin network policy controller, err: could not add Event Handler for anpInformer during admin network policy controller initialization, handler {0x1fcc6e0 0x1fcc3c0 0x1fcc360} was not added to shared informer because it has stopped already, failed to star\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-08T18:17:57Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 20s restarting failed container=ovnkube-controller pod=ovnkube-node-4g8gf_openshift-ovn-kubernetes(1dff864e-b75d-4e0b-b182-75f710eac8df)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s8jnn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2c67d2aed7126d76989e057d8e2d8e188a8902b2f3fe26ed6d1d45d3b013d897\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T18:17:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s8jnn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7a211d3da02b173867ff3a912eecd3b379483c7cc77e6b91122edd1ffd30fd53\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7a211d3da02b173867ff3a912eecd3b379483c7cc77e6b91122edd1ffd30fd53\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-08T18:17:32Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-08T18:17:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s8jnn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T18:17:32Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-4g8gf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T18:18:20Z is after 2025-08-24T17:21:41Z" Oct 08 18:18:20 crc kubenswrapper[4859]: I1008 18:18:20.200268 4859 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4efc16fb-e679-464d-8b3c-0f46f082593e\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:17:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:17:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:17:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:17:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:17:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://aa8f8ea895272cd79a08d1c47cd7863a1ec10e21bb60d12aa10862a56c99e4bf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T18:17:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5c4e95d8a712ab219455564539a90286ae48bbc46ddc322a5bfd96c1a8268110\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T18:17:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://bae6ecfd8b446ef04d393d823e5ffe37fe749337c0ca17fde79c5f88d5e31393\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T18:17:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6411d1e7e18d5ceadaaceed5a8a6b20a0cacc55cede793ca9fb4dcad70c64fc6\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T18:17:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T18:17:10Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T18:18:20Z is after 2025-08-24T17:21:41Z" Oct 08 18:18:20 crc kubenswrapper[4859]: I1008 18:18:20.215746 4859 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-08T18:17:29Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T18:17:29Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T18:18:20Z is after 2025-08-24T17:21:41Z" Oct 08 18:18:20 crc kubenswrapper[4859]: I1008 18:18:20.232336 4859 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-84k8x" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1c24078c-c9e8-4160-883a-f483b4b3dd23\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:17:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:17:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:17:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:17:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8661006834d0f50eaada7394369584ee95fc61ff54437eec35d755526afc5b48\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T18:17:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hk5xb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://18b291eb3344fb7c00ef5679dca9a6f7ed9d08cea684a803ff7dad2e8a18c585\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://18b291eb3344fb7c00ef5679dca9a6f7ed9d08cea684a803ff7dad2e8a18c585\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-08T18:17:32Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-08T18:17:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hk5xb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://160163f1a91d981583452db3c7a9244182d1c8ced45749a61b721d0d45c1dfe4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://160163f1a91d981583452db3c7a9244182d1c8ced45749a61b721d0d45c1dfe4\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-08T18:17:34Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-08T18:17:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hk5xb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://51e9b53f3847ac53f378ad83870debf8e71518d4edbc7fc0dc66644818e2103e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://51e9b53f3847ac53f378ad83870debf8e71518d4edbc7fc0dc66644818e2103e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-08T18:17:34Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-08T18:17:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hk5xb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1ff118919739759ac01b5cbe3240e6673db54eaa16353c609d670071ed7964cb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1ff118919739759ac01b5cbe3240e6673db54eaa16353c609d670071ed7964cb\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-08T18:17:35Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-08T18:17:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hk5xb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8522e803ff56eb57cd4d9fd036e0c3fb16240ea21d4a094c1db7aa7d50bfe2c2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8522e803ff56eb57cd4d9fd036e0c3fb16240ea21d4a094c1db7aa7d50bfe2c2\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-08T18:17:37Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-08T18:17:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hk5xb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://be61056edf515e037403ca2a0bfd5feb9b2537dd42212b58e133b23b294f327b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://be61056edf515e037403ca2a0bfd5feb9b2537dd42212b58e133b23b294f327b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-08T18:17:38Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-08T18:17:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hk5xb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T18:17:31Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-84k8x\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T18:18:20Z is after 2025-08-24T17:21:41Z" Oct 08 18:18:20 crc kubenswrapper[4859]: I1008 18:18:20.278429 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 18:18:20 crc kubenswrapper[4859]: I1008 18:18:20.278490 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 18:18:20 crc kubenswrapper[4859]: I1008 18:18:20.278506 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 18:18:20 crc kubenswrapper[4859]: I1008 18:18:20.278528 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 18:18:20 crc kubenswrapper[4859]: I1008 18:18:20.278544 4859 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T18:18:20Z","lastTransitionTime":"2025-10-08T18:18:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 18:18:20 crc kubenswrapper[4859]: I1008 18:18:20.381003 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 18:18:20 crc kubenswrapper[4859]: I1008 18:18:20.381038 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 18:18:20 crc kubenswrapper[4859]: I1008 18:18:20.381047 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 18:18:20 crc kubenswrapper[4859]: I1008 18:18:20.381065 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 18:18:20 crc kubenswrapper[4859]: I1008 18:18:20.381075 4859 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T18:18:20Z","lastTransitionTime":"2025-10-08T18:18:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 18:18:20 crc kubenswrapper[4859]: I1008 18:18:20.482947 4859 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-machine-config-operator/kube-rbac-proxy-crio-crc"] Oct 08 18:18:20 crc kubenswrapper[4859]: I1008 18:18:20.483739 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 18:18:20 crc kubenswrapper[4859]: I1008 18:18:20.483775 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 18:18:20 crc kubenswrapper[4859]: I1008 18:18:20.483785 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 18:18:20 crc kubenswrapper[4859]: I1008 18:18:20.483800 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 18:18:20 crc kubenswrapper[4859]: I1008 18:18:20.483812 4859 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T18:18:20Z","lastTransitionTime":"2025-10-08T18:18:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 18:18:20 crc kubenswrapper[4859]: I1008 18:18:20.485795 4859 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"63a623de-5608-4228-b1c2-77188c276986\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:17:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:17:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:17:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:17:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:17:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d4786a8fb8abd96fb22e60733c974f599017407d589ed815fb8541fd3b64f696\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T18:17:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://183d481502ed5108d5e19e483b714c289f355ea3f8b0891883e5335b802c83b6\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T18:17:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ec848c46fd5dde0469b00d3cab14ba4a652533a265490fadbee7a8b8843ef249\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T18:17:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4fe1828b4d63ff8106bfcd15e1dc1c1bd1ba7f77777cdb23ff1b291d0418774c\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2e700b2a20c81b414617f5b5c018a738278a3fa44e6c6d14b2170051d0725431\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-08T18:17:24Z\\\",\\\"message\\\":\\\"W1008 18:17:13.670625 1 cmd.go:257] Using insecure, self-signed certificates\\\\nI1008 18:17:13.671202 1 crypto.go:601] Generating new CA for check-endpoints-signer@1759947433 cert, and key in /tmp/serving-cert-3867218267/serving-signer.crt, /tmp/serving-cert-3867218267/serving-signer.key\\\\nI1008 18:17:13.944463 1 observer_polling.go:159] Starting file observer\\\\nW1008 18:17:13.951958 1 builder.go:272] unable to get owner reference (falling back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": dial tcp [::1]:6443: connect: connection refused\\\\nI1008 18:17:13.952088 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1008 18:17:13.956755 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3867218267/tls.crt::/tmp/serving-cert-3867218267/tls.key\\\\\\\"\\\\nF1008 18:17:24.290991 1 cmd.go:182] error initializing delegating authentication: unable to load configmap based request-header-client-ca-file: Get \\\\\\\"https://localhost:6443/api/v1/namespaces/kube-system/configmaps/extension-apiserver-authentication\\\\\\\": net/http: TLS handshake timeout\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-08T18:17:13Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T18:17:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c65b66f62d325ca10ef8d5d31f04529ccf4e691cf3303d231e7a414fb400e4d6\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T18:17:13Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://58c7bba4ace28215b062653a960e561b1dc1f890f2baf450145638a065af038b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://58c7bba4ace28215b062653a960e561b1dc1f890f2baf450145638a065af038b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-08T18:17:11Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-08T18:17:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T18:17:10Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T18:18:20Z is after 2025-08-24T17:21:41Z" Oct 08 18:18:20 crc kubenswrapper[4859]: I1008 18:18:20.500926 4859 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-08T18:17:31Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T18:17:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T18:17:31Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://99e2d51bd7757498dae60949eea6d732b5f7b481fa00f808dd5308a530a1daf1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T18:17:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e32aaba93699944665660ae41e9b5779178c1354c31e15ad8188bf8c3c8fa500\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T18:17:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T18:18:20Z is after 2025-08-24T17:21:41Z" Oct 08 18:18:20 crc kubenswrapper[4859]: I1008 18:18:20.513929 4859 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-08T18:17:29Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T18:17:29Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T18:18:20Z is after 2025-08-24T17:21:41Z" Oct 08 18:18:20 crc kubenswrapper[4859]: I1008 18:18:20.525673 4859 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-82s52" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b23a6a6c-9d92-4e7b-840e-55cfda873a2d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:17:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:17:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:17:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:17:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://71302c766485b9f55743d85f97440ff578cd44ba1e5a6f5b1c024c0e808072d2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T18:17:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6vx22\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c23ea3580998adf90621a0155c20ee4bd590c512629812174e95c28857bcfef9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T18:17:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6vx22\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T18:17:31Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-82s52\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T18:18:20Z is after 2025-08-24T17:21:41Z" Oct 08 18:18:20 crc kubenswrapper[4859]: I1008 18:18:20.547117 4859 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-4g8gf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1dff864e-b75d-4e0b-b182-75f710eac8df\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:17:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:17:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:17:32Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:17:32Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://768ca274dedcc6847fc21d15234d3665fa9ccae371b5b6a882cfcb9b182f527d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T18:17:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s8jnn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://669f5f3c6d3c1ec0c4b4870e090b0a853adc01727c23383e595715b476c32668\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T18:17:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s8jnn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b8009a42cc4c9a816457a0678bdb2f7b7c53d0c8604e007a6fb141f66d06a516\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T18:17:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s8jnn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://676511a11ed4102084b57ebdd5f3fda4969741b2bc8443b415c4bf94f39ae4c9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T18:17:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s8jnn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4a74c1f0d21c90cdd7e776ee14e1f48757a04e2220fbd1f884449181ec6d1c3a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T18:17:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s8jnn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6ddc0e842bb8f81663406070023b0e5ed9cc99c98db0123b1c84f20da59fb5b5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T18:17:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s8jnn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c646f0e0dd8e7f74125b687f0cb4114b1f34667e5a6325f394f0208d7fb89df6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c646f0e0dd8e7f74125b687f0cb4114b1f34667e5a6325f394f0208d7fb89df6\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-10-08T18:17:58Z\\\",\\\"message\\\":\\\"e *v1.Pod retry setup to complete in iterateRetryResources\\\\nI1008 18:17:58.424504 6520 obj_retry.go:365] Adding new object: *v1.Pod openshift-machine-config-operator/machine-config-daemon-82s52\\\\nI1008 18:17:58.424529 6520 ovn.go:134] Ensuring zone local for Pod openshift-machine-config-operator/machine-config-daemon-82s52 in node crc\\\\nI1008 18:17:58.424539 6520 obj_retry.go:303] Retry object setup: *v1.Pod openshift-network-diagnostics/network-check-source-55646444c4-trplf\\\\nI1008 18:17:58.424549 6520 obj_retry.go:365] Adding new object: *v1.Pod openshift-network-diagnostics/network-check-source-55646444c4-trplf\\\\nI1008 18:17:58.424549 6520 obj_retry.go:386] Retry successful for *v1.Pod openshift-machine-config-operator/machine-config-daemon-82s52 after 0 failed attempt(s)\\\\nI1008 18:17:58.424559 6520 ovn.go:134] Ensuring zone local for Pod openshift-network-diagnostics/network-check-source-55646444c4-trplf in node crc\\\\nF1008 18:17:58.424552 6520 ovnkube.go:137] failed to run ovnkube: [failed to start network controller: failed to start default network controller: unable to create admin network policy controller, err: could not add Event Handler for anpInformer during admin network policy controller initialization, handler {0x1fcc6e0 0x1fcc3c0 0x1fcc360} was not added to shared informer because it has stopped already, failed to star\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-08T18:17:57Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 20s restarting failed container=ovnkube-controller pod=ovnkube-node-4g8gf_openshift-ovn-kubernetes(1dff864e-b75d-4e0b-b182-75f710eac8df)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s8jnn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2c67d2aed7126d76989e057d8e2d8e188a8902b2f3fe26ed6d1d45d3b013d897\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T18:17:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s8jnn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7a211d3da02b173867ff3a912eecd3b379483c7cc77e6b91122edd1ffd30fd53\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7a211d3da02b173867ff3a912eecd3b379483c7cc77e6b91122edd1ffd30fd53\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-08T18:17:32Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-08T18:17:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s8jnn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T18:17:32Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-4g8gf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T18:18:20Z is after 2025-08-24T17:21:41Z" Oct 08 18:18:20 crc kubenswrapper[4859]: I1008 18:18:20.561330 4859 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-4b4sz" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2e53cd9b-64f7-4e07-8b96-fb77847c9ac6\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:17:45Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:17:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:17:45Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:17:45Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kxgbz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kxgbz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T18:17:45Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-4b4sz\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T18:18:20Z is after 2025-08-24T17:21:41Z" Oct 08 18:18:20 crc kubenswrapper[4859]: I1008 18:18:20.577859 4859 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4efc16fb-e679-464d-8b3c-0f46f082593e\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:17:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:17:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:17:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:17:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:17:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://aa8f8ea895272cd79a08d1c47cd7863a1ec10e21bb60d12aa10862a56c99e4bf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T18:17:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5c4e95d8a712ab219455564539a90286ae48bbc46ddc322a5bfd96c1a8268110\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T18:17:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://bae6ecfd8b446ef04d393d823e5ffe37fe749337c0ca17fde79c5f88d5e31393\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T18:17:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6411d1e7e18d5ceadaaceed5a8a6b20a0cacc55cede793ca9fb4dcad70c64fc6\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T18:17:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T18:17:10Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T18:18:20Z is after 2025-08-24T17:21:41Z" Oct 08 18:18:20 crc kubenswrapper[4859]: I1008 18:18:20.586933 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 18:18:20 crc kubenswrapper[4859]: I1008 18:18:20.586976 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 18:18:20 crc kubenswrapper[4859]: I1008 18:18:20.586990 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 18:18:20 crc kubenswrapper[4859]: I1008 18:18:20.587009 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 18:18:20 crc kubenswrapper[4859]: I1008 18:18:20.587023 4859 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T18:18:20Z","lastTransitionTime":"2025-10-08T18:18:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 18:18:20 crc kubenswrapper[4859]: I1008 18:18:20.592194 4859 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-08T18:17:29Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T18:17:29Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T18:18:20Z is after 2025-08-24T17:21:41Z" Oct 08 18:18:20 crc kubenswrapper[4859]: I1008 18:18:20.606755 4859 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-84k8x" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1c24078c-c9e8-4160-883a-f483b4b3dd23\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:17:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:17:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:17:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:17:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8661006834d0f50eaada7394369584ee95fc61ff54437eec35d755526afc5b48\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T18:17:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hk5xb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://18b291eb3344fb7c00ef5679dca9a6f7ed9d08cea684a803ff7dad2e8a18c585\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://18b291eb3344fb7c00ef5679dca9a6f7ed9d08cea684a803ff7dad2e8a18c585\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-08T18:17:32Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-08T18:17:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hk5xb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://160163f1a91d981583452db3c7a9244182d1c8ced45749a61b721d0d45c1dfe4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://160163f1a91d981583452db3c7a9244182d1c8ced45749a61b721d0d45c1dfe4\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-08T18:17:34Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-08T18:17:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hk5xb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://51e9b53f3847ac53f378ad83870debf8e71518d4edbc7fc0dc66644818e2103e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://51e9b53f3847ac53f378ad83870debf8e71518d4edbc7fc0dc66644818e2103e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-08T18:17:34Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-08T18:17:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hk5xb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1ff118919739759ac01b5cbe3240e6673db54eaa16353c609d670071ed7964cb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1ff118919739759ac01b5cbe3240e6673db54eaa16353c609d670071ed7964cb\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-08T18:17:35Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-08T18:17:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hk5xb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8522e803ff56eb57cd4d9fd036e0c3fb16240ea21d4a094c1db7aa7d50bfe2c2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8522e803ff56eb57cd4d9fd036e0c3fb16240ea21d4a094c1db7aa7d50bfe2c2\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-08T18:17:37Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-08T18:17:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hk5xb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://be61056edf515e037403ca2a0bfd5feb9b2537dd42212b58e133b23b294f327b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://be61056edf515e037403ca2a0bfd5feb9b2537dd42212b58e133b23b294f327b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-08T18:17:38Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-08T18:17:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hk5xb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T18:17:31Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-84k8x\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T18:18:20Z is after 2025-08-24T17:21:41Z" Oct 08 18:18:20 crc kubenswrapper[4859]: I1008 18:18:20.620240 4859 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3ca6ce76-77e8-4e46-a2f6-b1ed4aa0fd4b\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:17:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:17:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:18:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:18:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:17:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f587e149d57ea53626ed3beb83e421c3b89ba1fec9e949705e8ccb5b4f7cc79a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T18:17:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://428cbe63daefbfbcb2ec76dad0fe638cdc65934bac80b791cebe0b36b2d6909a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T18:17:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a41399a37b1ad86a063001224232ecb8f7ea56aeb5686d3858bc08102a0c8f49\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T18:17:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://008757fce552591f7c58193f605eb60792761d5943927cbd63184549618d2b43\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://008757fce552591f7c58193f605eb60792761d5943927cbd63184549618d2b43\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-08T18:17:11Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-08T18:17:11Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T18:17:10Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T18:18:20Z is after 2025-08-24T17:21:41Z" Oct 08 18:18:20 crc kubenswrapper[4859]: I1008 18:18:20.632151 4859 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-08T18:17:29Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T18:17:29Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T18:18:20Z is after 2025-08-24T17:21:41Z" Oct 08 18:18:20 crc kubenswrapper[4859]: I1008 18:18:20.642943 4859 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-btw28" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"97bc5598-3ead-4300-a938-3aebbff1e8c6\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:17:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:17:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:17:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:17:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d6d4d6ba771cb5fde637449bba05660f634df7e833a0c8dad2b2c7f3a422bb2b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T18:17:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qcxsd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T18:17:31Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-btw28\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T18:18:20Z is after 2025-08-24T17:21:41Z" Oct 08 18:18:20 crc kubenswrapper[4859]: I1008 18:18:20.655231 4859 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-8plkg" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"92b2d7de-31cc-4a91-95a5-ed5ea964b028\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:17:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:17:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:18:19Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:18:19Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fa00b8345d40b74a917ddc23cc1b2c5c684a9ca61b6119d090c92cc1439b71c4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://fa00b8345d40b74a917ddc23cc1b2c5c684a9ca61b6119d090c92cc1439b71c4\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-10-08T18:18:19Z\\\",\\\"message\\\":\\\"2025-10-08T18:17:33+00:00 [cnibincopy] Successfully copied files in /usr/src/multus-cni/rhel9/bin/ to /host/opt/cni/bin/upgrade_888b130f-8754-48c6-b73e-9ff6772793f9\\\\n2025-10-08T18:17:33+00:00 [cnibincopy] Successfully moved files in /host/opt/cni/bin/upgrade_888b130f-8754-48c6-b73e-9ff6772793f9 to /host/opt/cni/bin/\\\\n2025-10-08T18:17:34Z [verbose] multus-daemon started\\\\n2025-10-08T18:17:34Z [verbose] Readiness Indicator file check\\\\n2025-10-08T18:18:19Z [error] have you checked that your default network is ready? still waiting for readinessindicatorfile @ /host/run/multus/cni/net.d/10-ovn-kubernetes.conf. pollimmediate error: timed out waiting for the condition\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-08T18:17:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q9d7m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T18:17:31Z\\\"}}\" for pod \"openshift-multus\"/\"multus-8plkg\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T18:18:20Z is after 2025-08-24T17:21:41Z" Oct 08 18:18:20 crc kubenswrapper[4859]: I1008 18:18:20.671594 4859 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-btjmq" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b8605f7-fed7-493f-9b4f-6a6358adf907\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:17:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:17:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:17:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:17:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://14e9b6c4a4403f36adb78fcf1135a620ca6fc56ed974c9639dcdfc59e6917559\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T18:17:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xspg4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T18:17:35Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-btjmq\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T18:18:20Z is after 2025-08-24T17:21:41Z" Oct 08 18:18:20 crc kubenswrapper[4859]: I1008 18:18:20.685339 4859 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-g4t79" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3a92dbf1-21a8-4095-96bb-2d701d7f3d56\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:17:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:17:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:17:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:17:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f8772a25b6e605efe9a062dc32dbbc2cfe32b4fe164b4e84fbf3bcffab744179\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T18:17:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sv79p\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://92cb9986c09a7f9e1e552e3e3ac1f74f04128241e00ddee38882d2ba5c005a67\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T18:17:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sv79p\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T18:17:45Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-g4t79\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T18:18:20Z is after 2025-08-24T17:21:41Z" Oct 08 18:18:20 crc kubenswrapper[4859]: I1008 18:18:20.689595 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 18:18:20 crc kubenswrapper[4859]: I1008 18:18:20.689673 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 18:18:20 crc kubenswrapper[4859]: I1008 18:18:20.689708 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 18:18:20 crc kubenswrapper[4859]: I1008 18:18:20.689729 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 18:18:20 crc kubenswrapper[4859]: I1008 18:18:20.689744 4859 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T18:18:20Z","lastTransitionTime":"2025-10-08T18:18:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 18:18:20 crc kubenswrapper[4859]: I1008 18:18:20.701302 4859 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-08T18:17:31Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T18:17:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T18:17:31Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c6d15905694b6bbd4999d49dc8f7a53eb6bfaa7adb59efb0d7e4973c21d1c7be\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T18:17:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T18:18:20Z is after 2025-08-24T17:21:41Z" Oct 08 18:18:20 crc kubenswrapper[4859]: I1008 18:18:20.713404 4859 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-08T18:17:33Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T18:17:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T18:17:33Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5d99e071bd083914418d9c128f38955f26295c08b1520f2dc774c99b1c843693\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T18:17:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T18:18:20Z is after 2025-08-24T17:21:41Z" Oct 08 18:18:20 crc kubenswrapper[4859]: I1008 18:18:20.792966 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 18:18:20 crc kubenswrapper[4859]: I1008 18:18:20.793029 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 18:18:20 crc kubenswrapper[4859]: I1008 18:18:20.793042 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 18:18:20 crc kubenswrapper[4859]: I1008 18:18:20.793062 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 18:18:20 crc kubenswrapper[4859]: I1008 18:18:20.793075 4859 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T18:18:20Z","lastTransitionTime":"2025-10-08T18:18:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 18:18:20 crc kubenswrapper[4859]: I1008 18:18:20.895482 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 18:18:20 crc kubenswrapper[4859]: I1008 18:18:20.895519 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 18:18:20 crc kubenswrapper[4859]: I1008 18:18:20.895527 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 18:18:20 crc kubenswrapper[4859]: I1008 18:18:20.895542 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 18:18:20 crc kubenswrapper[4859]: I1008 18:18:20.895551 4859 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T18:18:20Z","lastTransitionTime":"2025-10-08T18:18:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 18:18:20 crc kubenswrapper[4859]: I1008 18:18:20.976874 4859 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-8plkg_92b2d7de-31cc-4a91-95a5-ed5ea964b028/kube-multus/0.log" Oct 08 18:18:20 crc kubenswrapper[4859]: I1008 18:18:20.976962 4859 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-8plkg" event={"ID":"92b2d7de-31cc-4a91-95a5-ed5ea964b028","Type":"ContainerStarted","Data":"729bd32fd01dcab495bffa64f2f5e44990fc1d78b05e7f131868009f92305831"} Oct 08 18:18:20 crc kubenswrapper[4859]: I1008 18:18:20.995658 4859 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-08T18:17:31Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T18:17:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T18:17:31Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c6d15905694b6bbd4999d49dc8f7a53eb6bfaa7adb59efb0d7e4973c21d1c7be\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T18:17:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T18:18:20Z is after 2025-08-24T17:21:41Z" Oct 08 18:18:20 crc kubenswrapper[4859]: I1008 18:18:20.997750 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 18:18:20 crc kubenswrapper[4859]: I1008 18:18:20.997782 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 18:18:20 crc kubenswrapper[4859]: I1008 18:18:20.997794 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 18:18:20 crc kubenswrapper[4859]: I1008 18:18:20.997811 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 18:18:20 crc kubenswrapper[4859]: I1008 18:18:20.997823 4859 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T18:18:20Z","lastTransitionTime":"2025-10-08T18:18:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 18:18:21 crc kubenswrapper[4859]: I1008 18:18:21.013050 4859 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-08T18:17:33Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T18:17:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T18:17:33Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5d99e071bd083914418d9c128f38955f26295c08b1520f2dc774c99b1c843693\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T18:17:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T18:18:21Z is after 2025-08-24T17:21:41Z" Oct 08 18:18:21 crc kubenswrapper[4859]: I1008 18:18:21.033481 4859 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-08T18:17:29Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T18:17:29Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T18:18:21Z is after 2025-08-24T17:21:41Z" Oct 08 18:18:21 crc kubenswrapper[4859]: I1008 18:18:21.045770 4859 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-82s52" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b23a6a6c-9d92-4e7b-840e-55cfda873a2d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:17:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:17:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:17:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:17:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://71302c766485b9f55743d85f97440ff578cd44ba1e5a6f5b1c024c0e808072d2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T18:17:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6vx22\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c23ea3580998adf90621a0155c20ee4bd590c512629812174e95c28857bcfef9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T18:17:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6vx22\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T18:17:31Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-82s52\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T18:18:21Z is after 2025-08-24T17:21:41Z" Oct 08 18:18:21 crc kubenswrapper[4859]: I1008 18:18:21.067549 4859 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-4g8gf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1dff864e-b75d-4e0b-b182-75f710eac8df\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:17:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:17:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:17:32Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:17:32Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://768ca274dedcc6847fc21d15234d3665fa9ccae371b5b6a882cfcb9b182f527d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T18:17:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s8jnn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://669f5f3c6d3c1ec0c4b4870e090b0a853adc01727c23383e595715b476c32668\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T18:17:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s8jnn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b8009a42cc4c9a816457a0678bdb2f7b7c53d0c8604e007a6fb141f66d06a516\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T18:17:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s8jnn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://676511a11ed4102084b57ebdd5f3fda4969741b2bc8443b415c4bf94f39ae4c9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T18:17:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s8jnn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4a74c1f0d21c90cdd7e776ee14e1f48757a04e2220fbd1f884449181ec6d1c3a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T18:17:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s8jnn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6ddc0e842bb8f81663406070023b0e5ed9cc99c98db0123b1c84f20da59fb5b5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T18:17:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s8jnn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c646f0e0dd8e7f74125b687f0cb4114b1f34667e5a6325f394f0208d7fb89df6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c646f0e0dd8e7f74125b687f0cb4114b1f34667e5a6325f394f0208d7fb89df6\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-10-08T18:17:58Z\\\",\\\"message\\\":\\\"e *v1.Pod retry setup to complete in iterateRetryResources\\\\nI1008 18:17:58.424504 6520 obj_retry.go:365] Adding new object: *v1.Pod openshift-machine-config-operator/machine-config-daemon-82s52\\\\nI1008 18:17:58.424529 6520 ovn.go:134] Ensuring zone local for Pod openshift-machine-config-operator/machine-config-daemon-82s52 in node crc\\\\nI1008 18:17:58.424539 6520 obj_retry.go:303] Retry object setup: *v1.Pod openshift-network-diagnostics/network-check-source-55646444c4-trplf\\\\nI1008 18:17:58.424549 6520 obj_retry.go:365] Adding new object: *v1.Pod openshift-network-diagnostics/network-check-source-55646444c4-trplf\\\\nI1008 18:17:58.424549 6520 obj_retry.go:386] Retry successful for *v1.Pod openshift-machine-config-operator/machine-config-daemon-82s52 after 0 failed attempt(s)\\\\nI1008 18:17:58.424559 6520 ovn.go:134] Ensuring zone local for Pod openshift-network-diagnostics/network-check-source-55646444c4-trplf in node crc\\\\nF1008 18:17:58.424552 6520 ovnkube.go:137] failed to run ovnkube: [failed to start network controller: failed to start default network controller: unable to create admin network policy controller, err: could not add Event Handler for anpInformer during admin network policy controller initialization, handler {0x1fcc6e0 0x1fcc3c0 0x1fcc360} was not added to shared informer because it has stopped already, failed to star\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-08T18:17:57Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 20s restarting failed container=ovnkube-controller pod=ovnkube-node-4g8gf_openshift-ovn-kubernetes(1dff864e-b75d-4e0b-b182-75f710eac8df)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s8jnn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2c67d2aed7126d76989e057d8e2d8e188a8902b2f3fe26ed6d1d45d3b013d897\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T18:17:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s8jnn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7a211d3da02b173867ff3a912eecd3b379483c7cc77e6b91122edd1ffd30fd53\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7a211d3da02b173867ff3a912eecd3b379483c7cc77e6b91122edd1ffd30fd53\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-08T18:17:32Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-08T18:17:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s8jnn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T18:17:32Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-4g8gf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T18:18:21Z is after 2025-08-24T17:21:41Z" Oct 08 18:18:21 crc kubenswrapper[4859]: I1008 18:18:21.080397 4859 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-4b4sz" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2e53cd9b-64f7-4e07-8b96-fb77847c9ac6\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:17:45Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:17:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:17:45Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:17:45Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kxgbz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kxgbz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T18:17:45Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-4b4sz\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T18:18:21Z is after 2025-08-24T17:21:41Z" Oct 08 18:18:21 crc kubenswrapper[4859]: I1008 18:18:21.091450 4859 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"262c958c-c49d-4c11-96ae-94106dd123d4\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:17:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:17:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:17:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:17:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:17:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a54a54b9a831ead1fbe6e03abbb622242dd16b4081a4482d559e03d47aebd2fe\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-crio\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T18:17:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kube\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3ab3bc5e8e2fd4941f30cbf8a90d9a9078b90fab91063c07da2cd48b63eef657\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3ab3bc5e8e2fd4941f30cbf8a90d9a9078b90fab91063c07da2cd48b63eef657\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-08T18:17:11Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-08T18:17:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T18:17:10Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"kube-rbac-proxy-crio-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T18:18:21Z is after 2025-08-24T17:21:41Z" Oct 08 18:18:21 crc kubenswrapper[4859]: I1008 18:18:21.101002 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 18:18:21 crc kubenswrapper[4859]: I1008 18:18:21.101472 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 18:18:21 crc kubenswrapper[4859]: I1008 18:18:21.101572 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 18:18:21 crc kubenswrapper[4859]: I1008 18:18:21.101744 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 18:18:21 crc kubenswrapper[4859]: I1008 18:18:21.101849 4859 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T18:18:21Z","lastTransitionTime":"2025-10-08T18:18:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 18:18:21 crc kubenswrapper[4859]: I1008 18:18:21.108406 4859 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"63a623de-5608-4228-b1c2-77188c276986\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:17:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:17:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:17:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:17:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:17:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d4786a8fb8abd96fb22e60733c974f599017407d589ed815fb8541fd3b64f696\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T18:17:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://183d481502ed5108d5e19e483b714c289f355ea3f8b0891883e5335b802c83b6\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T18:17:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ec848c46fd5dde0469b00d3cab14ba4a652533a265490fadbee7a8b8843ef249\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T18:17:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4fe1828b4d63ff8106bfcd15e1dc1c1bd1ba7f77777cdb23ff1b291d0418774c\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2e700b2a20c81b414617f5b5c018a738278a3fa44e6c6d14b2170051d0725431\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-08T18:17:24Z\\\",\\\"message\\\":\\\"W1008 18:17:13.670625 1 cmd.go:257] Using insecure, self-signed certificates\\\\nI1008 18:17:13.671202 1 crypto.go:601] Generating new CA for check-endpoints-signer@1759947433 cert, and key in /tmp/serving-cert-3867218267/serving-signer.crt, /tmp/serving-cert-3867218267/serving-signer.key\\\\nI1008 18:17:13.944463 1 observer_polling.go:159] Starting file observer\\\\nW1008 18:17:13.951958 1 builder.go:272] unable to get owner reference (falling back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": dial tcp [::1]:6443: connect: connection refused\\\\nI1008 18:17:13.952088 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1008 18:17:13.956755 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3867218267/tls.crt::/tmp/serving-cert-3867218267/tls.key\\\\\\\"\\\\nF1008 18:17:24.290991 1 cmd.go:182] error initializing delegating authentication: unable to load configmap based request-header-client-ca-file: Get \\\\\\\"https://localhost:6443/api/v1/namespaces/kube-system/configmaps/extension-apiserver-authentication\\\\\\\": net/http: TLS handshake timeout\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-08T18:17:13Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T18:17:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c65b66f62d325ca10ef8d5d31f04529ccf4e691cf3303d231e7a414fb400e4d6\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T18:17:13Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://58c7bba4ace28215b062653a960e561b1dc1f890f2baf450145638a065af038b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://58c7bba4ace28215b062653a960e561b1dc1f890f2baf450145638a065af038b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-08T18:17:11Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-08T18:17:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T18:17:10Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T18:18:21Z is after 2025-08-24T17:21:41Z" Oct 08 18:18:21 crc kubenswrapper[4859]: I1008 18:18:21.124608 4859 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-08T18:17:31Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T18:17:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T18:17:31Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://99e2d51bd7757498dae60949eea6d732b5f7b481fa00f808dd5308a530a1daf1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T18:17:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e32aaba93699944665660ae41e9b5779178c1354c31e15ad8188bf8c3c8fa500\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T18:17:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T18:18:21Z is after 2025-08-24T17:21:41Z" Oct 08 18:18:21 crc kubenswrapper[4859]: I1008 18:18:21.139558 4859 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4efc16fb-e679-464d-8b3c-0f46f082593e\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:17:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:17:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:17:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:17:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:17:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://aa8f8ea895272cd79a08d1c47cd7863a1ec10e21bb60d12aa10862a56c99e4bf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T18:17:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5c4e95d8a712ab219455564539a90286ae48bbc46ddc322a5bfd96c1a8268110\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T18:17:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://bae6ecfd8b446ef04d393d823e5ffe37fe749337c0ca17fde79c5f88d5e31393\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T18:17:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6411d1e7e18d5ceadaaceed5a8a6b20a0cacc55cede793ca9fb4dcad70c64fc6\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T18:17:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T18:17:10Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T18:18:21Z is after 2025-08-24T17:21:41Z" Oct 08 18:18:21 crc kubenswrapper[4859]: I1008 18:18:21.153912 4859 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-08T18:17:29Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T18:17:29Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T18:18:21Z is after 2025-08-24T17:21:41Z" Oct 08 18:18:21 crc kubenswrapper[4859]: I1008 18:18:21.174963 4859 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-84k8x" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1c24078c-c9e8-4160-883a-f483b4b3dd23\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:17:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:17:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:17:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:17:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8661006834d0f50eaada7394369584ee95fc61ff54437eec35d755526afc5b48\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T18:17:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hk5xb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://18b291eb3344fb7c00ef5679dca9a6f7ed9d08cea684a803ff7dad2e8a18c585\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://18b291eb3344fb7c00ef5679dca9a6f7ed9d08cea684a803ff7dad2e8a18c585\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-08T18:17:32Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-08T18:17:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hk5xb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://160163f1a91d981583452db3c7a9244182d1c8ced45749a61b721d0d45c1dfe4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://160163f1a91d981583452db3c7a9244182d1c8ced45749a61b721d0d45c1dfe4\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-08T18:17:34Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-08T18:17:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hk5xb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://51e9b53f3847ac53f378ad83870debf8e71518d4edbc7fc0dc66644818e2103e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://51e9b53f3847ac53f378ad83870debf8e71518d4edbc7fc0dc66644818e2103e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-08T18:17:34Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-08T18:17:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hk5xb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1ff118919739759ac01b5cbe3240e6673db54eaa16353c609d670071ed7964cb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1ff118919739759ac01b5cbe3240e6673db54eaa16353c609d670071ed7964cb\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-08T18:17:35Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-08T18:17:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hk5xb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8522e803ff56eb57cd4d9fd036e0c3fb16240ea21d4a094c1db7aa7d50bfe2c2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8522e803ff56eb57cd4d9fd036e0c3fb16240ea21d4a094c1db7aa7d50bfe2c2\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-08T18:17:37Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-08T18:17:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hk5xb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://be61056edf515e037403ca2a0bfd5feb9b2537dd42212b58e133b23b294f327b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://be61056edf515e037403ca2a0bfd5feb9b2537dd42212b58e133b23b294f327b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-08T18:17:38Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-08T18:17:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hk5xb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T18:17:31Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-84k8x\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T18:18:21Z is after 2025-08-24T17:21:41Z" Oct 08 18:18:21 crc kubenswrapper[4859]: I1008 18:18:21.188469 4859 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-btjmq" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b8605f7-fed7-493f-9b4f-6a6358adf907\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:17:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:17:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:17:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:17:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://14e9b6c4a4403f36adb78fcf1135a620ca6fc56ed974c9639dcdfc59e6917559\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T18:17:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xspg4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T18:17:35Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-btjmq\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T18:18:21Z is after 2025-08-24T17:21:41Z" Oct 08 18:18:21 crc kubenswrapper[4859]: I1008 18:18:21.200032 4859 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-g4t79" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3a92dbf1-21a8-4095-96bb-2d701d7f3d56\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:17:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:17:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:17:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:17:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f8772a25b6e605efe9a062dc32dbbc2cfe32b4fe164b4e84fbf3bcffab744179\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T18:17:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sv79p\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://92cb9986c09a7f9e1e552e3e3ac1f74f04128241e00ddee38882d2ba5c005a67\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T18:17:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sv79p\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T18:17:45Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-g4t79\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T18:18:21Z is after 2025-08-24T17:21:41Z" Oct 08 18:18:21 crc kubenswrapper[4859]: I1008 18:18:21.204149 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 18:18:21 crc kubenswrapper[4859]: I1008 18:18:21.204195 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 18:18:21 crc kubenswrapper[4859]: I1008 18:18:21.204206 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 18:18:21 crc kubenswrapper[4859]: I1008 18:18:21.204229 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 18:18:21 crc kubenswrapper[4859]: I1008 18:18:21.204245 4859 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T18:18:21Z","lastTransitionTime":"2025-10-08T18:18:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 18:18:21 crc kubenswrapper[4859]: I1008 18:18:21.211274 4859 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3ca6ce76-77e8-4e46-a2f6-b1ed4aa0fd4b\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:17:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:17:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:18:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:18:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:17:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f587e149d57ea53626ed3beb83e421c3b89ba1fec9e949705e8ccb5b4f7cc79a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T18:17:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://428cbe63daefbfbcb2ec76dad0fe638cdc65934bac80b791cebe0b36b2d6909a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T18:17:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a41399a37b1ad86a063001224232ecb8f7ea56aeb5686d3858bc08102a0c8f49\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T18:17:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://008757fce552591f7c58193f605eb60792761d5943927cbd63184549618d2b43\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://008757fce552591f7c58193f605eb60792761d5943927cbd63184549618d2b43\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-08T18:17:11Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-08T18:17:11Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T18:17:10Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T18:18:21Z is after 2025-08-24T17:21:41Z" Oct 08 18:18:21 crc kubenswrapper[4859]: I1008 18:18:21.224455 4859 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-08T18:17:29Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T18:17:29Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T18:18:21Z is after 2025-08-24T17:21:41Z" Oct 08 18:18:21 crc kubenswrapper[4859]: I1008 18:18:21.235462 4859 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-btw28" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"97bc5598-3ead-4300-a938-3aebbff1e8c6\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:17:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:17:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:17:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:17:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d6d4d6ba771cb5fde637449bba05660f634df7e833a0c8dad2b2c7f3a422bb2b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T18:17:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qcxsd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T18:17:31Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-btw28\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T18:18:21Z is after 2025-08-24T17:21:41Z" Oct 08 18:18:21 crc kubenswrapper[4859]: I1008 18:18:21.250010 4859 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-8plkg" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"92b2d7de-31cc-4a91-95a5-ed5ea964b028\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:17:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:17:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:18:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:18:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://729bd32fd01dcab495bffa64f2f5e44990fc1d78b05e7f131868009f92305831\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://fa00b8345d40b74a917ddc23cc1b2c5c684a9ca61b6119d090c92cc1439b71c4\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-10-08T18:18:19Z\\\",\\\"message\\\":\\\"2025-10-08T18:17:33+00:00 [cnibincopy] Successfully copied files in /usr/src/multus-cni/rhel9/bin/ to /host/opt/cni/bin/upgrade_888b130f-8754-48c6-b73e-9ff6772793f9\\\\n2025-10-08T18:17:33+00:00 [cnibincopy] Successfully moved files in /host/opt/cni/bin/upgrade_888b130f-8754-48c6-b73e-9ff6772793f9 to /host/opt/cni/bin/\\\\n2025-10-08T18:17:34Z [verbose] multus-daemon started\\\\n2025-10-08T18:17:34Z [verbose] Readiness Indicator file check\\\\n2025-10-08T18:18:19Z [error] have you checked that your default network is ready? still waiting for readinessindicatorfile @ /host/run/multus/cni/net.d/10-ovn-kubernetes.conf. pollimmediate error: timed out waiting for the condition\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-08T18:17:32Z\\\"}},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T18:18:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q9d7m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T18:17:31Z\\\"}}\" for pod \"openshift-multus\"/\"multus-8plkg\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T18:18:21Z is after 2025-08-24T17:21:41Z" Oct 08 18:18:21 crc kubenswrapper[4859]: I1008 18:18:21.306591 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 18:18:21 crc kubenswrapper[4859]: I1008 18:18:21.307014 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 18:18:21 crc kubenswrapper[4859]: I1008 18:18:21.307109 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 18:18:21 crc kubenswrapper[4859]: I1008 18:18:21.307197 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 18:18:21 crc kubenswrapper[4859]: I1008 18:18:21.307277 4859 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T18:18:21Z","lastTransitionTime":"2025-10-08T18:18:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 18:18:21 crc kubenswrapper[4859]: I1008 18:18:21.409431 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 18:18:21 crc kubenswrapper[4859]: I1008 18:18:21.409488 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 18:18:21 crc kubenswrapper[4859]: I1008 18:18:21.409500 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 18:18:21 crc kubenswrapper[4859]: I1008 18:18:21.409521 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 18:18:21 crc kubenswrapper[4859]: I1008 18:18:21.409534 4859 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T18:18:21Z","lastTransitionTime":"2025-10-08T18:18:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 18:18:21 crc kubenswrapper[4859]: I1008 18:18:21.469840 4859 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 08 18:18:21 crc kubenswrapper[4859]: I1008 18:18:21.469929 4859 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 08 18:18:21 crc kubenswrapper[4859]: I1008 18:18:21.469854 4859 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 08 18:18:21 crc kubenswrapper[4859]: E1008 18:18:21.469991 4859 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 08 18:18:21 crc kubenswrapper[4859]: I1008 18:18:21.470045 4859 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-4b4sz" Oct 08 18:18:21 crc kubenswrapper[4859]: E1008 18:18:21.470054 4859 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 08 18:18:21 crc kubenswrapper[4859]: E1008 18:18:21.470215 4859 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-4b4sz" podUID="2e53cd9b-64f7-4e07-8b96-fb77847c9ac6" Oct 08 18:18:21 crc kubenswrapper[4859]: E1008 18:18:21.470331 4859 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 08 18:18:21 crc kubenswrapper[4859]: I1008 18:18:21.512106 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 18:18:21 crc kubenswrapper[4859]: I1008 18:18:21.512157 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 18:18:21 crc kubenswrapper[4859]: I1008 18:18:21.512168 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 18:18:21 crc kubenswrapper[4859]: I1008 18:18:21.512186 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 18:18:21 crc kubenswrapper[4859]: I1008 18:18:21.512198 4859 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T18:18:21Z","lastTransitionTime":"2025-10-08T18:18:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 18:18:21 crc kubenswrapper[4859]: I1008 18:18:21.614986 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 18:18:21 crc kubenswrapper[4859]: I1008 18:18:21.615061 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 18:18:21 crc kubenswrapper[4859]: I1008 18:18:21.615080 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 18:18:21 crc kubenswrapper[4859]: I1008 18:18:21.615110 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 18:18:21 crc kubenswrapper[4859]: I1008 18:18:21.615130 4859 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T18:18:21Z","lastTransitionTime":"2025-10-08T18:18:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 18:18:21 crc kubenswrapper[4859]: I1008 18:18:21.717821 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 18:18:21 crc kubenswrapper[4859]: I1008 18:18:21.717875 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 18:18:21 crc kubenswrapper[4859]: I1008 18:18:21.717890 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 18:18:21 crc kubenswrapper[4859]: I1008 18:18:21.717912 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 18:18:21 crc kubenswrapper[4859]: I1008 18:18:21.717929 4859 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T18:18:21Z","lastTransitionTime":"2025-10-08T18:18:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 18:18:21 crc kubenswrapper[4859]: I1008 18:18:21.821272 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 18:18:21 crc kubenswrapper[4859]: I1008 18:18:21.821329 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 18:18:21 crc kubenswrapper[4859]: I1008 18:18:21.821339 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 18:18:21 crc kubenswrapper[4859]: I1008 18:18:21.821355 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 18:18:21 crc kubenswrapper[4859]: I1008 18:18:21.821365 4859 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T18:18:21Z","lastTransitionTime":"2025-10-08T18:18:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 18:18:21 crc kubenswrapper[4859]: I1008 18:18:21.923532 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 18:18:21 crc kubenswrapper[4859]: I1008 18:18:21.923580 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 18:18:21 crc kubenswrapper[4859]: I1008 18:18:21.923590 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 18:18:21 crc kubenswrapper[4859]: I1008 18:18:21.923610 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 18:18:21 crc kubenswrapper[4859]: I1008 18:18:21.923621 4859 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T18:18:21Z","lastTransitionTime":"2025-10-08T18:18:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 18:18:22 crc kubenswrapper[4859]: I1008 18:18:22.027677 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 18:18:22 crc kubenswrapper[4859]: I1008 18:18:22.027804 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 18:18:22 crc kubenswrapper[4859]: I1008 18:18:22.027828 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 18:18:22 crc kubenswrapper[4859]: I1008 18:18:22.027861 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 18:18:22 crc kubenswrapper[4859]: I1008 18:18:22.027882 4859 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T18:18:22Z","lastTransitionTime":"2025-10-08T18:18:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 18:18:22 crc kubenswrapper[4859]: I1008 18:18:22.131033 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 18:18:22 crc kubenswrapper[4859]: I1008 18:18:22.131073 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 18:18:22 crc kubenswrapper[4859]: I1008 18:18:22.131084 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 18:18:22 crc kubenswrapper[4859]: I1008 18:18:22.131102 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 18:18:22 crc kubenswrapper[4859]: I1008 18:18:22.131113 4859 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T18:18:22Z","lastTransitionTime":"2025-10-08T18:18:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 18:18:22 crc kubenswrapper[4859]: I1008 18:18:22.233725 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 18:18:22 crc kubenswrapper[4859]: I1008 18:18:22.233765 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 18:18:22 crc kubenswrapper[4859]: I1008 18:18:22.233775 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 18:18:22 crc kubenswrapper[4859]: I1008 18:18:22.233790 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 18:18:22 crc kubenswrapper[4859]: I1008 18:18:22.233800 4859 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T18:18:22Z","lastTransitionTime":"2025-10-08T18:18:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 18:18:22 crc kubenswrapper[4859]: I1008 18:18:22.336983 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 18:18:22 crc kubenswrapper[4859]: I1008 18:18:22.337028 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 18:18:22 crc kubenswrapper[4859]: I1008 18:18:22.337037 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 18:18:22 crc kubenswrapper[4859]: I1008 18:18:22.337054 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 18:18:22 crc kubenswrapper[4859]: I1008 18:18:22.337064 4859 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T18:18:22Z","lastTransitionTime":"2025-10-08T18:18:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 18:18:22 crc kubenswrapper[4859]: I1008 18:18:22.443744 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 18:18:22 crc kubenswrapper[4859]: I1008 18:18:22.444080 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 18:18:22 crc kubenswrapper[4859]: I1008 18:18:22.444160 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 18:18:22 crc kubenswrapper[4859]: I1008 18:18:22.444293 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 18:18:22 crc kubenswrapper[4859]: I1008 18:18:22.444375 4859 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T18:18:22Z","lastTransitionTime":"2025-10-08T18:18:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 18:18:22 crc kubenswrapper[4859]: I1008 18:18:22.547421 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 18:18:22 crc kubenswrapper[4859]: I1008 18:18:22.547464 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 18:18:22 crc kubenswrapper[4859]: I1008 18:18:22.547475 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 18:18:22 crc kubenswrapper[4859]: I1008 18:18:22.547492 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 18:18:22 crc kubenswrapper[4859]: I1008 18:18:22.547503 4859 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T18:18:22Z","lastTransitionTime":"2025-10-08T18:18:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 18:18:22 crc kubenswrapper[4859]: I1008 18:18:22.650098 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 18:18:22 crc kubenswrapper[4859]: I1008 18:18:22.650425 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 18:18:22 crc kubenswrapper[4859]: I1008 18:18:22.650870 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 18:18:22 crc kubenswrapper[4859]: I1008 18:18:22.650969 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 18:18:22 crc kubenswrapper[4859]: I1008 18:18:22.651048 4859 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T18:18:22Z","lastTransitionTime":"2025-10-08T18:18:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 18:18:22 crc kubenswrapper[4859]: I1008 18:18:22.754129 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 18:18:22 crc kubenswrapper[4859]: I1008 18:18:22.754188 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 18:18:22 crc kubenswrapper[4859]: I1008 18:18:22.754201 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 18:18:22 crc kubenswrapper[4859]: I1008 18:18:22.754220 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 18:18:22 crc kubenswrapper[4859]: I1008 18:18:22.754236 4859 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T18:18:22Z","lastTransitionTime":"2025-10-08T18:18:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 18:18:22 crc kubenswrapper[4859]: I1008 18:18:22.857237 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 18:18:22 crc kubenswrapper[4859]: I1008 18:18:22.857519 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 18:18:22 crc kubenswrapper[4859]: I1008 18:18:22.857605 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 18:18:22 crc kubenswrapper[4859]: I1008 18:18:22.857709 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 18:18:22 crc kubenswrapper[4859]: I1008 18:18:22.857837 4859 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T18:18:22Z","lastTransitionTime":"2025-10-08T18:18:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 18:18:22 crc kubenswrapper[4859]: I1008 18:18:22.961060 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 18:18:22 crc kubenswrapper[4859]: I1008 18:18:22.961120 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 18:18:22 crc kubenswrapper[4859]: I1008 18:18:22.961138 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 18:18:22 crc kubenswrapper[4859]: I1008 18:18:22.961163 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 18:18:22 crc kubenswrapper[4859]: I1008 18:18:22.961182 4859 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T18:18:22Z","lastTransitionTime":"2025-10-08T18:18:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 18:18:23 crc kubenswrapper[4859]: I1008 18:18:23.063499 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 18:18:23 crc kubenswrapper[4859]: I1008 18:18:23.063859 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 18:18:23 crc kubenswrapper[4859]: I1008 18:18:23.063932 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 18:18:23 crc kubenswrapper[4859]: I1008 18:18:23.064005 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 18:18:23 crc kubenswrapper[4859]: I1008 18:18:23.064097 4859 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T18:18:23Z","lastTransitionTime":"2025-10-08T18:18:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 18:18:23 crc kubenswrapper[4859]: I1008 18:18:23.167413 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 18:18:23 crc kubenswrapper[4859]: I1008 18:18:23.167928 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 18:18:23 crc kubenswrapper[4859]: I1008 18:18:23.168057 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 18:18:23 crc kubenswrapper[4859]: I1008 18:18:23.168180 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 18:18:23 crc kubenswrapper[4859]: I1008 18:18:23.168290 4859 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T18:18:23Z","lastTransitionTime":"2025-10-08T18:18:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 18:18:23 crc kubenswrapper[4859]: I1008 18:18:23.272468 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 18:18:23 crc kubenswrapper[4859]: I1008 18:18:23.272846 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 18:18:23 crc kubenswrapper[4859]: I1008 18:18:23.272956 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 18:18:23 crc kubenswrapper[4859]: I1008 18:18:23.273025 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 18:18:23 crc kubenswrapper[4859]: I1008 18:18:23.273110 4859 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T18:18:23Z","lastTransitionTime":"2025-10-08T18:18:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 18:18:23 crc kubenswrapper[4859]: I1008 18:18:23.375774 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 18:18:23 crc kubenswrapper[4859]: I1008 18:18:23.376178 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 18:18:23 crc kubenswrapper[4859]: I1008 18:18:23.376405 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 18:18:23 crc kubenswrapper[4859]: I1008 18:18:23.376598 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 18:18:23 crc kubenswrapper[4859]: I1008 18:18:23.376784 4859 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T18:18:23Z","lastTransitionTime":"2025-10-08T18:18:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 18:18:23 crc kubenswrapper[4859]: I1008 18:18:23.469037 4859 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 08 18:18:23 crc kubenswrapper[4859]: I1008 18:18:23.469091 4859 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 08 18:18:23 crc kubenswrapper[4859]: I1008 18:18:23.469208 4859 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 08 18:18:23 crc kubenswrapper[4859]: E1008 18:18:23.469206 4859 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 08 18:18:23 crc kubenswrapper[4859]: I1008 18:18:23.469237 4859 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-4b4sz" Oct 08 18:18:23 crc kubenswrapper[4859]: E1008 18:18:23.469367 4859 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 08 18:18:23 crc kubenswrapper[4859]: E1008 18:18:23.469466 4859 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 08 18:18:23 crc kubenswrapper[4859]: E1008 18:18:23.469536 4859 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-4b4sz" podUID="2e53cd9b-64f7-4e07-8b96-fb77847c9ac6" Oct 08 18:18:23 crc kubenswrapper[4859]: I1008 18:18:23.480465 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 18:18:23 crc kubenswrapper[4859]: I1008 18:18:23.480506 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 18:18:23 crc kubenswrapper[4859]: I1008 18:18:23.480522 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 18:18:23 crc kubenswrapper[4859]: I1008 18:18:23.480540 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 18:18:23 crc kubenswrapper[4859]: I1008 18:18:23.480552 4859 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T18:18:23Z","lastTransitionTime":"2025-10-08T18:18:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 18:18:23 crc kubenswrapper[4859]: I1008 18:18:23.584550 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 18:18:23 crc kubenswrapper[4859]: I1008 18:18:23.584903 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 18:18:23 crc kubenswrapper[4859]: I1008 18:18:23.584975 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 18:18:23 crc kubenswrapper[4859]: I1008 18:18:23.585056 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 18:18:23 crc kubenswrapper[4859]: I1008 18:18:23.585128 4859 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T18:18:23Z","lastTransitionTime":"2025-10-08T18:18:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 18:18:23 crc kubenswrapper[4859]: I1008 18:18:23.688138 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 18:18:23 crc kubenswrapper[4859]: I1008 18:18:23.688173 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 18:18:23 crc kubenswrapper[4859]: I1008 18:18:23.688185 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 18:18:23 crc kubenswrapper[4859]: I1008 18:18:23.688199 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 18:18:23 crc kubenswrapper[4859]: I1008 18:18:23.688211 4859 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T18:18:23Z","lastTransitionTime":"2025-10-08T18:18:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 18:18:23 crc kubenswrapper[4859]: I1008 18:18:23.790942 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 18:18:23 crc kubenswrapper[4859]: I1008 18:18:23.790996 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 18:18:23 crc kubenswrapper[4859]: I1008 18:18:23.791010 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 18:18:23 crc kubenswrapper[4859]: I1008 18:18:23.791030 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 18:18:23 crc kubenswrapper[4859]: I1008 18:18:23.791042 4859 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T18:18:23Z","lastTransitionTime":"2025-10-08T18:18:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 18:18:23 crc kubenswrapper[4859]: I1008 18:18:23.894171 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 18:18:23 crc kubenswrapper[4859]: I1008 18:18:23.894236 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 18:18:23 crc kubenswrapper[4859]: I1008 18:18:23.894248 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 18:18:23 crc kubenswrapper[4859]: I1008 18:18:23.894265 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 18:18:23 crc kubenswrapper[4859]: I1008 18:18:23.894276 4859 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T18:18:23Z","lastTransitionTime":"2025-10-08T18:18:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 18:18:23 crc kubenswrapper[4859]: I1008 18:18:23.996330 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 18:18:23 crc kubenswrapper[4859]: I1008 18:18:23.996372 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 18:18:23 crc kubenswrapper[4859]: I1008 18:18:23.996380 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 18:18:23 crc kubenswrapper[4859]: I1008 18:18:23.996397 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 18:18:23 crc kubenswrapper[4859]: I1008 18:18:23.996407 4859 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T18:18:23Z","lastTransitionTime":"2025-10-08T18:18:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 18:18:24 crc kubenswrapper[4859]: I1008 18:18:24.099905 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 18:18:24 crc kubenswrapper[4859]: I1008 18:18:24.099958 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 18:18:24 crc kubenswrapper[4859]: I1008 18:18:24.099971 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 18:18:24 crc kubenswrapper[4859]: I1008 18:18:24.099989 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 18:18:24 crc kubenswrapper[4859]: I1008 18:18:24.100000 4859 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T18:18:24Z","lastTransitionTime":"2025-10-08T18:18:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 18:18:24 crc kubenswrapper[4859]: I1008 18:18:24.202307 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 18:18:24 crc kubenswrapper[4859]: I1008 18:18:24.202621 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 18:18:24 crc kubenswrapper[4859]: I1008 18:18:24.202729 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 18:18:24 crc kubenswrapper[4859]: I1008 18:18:24.202819 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 18:18:24 crc kubenswrapper[4859]: I1008 18:18:24.202943 4859 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T18:18:24Z","lastTransitionTime":"2025-10-08T18:18:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 18:18:24 crc kubenswrapper[4859]: I1008 18:18:24.305354 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 18:18:24 crc kubenswrapper[4859]: I1008 18:18:24.305432 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 18:18:24 crc kubenswrapper[4859]: I1008 18:18:24.305451 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 18:18:24 crc kubenswrapper[4859]: I1008 18:18:24.305480 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 18:18:24 crc kubenswrapper[4859]: I1008 18:18:24.305504 4859 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T18:18:24Z","lastTransitionTime":"2025-10-08T18:18:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 18:18:24 crc kubenswrapper[4859]: I1008 18:18:24.409393 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 18:18:24 crc kubenswrapper[4859]: I1008 18:18:24.409451 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 18:18:24 crc kubenswrapper[4859]: I1008 18:18:24.409470 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 18:18:24 crc kubenswrapper[4859]: I1008 18:18:24.409536 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 18:18:24 crc kubenswrapper[4859]: I1008 18:18:24.409556 4859 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T18:18:24Z","lastTransitionTime":"2025-10-08T18:18:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 18:18:24 crc kubenswrapper[4859]: I1008 18:18:24.513324 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 18:18:24 crc kubenswrapper[4859]: I1008 18:18:24.513787 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 18:18:24 crc kubenswrapper[4859]: I1008 18:18:24.514009 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 18:18:24 crc kubenswrapper[4859]: I1008 18:18:24.514153 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 18:18:24 crc kubenswrapper[4859]: I1008 18:18:24.514396 4859 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T18:18:24Z","lastTransitionTime":"2025-10-08T18:18:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 18:18:24 crc kubenswrapper[4859]: I1008 18:18:24.617897 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 18:18:24 crc kubenswrapper[4859]: I1008 18:18:24.618353 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 18:18:24 crc kubenswrapper[4859]: I1008 18:18:24.618503 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 18:18:24 crc kubenswrapper[4859]: I1008 18:18:24.618654 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 18:18:24 crc kubenswrapper[4859]: I1008 18:18:24.618875 4859 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T18:18:24Z","lastTransitionTime":"2025-10-08T18:18:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 18:18:24 crc kubenswrapper[4859]: I1008 18:18:24.733096 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 18:18:24 crc kubenswrapper[4859]: I1008 18:18:24.733159 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 18:18:24 crc kubenswrapper[4859]: I1008 18:18:24.733177 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 18:18:24 crc kubenswrapper[4859]: I1008 18:18:24.733199 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 18:18:24 crc kubenswrapper[4859]: I1008 18:18:24.733214 4859 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T18:18:24Z","lastTransitionTime":"2025-10-08T18:18:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 18:18:24 crc kubenswrapper[4859]: I1008 18:18:24.835581 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 18:18:24 crc kubenswrapper[4859]: I1008 18:18:24.836187 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 18:18:24 crc kubenswrapper[4859]: I1008 18:18:24.836254 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 18:18:24 crc kubenswrapper[4859]: I1008 18:18:24.836320 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 18:18:24 crc kubenswrapper[4859]: I1008 18:18:24.836378 4859 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T18:18:24Z","lastTransitionTime":"2025-10-08T18:18:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 18:18:24 crc kubenswrapper[4859]: I1008 18:18:24.939557 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 18:18:24 crc kubenswrapper[4859]: I1008 18:18:24.939616 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 18:18:24 crc kubenswrapper[4859]: I1008 18:18:24.939631 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 18:18:24 crc kubenswrapper[4859]: I1008 18:18:24.939656 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 18:18:24 crc kubenswrapper[4859]: I1008 18:18:24.939673 4859 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T18:18:24Z","lastTransitionTime":"2025-10-08T18:18:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 18:18:25 crc kubenswrapper[4859]: I1008 18:18:25.043326 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 18:18:25 crc kubenswrapper[4859]: I1008 18:18:25.043414 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 18:18:25 crc kubenswrapper[4859]: I1008 18:18:25.043432 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 18:18:25 crc kubenswrapper[4859]: I1008 18:18:25.043459 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 18:18:25 crc kubenswrapper[4859]: I1008 18:18:25.043478 4859 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T18:18:25Z","lastTransitionTime":"2025-10-08T18:18:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 18:18:25 crc kubenswrapper[4859]: I1008 18:18:25.146522 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 18:18:25 crc kubenswrapper[4859]: I1008 18:18:25.146583 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 18:18:25 crc kubenswrapper[4859]: I1008 18:18:25.146596 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 18:18:25 crc kubenswrapper[4859]: I1008 18:18:25.146617 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 18:18:25 crc kubenswrapper[4859]: I1008 18:18:25.146630 4859 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T18:18:25Z","lastTransitionTime":"2025-10-08T18:18:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 18:18:25 crc kubenswrapper[4859]: I1008 18:18:25.249484 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 18:18:25 crc kubenswrapper[4859]: I1008 18:18:25.249532 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 18:18:25 crc kubenswrapper[4859]: I1008 18:18:25.249546 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 18:18:25 crc kubenswrapper[4859]: I1008 18:18:25.249566 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 18:18:25 crc kubenswrapper[4859]: I1008 18:18:25.249581 4859 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T18:18:25Z","lastTransitionTime":"2025-10-08T18:18:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 18:18:25 crc kubenswrapper[4859]: I1008 18:18:25.352954 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 18:18:25 crc kubenswrapper[4859]: I1008 18:18:25.353054 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 18:18:25 crc kubenswrapper[4859]: I1008 18:18:25.353075 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 18:18:25 crc kubenswrapper[4859]: I1008 18:18:25.353098 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 18:18:25 crc kubenswrapper[4859]: I1008 18:18:25.353117 4859 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T18:18:25Z","lastTransitionTime":"2025-10-08T18:18:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 18:18:25 crc kubenswrapper[4859]: I1008 18:18:25.455837 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 18:18:25 crc kubenswrapper[4859]: I1008 18:18:25.456290 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 18:18:25 crc kubenswrapper[4859]: I1008 18:18:25.456467 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 18:18:25 crc kubenswrapper[4859]: I1008 18:18:25.456638 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 18:18:25 crc kubenswrapper[4859]: I1008 18:18:25.456884 4859 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T18:18:25Z","lastTransitionTime":"2025-10-08T18:18:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 18:18:25 crc kubenswrapper[4859]: I1008 18:18:25.469413 4859 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-4b4sz" Oct 08 18:18:25 crc kubenswrapper[4859]: I1008 18:18:25.469461 4859 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 08 18:18:25 crc kubenswrapper[4859]: E1008 18:18:25.469945 4859 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-4b4sz" podUID="2e53cd9b-64f7-4e07-8b96-fb77847c9ac6" Oct 08 18:18:25 crc kubenswrapper[4859]: I1008 18:18:25.469495 4859 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 08 18:18:25 crc kubenswrapper[4859]: I1008 18:18:25.469484 4859 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 08 18:18:25 crc kubenswrapper[4859]: E1008 18:18:25.470019 4859 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 08 18:18:25 crc kubenswrapper[4859]: E1008 18:18:25.470324 4859 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 08 18:18:25 crc kubenswrapper[4859]: E1008 18:18:25.470596 4859 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 08 18:18:25 crc kubenswrapper[4859]: I1008 18:18:25.559202 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 18:18:25 crc kubenswrapper[4859]: I1008 18:18:25.559265 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 18:18:25 crc kubenswrapper[4859]: I1008 18:18:25.559282 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 18:18:25 crc kubenswrapper[4859]: I1008 18:18:25.559303 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 18:18:25 crc kubenswrapper[4859]: I1008 18:18:25.559316 4859 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T18:18:25Z","lastTransitionTime":"2025-10-08T18:18:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 18:18:25 crc kubenswrapper[4859]: I1008 18:18:25.661978 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 18:18:25 crc kubenswrapper[4859]: I1008 18:18:25.662031 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 18:18:25 crc kubenswrapper[4859]: I1008 18:18:25.662044 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 18:18:25 crc kubenswrapper[4859]: I1008 18:18:25.662064 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 18:18:25 crc kubenswrapper[4859]: I1008 18:18:25.662077 4859 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T18:18:25Z","lastTransitionTime":"2025-10-08T18:18:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 18:18:25 crc kubenswrapper[4859]: I1008 18:18:25.765186 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 18:18:25 crc kubenswrapper[4859]: I1008 18:18:25.765230 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 18:18:25 crc kubenswrapper[4859]: I1008 18:18:25.765242 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 18:18:25 crc kubenswrapper[4859]: I1008 18:18:25.765262 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 18:18:25 crc kubenswrapper[4859]: I1008 18:18:25.765274 4859 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T18:18:25Z","lastTransitionTime":"2025-10-08T18:18:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 18:18:25 crc kubenswrapper[4859]: I1008 18:18:25.868717 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 18:18:25 crc kubenswrapper[4859]: I1008 18:18:25.869669 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 18:18:25 crc kubenswrapper[4859]: I1008 18:18:25.869883 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 18:18:25 crc kubenswrapper[4859]: I1008 18:18:25.870046 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 18:18:25 crc kubenswrapper[4859]: I1008 18:18:25.870281 4859 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T18:18:25Z","lastTransitionTime":"2025-10-08T18:18:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 18:18:25 crc kubenswrapper[4859]: I1008 18:18:25.973607 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 18:18:25 crc kubenswrapper[4859]: I1008 18:18:25.973651 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 18:18:25 crc kubenswrapper[4859]: I1008 18:18:25.973662 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 18:18:25 crc kubenswrapper[4859]: I1008 18:18:25.973703 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 18:18:25 crc kubenswrapper[4859]: I1008 18:18:25.973718 4859 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T18:18:25Z","lastTransitionTime":"2025-10-08T18:18:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 18:18:26 crc kubenswrapper[4859]: I1008 18:18:26.077507 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 18:18:26 crc kubenswrapper[4859]: I1008 18:18:26.077574 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 18:18:26 crc kubenswrapper[4859]: I1008 18:18:26.077593 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 18:18:26 crc kubenswrapper[4859]: I1008 18:18:26.077673 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 18:18:26 crc kubenswrapper[4859]: I1008 18:18:26.077730 4859 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T18:18:26Z","lastTransitionTime":"2025-10-08T18:18:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 18:18:26 crc kubenswrapper[4859]: I1008 18:18:26.180203 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 18:18:26 crc kubenswrapper[4859]: I1008 18:18:26.180266 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 18:18:26 crc kubenswrapper[4859]: I1008 18:18:26.180284 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 18:18:26 crc kubenswrapper[4859]: I1008 18:18:26.180311 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 18:18:26 crc kubenswrapper[4859]: I1008 18:18:26.180331 4859 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T18:18:26Z","lastTransitionTime":"2025-10-08T18:18:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 18:18:26 crc kubenswrapper[4859]: I1008 18:18:26.283883 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 18:18:26 crc kubenswrapper[4859]: I1008 18:18:26.283951 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 18:18:26 crc kubenswrapper[4859]: I1008 18:18:26.283971 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 18:18:26 crc kubenswrapper[4859]: I1008 18:18:26.283997 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 18:18:26 crc kubenswrapper[4859]: I1008 18:18:26.284020 4859 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T18:18:26Z","lastTransitionTime":"2025-10-08T18:18:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 18:18:26 crc kubenswrapper[4859]: I1008 18:18:26.387620 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 18:18:26 crc kubenswrapper[4859]: I1008 18:18:26.387728 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 18:18:26 crc kubenswrapper[4859]: I1008 18:18:26.387749 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 18:18:26 crc kubenswrapper[4859]: I1008 18:18:26.387777 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 18:18:26 crc kubenswrapper[4859]: I1008 18:18:26.387795 4859 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T18:18:26Z","lastTransitionTime":"2025-10-08T18:18:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 18:18:26 crc kubenswrapper[4859]: I1008 18:18:26.490616 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 18:18:26 crc kubenswrapper[4859]: I1008 18:18:26.490666 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 18:18:26 crc kubenswrapper[4859]: I1008 18:18:26.490679 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 18:18:26 crc kubenswrapper[4859]: I1008 18:18:26.490716 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 18:18:26 crc kubenswrapper[4859]: I1008 18:18:26.490729 4859 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T18:18:26Z","lastTransitionTime":"2025-10-08T18:18:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 18:18:26 crc kubenswrapper[4859]: I1008 18:18:26.594427 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 18:18:26 crc kubenswrapper[4859]: I1008 18:18:26.594480 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 18:18:26 crc kubenswrapper[4859]: I1008 18:18:26.594490 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 18:18:26 crc kubenswrapper[4859]: I1008 18:18:26.594508 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 18:18:26 crc kubenswrapper[4859]: I1008 18:18:26.594526 4859 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T18:18:26Z","lastTransitionTime":"2025-10-08T18:18:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 18:18:26 crc kubenswrapper[4859]: I1008 18:18:26.697848 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 18:18:26 crc kubenswrapper[4859]: I1008 18:18:26.697938 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 18:18:26 crc kubenswrapper[4859]: I1008 18:18:26.697959 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 18:18:26 crc kubenswrapper[4859]: I1008 18:18:26.697988 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 18:18:26 crc kubenswrapper[4859]: I1008 18:18:26.698011 4859 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T18:18:26Z","lastTransitionTime":"2025-10-08T18:18:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 18:18:26 crc kubenswrapper[4859]: I1008 18:18:26.800938 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 18:18:26 crc kubenswrapper[4859]: I1008 18:18:26.800998 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 18:18:26 crc kubenswrapper[4859]: I1008 18:18:26.801010 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 18:18:26 crc kubenswrapper[4859]: I1008 18:18:26.801031 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 18:18:26 crc kubenswrapper[4859]: I1008 18:18:26.801044 4859 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T18:18:26Z","lastTransitionTime":"2025-10-08T18:18:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 18:18:26 crc kubenswrapper[4859]: I1008 18:18:26.904122 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 18:18:26 crc kubenswrapper[4859]: I1008 18:18:26.904164 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 18:18:26 crc kubenswrapper[4859]: I1008 18:18:26.904209 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 18:18:26 crc kubenswrapper[4859]: I1008 18:18:26.904227 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 18:18:26 crc kubenswrapper[4859]: I1008 18:18:26.904238 4859 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T18:18:26Z","lastTransitionTime":"2025-10-08T18:18:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 18:18:27 crc kubenswrapper[4859]: I1008 18:18:27.006758 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 18:18:27 crc kubenswrapper[4859]: I1008 18:18:27.006816 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 18:18:27 crc kubenswrapper[4859]: I1008 18:18:27.006829 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 18:18:27 crc kubenswrapper[4859]: I1008 18:18:27.006846 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 18:18:27 crc kubenswrapper[4859]: I1008 18:18:27.006855 4859 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T18:18:27Z","lastTransitionTime":"2025-10-08T18:18:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 18:18:27 crc kubenswrapper[4859]: I1008 18:18:27.109785 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 18:18:27 crc kubenswrapper[4859]: I1008 18:18:27.110212 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 18:18:27 crc kubenswrapper[4859]: I1008 18:18:27.110363 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 18:18:27 crc kubenswrapper[4859]: I1008 18:18:27.110567 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 18:18:27 crc kubenswrapper[4859]: I1008 18:18:27.110740 4859 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T18:18:27Z","lastTransitionTime":"2025-10-08T18:18:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 18:18:27 crc kubenswrapper[4859]: I1008 18:18:27.214068 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 18:18:27 crc kubenswrapper[4859]: I1008 18:18:27.214553 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 18:18:27 crc kubenswrapper[4859]: I1008 18:18:27.214625 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 18:18:27 crc kubenswrapper[4859]: I1008 18:18:27.214719 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 18:18:27 crc kubenswrapper[4859]: I1008 18:18:27.214809 4859 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T18:18:27Z","lastTransitionTime":"2025-10-08T18:18:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 18:18:27 crc kubenswrapper[4859]: I1008 18:18:27.317856 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 18:18:27 crc kubenswrapper[4859]: I1008 18:18:27.317914 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 18:18:27 crc kubenswrapper[4859]: I1008 18:18:27.317934 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 18:18:27 crc kubenswrapper[4859]: I1008 18:18:27.317958 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 18:18:27 crc kubenswrapper[4859]: I1008 18:18:27.317975 4859 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T18:18:27Z","lastTransitionTime":"2025-10-08T18:18:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 18:18:27 crc kubenswrapper[4859]: I1008 18:18:27.421015 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 18:18:27 crc kubenswrapper[4859]: I1008 18:18:27.421434 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 18:18:27 crc kubenswrapper[4859]: I1008 18:18:27.421589 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 18:18:27 crc kubenswrapper[4859]: I1008 18:18:27.421818 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 18:18:27 crc kubenswrapper[4859]: I1008 18:18:27.422003 4859 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T18:18:27Z","lastTransitionTime":"2025-10-08T18:18:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 18:18:27 crc kubenswrapper[4859]: I1008 18:18:27.469680 4859 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-4b4sz" Oct 08 18:18:27 crc kubenswrapper[4859]: I1008 18:18:27.469732 4859 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 08 18:18:27 crc kubenswrapper[4859]: E1008 18:18:27.469860 4859 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-4b4sz" podUID="2e53cd9b-64f7-4e07-8b96-fb77847c9ac6" Oct 08 18:18:27 crc kubenswrapper[4859]: I1008 18:18:27.469880 4859 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 08 18:18:27 crc kubenswrapper[4859]: E1008 18:18:27.470031 4859 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 08 18:18:27 crc kubenswrapper[4859]: E1008 18:18:27.470102 4859 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 08 18:18:27 crc kubenswrapper[4859]: I1008 18:18:27.469682 4859 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 08 18:18:27 crc kubenswrapper[4859]: E1008 18:18:27.470314 4859 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 08 18:18:27 crc kubenswrapper[4859]: I1008 18:18:27.525069 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 18:18:27 crc kubenswrapper[4859]: I1008 18:18:27.525122 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 18:18:27 crc kubenswrapper[4859]: I1008 18:18:27.525137 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 18:18:27 crc kubenswrapper[4859]: I1008 18:18:27.525158 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 18:18:27 crc kubenswrapper[4859]: I1008 18:18:27.525172 4859 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T18:18:27Z","lastTransitionTime":"2025-10-08T18:18:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 18:18:27 crc kubenswrapper[4859]: I1008 18:18:27.629900 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 18:18:27 crc kubenswrapper[4859]: I1008 18:18:27.629972 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 18:18:27 crc kubenswrapper[4859]: I1008 18:18:27.629991 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 18:18:27 crc kubenswrapper[4859]: I1008 18:18:27.630017 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 18:18:27 crc kubenswrapper[4859]: I1008 18:18:27.630037 4859 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T18:18:27Z","lastTransitionTime":"2025-10-08T18:18:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 18:18:27 crc kubenswrapper[4859]: I1008 18:18:27.732469 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 18:18:27 crc kubenswrapper[4859]: I1008 18:18:27.732540 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 18:18:27 crc kubenswrapper[4859]: I1008 18:18:27.732557 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 18:18:27 crc kubenswrapper[4859]: I1008 18:18:27.732582 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 18:18:27 crc kubenswrapper[4859]: I1008 18:18:27.732600 4859 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T18:18:27Z","lastTransitionTime":"2025-10-08T18:18:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 18:18:27 crc kubenswrapper[4859]: I1008 18:18:27.835594 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 18:18:27 crc kubenswrapper[4859]: I1008 18:18:27.836026 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 18:18:27 crc kubenswrapper[4859]: I1008 18:18:27.836156 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 18:18:27 crc kubenswrapper[4859]: I1008 18:18:27.836339 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 18:18:27 crc kubenswrapper[4859]: I1008 18:18:27.836596 4859 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T18:18:27Z","lastTransitionTime":"2025-10-08T18:18:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 18:18:27 crc kubenswrapper[4859]: I1008 18:18:27.940391 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 18:18:27 crc kubenswrapper[4859]: I1008 18:18:27.941111 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 18:18:27 crc kubenswrapper[4859]: I1008 18:18:27.941136 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 18:18:27 crc kubenswrapper[4859]: I1008 18:18:27.941158 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 18:18:27 crc kubenswrapper[4859]: I1008 18:18:27.941172 4859 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T18:18:27Z","lastTransitionTime":"2025-10-08T18:18:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 18:18:28 crc kubenswrapper[4859]: I1008 18:18:28.044194 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 18:18:28 crc kubenswrapper[4859]: I1008 18:18:28.044251 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 18:18:28 crc kubenswrapper[4859]: I1008 18:18:28.044263 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 18:18:28 crc kubenswrapper[4859]: I1008 18:18:28.044296 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 18:18:28 crc kubenswrapper[4859]: I1008 18:18:28.044345 4859 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T18:18:28Z","lastTransitionTime":"2025-10-08T18:18:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 18:18:28 crc kubenswrapper[4859]: I1008 18:18:28.147658 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 18:18:28 crc kubenswrapper[4859]: I1008 18:18:28.147776 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 18:18:28 crc kubenswrapper[4859]: I1008 18:18:28.147794 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 18:18:28 crc kubenswrapper[4859]: I1008 18:18:28.147821 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 18:18:28 crc kubenswrapper[4859]: I1008 18:18:28.147839 4859 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T18:18:28Z","lastTransitionTime":"2025-10-08T18:18:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 18:18:28 crc kubenswrapper[4859]: I1008 18:18:28.251341 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 18:18:28 crc kubenswrapper[4859]: I1008 18:18:28.251928 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 18:18:28 crc kubenswrapper[4859]: I1008 18:18:28.252050 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 18:18:28 crc kubenswrapper[4859]: I1008 18:18:28.252204 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 18:18:28 crc kubenswrapper[4859]: I1008 18:18:28.252325 4859 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T18:18:28Z","lastTransitionTime":"2025-10-08T18:18:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 18:18:28 crc kubenswrapper[4859]: I1008 18:18:28.355106 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 18:18:28 crc kubenswrapper[4859]: I1008 18:18:28.355155 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 18:18:28 crc kubenswrapper[4859]: I1008 18:18:28.355173 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 18:18:28 crc kubenswrapper[4859]: I1008 18:18:28.355203 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 18:18:28 crc kubenswrapper[4859]: I1008 18:18:28.355225 4859 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T18:18:28Z","lastTransitionTime":"2025-10-08T18:18:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 18:18:28 crc kubenswrapper[4859]: I1008 18:18:28.459207 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 18:18:28 crc kubenswrapper[4859]: I1008 18:18:28.459284 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 18:18:28 crc kubenswrapper[4859]: I1008 18:18:28.459307 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 18:18:28 crc kubenswrapper[4859]: I1008 18:18:28.459338 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 18:18:28 crc kubenswrapper[4859]: I1008 18:18:28.459358 4859 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T18:18:28Z","lastTransitionTime":"2025-10-08T18:18:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 18:18:28 crc kubenswrapper[4859]: I1008 18:18:28.563246 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 18:18:28 crc kubenswrapper[4859]: I1008 18:18:28.563317 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 18:18:28 crc kubenswrapper[4859]: I1008 18:18:28.563329 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 18:18:28 crc kubenswrapper[4859]: I1008 18:18:28.563352 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 18:18:28 crc kubenswrapper[4859]: I1008 18:18:28.563384 4859 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T18:18:28Z","lastTransitionTime":"2025-10-08T18:18:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 18:18:28 crc kubenswrapper[4859]: I1008 18:18:28.571274 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 18:18:28 crc kubenswrapper[4859]: I1008 18:18:28.571446 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 18:18:28 crc kubenswrapper[4859]: I1008 18:18:28.571539 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 18:18:28 crc kubenswrapper[4859]: I1008 18:18:28.571624 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 18:18:28 crc kubenswrapper[4859]: I1008 18:18:28.571746 4859 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T18:18:28Z","lastTransitionTime":"2025-10-08T18:18:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 18:18:28 crc kubenswrapper[4859]: E1008 18:18:28.591317 4859 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404552Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865352Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-10-08T18:18:28Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-08T18:18:28Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-08T18:18:28Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-08T18:18:28Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-08T18:18:28Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-08T18:18:28Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-08T18:18:28Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-08T18:18:28Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"edb3ee6d-476b-407c-bb7d-4480786e8777\\\",\\\"systemUUID\\\":\\\"8f73c7ed-44aa-4d16-bae1-d4e684803cfc\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T18:18:28Z is after 2025-08-24T17:21:41Z" Oct 08 18:18:28 crc kubenswrapper[4859]: I1008 18:18:28.596866 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 18:18:28 crc kubenswrapper[4859]: I1008 18:18:28.597118 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 18:18:28 crc kubenswrapper[4859]: I1008 18:18:28.597298 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 18:18:28 crc kubenswrapper[4859]: I1008 18:18:28.597459 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 18:18:28 crc kubenswrapper[4859]: I1008 18:18:28.597613 4859 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T18:18:28Z","lastTransitionTime":"2025-10-08T18:18:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 18:18:28 crc kubenswrapper[4859]: E1008 18:18:28.618308 4859 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404552Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865352Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-10-08T18:18:28Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-08T18:18:28Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-08T18:18:28Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-08T18:18:28Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-08T18:18:28Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-08T18:18:28Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-08T18:18:28Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-08T18:18:28Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"edb3ee6d-476b-407c-bb7d-4480786e8777\\\",\\\"systemUUID\\\":\\\"8f73c7ed-44aa-4d16-bae1-d4e684803cfc\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T18:18:28Z is after 2025-08-24T17:21:41Z" Oct 08 18:18:28 crc kubenswrapper[4859]: I1008 18:18:28.624337 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 18:18:28 crc kubenswrapper[4859]: I1008 18:18:28.624491 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 18:18:28 crc kubenswrapper[4859]: I1008 18:18:28.624581 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 18:18:28 crc kubenswrapper[4859]: I1008 18:18:28.624673 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 18:18:28 crc kubenswrapper[4859]: I1008 18:18:28.624809 4859 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T18:18:28Z","lastTransitionTime":"2025-10-08T18:18:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 18:18:28 crc kubenswrapper[4859]: E1008 18:18:28.640849 4859 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404552Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865352Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-10-08T18:18:28Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-08T18:18:28Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-08T18:18:28Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-08T18:18:28Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-08T18:18:28Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-08T18:18:28Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-08T18:18:28Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-08T18:18:28Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"edb3ee6d-476b-407c-bb7d-4480786e8777\\\",\\\"systemUUID\\\":\\\"8f73c7ed-44aa-4d16-bae1-d4e684803cfc\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T18:18:28Z is after 2025-08-24T17:21:41Z" Oct 08 18:18:28 crc kubenswrapper[4859]: I1008 18:18:28.646346 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 18:18:28 crc kubenswrapper[4859]: I1008 18:18:28.646428 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 18:18:28 crc kubenswrapper[4859]: I1008 18:18:28.646453 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 18:18:28 crc kubenswrapper[4859]: I1008 18:18:28.646482 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 18:18:28 crc kubenswrapper[4859]: I1008 18:18:28.646503 4859 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T18:18:28Z","lastTransitionTime":"2025-10-08T18:18:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 18:18:28 crc kubenswrapper[4859]: E1008 18:18:28.665616 4859 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404552Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865352Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-10-08T18:18:28Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-08T18:18:28Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-08T18:18:28Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-08T18:18:28Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-08T18:18:28Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-08T18:18:28Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-08T18:18:28Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-08T18:18:28Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"edb3ee6d-476b-407c-bb7d-4480786e8777\\\",\\\"systemUUID\\\":\\\"8f73c7ed-44aa-4d16-bae1-d4e684803cfc\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T18:18:28Z is after 2025-08-24T17:21:41Z" Oct 08 18:18:28 crc kubenswrapper[4859]: I1008 18:18:28.670329 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 18:18:28 crc kubenswrapper[4859]: I1008 18:18:28.670416 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 18:18:28 crc kubenswrapper[4859]: I1008 18:18:28.670438 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 18:18:28 crc kubenswrapper[4859]: I1008 18:18:28.670469 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 18:18:28 crc kubenswrapper[4859]: I1008 18:18:28.670490 4859 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T18:18:28Z","lastTransitionTime":"2025-10-08T18:18:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 18:18:28 crc kubenswrapper[4859]: E1008 18:18:28.686984 4859 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404552Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865352Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-10-08T18:18:28Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-08T18:18:28Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-08T18:18:28Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-08T18:18:28Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-08T18:18:28Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-08T18:18:28Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-08T18:18:28Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-08T18:18:28Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"edb3ee6d-476b-407c-bb7d-4480786e8777\\\",\\\"systemUUID\\\":\\\"8f73c7ed-44aa-4d16-bae1-d4e684803cfc\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T18:18:28Z is after 2025-08-24T17:21:41Z" Oct 08 18:18:28 crc kubenswrapper[4859]: E1008 18:18:28.687095 4859 kubelet_node_status.go:572] "Unable to update node status" err="update node status exceeds retry count" Oct 08 18:18:28 crc kubenswrapper[4859]: I1008 18:18:28.688705 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 18:18:28 crc kubenswrapper[4859]: I1008 18:18:28.688738 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 18:18:28 crc kubenswrapper[4859]: I1008 18:18:28.688747 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 18:18:28 crc kubenswrapper[4859]: I1008 18:18:28.688761 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 18:18:28 crc kubenswrapper[4859]: I1008 18:18:28.688771 4859 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T18:18:28Z","lastTransitionTime":"2025-10-08T18:18:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 18:18:28 crc kubenswrapper[4859]: I1008 18:18:28.791772 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 18:18:28 crc kubenswrapper[4859]: I1008 18:18:28.791815 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 18:18:28 crc kubenswrapper[4859]: I1008 18:18:28.791824 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 18:18:28 crc kubenswrapper[4859]: I1008 18:18:28.791839 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 18:18:28 crc kubenswrapper[4859]: I1008 18:18:28.791849 4859 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T18:18:28Z","lastTransitionTime":"2025-10-08T18:18:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 18:18:28 crc kubenswrapper[4859]: I1008 18:18:28.895077 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 18:18:28 crc kubenswrapper[4859]: I1008 18:18:28.895160 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 18:18:28 crc kubenswrapper[4859]: I1008 18:18:28.895181 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 18:18:28 crc kubenswrapper[4859]: I1008 18:18:28.895211 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 18:18:28 crc kubenswrapper[4859]: I1008 18:18:28.895233 4859 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T18:18:28Z","lastTransitionTime":"2025-10-08T18:18:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 18:18:28 crc kubenswrapper[4859]: I1008 18:18:28.998147 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 18:18:28 crc kubenswrapper[4859]: I1008 18:18:28.998237 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 18:18:28 crc kubenswrapper[4859]: I1008 18:18:28.998268 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 18:18:28 crc kubenswrapper[4859]: I1008 18:18:28.998301 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 18:18:28 crc kubenswrapper[4859]: I1008 18:18:28.998325 4859 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T18:18:28Z","lastTransitionTime":"2025-10-08T18:18:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 18:18:29 crc kubenswrapper[4859]: I1008 18:18:29.102442 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 18:18:29 crc kubenswrapper[4859]: I1008 18:18:29.102533 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 18:18:29 crc kubenswrapper[4859]: I1008 18:18:29.102560 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 18:18:29 crc kubenswrapper[4859]: I1008 18:18:29.102603 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 18:18:29 crc kubenswrapper[4859]: I1008 18:18:29.102629 4859 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T18:18:29Z","lastTransitionTime":"2025-10-08T18:18:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 18:18:29 crc kubenswrapper[4859]: I1008 18:18:29.205636 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 18:18:29 crc kubenswrapper[4859]: I1008 18:18:29.205710 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 18:18:29 crc kubenswrapper[4859]: I1008 18:18:29.205723 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 18:18:29 crc kubenswrapper[4859]: I1008 18:18:29.205742 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 18:18:29 crc kubenswrapper[4859]: I1008 18:18:29.205757 4859 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T18:18:29Z","lastTransitionTime":"2025-10-08T18:18:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 18:18:29 crc kubenswrapper[4859]: I1008 18:18:29.308290 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 18:18:29 crc kubenswrapper[4859]: I1008 18:18:29.308339 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 18:18:29 crc kubenswrapper[4859]: I1008 18:18:29.308389 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 18:18:29 crc kubenswrapper[4859]: I1008 18:18:29.308414 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 18:18:29 crc kubenswrapper[4859]: I1008 18:18:29.308426 4859 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T18:18:29Z","lastTransitionTime":"2025-10-08T18:18:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 18:18:29 crc kubenswrapper[4859]: I1008 18:18:29.412092 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 18:18:29 crc kubenswrapper[4859]: I1008 18:18:29.412162 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 18:18:29 crc kubenswrapper[4859]: I1008 18:18:29.412181 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 18:18:29 crc kubenswrapper[4859]: I1008 18:18:29.412207 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 18:18:29 crc kubenswrapper[4859]: I1008 18:18:29.412227 4859 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T18:18:29Z","lastTransitionTime":"2025-10-08T18:18:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 18:18:29 crc kubenswrapper[4859]: I1008 18:18:29.469324 4859 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 08 18:18:29 crc kubenswrapper[4859]: I1008 18:18:29.469319 4859 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 08 18:18:29 crc kubenswrapper[4859]: E1008 18:18:29.469518 4859 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 08 18:18:29 crc kubenswrapper[4859]: E1008 18:18:29.469544 4859 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 08 18:18:29 crc kubenswrapper[4859]: I1008 18:18:29.469377 4859 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 08 18:18:29 crc kubenswrapper[4859]: E1008 18:18:29.469620 4859 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 08 18:18:29 crc kubenswrapper[4859]: I1008 18:18:29.469347 4859 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-4b4sz" Oct 08 18:18:29 crc kubenswrapper[4859]: E1008 18:18:29.469721 4859 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-4b4sz" podUID="2e53cd9b-64f7-4e07-8b96-fb77847c9ac6" Oct 08 18:18:29 crc kubenswrapper[4859]: I1008 18:18:29.515002 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 18:18:29 crc kubenswrapper[4859]: I1008 18:18:29.515093 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 18:18:29 crc kubenswrapper[4859]: I1008 18:18:29.515123 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 18:18:29 crc kubenswrapper[4859]: I1008 18:18:29.515161 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 18:18:29 crc kubenswrapper[4859]: I1008 18:18:29.515188 4859 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T18:18:29Z","lastTransitionTime":"2025-10-08T18:18:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 18:18:29 crc kubenswrapper[4859]: I1008 18:18:29.618337 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 18:18:29 crc kubenswrapper[4859]: I1008 18:18:29.618398 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 18:18:29 crc kubenswrapper[4859]: I1008 18:18:29.618416 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 18:18:29 crc kubenswrapper[4859]: I1008 18:18:29.618446 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 18:18:29 crc kubenswrapper[4859]: I1008 18:18:29.618466 4859 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T18:18:29Z","lastTransitionTime":"2025-10-08T18:18:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 18:18:29 crc kubenswrapper[4859]: I1008 18:18:29.720628 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 18:18:29 crc kubenswrapper[4859]: I1008 18:18:29.720664 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 18:18:29 crc kubenswrapper[4859]: I1008 18:18:29.720674 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 18:18:29 crc kubenswrapper[4859]: I1008 18:18:29.720709 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 18:18:29 crc kubenswrapper[4859]: I1008 18:18:29.720722 4859 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T18:18:29Z","lastTransitionTime":"2025-10-08T18:18:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 18:18:29 crc kubenswrapper[4859]: I1008 18:18:29.823610 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 18:18:29 crc kubenswrapper[4859]: I1008 18:18:29.823660 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 18:18:29 crc kubenswrapper[4859]: I1008 18:18:29.823673 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 18:18:29 crc kubenswrapper[4859]: I1008 18:18:29.823711 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 18:18:29 crc kubenswrapper[4859]: I1008 18:18:29.823729 4859 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T18:18:29Z","lastTransitionTime":"2025-10-08T18:18:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 18:18:29 crc kubenswrapper[4859]: I1008 18:18:29.927064 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 18:18:29 crc kubenswrapper[4859]: I1008 18:18:29.927131 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 18:18:29 crc kubenswrapper[4859]: I1008 18:18:29.927145 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 18:18:29 crc kubenswrapper[4859]: I1008 18:18:29.927192 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 18:18:29 crc kubenswrapper[4859]: I1008 18:18:29.927208 4859 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T18:18:29Z","lastTransitionTime":"2025-10-08T18:18:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 18:18:30 crc kubenswrapper[4859]: I1008 18:18:30.030379 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 18:18:30 crc kubenswrapper[4859]: I1008 18:18:30.030431 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 18:18:30 crc kubenswrapper[4859]: I1008 18:18:30.030442 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 18:18:30 crc kubenswrapper[4859]: I1008 18:18:30.030463 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 18:18:30 crc kubenswrapper[4859]: I1008 18:18:30.030475 4859 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T18:18:30Z","lastTransitionTime":"2025-10-08T18:18:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 18:18:30 crc kubenswrapper[4859]: I1008 18:18:30.133184 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 18:18:30 crc kubenswrapper[4859]: I1008 18:18:30.133250 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 18:18:30 crc kubenswrapper[4859]: I1008 18:18:30.133264 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 18:18:30 crc kubenswrapper[4859]: I1008 18:18:30.133287 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 18:18:30 crc kubenswrapper[4859]: I1008 18:18:30.133300 4859 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T18:18:30Z","lastTransitionTime":"2025-10-08T18:18:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 18:18:30 crc kubenswrapper[4859]: I1008 18:18:30.235941 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 18:18:30 crc kubenswrapper[4859]: I1008 18:18:30.235991 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 18:18:30 crc kubenswrapper[4859]: I1008 18:18:30.236001 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 18:18:30 crc kubenswrapper[4859]: I1008 18:18:30.236019 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 18:18:30 crc kubenswrapper[4859]: I1008 18:18:30.236029 4859 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T18:18:30Z","lastTransitionTime":"2025-10-08T18:18:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 18:18:30 crc kubenswrapper[4859]: I1008 18:18:30.339108 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 18:18:30 crc kubenswrapper[4859]: I1008 18:18:30.339193 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 18:18:30 crc kubenswrapper[4859]: I1008 18:18:30.339209 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 18:18:30 crc kubenswrapper[4859]: I1008 18:18:30.339233 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 18:18:30 crc kubenswrapper[4859]: I1008 18:18:30.339249 4859 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T18:18:30Z","lastTransitionTime":"2025-10-08T18:18:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 18:18:30 crc kubenswrapper[4859]: I1008 18:18:30.443275 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 18:18:30 crc kubenswrapper[4859]: I1008 18:18:30.443325 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 18:18:30 crc kubenswrapper[4859]: I1008 18:18:30.443337 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 18:18:30 crc kubenswrapper[4859]: I1008 18:18:30.443354 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 18:18:30 crc kubenswrapper[4859]: I1008 18:18:30.443364 4859 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T18:18:30Z","lastTransitionTime":"2025-10-08T18:18:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 18:18:30 crc kubenswrapper[4859]: I1008 18:18:30.470276 4859 scope.go:117] "RemoveContainer" containerID="c646f0e0dd8e7f74125b687f0cb4114b1f34667e5a6325f394f0208d7fb89df6" Oct 08 18:18:30 crc kubenswrapper[4859]: I1008 18:18:30.484776 4859 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4efc16fb-e679-464d-8b3c-0f46f082593e\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:17:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:17:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:17:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:17:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:17:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://aa8f8ea895272cd79a08d1c47cd7863a1ec10e21bb60d12aa10862a56c99e4bf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T18:17:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5c4e95d8a712ab219455564539a90286ae48bbc46ddc322a5bfd96c1a8268110\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T18:17:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://bae6ecfd8b446ef04d393d823e5ffe37fe749337c0ca17fde79c5f88d5e31393\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T18:17:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6411d1e7e18d5ceadaaceed5a8a6b20a0cacc55cede793ca9fb4dcad70c64fc6\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T18:17:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T18:17:10Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T18:18:30Z is after 2025-08-24T17:21:41Z" Oct 08 18:18:30 crc kubenswrapper[4859]: I1008 18:18:30.502269 4859 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-08T18:17:29Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T18:17:29Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T18:18:30Z is after 2025-08-24T17:21:41Z" Oct 08 18:18:30 crc kubenswrapper[4859]: I1008 18:18:30.520766 4859 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-84k8x" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1c24078c-c9e8-4160-883a-f483b4b3dd23\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:17:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:17:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:17:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:17:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8661006834d0f50eaada7394369584ee95fc61ff54437eec35d755526afc5b48\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T18:17:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hk5xb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://18b291eb3344fb7c00ef5679dca9a6f7ed9d08cea684a803ff7dad2e8a18c585\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://18b291eb3344fb7c00ef5679dca9a6f7ed9d08cea684a803ff7dad2e8a18c585\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-08T18:17:32Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-08T18:17:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hk5xb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://160163f1a91d981583452db3c7a9244182d1c8ced45749a61b721d0d45c1dfe4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://160163f1a91d981583452db3c7a9244182d1c8ced45749a61b721d0d45c1dfe4\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-08T18:17:34Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-08T18:17:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hk5xb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://51e9b53f3847ac53f378ad83870debf8e71518d4edbc7fc0dc66644818e2103e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://51e9b53f3847ac53f378ad83870debf8e71518d4edbc7fc0dc66644818e2103e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-08T18:17:34Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-08T18:17:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hk5xb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1ff118919739759ac01b5cbe3240e6673db54eaa16353c609d670071ed7964cb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1ff118919739759ac01b5cbe3240e6673db54eaa16353c609d670071ed7964cb\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-08T18:17:35Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-08T18:17:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hk5xb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8522e803ff56eb57cd4d9fd036e0c3fb16240ea21d4a094c1db7aa7d50bfe2c2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8522e803ff56eb57cd4d9fd036e0c3fb16240ea21d4a094c1db7aa7d50bfe2c2\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-08T18:17:37Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-08T18:17:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hk5xb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://be61056edf515e037403ca2a0bfd5feb9b2537dd42212b58e133b23b294f327b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://be61056edf515e037403ca2a0bfd5feb9b2537dd42212b58e133b23b294f327b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-08T18:17:38Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-08T18:17:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hk5xb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T18:17:31Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-84k8x\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T18:18:30Z is after 2025-08-24T17:21:41Z" Oct 08 18:18:30 crc kubenswrapper[4859]: I1008 18:18:30.533281 4859 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3ca6ce76-77e8-4e46-a2f6-b1ed4aa0fd4b\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:17:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:17:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:18:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:18:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:17:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f587e149d57ea53626ed3beb83e421c3b89ba1fec9e949705e8ccb5b4f7cc79a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T18:17:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://428cbe63daefbfbcb2ec76dad0fe638cdc65934bac80b791cebe0b36b2d6909a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T18:17:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a41399a37b1ad86a063001224232ecb8f7ea56aeb5686d3858bc08102a0c8f49\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T18:17:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://008757fce552591f7c58193f605eb60792761d5943927cbd63184549618d2b43\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://008757fce552591f7c58193f605eb60792761d5943927cbd63184549618d2b43\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-08T18:17:11Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-08T18:17:11Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T18:17:10Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T18:18:30Z is after 2025-08-24T17:21:41Z" Oct 08 18:18:30 crc kubenswrapper[4859]: I1008 18:18:30.546525 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 18:18:30 crc kubenswrapper[4859]: I1008 18:18:30.546583 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 18:18:30 crc kubenswrapper[4859]: I1008 18:18:30.546594 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 18:18:30 crc kubenswrapper[4859]: I1008 18:18:30.546612 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 18:18:30 crc kubenswrapper[4859]: I1008 18:18:30.546623 4859 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T18:18:30Z","lastTransitionTime":"2025-10-08T18:18:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 18:18:30 crc kubenswrapper[4859]: I1008 18:18:30.547028 4859 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-08T18:17:29Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T18:17:29Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T18:18:30Z is after 2025-08-24T17:21:41Z" Oct 08 18:18:30 crc kubenswrapper[4859]: I1008 18:18:30.558536 4859 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-btw28" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"97bc5598-3ead-4300-a938-3aebbff1e8c6\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:17:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:17:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:17:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:17:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d6d4d6ba771cb5fde637449bba05660f634df7e833a0c8dad2b2c7f3a422bb2b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T18:17:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qcxsd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T18:17:31Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-btw28\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T18:18:30Z is after 2025-08-24T17:21:41Z" Oct 08 18:18:30 crc kubenswrapper[4859]: I1008 18:18:30.576122 4859 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-8plkg" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"92b2d7de-31cc-4a91-95a5-ed5ea964b028\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:17:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:17:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:18:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:18:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://729bd32fd01dcab495bffa64f2f5e44990fc1d78b05e7f131868009f92305831\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://fa00b8345d40b74a917ddc23cc1b2c5c684a9ca61b6119d090c92cc1439b71c4\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-10-08T18:18:19Z\\\",\\\"message\\\":\\\"2025-10-08T18:17:33+00:00 [cnibincopy] Successfully copied files in /usr/src/multus-cni/rhel9/bin/ to /host/opt/cni/bin/upgrade_888b130f-8754-48c6-b73e-9ff6772793f9\\\\n2025-10-08T18:17:33+00:00 [cnibincopy] Successfully moved files in /host/opt/cni/bin/upgrade_888b130f-8754-48c6-b73e-9ff6772793f9 to /host/opt/cni/bin/\\\\n2025-10-08T18:17:34Z [verbose] multus-daemon started\\\\n2025-10-08T18:17:34Z [verbose] Readiness Indicator file check\\\\n2025-10-08T18:18:19Z [error] have you checked that your default network is ready? still waiting for readinessindicatorfile @ /host/run/multus/cni/net.d/10-ovn-kubernetes.conf. pollimmediate error: timed out waiting for the condition\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-08T18:17:32Z\\\"}},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T18:18:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q9d7m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T18:17:31Z\\\"}}\" for pod \"openshift-multus\"/\"multus-8plkg\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T18:18:30Z is after 2025-08-24T17:21:41Z" Oct 08 18:18:30 crc kubenswrapper[4859]: I1008 18:18:30.586015 4859 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-btjmq" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b8605f7-fed7-493f-9b4f-6a6358adf907\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:17:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:17:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:17:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:17:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://14e9b6c4a4403f36adb78fcf1135a620ca6fc56ed974c9639dcdfc59e6917559\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T18:17:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xspg4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T18:17:35Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-btjmq\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T18:18:30Z is after 2025-08-24T17:21:41Z" Oct 08 18:18:30 crc kubenswrapper[4859]: I1008 18:18:30.595532 4859 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-g4t79" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3a92dbf1-21a8-4095-96bb-2d701d7f3d56\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:17:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:17:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:17:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:17:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f8772a25b6e605efe9a062dc32dbbc2cfe32b4fe164b4e84fbf3bcffab744179\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T18:17:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sv79p\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://92cb9986c09a7f9e1e552e3e3ac1f74f04128241e00ddee38882d2ba5c005a67\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T18:17:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sv79p\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T18:17:45Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-g4t79\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T18:18:30Z is after 2025-08-24T17:21:41Z" Oct 08 18:18:30 crc kubenswrapper[4859]: I1008 18:18:30.608820 4859 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-08T18:17:31Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T18:17:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T18:17:31Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c6d15905694b6bbd4999d49dc8f7a53eb6bfaa7adb59efb0d7e4973c21d1c7be\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T18:17:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T18:18:30Z is after 2025-08-24T17:21:41Z" Oct 08 18:18:30 crc kubenswrapper[4859]: I1008 18:18:30.621471 4859 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-08T18:17:33Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T18:17:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T18:17:33Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5d99e071bd083914418d9c128f38955f26295c08b1520f2dc774c99b1c843693\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T18:17:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T18:18:30Z is after 2025-08-24T17:21:41Z" Oct 08 18:18:30 crc kubenswrapper[4859]: I1008 18:18:30.641910 4859 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-4g8gf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1dff864e-b75d-4e0b-b182-75f710eac8df\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:17:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:17:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:17:32Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:17:32Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://768ca274dedcc6847fc21d15234d3665fa9ccae371b5b6a882cfcb9b182f527d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T18:17:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s8jnn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://669f5f3c6d3c1ec0c4b4870e090b0a853adc01727c23383e595715b476c32668\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T18:17:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s8jnn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b8009a42cc4c9a816457a0678bdb2f7b7c53d0c8604e007a6fb141f66d06a516\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T18:17:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s8jnn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://676511a11ed4102084b57ebdd5f3fda4969741b2bc8443b415c4bf94f39ae4c9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T18:17:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s8jnn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4a74c1f0d21c90cdd7e776ee14e1f48757a04e2220fbd1f884449181ec6d1c3a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T18:17:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s8jnn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6ddc0e842bb8f81663406070023b0e5ed9cc99c98db0123b1c84f20da59fb5b5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T18:17:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s8jnn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c646f0e0dd8e7f74125b687f0cb4114b1f34667e5a6325f394f0208d7fb89df6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c646f0e0dd8e7f74125b687f0cb4114b1f34667e5a6325f394f0208d7fb89df6\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-10-08T18:17:58Z\\\",\\\"message\\\":\\\"e *v1.Pod retry setup to complete in iterateRetryResources\\\\nI1008 18:17:58.424504 6520 obj_retry.go:365] Adding new object: *v1.Pod openshift-machine-config-operator/machine-config-daemon-82s52\\\\nI1008 18:17:58.424529 6520 ovn.go:134] Ensuring zone local for Pod openshift-machine-config-operator/machine-config-daemon-82s52 in node crc\\\\nI1008 18:17:58.424539 6520 obj_retry.go:303] Retry object setup: *v1.Pod openshift-network-diagnostics/network-check-source-55646444c4-trplf\\\\nI1008 18:17:58.424549 6520 obj_retry.go:365] Adding new object: *v1.Pod openshift-network-diagnostics/network-check-source-55646444c4-trplf\\\\nI1008 18:17:58.424549 6520 obj_retry.go:386] Retry successful for *v1.Pod openshift-machine-config-operator/machine-config-daemon-82s52 after 0 failed attempt(s)\\\\nI1008 18:17:58.424559 6520 ovn.go:134] Ensuring zone local for Pod openshift-network-diagnostics/network-check-source-55646444c4-trplf in node crc\\\\nF1008 18:17:58.424552 6520 ovnkube.go:137] failed to run ovnkube: [failed to start network controller: failed to start default network controller: unable to create admin network policy controller, err: could not add Event Handler for anpInformer during admin network policy controller initialization, handler {0x1fcc6e0 0x1fcc3c0 0x1fcc360} was not added to shared informer because it has stopped already, failed to star\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-08T18:17:57Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 20s restarting failed container=ovnkube-controller pod=ovnkube-node-4g8gf_openshift-ovn-kubernetes(1dff864e-b75d-4e0b-b182-75f710eac8df)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s8jnn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2c67d2aed7126d76989e057d8e2d8e188a8902b2f3fe26ed6d1d45d3b013d897\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T18:17:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s8jnn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7a211d3da02b173867ff3a912eecd3b379483c7cc77e6b91122edd1ffd30fd53\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7a211d3da02b173867ff3a912eecd3b379483c7cc77e6b91122edd1ffd30fd53\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-08T18:17:32Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-08T18:17:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s8jnn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T18:17:32Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-4g8gf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T18:18:30Z is after 2025-08-24T17:21:41Z" Oct 08 18:18:30 crc kubenswrapper[4859]: I1008 18:18:30.650028 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 18:18:30 crc kubenswrapper[4859]: I1008 18:18:30.650088 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 18:18:30 crc kubenswrapper[4859]: I1008 18:18:30.650107 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 18:18:30 crc kubenswrapper[4859]: I1008 18:18:30.650129 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 18:18:30 crc kubenswrapper[4859]: I1008 18:18:30.650144 4859 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T18:18:30Z","lastTransitionTime":"2025-10-08T18:18:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 18:18:30 crc kubenswrapper[4859]: I1008 18:18:30.655676 4859 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-4b4sz" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2e53cd9b-64f7-4e07-8b96-fb77847c9ac6\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:17:45Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:17:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:17:45Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:17:45Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kxgbz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kxgbz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T18:17:45Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-4b4sz\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T18:18:30Z is after 2025-08-24T17:21:41Z" Oct 08 18:18:30 crc kubenswrapper[4859]: I1008 18:18:30.669899 4859 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"262c958c-c49d-4c11-96ae-94106dd123d4\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:17:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:17:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:17:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:17:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:17:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a54a54b9a831ead1fbe6e03abbb622242dd16b4081a4482d559e03d47aebd2fe\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-crio\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T18:17:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kube\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3ab3bc5e8e2fd4941f30cbf8a90d9a9078b90fab91063c07da2cd48b63eef657\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3ab3bc5e8e2fd4941f30cbf8a90d9a9078b90fab91063c07da2cd48b63eef657\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-08T18:17:11Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-08T18:17:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T18:17:10Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"kube-rbac-proxy-crio-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T18:18:30Z is after 2025-08-24T17:21:41Z" Oct 08 18:18:30 crc kubenswrapper[4859]: I1008 18:18:30.687247 4859 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"63a623de-5608-4228-b1c2-77188c276986\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:17:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:17:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:17:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:17:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:17:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d4786a8fb8abd96fb22e60733c974f599017407d589ed815fb8541fd3b64f696\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T18:17:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://183d481502ed5108d5e19e483b714c289f355ea3f8b0891883e5335b802c83b6\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T18:17:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ec848c46fd5dde0469b00d3cab14ba4a652533a265490fadbee7a8b8843ef249\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T18:17:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4fe1828b4d63ff8106bfcd15e1dc1c1bd1ba7f77777cdb23ff1b291d0418774c\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2e700b2a20c81b414617f5b5c018a738278a3fa44e6c6d14b2170051d0725431\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-08T18:17:24Z\\\",\\\"message\\\":\\\"W1008 18:17:13.670625 1 cmd.go:257] Using insecure, self-signed certificates\\\\nI1008 18:17:13.671202 1 crypto.go:601] Generating new CA for check-endpoints-signer@1759947433 cert, and key in /tmp/serving-cert-3867218267/serving-signer.crt, /tmp/serving-cert-3867218267/serving-signer.key\\\\nI1008 18:17:13.944463 1 observer_polling.go:159] Starting file observer\\\\nW1008 18:17:13.951958 1 builder.go:272] unable to get owner reference (falling back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": dial tcp [::1]:6443: connect: connection refused\\\\nI1008 18:17:13.952088 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1008 18:17:13.956755 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3867218267/tls.crt::/tmp/serving-cert-3867218267/tls.key\\\\\\\"\\\\nF1008 18:17:24.290991 1 cmd.go:182] error initializing delegating authentication: unable to load configmap based request-header-client-ca-file: Get \\\\\\\"https://localhost:6443/api/v1/namespaces/kube-system/configmaps/extension-apiserver-authentication\\\\\\\": net/http: TLS handshake timeout\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-08T18:17:13Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T18:17:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c65b66f62d325ca10ef8d5d31f04529ccf4e691cf3303d231e7a414fb400e4d6\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T18:17:13Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://58c7bba4ace28215b062653a960e561b1dc1f890f2baf450145638a065af038b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://58c7bba4ace28215b062653a960e561b1dc1f890f2baf450145638a065af038b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-08T18:17:11Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-08T18:17:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T18:17:10Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T18:18:30Z is after 2025-08-24T17:21:41Z" Oct 08 18:18:30 crc kubenswrapper[4859]: I1008 18:18:30.701859 4859 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-08T18:17:31Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T18:17:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T18:17:31Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://99e2d51bd7757498dae60949eea6d732b5f7b481fa00f808dd5308a530a1daf1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T18:17:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e32aaba93699944665660ae41e9b5779178c1354c31e15ad8188bf8c3c8fa500\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T18:17:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T18:18:30Z is after 2025-08-24T17:21:41Z" Oct 08 18:18:30 crc kubenswrapper[4859]: I1008 18:18:30.717971 4859 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-08T18:17:29Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T18:17:29Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T18:18:30Z is after 2025-08-24T17:21:41Z" Oct 08 18:18:30 crc kubenswrapper[4859]: I1008 18:18:30.730836 4859 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-82s52" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b23a6a6c-9d92-4e7b-840e-55cfda873a2d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:17:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:17:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:17:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:17:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://71302c766485b9f55743d85f97440ff578cd44ba1e5a6f5b1c024c0e808072d2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T18:17:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6vx22\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c23ea3580998adf90621a0155c20ee4bd590c512629812174e95c28857bcfef9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T18:17:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6vx22\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T18:17:31Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-82s52\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T18:18:30Z is after 2025-08-24T17:21:41Z" Oct 08 18:18:30 crc kubenswrapper[4859]: I1008 18:18:30.752667 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 18:18:30 crc kubenswrapper[4859]: I1008 18:18:30.752750 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 18:18:30 crc kubenswrapper[4859]: I1008 18:18:30.752767 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 18:18:30 crc kubenswrapper[4859]: I1008 18:18:30.752792 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 18:18:30 crc kubenswrapper[4859]: I1008 18:18:30.752812 4859 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T18:18:30Z","lastTransitionTime":"2025-10-08T18:18:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 18:18:30 crc kubenswrapper[4859]: I1008 18:18:30.855091 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 18:18:30 crc kubenswrapper[4859]: I1008 18:18:30.855131 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 18:18:30 crc kubenswrapper[4859]: I1008 18:18:30.855142 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 18:18:30 crc kubenswrapper[4859]: I1008 18:18:30.855159 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 18:18:30 crc kubenswrapper[4859]: I1008 18:18:30.855169 4859 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T18:18:30Z","lastTransitionTime":"2025-10-08T18:18:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 18:18:30 crc kubenswrapper[4859]: I1008 18:18:30.959708 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 18:18:30 crc kubenswrapper[4859]: I1008 18:18:30.959752 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 18:18:30 crc kubenswrapper[4859]: I1008 18:18:30.959764 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 18:18:30 crc kubenswrapper[4859]: I1008 18:18:30.959785 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 18:18:30 crc kubenswrapper[4859]: I1008 18:18:30.959801 4859 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T18:18:30Z","lastTransitionTime":"2025-10-08T18:18:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 18:18:31 crc kubenswrapper[4859]: I1008 18:18:31.013948 4859 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-4g8gf_1dff864e-b75d-4e0b-b182-75f710eac8df/ovnkube-controller/2.log" Oct 08 18:18:31 crc kubenswrapper[4859]: I1008 18:18:31.021667 4859 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-4g8gf" event={"ID":"1dff864e-b75d-4e0b-b182-75f710eac8df","Type":"ContainerStarted","Data":"f3a760c557ef21fbc6b54e78c3f6c5b325f296f93a022425cca7b1399eed8661"} Oct 08 18:18:31 crc kubenswrapper[4859]: I1008 18:18:31.022278 4859 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ovn-kubernetes/ovnkube-node-4g8gf" Oct 08 18:18:31 crc kubenswrapper[4859]: I1008 18:18:31.040992 4859 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4efc16fb-e679-464d-8b3c-0f46f082593e\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:17:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:17:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:17:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:17:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:17:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://aa8f8ea895272cd79a08d1c47cd7863a1ec10e21bb60d12aa10862a56c99e4bf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T18:17:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5c4e95d8a712ab219455564539a90286ae48bbc46ddc322a5bfd96c1a8268110\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T18:17:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://bae6ecfd8b446ef04d393d823e5ffe37fe749337c0ca17fde79c5f88d5e31393\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T18:17:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6411d1e7e18d5ceadaaceed5a8a6b20a0cacc55cede793ca9fb4dcad70c64fc6\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T18:17:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T18:17:10Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T18:18:31Z is after 2025-08-24T17:21:41Z" Oct 08 18:18:31 crc kubenswrapper[4859]: I1008 18:18:31.062103 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 18:18:31 crc kubenswrapper[4859]: I1008 18:18:31.062154 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 18:18:31 crc kubenswrapper[4859]: I1008 18:18:31.062164 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 18:18:31 crc kubenswrapper[4859]: I1008 18:18:31.062184 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 18:18:31 crc kubenswrapper[4859]: I1008 18:18:31.062198 4859 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T18:18:31Z","lastTransitionTime":"2025-10-08T18:18:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 18:18:31 crc kubenswrapper[4859]: I1008 18:18:31.062958 4859 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-08T18:17:29Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T18:17:29Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T18:18:31Z is after 2025-08-24T17:21:41Z" Oct 08 18:18:31 crc kubenswrapper[4859]: I1008 18:18:31.086586 4859 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-84k8x" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1c24078c-c9e8-4160-883a-f483b4b3dd23\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:17:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:17:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:17:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:17:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8661006834d0f50eaada7394369584ee95fc61ff54437eec35d755526afc5b48\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T18:17:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hk5xb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://18b291eb3344fb7c00ef5679dca9a6f7ed9d08cea684a803ff7dad2e8a18c585\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://18b291eb3344fb7c00ef5679dca9a6f7ed9d08cea684a803ff7dad2e8a18c585\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-08T18:17:32Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-08T18:17:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hk5xb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://160163f1a91d981583452db3c7a9244182d1c8ced45749a61b721d0d45c1dfe4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://160163f1a91d981583452db3c7a9244182d1c8ced45749a61b721d0d45c1dfe4\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-08T18:17:34Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-08T18:17:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hk5xb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://51e9b53f3847ac53f378ad83870debf8e71518d4edbc7fc0dc66644818e2103e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://51e9b53f3847ac53f378ad83870debf8e71518d4edbc7fc0dc66644818e2103e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-08T18:17:34Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-08T18:17:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hk5xb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1ff118919739759ac01b5cbe3240e6673db54eaa16353c609d670071ed7964cb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1ff118919739759ac01b5cbe3240e6673db54eaa16353c609d670071ed7964cb\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-08T18:17:35Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-08T18:17:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hk5xb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8522e803ff56eb57cd4d9fd036e0c3fb16240ea21d4a094c1db7aa7d50bfe2c2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8522e803ff56eb57cd4d9fd036e0c3fb16240ea21d4a094c1db7aa7d50bfe2c2\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-08T18:17:37Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-08T18:17:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hk5xb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://be61056edf515e037403ca2a0bfd5feb9b2537dd42212b58e133b23b294f327b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://be61056edf515e037403ca2a0bfd5feb9b2537dd42212b58e133b23b294f327b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-08T18:17:38Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-08T18:17:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hk5xb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T18:17:31Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-84k8x\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T18:18:31Z is after 2025-08-24T17:21:41Z" Oct 08 18:18:31 crc kubenswrapper[4859]: I1008 18:18:31.108561 4859 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3ca6ce76-77e8-4e46-a2f6-b1ed4aa0fd4b\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:17:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:17:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:18:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:18:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:17:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f587e149d57ea53626ed3beb83e421c3b89ba1fec9e949705e8ccb5b4f7cc79a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T18:17:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://428cbe63daefbfbcb2ec76dad0fe638cdc65934bac80b791cebe0b36b2d6909a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T18:17:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a41399a37b1ad86a063001224232ecb8f7ea56aeb5686d3858bc08102a0c8f49\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T18:17:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://008757fce552591f7c58193f605eb60792761d5943927cbd63184549618d2b43\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://008757fce552591f7c58193f605eb60792761d5943927cbd63184549618d2b43\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-08T18:17:11Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-08T18:17:11Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T18:17:10Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T18:18:31Z is after 2025-08-24T17:21:41Z" Oct 08 18:18:31 crc kubenswrapper[4859]: I1008 18:18:31.124504 4859 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-08T18:17:29Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T18:17:29Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T18:18:31Z is after 2025-08-24T17:21:41Z" Oct 08 18:18:31 crc kubenswrapper[4859]: I1008 18:18:31.138008 4859 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-btw28" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"97bc5598-3ead-4300-a938-3aebbff1e8c6\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:17:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:17:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:17:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:17:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d6d4d6ba771cb5fde637449bba05660f634df7e833a0c8dad2b2c7f3a422bb2b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T18:17:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qcxsd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T18:17:31Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-btw28\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T18:18:31Z is after 2025-08-24T17:21:41Z" Oct 08 18:18:31 crc kubenswrapper[4859]: I1008 18:18:31.153203 4859 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-8plkg" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"92b2d7de-31cc-4a91-95a5-ed5ea964b028\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:17:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:17:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:18:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:18:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://729bd32fd01dcab495bffa64f2f5e44990fc1d78b05e7f131868009f92305831\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://fa00b8345d40b74a917ddc23cc1b2c5c684a9ca61b6119d090c92cc1439b71c4\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-10-08T18:18:19Z\\\",\\\"message\\\":\\\"2025-10-08T18:17:33+00:00 [cnibincopy] Successfully copied files in /usr/src/multus-cni/rhel9/bin/ to /host/opt/cni/bin/upgrade_888b130f-8754-48c6-b73e-9ff6772793f9\\\\n2025-10-08T18:17:33+00:00 [cnibincopy] Successfully moved files in /host/opt/cni/bin/upgrade_888b130f-8754-48c6-b73e-9ff6772793f9 to /host/opt/cni/bin/\\\\n2025-10-08T18:17:34Z [verbose] multus-daemon started\\\\n2025-10-08T18:17:34Z [verbose] Readiness Indicator file check\\\\n2025-10-08T18:18:19Z [error] have you checked that your default network is ready? still waiting for readinessindicatorfile @ /host/run/multus/cni/net.d/10-ovn-kubernetes.conf. pollimmediate error: timed out waiting for the condition\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-08T18:17:32Z\\\"}},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T18:18:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q9d7m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T18:17:31Z\\\"}}\" for pod \"openshift-multus\"/\"multus-8plkg\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T18:18:31Z is after 2025-08-24T17:21:41Z" Oct 08 18:18:31 crc kubenswrapper[4859]: I1008 18:18:31.165661 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 18:18:31 crc kubenswrapper[4859]: I1008 18:18:31.165718 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 18:18:31 crc kubenswrapper[4859]: I1008 18:18:31.165730 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 18:18:31 crc kubenswrapper[4859]: I1008 18:18:31.165745 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 18:18:31 crc kubenswrapper[4859]: I1008 18:18:31.165758 4859 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T18:18:31Z","lastTransitionTime":"2025-10-08T18:18:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 18:18:31 crc kubenswrapper[4859]: I1008 18:18:31.245858 4859 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-btjmq" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b8605f7-fed7-493f-9b4f-6a6358adf907\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:17:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:17:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:17:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:17:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://14e9b6c4a4403f36adb78fcf1135a620ca6fc56ed974c9639dcdfc59e6917559\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T18:17:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xspg4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T18:17:35Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-btjmq\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T18:18:31Z is after 2025-08-24T17:21:41Z" Oct 08 18:18:31 crc kubenswrapper[4859]: I1008 18:18:31.261247 4859 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-g4t79" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3a92dbf1-21a8-4095-96bb-2d701d7f3d56\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:17:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:17:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:17:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:17:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f8772a25b6e605efe9a062dc32dbbc2cfe32b4fe164b4e84fbf3bcffab744179\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T18:17:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sv79p\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://92cb9986c09a7f9e1e552e3e3ac1f74f04128241e00ddee38882d2ba5c005a67\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T18:17:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sv79p\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T18:17:45Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-g4t79\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T18:18:31Z is after 2025-08-24T17:21:41Z" Oct 08 18:18:31 crc kubenswrapper[4859]: I1008 18:18:31.268849 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 18:18:31 crc kubenswrapper[4859]: I1008 18:18:31.268908 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 18:18:31 crc kubenswrapper[4859]: I1008 18:18:31.268929 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 18:18:31 crc kubenswrapper[4859]: I1008 18:18:31.268956 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 18:18:31 crc kubenswrapper[4859]: I1008 18:18:31.268975 4859 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T18:18:31Z","lastTransitionTime":"2025-10-08T18:18:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 18:18:31 crc kubenswrapper[4859]: I1008 18:18:31.276331 4859 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-08T18:17:31Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T18:17:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T18:17:31Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c6d15905694b6bbd4999d49dc8f7a53eb6bfaa7adb59efb0d7e4973c21d1c7be\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T18:17:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T18:18:31Z is after 2025-08-24T17:21:41Z" Oct 08 18:18:31 crc kubenswrapper[4859]: I1008 18:18:31.293993 4859 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-08T18:17:33Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T18:17:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T18:17:33Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5d99e071bd083914418d9c128f38955f26295c08b1520f2dc774c99b1c843693\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T18:17:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T18:18:31Z is after 2025-08-24T17:21:41Z" Oct 08 18:18:31 crc kubenswrapper[4859]: I1008 18:18:31.311062 4859 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-4g8gf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1dff864e-b75d-4e0b-b182-75f710eac8df\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:17:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:17:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:17:32Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:17:32Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://768ca274dedcc6847fc21d15234d3665fa9ccae371b5b6a882cfcb9b182f527d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T18:17:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s8jnn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://669f5f3c6d3c1ec0c4b4870e090b0a853adc01727c23383e595715b476c32668\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T18:17:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s8jnn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b8009a42cc4c9a816457a0678bdb2f7b7c53d0c8604e007a6fb141f66d06a516\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T18:17:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s8jnn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://676511a11ed4102084b57ebdd5f3fda4969741b2bc8443b415c4bf94f39ae4c9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T18:17:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s8jnn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4a74c1f0d21c90cdd7e776ee14e1f48757a04e2220fbd1f884449181ec6d1c3a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T18:17:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s8jnn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6ddc0e842bb8f81663406070023b0e5ed9cc99c98db0123b1c84f20da59fb5b5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T18:17:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s8jnn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f3a760c557ef21fbc6b54e78c3f6c5b325f296f93a022425cca7b1399eed8661\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c646f0e0dd8e7f74125b687f0cb4114b1f34667e5a6325f394f0208d7fb89df6\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-10-08T18:17:58Z\\\",\\\"message\\\":\\\"e *v1.Pod retry setup to complete in iterateRetryResources\\\\nI1008 18:17:58.424504 6520 obj_retry.go:365] Adding new object: *v1.Pod openshift-machine-config-operator/machine-config-daemon-82s52\\\\nI1008 18:17:58.424529 6520 ovn.go:134] Ensuring zone local for Pod openshift-machine-config-operator/machine-config-daemon-82s52 in node crc\\\\nI1008 18:17:58.424539 6520 obj_retry.go:303] Retry object setup: *v1.Pod openshift-network-diagnostics/network-check-source-55646444c4-trplf\\\\nI1008 18:17:58.424549 6520 obj_retry.go:365] Adding new object: *v1.Pod openshift-network-diagnostics/network-check-source-55646444c4-trplf\\\\nI1008 18:17:58.424549 6520 obj_retry.go:386] Retry successful for *v1.Pod openshift-machine-config-operator/machine-config-daemon-82s52 after 0 failed attempt(s)\\\\nI1008 18:17:58.424559 6520 ovn.go:134] Ensuring zone local for Pod openshift-network-diagnostics/network-check-source-55646444c4-trplf in node crc\\\\nF1008 18:17:58.424552 6520 ovnkube.go:137] failed to run ovnkube: [failed to start network controller: failed to start default network controller: unable to create admin network policy controller, err: could not add Event Handler for anpInformer during admin network policy controller initialization, handler {0x1fcc6e0 0x1fcc3c0 0x1fcc360} was not added to shared informer because it has stopped already, failed to star\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-08T18:17:57Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T18:18:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s8jnn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2c67d2aed7126d76989e057d8e2d8e188a8902b2f3fe26ed6d1d45d3b013d897\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T18:17:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s8jnn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7a211d3da02b173867ff3a912eecd3b379483c7cc77e6b91122edd1ffd30fd53\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7a211d3da02b173867ff3a912eecd3b379483c7cc77e6b91122edd1ffd30fd53\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-08T18:17:32Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-08T18:17:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s8jnn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T18:17:32Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-4g8gf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T18:18:31Z is after 2025-08-24T17:21:41Z" Oct 08 18:18:31 crc kubenswrapper[4859]: I1008 18:18:31.325213 4859 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-4b4sz" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2e53cd9b-64f7-4e07-8b96-fb77847c9ac6\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:17:45Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:17:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:17:45Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:17:45Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kxgbz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kxgbz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T18:17:45Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-4b4sz\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T18:18:31Z is after 2025-08-24T17:21:41Z" Oct 08 18:18:31 crc kubenswrapper[4859]: I1008 18:18:31.338713 4859 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"262c958c-c49d-4c11-96ae-94106dd123d4\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:17:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:17:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:17:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:17:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:17:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a54a54b9a831ead1fbe6e03abbb622242dd16b4081a4482d559e03d47aebd2fe\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-crio\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T18:17:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kube\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3ab3bc5e8e2fd4941f30cbf8a90d9a9078b90fab91063c07da2cd48b63eef657\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3ab3bc5e8e2fd4941f30cbf8a90d9a9078b90fab91063c07da2cd48b63eef657\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-08T18:17:11Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-08T18:17:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T18:17:10Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"kube-rbac-proxy-crio-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T18:18:31Z is after 2025-08-24T17:21:41Z" Oct 08 18:18:31 crc kubenswrapper[4859]: I1008 18:18:31.353819 4859 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"63a623de-5608-4228-b1c2-77188c276986\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:17:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:17:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:17:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:17:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:17:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d4786a8fb8abd96fb22e60733c974f599017407d589ed815fb8541fd3b64f696\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T18:17:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://183d481502ed5108d5e19e483b714c289f355ea3f8b0891883e5335b802c83b6\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T18:17:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ec848c46fd5dde0469b00d3cab14ba4a652533a265490fadbee7a8b8843ef249\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T18:17:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4fe1828b4d63ff8106bfcd15e1dc1c1bd1ba7f77777cdb23ff1b291d0418774c\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2e700b2a20c81b414617f5b5c018a738278a3fa44e6c6d14b2170051d0725431\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-08T18:17:24Z\\\",\\\"message\\\":\\\"W1008 18:17:13.670625 1 cmd.go:257] Using insecure, self-signed certificates\\\\nI1008 18:17:13.671202 1 crypto.go:601] Generating new CA for check-endpoints-signer@1759947433 cert, and key in /tmp/serving-cert-3867218267/serving-signer.crt, /tmp/serving-cert-3867218267/serving-signer.key\\\\nI1008 18:17:13.944463 1 observer_polling.go:159] Starting file observer\\\\nW1008 18:17:13.951958 1 builder.go:272] unable to get owner reference (falling back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": dial tcp [::1]:6443: connect: connection refused\\\\nI1008 18:17:13.952088 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1008 18:17:13.956755 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3867218267/tls.crt::/tmp/serving-cert-3867218267/tls.key\\\\\\\"\\\\nF1008 18:17:24.290991 1 cmd.go:182] error initializing delegating authentication: unable to load configmap based request-header-client-ca-file: Get \\\\\\\"https://localhost:6443/api/v1/namespaces/kube-system/configmaps/extension-apiserver-authentication\\\\\\\": net/http: TLS handshake timeout\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-08T18:17:13Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T18:17:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c65b66f62d325ca10ef8d5d31f04529ccf4e691cf3303d231e7a414fb400e4d6\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T18:17:13Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://58c7bba4ace28215b062653a960e561b1dc1f890f2baf450145638a065af038b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://58c7bba4ace28215b062653a960e561b1dc1f890f2baf450145638a065af038b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-08T18:17:11Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-08T18:17:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T18:17:10Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T18:18:31Z is after 2025-08-24T17:21:41Z" Oct 08 18:18:31 crc kubenswrapper[4859]: I1008 18:18:31.368318 4859 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-08T18:17:31Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T18:17:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T18:17:31Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://99e2d51bd7757498dae60949eea6d732b5f7b481fa00f808dd5308a530a1daf1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T18:17:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e32aaba93699944665660ae41e9b5779178c1354c31e15ad8188bf8c3c8fa500\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T18:17:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T18:18:31Z is after 2025-08-24T17:21:41Z" Oct 08 18:18:31 crc kubenswrapper[4859]: I1008 18:18:31.371195 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 18:18:31 crc kubenswrapper[4859]: I1008 18:18:31.371238 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 18:18:31 crc kubenswrapper[4859]: I1008 18:18:31.371250 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 18:18:31 crc kubenswrapper[4859]: I1008 18:18:31.371269 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 18:18:31 crc kubenswrapper[4859]: I1008 18:18:31.371282 4859 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T18:18:31Z","lastTransitionTime":"2025-10-08T18:18:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 18:18:31 crc kubenswrapper[4859]: I1008 18:18:31.382619 4859 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-08T18:17:29Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T18:17:29Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T18:18:31Z is after 2025-08-24T17:21:41Z" Oct 08 18:18:31 crc kubenswrapper[4859]: I1008 18:18:31.395876 4859 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-82s52" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b23a6a6c-9d92-4e7b-840e-55cfda873a2d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:17:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:17:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:17:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:17:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://71302c766485b9f55743d85f97440ff578cd44ba1e5a6f5b1c024c0e808072d2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T18:17:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6vx22\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c23ea3580998adf90621a0155c20ee4bd590c512629812174e95c28857bcfef9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T18:17:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6vx22\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T18:17:31Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-82s52\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T18:18:31Z is after 2025-08-24T17:21:41Z" Oct 08 18:18:31 crc kubenswrapper[4859]: I1008 18:18:31.468965 4859 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-4b4sz" Oct 08 18:18:31 crc kubenswrapper[4859]: E1008 18:18:31.469100 4859 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-4b4sz" podUID="2e53cd9b-64f7-4e07-8b96-fb77847c9ac6" Oct 08 18:18:31 crc kubenswrapper[4859]: I1008 18:18:31.469270 4859 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 08 18:18:31 crc kubenswrapper[4859]: E1008 18:18:31.469315 4859 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 08 18:18:31 crc kubenswrapper[4859]: I1008 18:18:31.469421 4859 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 08 18:18:31 crc kubenswrapper[4859]: E1008 18:18:31.469462 4859 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 08 18:18:31 crc kubenswrapper[4859]: I1008 18:18:31.469561 4859 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 08 18:18:31 crc kubenswrapper[4859]: E1008 18:18:31.469601 4859 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 08 18:18:31 crc kubenswrapper[4859]: I1008 18:18:31.473490 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 18:18:31 crc kubenswrapper[4859]: I1008 18:18:31.473557 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 18:18:31 crc kubenswrapper[4859]: I1008 18:18:31.473572 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 18:18:31 crc kubenswrapper[4859]: I1008 18:18:31.473591 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 18:18:31 crc kubenswrapper[4859]: I1008 18:18:31.473604 4859 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T18:18:31Z","lastTransitionTime":"2025-10-08T18:18:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 18:18:31 crc kubenswrapper[4859]: I1008 18:18:31.576504 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 18:18:31 crc kubenswrapper[4859]: I1008 18:18:31.576558 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 18:18:31 crc kubenswrapper[4859]: I1008 18:18:31.576572 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 18:18:31 crc kubenswrapper[4859]: I1008 18:18:31.576595 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 18:18:31 crc kubenswrapper[4859]: I1008 18:18:31.576609 4859 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T18:18:31Z","lastTransitionTime":"2025-10-08T18:18:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 18:18:31 crc kubenswrapper[4859]: I1008 18:18:31.679753 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 18:18:31 crc kubenswrapper[4859]: I1008 18:18:31.679794 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 18:18:31 crc kubenswrapper[4859]: I1008 18:18:31.679804 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 18:18:31 crc kubenswrapper[4859]: I1008 18:18:31.679821 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 18:18:31 crc kubenswrapper[4859]: I1008 18:18:31.679833 4859 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T18:18:31Z","lastTransitionTime":"2025-10-08T18:18:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 18:18:31 crc kubenswrapper[4859]: I1008 18:18:31.783087 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 18:18:31 crc kubenswrapper[4859]: I1008 18:18:31.783197 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 18:18:31 crc kubenswrapper[4859]: I1008 18:18:31.783221 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 18:18:31 crc kubenswrapper[4859]: I1008 18:18:31.783252 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 18:18:31 crc kubenswrapper[4859]: I1008 18:18:31.783271 4859 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T18:18:31Z","lastTransitionTime":"2025-10-08T18:18:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 18:18:31 crc kubenswrapper[4859]: I1008 18:18:31.886773 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 18:18:31 crc kubenswrapper[4859]: I1008 18:18:31.887124 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 18:18:31 crc kubenswrapper[4859]: I1008 18:18:31.887142 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 18:18:31 crc kubenswrapper[4859]: I1008 18:18:31.887169 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 18:18:31 crc kubenswrapper[4859]: I1008 18:18:31.887186 4859 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T18:18:31Z","lastTransitionTime":"2025-10-08T18:18:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 18:18:31 crc kubenswrapper[4859]: I1008 18:18:31.989898 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 18:18:31 crc kubenswrapper[4859]: I1008 18:18:31.989937 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 18:18:31 crc kubenswrapper[4859]: I1008 18:18:31.989948 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 18:18:31 crc kubenswrapper[4859]: I1008 18:18:31.989968 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 18:18:31 crc kubenswrapper[4859]: I1008 18:18:31.989981 4859 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T18:18:31Z","lastTransitionTime":"2025-10-08T18:18:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 18:18:32 crc kubenswrapper[4859]: I1008 18:18:32.026781 4859 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-4g8gf_1dff864e-b75d-4e0b-b182-75f710eac8df/ovnkube-controller/3.log" Oct 08 18:18:32 crc kubenswrapper[4859]: I1008 18:18:32.027501 4859 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-4g8gf_1dff864e-b75d-4e0b-b182-75f710eac8df/ovnkube-controller/2.log" Oct 08 18:18:32 crc kubenswrapper[4859]: I1008 18:18:32.030131 4859 generic.go:334] "Generic (PLEG): container finished" podID="1dff864e-b75d-4e0b-b182-75f710eac8df" containerID="f3a760c557ef21fbc6b54e78c3f6c5b325f296f93a022425cca7b1399eed8661" exitCode=1 Oct 08 18:18:32 crc kubenswrapper[4859]: I1008 18:18:32.030184 4859 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-4g8gf" event={"ID":"1dff864e-b75d-4e0b-b182-75f710eac8df","Type":"ContainerDied","Data":"f3a760c557ef21fbc6b54e78c3f6c5b325f296f93a022425cca7b1399eed8661"} Oct 08 18:18:32 crc kubenswrapper[4859]: I1008 18:18:32.030235 4859 scope.go:117] "RemoveContainer" containerID="c646f0e0dd8e7f74125b687f0cb4114b1f34667e5a6325f394f0208d7fb89df6" Oct 08 18:18:32 crc kubenswrapper[4859]: I1008 18:18:32.031448 4859 scope.go:117] "RemoveContainer" containerID="f3a760c557ef21fbc6b54e78c3f6c5b325f296f93a022425cca7b1399eed8661" Oct 08 18:18:32 crc kubenswrapper[4859]: E1008 18:18:32.031726 4859 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ovnkube-controller\" with CrashLoopBackOff: \"back-off 40s restarting failed container=ovnkube-controller pod=ovnkube-node-4g8gf_openshift-ovn-kubernetes(1dff864e-b75d-4e0b-b182-75f710eac8df)\"" pod="openshift-ovn-kubernetes/ovnkube-node-4g8gf" podUID="1dff864e-b75d-4e0b-b182-75f710eac8df" Oct 08 18:18:32 crc kubenswrapper[4859]: I1008 18:18:32.053567 4859 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-08T18:17:31Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T18:17:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T18:17:31Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c6d15905694b6bbd4999d49dc8f7a53eb6bfaa7adb59efb0d7e4973c21d1c7be\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T18:17:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T18:18:32Z is after 2025-08-24T17:21:41Z" Oct 08 18:18:32 crc kubenswrapper[4859]: I1008 18:18:32.071233 4859 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-08T18:17:33Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T18:17:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T18:17:33Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5d99e071bd083914418d9c128f38955f26295c08b1520f2dc774c99b1c843693\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T18:17:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T18:18:32Z is after 2025-08-24T17:21:41Z" Oct 08 18:18:32 crc kubenswrapper[4859]: I1008 18:18:32.084122 4859 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-4b4sz" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2e53cd9b-64f7-4e07-8b96-fb77847c9ac6\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:17:45Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:17:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:17:45Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:17:45Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kxgbz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kxgbz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T18:17:45Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-4b4sz\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T18:18:32Z is after 2025-08-24T17:21:41Z" Oct 08 18:18:32 crc kubenswrapper[4859]: I1008 18:18:32.092896 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 18:18:32 crc kubenswrapper[4859]: I1008 18:18:32.092947 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 18:18:32 crc kubenswrapper[4859]: I1008 18:18:32.092963 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 18:18:32 crc kubenswrapper[4859]: I1008 18:18:32.092986 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 18:18:32 crc kubenswrapper[4859]: I1008 18:18:32.093000 4859 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T18:18:32Z","lastTransitionTime":"2025-10-08T18:18:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 18:18:32 crc kubenswrapper[4859]: I1008 18:18:32.096953 4859 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"262c958c-c49d-4c11-96ae-94106dd123d4\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:17:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:17:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:17:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:17:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:17:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a54a54b9a831ead1fbe6e03abbb622242dd16b4081a4482d559e03d47aebd2fe\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-crio\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T18:17:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kube\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3ab3bc5e8e2fd4941f30cbf8a90d9a9078b90fab91063c07da2cd48b63eef657\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3ab3bc5e8e2fd4941f30cbf8a90d9a9078b90fab91063c07da2cd48b63eef657\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-08T18:17:11Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-08T18:17:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T18:17:10Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"kube-rbac-proxy-crio-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T18:18:32Z is after 2025-08-24T17:21:41Z" Oct 08 18:18:32 crc kubenswrapper[4859]: I1008 18:18:32.113440 4859 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"63a623de-5608-4228-b1c2-77188c276986\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:17:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:17:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:17:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:17:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:17:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d4786a8fb8abd96fb22e60733c974f599017407d589ed815fb8541fd3b64f696\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T18:17:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://183d481502ed5108d5e19e483b714c289f355ea3f8b0891883e5335b802c83b6\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T18:17:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ec848c46fd5dde0469b00d3cab14ba4a652533a265490fadbee7a8b8843ef249\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T18:17:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4fe1828b4d63ff8106bfcd15e1dc1c1bd1ba7f77777cdb23ff1b291d0418774c\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2e700b2a20c81b414617f5b5c018a738278a3fa44e6c6d14b2170051d0725431\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-08T18:17:24Z\\\",\\\"message\\\":\\\"W1008 18:17:13.670625 1 cmd.go:257] Using insecure, self-signed certificates\\\\nI1008 18:17:13.671202 1 crypto.go:601] Generating new CA for check-endpoints-signer@1759947433 cert, and key in /tmp/serving-cert-3867218267/serving-signer.crt, /tmp/serving-cert-3867218267/serving-signer.key\\\\nI1008 18:17:13.944463 1 observer_polling.go:159] Starting file observer\\\\nW1008 18:17:13.951958 1 builder.go:272] unable to get owner reference (falling back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": dial tcp [::1]:6443: connect: connection refused\\\\nI1008 18:17:13.952088 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1008 18:17:13.956755 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3867218267/tls.crt::/tmp/serving-cert-3867218267/tls.key\\\\\\\"\\\\nF1008 18:17:24.290991 1 cmd.go:182] error initializing delegating authentication: unable to load configmap based request-header-client-ca-file: Get \\\\\\\"https://localhost:6443/api/v1/namespaces/kube-system/configmaps/extension-apiserver-authentication\\\\\\\": net/http: TLS handshake timeout\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-08T18:17:13Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T18:17:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c65b66f62d325ca10ef8d5d31f04529ccf4e691cf3303d231e7a414fb400e4d6\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T18:17:13Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://58c7bba4ace28215b062653a960e561b1dc1f890f2baf450145638a065af038b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://58c7bba4ace28215b062653a960e561b1dc1f890f2baf450145638a065af038b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-08T18:17:11Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-08T18:17:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T18:17:10Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T18:18:32Z is after 2025-08-24T17:21:41Z" Oct 08 18:18:32 crc kubenswrapper[4859]: I1008 18:18:32.127037 4859 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-08T18:17:31Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T18:17:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T18:17:31Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://99e2d51bd7757498dae60949eea6d732b5f7b481fa00f808dd5308a530a1daf1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T18:17:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e32aaba93699944665660ae41e9b5779178c1354c31e15ad8188bf8c3c8fa500\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T18:17:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T18:18:32Z is after 2025-08-24T17:21:41Z" Oct 08 18:18:32 crc kubenswrapper[4859]: I1008 18:18:32.141993 4859 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-08T18:17:29Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T18:17:29Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T18:18:32Z is after 2025-08-24T17:21:41Z" Oct 08 18:18:32 crc kubenswrapper[4859]: I1008 18:18:32.154524 4859 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-82s52" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b23a6a6c-9d92-4e7b-840e-55cfda873a2d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:17:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:17:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:17:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:17:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://71302c766485b9f55743d85f97440ff578cd44ba1e5a6f5b1c024c0e808072d2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T18:17:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6vx22\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c23ea3580998adf90621a0155c20ee4bd590c512629812174e95c28857bcfef9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T18:17:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6vx22\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T18:17:31Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-82s52\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T18:18:32Z is after 2025-08-24T17:21:41Z" Oct 08 18:18:32 crc kubenswrapper[4859]: I1008 18:18:32.174616 4859 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-4g8gf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1dff864e-b75d-4e0b-b182-75f710eac8df\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:17:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:17:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:17:32Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:17:32Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://768ca274dedcc6847fc21d15234d3665fa9ccae371b5b6a882cfcb9b182f527d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T18:17:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s8jnn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://669f5f3c6d3c1ec0c4b4870e090b0a853adc01727c23383e595715b476c32668\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T18:17:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s8jnn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b8009a42cc4c9a816457a0678bdb2f7b7c53d0c8604e007a6fb141f66d06a516\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T18:17:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s8jnn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://676511a11ed4102084b57ebdd5f3fda4969741b2bc8443b415c4bf94f39ae4c9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T18:17:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s8jnn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4a74c1f0d21c90cdd7e776ee14e1f48757a04e2220fbd1f884449181ec6d1c3a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T18:17:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s8jnn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6ddc0e842bb8f81663406070023b0e5ed9cc99c98db0123b1c84f20da59fb5b5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T18:17:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s8jnn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f3a760c557ef21fbc6b54e78c3f6c5b325f296f93a022425cca7b1399eed8661\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c646f0e0dd8e7f74125b687f0cb4114b1f34667e5a6325f394f0208d7fb89df6\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-10-08T18:17:58Z\\\",\\\"message\\\":\\\"e *v1.Pod retry setup to complete in iterateRetryResources\\\\nI1008 18:17:58.424504 6520 obj_retry.go:365] Adding new object: *v1.Pod openshift-machine-config-operator/machine-config-daemon-82s52\\\\nI1008 18:17:58.424529 6520 ovn.go:134] Ensuring zone local for Pod openshift-machine-config-operator/machine-config-daemon-82s52 in node crc\\\\nI1008 18:17:58.424539 6520 obj_retry.go:303] Retry object setup: *v1.Pod openshift-network-diagnostics/network-check-source-55646444c4-trplf\\\\nI1008 18:17:58.424549 6520 obj_retry.go:365] Adding new object: *v1.Pod openshift-network-diagnostics/network-check-source-55646444c4-trplf\\\\nI1008 18:17:58.424549 6520 obj_retry.go:386] Retry successful for *v1.Pod openshift-machine-config-operator/machine-config-daemon-82s52 after 0 failed attempt(s)\\\\nI1008 18:17:58.424559 6520 ovn.go:134] Ensuring zone local for Pod openshift-network-diagnostics/network-check-source-55646444c4-trplf in node crc\\\\nF1008 18:17:58.424552 6520 ovnkube.go:137] failed to run ovnkube: [failed to start network controller: failed to start default network controller: unable to create admin network policy controller, err: could not add Event Handler for anpInformer during admin network policy controller initialization, handler {0x1fcc6e0 0x1fcc3c0 0x1fcc360} was not added to shared informer because it has stopped already, failed to star\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-08T18:17:57Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f3a760c557ef21fbc6b54e78c3f6c5b325f296f93a022425cca7b1399eed8661\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-10-08T18:18:31Z\\\",\\\"message\\\":\\\"etrics-daemon-4b4sz\\\\nI1008 18:18:31.553175 6912 obj_retry.go:303] Retry object setup: *v1.Pod openshift-machine-config-operator/machine-config-daemon-82s52\\\\nI1008 18:18:31.553182 6912 obj_retry.go:365] Adding new object: *v1.Pod openshift-machine-config-operator/machine-config-daemon-82s52\\\\nI1008 18:18:31.553190 6912 ovn.go:134] Ensuring zone local for Pod openshift-machine-config-operator/machine-config-daemon-82s52 in node crc\\\\nI1008 18:18:31.553195 6912 obj_retry.go:386] Retry successful for *v1.Pod openshift-machine-config-operator/machine-config-daemon-82s52 after 0 failed attempt(s)\\\\nI1008 18:18:31.553199 6912 default_network_controller.go:776] Recording success event on pod openshift-machine-config-operator/machine-config-daemon-82s52\\\\nI1008 18:18:31.553197 6912 obj_retry.go:303] Retry object setup: *v1.Pod openshift-network-diagnostics/network-check-source-55646444c4-trplf\\\\nI1008 18:18:31.553210 6912 obj_retry.go:365] Adding new object: *v1.Pod openshift-network-diagnostics/network-check-source-55646444c4-trplf\\\\nF1008 18:18:31.553216 6912 ovnkube.go:137] failed to run ovnkube: [failed to start network controller: failed to start default network controller: unable to create admin network policy controller, err: could not add Event Handler for anpInformer during admin network policy controller initialization, handler {0x1\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-08T18:18:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s8jnn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2c67d2aed7126d76989e057d8e2d8e188a8902b2f3fe26ed6d1d45d3b013d897\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T18:17:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s8jnn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7a211d3da02b173867ff3a912eecd3b379483c7cc77e6b91122edd1ffd30fd53\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7a211d3da02b173867ff3a912eecd3b379483c7cc77e6b91122edd1ffd30fd53\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-08T18:17:32Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-08T18:17:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s8jnn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T18:17:32Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-4g8gf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T18:18:32Z is after 2025-08-24T17:21:41Z" Oct 08 18:18:32 crc kubenswrapper[4859]: I1008 18:18:32.190056 4859 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4efc16fb-e679-464d-8b3c-0f46f082593e\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:17:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:17:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:17:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:17:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:17:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://aa8f8ea895272cd79a08d1c47cd7863a1ec10e21bb60d12aa10862a56c99e4bf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T18:17:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5c4e95d8a712ab219455564539a90286ae48bbc46ddc322a5bfd96c1a8268110\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T18:17:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://bae6ecfd8b446ef04d393d823e5ffe37fe749337c0ca17fde79c5f88d5e31393\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T18:17:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6411d1e7e18d5ceadaaceed5a8a6b20a0cacc55cede793ca9fb4dcad70c64fc6\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T18:17:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T18:17:10Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T18:18:32Z is after 2025-08-24T17:21:41Z" Oct 08 18:18:32 crc kubenswrapper[4859]: I1008 18:18:32.195801 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 18:18:32 crc kubenswrapper[4859]: I1008 18:18:32.195841 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 18:18:32 crc kubenswrapper[4859]: I1008 18:18:32.195854 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 18:18:32 crc kubenswrapper[4859]: I1008 18:18:32.195873 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 18:18:32 crc kubenswrapper[4859]: I1008 18:18:32.195887 4859 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T18:18:32Z","lastTransitionTime":"2025-10-08T18:18:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 18:18:32 crc kubenswrapper[4859]: I1008 18:18:32.208029 4859 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-08T18:17:29Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T18:17:29Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T18:18:32Z is after 2025-08-24T17:21:41Z" Oct 08 18:18:32 crc kubenswrapper[4859]: I1008 18:18:32.230662 4859 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-84k8x" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1c24078c-c9e8-4160-883a-f483b4b3dd23\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:17:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:17:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:17:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:17:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8661006834d0f50eaada7394369584ee95fc61ff54437eec35d755526afc5b48\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T18:17:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hk5xb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://18b291eb3344fb7c00ef5679dca9a6f7ed9d08cea684a803ff7dad2e8a18c585\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://18b291eb3344fb7c00ef5679dca9a6f7ed9d08cea684a803ff7dad2e8a18c585\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-08T18:17:32Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-08T18:17:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hk5xb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://160163f1a91d981583452db3c7a9244182d1c8ced45749a61b721d0d45c1dfe4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://160163f1a91d981583452db3c7a9244182d1c8ced45749a61b721d0d45c1dfe4\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-08T18:17:34Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-08T18:17:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hk5xb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://51e9b53f3847ac53f378ad83870debf8e71518d4edbc7fc0dc66644818e2103e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://51e9b53f3847ac53f378ad83870debf8e71518d4edbc7fc0dc66644818e2103e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-08T18:17:34Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-08T18:17:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hk5xb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1ff118919739759ac01b5cbe3240e6673db54eaa16353c609d670071ed7964cb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1ff118919739759ac01b5cbe3240e6673db54eaa16353c609d670071ed7964cb\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-08T18:17:35Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-08T18:17:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hk5xb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8522e803ff56eb57cd4d9fd036e0c3fb16240ea21d4a094c1db7aa7d50bfe2c2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8522e803ff56eb57cd4d9fd036e0c3fb16240ea21d4a094c1db7aa7d50bfe2c2\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-08T18:17:37Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-08T18:17:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hk5xb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://be61056edf515e037403ca2a0bfd5feb9b2537dd42212b58e133b23b294f327b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://be61056edf515e037403ca2a0bfd5feb9b2537dd42212b58e133b23b294f327b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-08T18:17:38Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-08T18:17:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hk5xb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T18:17:31Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-84k8x\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T18:18:32Z is after 2025-08-24T17:21:41Z" Oct 08 18:18:32 crc kubenswrapper[4859]: I1008 18:18:32.246077 4859 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3ca6ce76-77e8-4e46-a2f6-b1ed4aa0fd4b\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:17:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:17:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:18:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:18:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:17:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f587e149d57ea53626ed3beb83e421c3b89ba1fec9e949705e8ccb5b4f7cc79a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T18:17:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://428cbe63daefbfbcb2ec76dad0fe638cdc65934bac80b791cebe0b36b2d6909a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T18:17:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a41399a37b1ad86a063001224232ecb8f7ea56aeb5686d3858bc08102a0c8f49\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T18:17:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://008757fce552591f7c58193f605eb60792761d5943927cbd63184549618d2b43\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://008757fce552591f7c58193f605eb60792761d5943927cbd63184549618d2b43\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-08T18:17:11Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-08T18:17:11Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T18:17:10Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T18:18:32Z is after 2025-08-24T17:21:41Z" Oct 08 18:18:32 crc kubenswrapper[4859]: I1008 18:18:32.265228 4859 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-08T18:17:29Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T18:17:29Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T18:18:32Z is after 2025-08-24T17:21:41Z" Oct 08 18:18:32 crc kubenswrapper[4859]: I1008 18:18:32.281329 4859 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-btw28" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"97bc5598-3ead-4300-a938-3aebbff1e8c6\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:17:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:17:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:17:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:17:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d6d4d6ba771cb5fde637449bba05660f634df7e833a0c8dad2b2c7f3a422bb2b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T18:17:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qcxsd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T18:17:31Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-btw28\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T18:18:32Z is after 2025-08-24T17:21:41Z" Oct 08 18:18:32 crc kubenswrapper[4859]: I1008 18:18:32.298959 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 18:18:32 crc kubenswrapper[4859]: I1008 18:18:32.299008 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 18:18:32 crc kubenswrapper[4859]: I1008 18:18:32.299024 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 18:18:32 crc kubenswrapper[4859]: I1008 18:18:32.299041 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 18:18:32 crc kubenswrapper[4859]: I1008 18:18:32.299054 4859 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T18:18:32Z","lastTransitionTime":"2025-10-08T18:18:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 18:18:32 crc kubenswrapper[4859]: I1008 18:18:32.302175 4859 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-8plkg" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"92b2d7de-31cc-4a91-95a5-ed5ea964b028\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:17:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:17:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:18:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:18:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://729bd32fd01dcab495bffa64f2f5e44990fc1d78b05e7f131868009f92305831\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://fa00b8345d40b74a917ddc23cc1b2c5c684a9ca61b6119d090c92cc1439b71c4\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-10-08T18:18:19Z\\\",\\\"message\\\":\\\"2025-10-08T18:17:33+00:00 [cnibincopy] Successfully copied files in /usr/src/multus-cni/rhel9/bin/ to /host/opt/cni/bin/upgrade_888b130f-8754-48c6-b73e-9ff6772793f9\\\\n2025-10-08T18:17:33+00:00 [cnibincopy] Successfully moved files in /host/opt/cni/bin/upgrade_888b130f-8754-48c6-b73e-9ff6772793f9 to /host/opt/cni/bin/\\\\n2025-10-08T18:17:34Z [verbose] multus-daemon started\\\\n2025-10-08T18:17:34Z [verbose] Readiness Indicator file check\\\\n2025-10-08T18:18:19Z [error] have you checked that your default network is ready? still waiting for readinessindicatorfile @ /host/run/multus/cni/net.d/10-ovn-kubernetes.conf. pollimmediate error: timed out waiting for the condition\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-08T18:17:32Z\\\"}},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T18:18:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q9d7m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T18:17:31Z\\\"}}\" for pod \"openshift-multus\"/\"multus-8plkg\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T18:18:32Z is after 2025-08-24T17:21:41Z" Oct 08 18:18:32 crc kubenswrapper[4859]: I1008 18:18:32.326183 4859 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-btjmq" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b8605f7-fed7-493f-9b4f-6a6358adf907\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:17:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:17:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:17:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:17:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://14e9b6c4a4403f36adb78fcf1135a620ca6fc56ed974c9639dcdfc59e6917559\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T18:17:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xspg4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T18:17:35Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-btjmq\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T18:18:32Z is after 2025-08-24T17:21:41Z" Oct 08 18:18:32 crc kubenswrapper[4859]: I1008 18:18:32.341539 4859 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-g4t79" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3a92dbf1-21a8-4095-96bb-2d701d7f3d56\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:17:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:17:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:17:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:17:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f8772a25b6e605efe9a062dc32dbbc2cfe32b4fe164b4e84fbf3bcffab744179\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T18:17:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sv79p\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://92cb9986c09a7f9e1e552e3e3ac1f74f04128241e00ddee38882d2ba5c005a67\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T18:17:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sv79p\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T18:17:45Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-g4t79\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T18:18:32Z is after 2025-08-24T17:21:41Z" Oct 08 18:18:32 crc kubenswrapper[4859]: I1008 18:18:32.402434 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 18:18:32 crc kubenswrapper[4859]: I1008 18:18:32.402484 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 18:18:32 crc kubenswrapper[4859]: I1008 18:18:32.402497 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 18:18:32 crc kubenswrapper[4859]: I1008 18:18:32.402516 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 18:18:32 crc kubenswrapper[4859]: I1008 18:18:32.402529 4859 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T18:18:32Z","lastTransitionTime":"2025-10-08T18:18:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 18:18:32 crc kubenswrapper[4859]: I1008 18:18:32.505668 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 18:18:32 crc kubenswrapper[4859]: I1008 18:18:32.505733 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 18:18:32 crc kubenswrapper[4859]: I1008 18:18:32.505762 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 18:18:32 crc kubenswrapper[4859]: I1008 18:18:32.505778 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 18:18:32 crc kubenswrapper[4859]: I1008 18:18:32.505788 4859 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T18:18:32Z","lastTransitionTime":"2025-10-08T18:18:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 18:18:32 crc kubenswrapper[4859]: I1008 18:18:32.609482 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 18:18:32 crc kubenswrapper[4859]: I1008 18:18:32.609618 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 18:18:32 crc kubenswrapper[4859]: I1008 18:18:32.609642 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 18:18:32 crc kubenswrapper[4859]: I1008 18:18:32.609669 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 18:18:32 crc kubenswrapper[4859]: I1008 18:18:32.609745 4859 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T18:18:32Z","lastTransitionTime":"2025-10-08T18:18:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 18:18:32 crc kubenswrapper[4859]: I1008 18:18:32.712874 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 18:18:32 crc kubenswrapper[4859]: I1008 18:18:32.712955 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 18:18:32 crc kubenswrapper[4859]: I1008 18:18:32.712972 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 18:18:32 crc kubenswrapper[4859]: I1008 18:18:32.713003 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 18:18:32 crc kubenswrapper[4859]: I1008 18:18:32.713020 4859 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T18:18:32Z","lastTransitionTime":"2025-10-08T18:18:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 18:18:32 crc kubenswrapper[4859]: I1008 18:18:32.816793 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 18:18:32 crc kubenswrapper[4859]: I1008 18:18:32.816880 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 18:18:32 crc kubenswrapper[4859]: I1008 18:18:32.816903 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 18:18:32 crc kubenswrapper[4859]: I1008 18:18:32.816934 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 18:18:32 crc kubenswrapper[4859]: I1008 18:18:32.816956 4859 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T18:18:32Z","lastTransitionTime":"2025-10-08T18:18:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 18:18:32 crc kubenswrapper[4859]: I1008 18:18:32.920612 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 18:18:32 crc kubenswrapper[4859]: I1008 18:18:32.920706 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 18:18:32 crc kubenswrapper[4859]: I1008 18:18:32.920722 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 18:18:32 crc kubenswrapper[4859]: I1008 18:18:32.920741 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 18:18:32 crc kubenswrapper[4859]: I1008 18:18:32.920753 4859 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T18:18:32Z","lastTransitionTime":"2025-10-08T18:18:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 18:18:33 crc kubenswrapper[4859]: I1008 18:18:33.023312 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 18:18:33 crc kubenswrapper[4859]: I1008 18:18:33.023349 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 18:18:33 crc kubenswrapper[4859]: I1008 18:18:33.023361 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 18:18:33 crc kubenswrapper[4859]: I1008 18:18:33.023376 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 18:18:33 crc kubenswrapper[4859]: I1008 18:18:33.023388 4859 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T18:18:33Z","lastTransitionTime":"2025-10-08T18:18:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 18:18:33 crc kubenswrapper[4859]: I1008 18:18:33.034231 4859 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-4g8gf_1dff864e-b75d-4e0b-b182-75f710eac8df/ovnkube-controller/3.log" Oct 08 18:18:33 crc kubenswrapper[4859]: I1008 18:18:33.038055 4859 scope.go:117] "RemoveContainer" containerID="f3a760c557ef21fbc6b54e78c3f6c5b325f296f93a022425cca7b1399eed8661" Oct 08 18:18:33 crc kubenswrapper[4859]: E1008 18:18:33.038211 4859 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ovnkube-controller\" with CrashLoopBackOff: \"back-off 40s restarting failed container=ovnkube-controller pod=ovnkube-node-4g8gf_openshift-ovn-kubernetes(1dff864e-b75d-4e0b-b182-75f710eac8df)\"" pod="openshift-ovn-kubernetes/ovnkube-node-4g8gf" podUID="1dff864e-b75d-4e0b-b182-75f710eac8df" Oct 08 18:18:33 crc kubenswrapper[4859]: I1008 18:18:33.049599 4859 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-g4t79" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3a92dbf1-21a8-4095-96bb-2d701d7f3d56\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:17:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:17:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:17:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:17:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f8772a25b6e605efe9a062dc32dbbc2cfe32b4fe164b4e84fbf3bcffab744179\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T18:17:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sv79p\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://92cb9986c09a7f9e1e552e3e3ac1f74f04128241e00ddee38882d2ba5c005a67\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T18:17:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sv79p\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T18:17:45Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-g4t79\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T18:18:33Z is after 2025-08-24T17:21:41Z" Oct 08 18:18:33 crc kubenswrapper[4859]: I1008 18:18:33.062276 4859 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3ca6ce76-77e8-4e46-a2f6-b1ed4aa0fd4b\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:17:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:17:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:18:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:18:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:17:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f587e149d57ea53626ed3beb83e421c3b89ba1fec9e949705e8ccb5b4f7cc79a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T18:17:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://428cbe63daefbfbcb2ec76dad0fe638cdc65934bac80b791cebe0b36b2d6909a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T18:17:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a41399a37b1ad86a063001224232ecb8f7ea56aeb5686d3858bc08102a0c8f49\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T18:17:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://008757fce552591f7c58193f605eb60792761d5943927cbd63184549618d2b43\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://008757fce552591f7c58193f605eb60792761d5943927cbd63184549618d2b43\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-08T18:17:11Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-08T18:17:11Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T18:17:10Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T18:18:33Z is after 2025-08-24T17:21:41Z" Oct 08 18:18:33 crc kubenswrapper[4859]: I1008 18:18:33.078163 4859 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-08T18:17:29Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T18:17:29Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T18:18:33Z is after 2025-08-24T17:21:41Z" Oct 08 18:18:33 crc kubenswrapper[4859]: I1008 18:18:33.090585 4859 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-btw28" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"97bc5598-3ead-4300-a938-3aebbff1e8c6\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:17:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:17:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:17:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:17:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d6d4d6ba771cb5fde637449bba05660f634df7e833a0c8dad2b2c7f3a422bb2b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T18:17:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qcxsd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T18:17:31Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-btw28\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T18:18:33Z is after 2025-08-24T17:21:41Z" Oct 08 18:18:33 crc kubenswrapper[4859]: I1008 18:18:33.106333 4859 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-8plkg" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"92b2d7de-31cc-4a91-95a5-ed5ea964b028\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:17:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:17:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:18:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:18:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://729bd32fd01dcab495bffa64f2f5e44990fc1d78b05e7f131868009f92305831\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://fa00b8345d40b74a917ddc23cc1b2c5c684a9ca61b6119d090c92cc1439b71c4\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-10-08T18:18:19Z\\\",\\\"message\\\":\\\"2025-10-08T18:17:33+00:00 [cnibincopy] Successfully copied files in /usr/src/multus-cni/rhel9/bin/ to /host/opt/cni/bin/upgrade_888b130f-8754-48c6-b73e-9ff6772793f9\\\\n2025-10-08T18:17:33+00:00 [cnibincopy] Successfully moved files in /host/opt/cni/bin/upgrade_888b130f-8754-48c6-b73e-9ff6772793f9 to /host/opt/cni/bin/\\\\n2025-10-08T18:17:34Z [verbose] multus-daemon started\\\\n2025-10-08T18:17:34Z [verbose] Readiness Indicator file check\\\\n2025-10-08T18:18:19Z [error] have you checked that your default network is ready? still waiting for readinessindicatorfile @ /host/run/multus/cni/net.d/10-ovn-kubernetes.conf. pollimmediate error: timed out waiting for the condition\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-08T18:17:32Z\\\"}},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T18:18:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q9d7m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T18:17:31Z\\\"}}\" for pod \"openshift-multus\"/\"multus-8plkg\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T18:18:33Z is after 2025-08-24T17:21:41Z" Oct 08 18:18:33 crc kubenswrapper[4859]: I1008 18:18:33.117781 4859 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-btjmq" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b8605f7-fed7-493f-9b4f-6a6358adf907\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:17:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:17:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:17:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:17:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://14e9b6c4a4403f36adb78fcf1135a620ca6fc56ed974c9639dcdfc59e6917559\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T18:17:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xspg4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T18:17:35Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-btjmq\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T18:18:33Z is after 2025-08-24T17:21:41Z" Oct 08 18:18:33 crc kubenswrapper[4859]: I1008 18:18:33.126015 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 18:18:33 crc kubenswrapper[4859]: I1008 18:18:33.126054 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 18:18:33 crc kubenswrapper[4859]: I1008 18:18:33.126067 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 18:18:33 crc kubenswrapper[4859]: I1008 18:18:33.126090 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 18:18:33 crc kubenswrapper[4859]: I1008 18:18:33.126137 4859 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T18:18:33Z","lastTransitionTime":"2025-10-08T18:18:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 18:18:33 crc kubenswrapper[4859]: I1008 18:18:33.131772 4859 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-08T18:17:31Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T18:17:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T18:17:31Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c6d15905694b6bbd4999d49dc8f7a53eb6bfaa7adb59efb0d7e4973c21d1c7be\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T18:17:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T18:18:33Z is after 2025-08-24T17:21:41Z" Oct 08 18:18:33 crc kubenswrapper[4859]: I1008 18:18:33.144055 4859 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-08T18:17:33Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T18:17:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T18:17:33Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5d99e071bd083914418d9c128f38955f26295c08b1520f2dc774c99b1c843693\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T18:17:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T18:18:33Z is after 2025-08-24T17:21:41Z" Oct 08 18:18:33 crc kubenswrapper[4859]: I1008 18:18:33.160749 4859 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-82s52" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b23a6a6c-9d92-4e7b-840e-55cfda873a2d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:17:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:17:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:17:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:17:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://71302c766485b9f55743d85f97440ff578cd44ba1e5a6f5b1c024c0e808072d2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T18:17:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6vx22\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c23ea3580998adf90621a0155c20ee4bd590c512629812174e95c28857bcfef9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T18:17:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6vx22\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T18:17:31Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-82s52\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T18:18:33Z is after 2025-08-24T17:21:41Z" Oct 08 18:18:33 crc kubenswrapper[4859]: I1008 18:18:33.184149 4859 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-4g8gf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1dff864e-b75d-4e0b-b182-75f710eac8df\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:17:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:17:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:17:32Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:17:32Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://768ca274dedcc6847fc21d15234d3665fa9ccae371b5b6a882cfcb9b182f527d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T18:17:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s8jnn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://669f5f3c6d3c1ec0c4b4870e090b0a853adc01727c23383e595715b476c32668\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T18:17:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s8jnn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b8009a42cc4c9a816457a0678bdb2f7b7c53d0c8604e007a6fb141f66d06a516\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T18:17:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s8jnn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://676511a11ed4102084b57ebdd5f3fda4969741b2bc8443b415c4bf94f39ae4c9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T18:17:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s8jnn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4a74c1f0d21c90cdd7e776ee14e1f48757a04e2220fbd1f884449181ec6d1c3a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T18:17:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s8jnn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6ddc0e842bb8f81663406070023b0e5ed9cc99c98db0123b1c84f20da59fb5b5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T18:17:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s8jnn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f3a760c557ef21fbc6b54e78c3f6c5b325f296f93a022425cca7b1399eed8661\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f3a760c557ef21fbc6b54e78c3f6c5b325f296f93a022425cca7b1399eed8661\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-10-08T18:18:31Z\\\",\\\"message\\\":\\\"etrics-daemon-4b4sz\\\\nI1008 18:18:31.553175 6912 obj_retry.go:303] Retry object setup: *v1.Pod openshift-machine-config-operator/machine-config-daemon-82s52\\\\nI1008 18:18:31.553182 6912 obj_retry.go:365] Adding new object: *v1.Pod openshift-machine-config-operator/machine-config-daemon-82s52\\\\nI1008 18:18:31.553190 6912 ovn.go:134] Ensuring zone local for Pod openshift-machine-config-operator/machine-config-daemon-82s52 in node crc\\\\nI1008 18:18:31.553195 6912 obj_retry.go:386] Retry successful for *v1.Pod openshift-machine-config-operator/machine-config-daemon-82s52 after 0 failed attempt(s)\\\\nI1008 18:18:31.553199 6912 default_network_controller.go:776] Recording success event on pod openshift-machine-config-operator/machine-config-daemon-82s52\\\\nI1008 18:18:31.553197 6912 obj_retry.go:303] Retry object setup: *v1.Pod openshift-network-diagnostics/network-check-source-55646444c4-trplf\\\\nI1008 18:18:31.553210 6912 obj_retry.go:365] Adding new object: *v1.Pod openshift-network-diagnostics/network-check-source-55646444c4-trplf\\\\nF1008 18:18:31.553216 6912 ovnkube.go:137] failed to run ovnkube: [failed to start network controller: failed to start default network controller: unable to create admin network policy controller, err: could not add Event Handler for anpInformer during admin network policy controller initialization, handler {0x1\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-08T18:18:30Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 40s restarting failed container=ovnkube-controller pod=ovnkube-node-4g8gf_openshift-ovn-kubernetes(1dff864e-b75d-4e0b-b182-75f710eac8df)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s8jnn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2c67d2aed7126d76989e057d8e2d8e188a8902b2f3fe26ed6d1d45d3b013d897\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T18:17:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s8jnn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7a211d3da02b173867ff3a912eecd3b379483c7cc77e6b91122edd1ffd30fd53\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7a211d3da02b173867ff3a912eecd3b379483c7cc77e6b91122edd1ffd30fd53\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-08T18:17:32Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-08T18:17:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s8jnn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T18:17:32Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-4g8gf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T18:18:33Z is after 2025-08-24T17:21:41Z" Oct 08 18:18:33 crc kubenswrapper[4859]: I1008 18:18:33.197426 4859 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-4b4sz" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2e53cd9b-64f7-4e07-8b96-fb77847c9ac6\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:17:45Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:17:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:17:45Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:17:45Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kxgbz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kxgbz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T18:17:45Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-4b4sz\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T18:18:33Z is after 2025-08-24T17:21:41Z" Oct 08 18:18:33 crc kubenswrapper[4859]: I1008 18:18:33.208767 4859 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"262c958c-c49d-4c11-96ae-94106dd123d4\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:17:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:17:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:17:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:17:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:17:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a54a54b9a831ead1fbe6e03abbb622242dd16b4081a4482d559e03d47aebd2fe\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-crio\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T18:17:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kube\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3ab3bc5e8e2fd4941f30cbf8a90d9a9078b90fab91063c07da2cd48b63eef657\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3ab3bc5e8e2fd4941f30cbf8a90d9a9078b90fab91063c07da2cd48b63eef657\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-08T18:17:11Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-08T18:17:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T18:17:10Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"kube-rbac-proxy-crio-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T18:18:33Z is after 2025-08-24T17:21:41Z" Oct 08 18:18:33 crc kubenswrapper[4859]: I1008 18:18:33.222058 4859 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"63a623de-5608-4228-b1c2-77188c276986\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:17:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:17:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:17:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:17:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:17:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d4786a8fb8abd96fb22e60733c974f599017407d589ed815fb8541fd3b64f696\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T18:17:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://183d481502ed5108d5e19e483b714c289f355ea3f8b0891883e5335b802c83b6\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T18:17:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ec848c46fd5dde0469b00d3cab14ba4a652533a265490fadbee7a8b8843ef249\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T18:17:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4fe1828b4d63ff8106bfcd15e1dc1c1bd1ba7f77777cdb23ff1b291d0418774c\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2e700b2a20c81b414617f5b5c018a738278a3fa44e6c6d14b2170051d0725431\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-08T18:17:24Z\\\",\\\"message\\\":\\\"W1008 18:17:13.670625 1 cmd.go:257] Using insecure, self-signed certificates\\\\nI1008 18:17:13.671202 1 crypto.go:601] Generating new CA for check-endpoints-signer@1759947433 cert, and key in /tmp/serving-cert-3867218267/serving-signer.crt, /tmp/serving-cert-3867218267/serving-signer.key\\\\nI1008 18:17:13.944463 1 observer_polling.go:159] Starting file observer\\\\nW1008 18:17:13.951958 1 builder.go:272] unable to get owner reference (falling back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": dial tcp [::1]:6443: connect: connection refused\\\\nI1008 18:17:13.952088 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1008 18:17:13.956755 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3867218267/tls.crt::/tmp/serving-cert-3867218267/tls.key\\\\\\\"\\\\nF1008 18:17:24.290991 1 cmd.go:182] error initializing delegating authentication: unable to load configmap based request-header-client-ca-file: Get \\\\\\\"https://localhost:6443/api/v1/namespaces/kube-system/configmaps/extension-apiserver-authentication\\\\\\\": net/http: TLS handshake timeout\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-08T18:17:13Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T18:17:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c65b66f62d325ca10ef8d5d31f04529ccf4e691cf3303d231e7a414fb400e4d6\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T18:17:13Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://58c7bba4ace28215b062653a960e561b1dc1f890f2baf450145638a065af038b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://58c7bba4ace28215b062653a960e561b1dc1f890f2baf450145638a065af038b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-08T18:17:11Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-08T18:17:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T18:17:10Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T18:18:33Z is after 2025-08-24T17:21:41Z" Oct 08 18:18:33 crc kubenswrapper[4859]: I1008 18:18:33.229535 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 18:18:33 crc kubenswrapper[4859]: I1008 18:18:33.229575 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 18:18:33 crc kubenswrapper[4859]: I1008 18:18:33.229589 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 18:18:33 crc kubenswrapper[4859]: I1008 18:18:33.229611 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 18:18:33 crc kubenswrapper[4859]: I1008 18:18:33.229625 4859 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T18:18:33Z","lastTransitionTime":"2025-10-08T18:18:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 18:18:33 crc kubenswrapper[4859]: I1008 18:18:33.233528 4859 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-08T18:17:31Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T18:17:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T18:17:31Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://99e2d51bd7757498dae60949eea6d732b5f7b481fa00f808dd5308a530a1daf1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T18:17:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e32aaba93699944665660ae41e9b5779178c1354c31e15ad8188bf8c3c8fa500\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T18:17:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T18:18:33Z is after 2025-08-24T17:21:41Z" Oct 08 18:18:33 crc kubenswrapper[4859]: I1008 18:18:33.246934 4859 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-08T18:17:29Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T18:17:29Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T18:18:33Z is after 2025-08-24T17:21:41Z" Oct 08 18:18:33 crc kubenswrapper[4859]: I1008 18:18:33.259452 4859 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4efc16fb-e679-464d-8b3c-0f46f082593e\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:17:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:17:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:17:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:17:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:17:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://aa8f8ea895272cd79a08d1c47cd7863a1ec10e21bb60d12aa10862a56c99e4bf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T18:17:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5c4e95d8a712ab219455564539a90286ae48bbc46ddc322a5bfd96c1a8268110\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T18:17:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://bae6ecfd8b446ef04d393d823e5ffe37fe749337c0ca17fde79c5f88d5e31393\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T18:17:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6411d1e7e18d5ceadaaceed5a8a6b20a0cacc55cede793ca9fb4dcad70c64fc6\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T18:17:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T18:17:10Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T18:18:33Z is after 2025-08-24T17:21:41Z" Oct 08 18:18:33 crc kubenswrapper[4859]: I1008 18:18:33.272521 4859 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-08T18:17:29Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T18:17:29Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T18:18:33Z is after 2025-08-24T17:21:41Z" Oct 08 18:18:33 crc kubenswrapper[4859]: I1008 18:18:33.286353 4859 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-84k8x" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1c24078c-c9e8-4160-883a-f483b4b3dd23\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:17:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:17:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:17:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:17:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8661006834d0f50eaada7394369584ee95fc61ff54437eec35d755526afc5b48\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T18:17:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hk5xb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://18b291eb3344fb7c00ef5679dca9a6f7ed9d08cea684a803ff7dad2e8a18c585\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://18b291eb3344fb7c00ef5679dca9a6f7ed9d08cea684a803ff7dad2e8a18c585\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-08T18:17:32Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-08T18:17:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hk5xb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://160163f1a91d981583452db3c7a9244182d1c8ced45749a61b721d0d45c1dfe4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://160163f1a91d981583452db3c7a9244182d1c8ced45749a61b721d0d45c1dfe4\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-08T18:17:34Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-08T18:17:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hk5xb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://51e9b53f3847ac53f378ad83870debf8e71518d4edbc7fc0dc66644818e2103e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://51e9b53f3847ac53f378ad83870debf8e71518d4edbc7fc0dc66644818e2103e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-08T18:17:34Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-08T18:17:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hk5xb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1ff118919739759ac01b5cbe3240e6673db54eaa16353c609d670071ed7964cb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1ff118919739759ac01b5cbe3240e6673db54eaa16353c609d670071ed7964cb\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-08T18:17:35Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-08T18:17:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hk5xb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8522e803ff56eb57cd4d9fd036e0c3fb16240ea21d4a094c1db7aa7d50bfe2c2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8522e803ff56eb57cd4d9fd036e0c3fb16240ea21d4a094c1db7aa7d50bfe2c2\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-08T18:17:37Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-08T18:17:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hk5xb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://be61056edf515e037403ca2a0bfd5feb9b2537dd42212b58e133b23b294f327b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://be61056edf515e037403ca2a0bfd5feb9b2537dd42212b58e133b23b294f327b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-08T18:17:38Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-08T18:17:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hk5xb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T18:17:31Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-84k8x\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T18:18:33Z is after 2025-08-24T17:21:41Z" Oct 08 18:18:33 crc kubenswrapper[4859]: I1008 18:18:33.331887 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 18:18:33 crc kubenswrapper[4859]: I1008 18:18:33.331948 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 18:18:33 crc kubenswrapper[4859]: I1008 18:18:33.331962 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 18:18:33 crc kubenswrapper[4859]: I1008 18:18:33.331986 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 18:18:33 crc kubenswrapper[4859]: I1008 18:18:33.332000 4859 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T18:18:33Z","lastTransitionTime":"2025-10-08T18:18:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 18:18:33 crc kubenswrapper[4859]: I1008 18:18:33.435813 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 18:18:33 crc kubenswrapper[4859]: I1008 18:18:33.435886 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 18:18:33 crc kubenswrapper[4859]: I1008 18:18:33.435913 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 18:18:33 crc kubenswrapper[4859]: I1008 18:18:33.435949 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 18:18:33 crc kubenswrapper[4859]: I1008 18:18:33.435974 4859 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T18:18:33Z","lastTransitionTime":"2025-10-08T18:18:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 18:18:33 crc kubenswrapper[4859]: I1008 18:18:33.469084 4859 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 08 18:18:33 crc kubenswrapper[4859]: I1008 18:18:33.469120 4859 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 08 18:18:33 crc kubenswrapper[4859]: I1008 18:18:33.469121 4859 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 08 18:18:33 crc kubenswrapper[4859]: E1008 18:18:33.469266 4859 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 08 18:18:33 crc kubenswrapper[4859]: I1008 18:18:33.469312 4859 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-4b4sz" Oct 08 18:18:33 crc kubenswrapper[4859]: E1008 18:18:33.469548 4859 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 08 18:18:33 crc kubenswrapper[4859]: E1008 18:18:33.469578 4859 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-4b4sz" podUID="2e53cd9b-64f7-4e07-8b96-fb77847c9ac6" Oct 08 18:18:33 crc kubenswrapper[4859]: E1008 18:18:33.469633 4859 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 08 18:18:33 crc kubenswrapper[4859]: I1008 18:18:33.538528 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 18:18:33 crc kubenswrapper[4859]: I1008 18:18:33.538576 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 18:18:33 crc kubenswrapper[4859]: I1008 18:18:33.538586 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 18:18:33 crc kubenswrapper[4859]: I1008 18:18:33.538603 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 18:18:33 crc kubenswrapper[4859]: I1008 18:18:33.538615 4859 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T18:18:33Z","lastTransitionTime":"2025-10-08T18:18:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 18:18:33 crc kubenswrapper[4859]: I1008 18:18:33.642266 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 18:18:33 crc kubenswrapper[4859]: I1008 18:18:33.642355 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 18:18:33 crc kubenswrapper[4859]: I1008 18:18:33.642394 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 18:18:33 crc kubenswrapper[4859]: I1008 18:18:33.642427 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 18:18:33 crc kubenswrapper[4859]: I1008 18:18:33.642450 4859 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T18:18:33Z","lastTransitionTime":"2025-10-08T18:18:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 18:18:33 crc kubenswrapper[4859]: I1008 18:18:33.746080 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 18:18:33 crc kubenswrapper[4859]: I1008 18:18:33.746133 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 18:18:33 crc kubenswrapper[4859]: I1008 18:18:33.746145 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 18:18:33 crc kubenswrapper[4859]: I1008 18:18:33.746166 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 18:18:33 crc kubenswrapper[4859]: I1008 18:18:33.746177 4859 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T18:18:33Z","lastTransitionTime":"2025-10-08T18:18:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 18:18:33 crc kubenswrapper[4859]: I1008 18:18:33.849663 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 18:18:33 crc kubenswrapper[4859]: I1008 18:18:33.849761 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 18:18:33 crc kubenswrapper[4859]: I1008 18:18:33.849789 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 18:18:33 crc kubenswrapper[4859]: I1008 18:18:33.849821 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 18:18:33 crc kubenswrapper[4859]: I1008 18:18:33.849843 4859 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T18:18:33Z","lastTransitionTime":"2025-10-08T18:18:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 18:18:33 crc kubenswrapper[4859]: I1008 18:18:33.953618 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 18:18:33 crc kubenswrapper[4859]: I1008 18:18:33.953680 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 18:18:33 crc kubenswrapper[4859]: I1008 18:18:33.953763 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 18:18:33 crc kubenswrapper[4859]: I1008 18:18:33.953802 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 18:18:33 crc kubenswrapper[4859]: I1008 18:18:33.953827 4859 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T18:18:33Z","lastTransitionTime":"2025-10-08T18:18:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 18:18:34 crc kubenswrapper[4859]: I1008 18:18:34.056259 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 18:18:34 crc kubenswrapper[4859]: I1008 18:18:34.056336 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 18:18:34 crc kubenswrapper[4859]: I1008 18:18:34.056357 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 18:18:34 crc kubenswrapper[4859]: I1008 18:18:34.056388 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 18:18:34 crc kubenswrapper[4859]: I1008 18:18:34.056411 4859 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T18:18:34Z","lastTransitionTime":"2025-10-08T18:18:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 18:18:34 crc kubenswrapper[4859]: I1008 18:18:34.158667 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 18:18:34 crc kubenswrapper[4859]: I1008 18:18:34.158739 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 18:18:34 crc kubenswrapper[4859]: I1008 18:18:34.158750 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 18:18:34 crc kubenswrapper[4859]: I1008 18:18:34.158770 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 18:18:34 crc kubenswrapper[4859]: I1008 18:18:34.158782 4859 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T18:18:34Z","lastTransitionTime":"2025-10-08T18:18:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 18:18:34 crc kubenswrapper[4859]: I1008 18:18:34.261667 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 18:18:34 crc kubenswrapper[4859]: I1008 18:18:34.261742 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 18:18:34 crc kubenswrapper[4859]: I1008 18:18:34.261755 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 18:18:34 crc kubenswrapper[4859]: I1008 18:18:34.261779 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 18:18:34 crc kubenswrapper[4859]: I1008 18:18:34.261795 4859 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T18:18:34Z","lastTransitionTime":"2025-10-08T18:18:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 18:18:34 crc kubenswrapper[4859]: I1008 18:18:34.364631 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 18:18:34 crc kubenswrapper[4859]: I1008 18:18:34.364701 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 18:18:34 crc kubenswrapper[4859]: I1008 18:18:34.364711 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 18:18:34 crc kubenswrapper[4859]: I1008 18:18:34.364728 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 18:18:34 crc kubenswrapper[4859]: I1008 18:18:34.364738 4859 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T18:18:34Z","lastTransitionTime":"2025-10-08T18:18:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 18:18:34 crc kubenswrapper[4859]: I1008 18:18:34.458649 4859 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 08 18:18:34 crc kubenswrapper[4859]: E1008 18:18:34.458875 4859 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-08 18:19:38.45884129 +0000 UTC m=+148.705680699 (durationBeforeRetry 1m4s). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 08 18:18:34 crc kubenswrapper[4859]: I1008 18:18:34.467991 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 18:18:34 crc kubenswrapper[4859]: I1008 18:18:34.468062 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 18:18:34 crc kubenswrapper[4859]: I1008 18:18:34.468078 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 18:18:34 crc kubenswrapper[4859]: I1008 18:18:34.468108 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 18:18:34 crc kubenswrapper[4859]: I1008 18:18:34.468123 4859 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T18:18:34Z","lastTransitionTime":"2025-10-08T18:18:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 18:18:34 crc kubenswrapper[4859]: I1008 18:18:34.560308 4859 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 08 18:18:34 crc kubenswrapper[4859]: I1008 18:18:34.560397 4859 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 08 18:18:34 crc kubenswrapper[4859]: I1008 18:18:34.560459 4859 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 08 18:18:34 crc kubenswrapper[4859]: I1008 18:18:34.560550 4859 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 08 18:18:34 crc kubenswrapper[4859]: E1008 18:18:34.560611 4859 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Oct 08 18:18:34 crc kubenswrapper[4859]: E1008 18:18:34.560660 4859 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Oct 08 18:18:34 crc kubenswrapper[4859]: E1008 18:18:34.560723 4859 projected.go:194] Error preparing data for projected volume kube-api-access-cqllr for pod openshift-network-diagnostics/network-check-target-xd92c: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Oct 08 18:18:34 crc kubenswrapper[4859]: E1008 18:18:34.560726 4859 configmap.go:193] Couldn't get configMap openshift-network-console/networking-console-plugin: object "openshift-network-console"/"networking-console-plugin" not registered Oct 08 18:18:34 crc kubenswrapper[4859]: E1008 18:18:34.560613 4859 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Oct 08 18:18:34 crc kubenswrapper[4859]: E1008 18:18:34.560806 4859 secret.go:188] Couldn't get secret openshift-network-console/networking-console-plugin-cert: object "openshift-network-console"/"networking-console-plugin-cert" not registered Oct 08 18:18:34 crc kubenswrapper[4859]: E1008 18:18:34.560819 4859 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Oct 08 18:18:34 crc kubenswrapper[4859]: E1008 18:18:34.560850 4859 projected.go:194] Error preparing data for projected volume kube-api-access-s2dwl for pod openshift-network-diagnostics/network-check-source-55646444c4-trplf: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Oct 08 18:18:34 crc kubenswrapper[4859]: E1008 18:18:34.560821 4859 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-10-08 18:19:38.560791503 +0000 UTC m=+148.807630922 (durationBeforeRetry 1m4s). Error: MountVolume.SetUp failed for volume "nginx-conf" (UniqueName: "kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin" not registered Oct 08 18:18:34 crc kubenswrapper[4859]: E1008 18:18:34.560938 4859 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr podName:3b6479f0-333b-4a96-9adf-2099afdc2447 nodeName:}" failed. No retries permitted until 2025-10-08 18:19:38.560915187 +0000 UTC m=+148.807754606 (durationBeforeRetry 1m4s). Error: MountVolume.SetUp failed for volume "kube-api-access-cqllr" (UniqueName: "kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr") pod "network-check-target-xd92c" (UID: "3b6479f0-333b-4a96-9adf-2099afdc2447") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Oct 08 18:18:34 crc kubenswrapper[4859]: E1008 18:18:34.561013 4859 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-10-08 18:19:38.560996319 +0000 UTC m=+148.807835748 (durationBeforeRetry 1m4s). Error: MountVolume.SetUp failed for volume "networking-console-plugin-cert" (UniqueName: "kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin-cert" not registered Oct 08 18:18:34 crc kubenswrapper[4859]: E1008 18:18:34.561054 4859 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl podName:9d751cbb-f2e2-430d-9754-c882a5e924a5 nodeName:}" failed. No retries permitted until 2025-10-08 18:19:38.56103842 +0000 UTC m=+148.807877839 (durationBeforeRetry 1m4s). Error: MountVolume.SetUp failed for volume "kube-api-access-s2dwl" (UniqueName: "kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl") pod "network-check-source-55646444c4-trplf" (UID: "9d751cbb-f2e2-430d-9754-c882a5e924a5") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Oct 08 18:18:34 crc kubenswrapper[4859]: I1008 18:18:34.571409 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 18:18:34 crc kubenswrapper[4859]: I1008 18:18:34.571449 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 18:18:34 crc kubenswrapper[4859]: I1008 18:18:34.571460 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 18:18:34 crc kubenswrapper[4859]: I1008 18:18:34.571476 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 18:18:34 crc kubenswrapper[4859]: I1008 18:18:34.571488 4859 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T18:18:34Z","lastTransitionTime":"2025-10-08T18:18:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 18:18:34 crc kubenswrapper[4859]: I1008 18:18:34.673909 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 18:18:34 crc kubenswrapper[4859]: I1008 18:18:34.673955 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 18:18:34 crc kubenswrapper[4859]: I1008 18:18:34.673971 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 18:18:34 crc kubenswrapper[4859]: I1008 18:18:34.673994 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 18:18:34 crc kubenswrapper[4859]: I1008 18:18:34.674006 4859 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T18:18:34Z","lastTransitionTime":"2025-10-08T18:18:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 18:18:34 crc kubenswrapper[4859]: I1008 18:18:34.776918 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 18:18:34 crc kubenswrapper[4859]: I1008 18:18:34.776988 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 18:18:34 crc kubenswrapper[4859]: I1008 18:18:34.777012 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 18:18:34 crc kubenswrapper[4859]: I1008 18:18:34.777043 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 18:18:34 crc kubenswrapper[4859]: I1008 18:18:34.777070 4859 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T18:18:34Z","lastTransitionTime":"2025-10-08T18:18:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 18:18:34 crc kubenswrapper[4859]: I1008 18:18:34.880264 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 18:18:34 crc kubenswrapper[4859]: I1008 18:18:34.880907 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 18:18:34 crc kubenswrapper[4859]: I1008 18:18:34.880925 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 18:18:34 crc kubenswrapper[4859]: I1008 18:18:34.880944 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 18:18:34 crc kubenswrapper[4859]: I1008 18:18:34.880957 4859 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T18:18:34Z","lastTransitionTime":"2025-10-08T18:18:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 18:18:34 crc kubenswrapper[4859]: I1008 18:18:34.984166 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 18:18:34 crc kubenswrapper[4859]: I1008 18:18:34.984236 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 18:18:34 crc kubenswrapper[4859]: I1008 18:18:34.984247 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 18:18:34 crc kubenswrapper[4859]: I1008 18:18:34.984267 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 18:18:34 crc kubenswrapper[4859]: I1008 18:18:34.984279 4859 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T18:18:34Z","lastTransitionTime":"2025-10-08T18:18:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 18:18:35 crc kubenswrapper[4859]: I1008 18:18:35.087306 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 18:18:35 crc kubenswrapper[4859]: I1008 18:18:35.087355 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 18:18:35 crc kubenswrapper[4859]: I1008 18:18:35.087367 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 18:18:35 crc kubenswrapper[4859]: I1008 18:18:35.087390 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 18:18:35 crc kubenswrapper[4859]: I1008 18:18:35.087403 4859 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T18:18:35Z","lastTransitionTime":"2025-10-08T18:18:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 18:18:35 crc kubenswrapper[4859]: I1008 18:18:35.190565 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 18:18:35 crc kubenswrapper[4859]: I1008 18:18:35.190606 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 18:18:35 crc kubenswrapper[4859]: I1008 18:18:35.190617 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 18:18:35 crc kubenswrapper[4859]: I1008 18:18:35.190650 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 18:18:35 crc kubenswrapper[4859]: I1008 18:18:35.190662 4859 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T18:18:35Z","lastTransitionTime":"2025-10-08T18:18:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 18:18:35 crc kubenswrapper[4859]: I1008 18:18:35.294373 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 18:18:35 crc kubenswrapper[4859]: I1008 18:18:35.294427 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 18:18:35 crc kubenswrapper[4859]: I1008 18:18:35.294445 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 18:18:35 crc kubenswrapper[4859]: I1008 18:18:35.294494 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 18:18:35 crc kubenswrapper[4859]: I1008 18:18:35.294515 4859 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T18:18:35Z","lastTransitionTime":"2025-10-08T18:18:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 18:18:35 crc kubenswrapper[4859]: I1008 18:18:35.401784 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 18:18:35 crc kubenswrapper[4859]: I1008 18:18:35.401829 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 18:18:35 crc kubenswrapper[4859]: I1008 18:18:35.401841 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 18:18:35 crc kubenswrapper[4859]: I1008 18:18:35.401864 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 18:18:35 crc kubenswrapper[4859]: I1008 18:18:35.401876 4859 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T18:18:35Z","lastTransitionTime":"2025-10-08T18:18:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 18:18:35 crc kubenswrapper[4859]: I1008 18:18:35.469274 4859 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 08 18:18:35 crc kubenswrapper[4859]: E1008 18:18:35.470315 4859 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 08 18:18:35 crc kubenswrapper[4859]: I1008 18:18:35.469405 4859 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 08 18:18:35 crc kubenswrapper[4859]: I1008 18:18:35.469390 4859 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-4b4sz" Oct 08 18:18:35 crc kubenswrapper[4859]: E1008 18:18:35.470398 4859 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 08 18:18:35 crc kubenswrapper[4859]: I1008 18:18:35.469442 4859 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 08 18:18:35 crc kubenswrapper[4859]: E1008 18:18:35.470574 4859 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-4b4sz" podUID="2e53cd9b-64f7-4e07-8b96-fb77847c9ac6" Oct 08 18:18:35 crc kubenswrapper[4859]: E1008 18:18:35.470744 4859 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 08 18:18:35 crc kubenswrapper[4859]: I1008 18:18:35.505298 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 18:18:35 crc kubenswrapper[4859]: I1008 18:18:35.505375 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 18:18:35 crc kubenswrapper[4859]: I1008 18:18:35.505408 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 18:18:35 crc kubenswrapper[4859]: I1008 18:18:35.505425 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 18:18:35 crc kubenswrapper[4859]: I1008 18:18:35.505438 4859 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T18:18:35Z","lastTransitionTime":"2025-10-08T18:18:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 18:18:35 crc kubenswrapper[4859]: I1008 18:18:35.608209 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 18:18:35 crc kubenswrapper[4859]: I1008 18:18:35.608650 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 18:18:35 crc kubenswrapper[4859]: I1008 18:18:35.608673 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 18:18:35 crc kubenswrapper[4859]: I1008 18:18:35.608743 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 18:18:35 crc kubenswrapper[4859]: I1008 18:18:35.608771 4859 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T18:18:35Z","lastTransitionTime":"2025-10-08T18:18:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 18:18:35 crc kubenswrapper[4859]: I1008 18:18:35.712733 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 18:18:35 crc kubenswrapper[4859]: I1008 18:18:35.712810 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 18:18:35 crc kubenswrapper[4859]: I1008 18:18:35.712820 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 18:18:35 crc kubenswrapper[4859]: I1008 18:18:35.712843 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 18:18:35 crc kubenswrapper[4859]: I1008 18:18:35.712854 4859 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T18:18:35Z","lastTransitionTime":"2025-10-08T18:18:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 18:18:35 crc kubenswrapper[4859]: I1008 18:18:35.815996 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 18:18:35 crc kubenswrapper[4859]: I1008 18:18:35.816071 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 18:18:35 crc kubenswrapper[4859]: I1008 18:18:35.816090 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 18:18:35 crc kubenswrapper[4859]: I1008 18:18:35.816122 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 18:18:35 crc kubenswrapper[4859]: I1008 18:18:35.816141 4859 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T18:18:35Z","lastTransitionTime":"2025-10-08T18:18:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 18:18:35 crc kubenswrapper[4859]: I1008 18:18:35.919031 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 18:18:35 crc kubenswrapper[4859]: I1008 18:18:35.919083 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 18:18:35 crc kubenswrapper[4859]: I1008 18:18:35.919097 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 18:18:35 crc kubenswrapper[4859]: I1008 18:18:35.919120 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 18:18:35 crc kubenswrapper[4859]: I1008 18:18:35.919134 4859 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T18:18:35Z","lastTransitionTime":"2025-10-08T18:18:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 18:18:36 crc kubenswrapper[4859]: I1008 18:18:36.021717 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 18:18:36 crc kubenswrapper[4859]: I1008 18:18:36.021777 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 18:18:36 crc kubenswrapper[4859]: I1008 18:18:36.021793 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 18:18:36 crc kubenswrapper[4859]: I1008 18:18:36.021816 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 18:18:36 crc kubenswrapper[4859]: I1008 18:18:36.021830 4859 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T18:18:36Z","lastTransitionTime":"2025-10-08T18:18:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 18:18:36 crc kubenswrapper[4859]: I1008 18:18:36.124994 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 18:18:36 crc kubenswrapper[4859]: I1008 18:18:36.125086 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 18:18:36 crc kubenswrapper[4859]: I1008 18:18:36.125116 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 18:18:36 crc kubenswrapper[4859]: I1008 18:18:36.125147 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 18:18:36 crc kubenswrapper[4859]: I1008 18:18:36.125166 4859 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T18:18:36Z","lastTransitionTime":"2025-10-08T18:18:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 18:18:36 crc kubenswrapper[4859]: I1008 18:18:36.228250 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 18:18:36 crc kubenswrapper[4859]: I1008 18:18:36.228290 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 18:18:36 crc kubenswrapper[4859]: I1008 18:18:36.228299 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 18:18:36 crc kubenswrapper[4859]: I1008 18:18:36.228317 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 18:18:36 crc kubenswrapper[4859]: I1008 18:18:36.228327 4859 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T18:18:36Z","lastTransitionTime":"2025-10-08T18:18:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 18:18:36 crc kubenswrapper[4859]: I1008 18:18:36.331182 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 18:18:36 crc kubenswrapper[4859]: I1008 18:18:36.331267 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 18:18:36 crc kubenswrapper[4859]: I1008 18:18:36.331288 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 18:18:36 crc kubenswrapper[4859]: I1008 18:18:36.331314 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 18:18:36 crc kubenswrapper[4859]: I1008 18:18:36.331333 4859 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T18:18:36Z","lastTransitionTime":"2025-10-08T18:18:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 18:18:36 crc kubenswrapper[4859]: I1008 18:18:36.434765 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 18:18:36 crc kubenswrapper[4859]: I1008 18:18:36.434835 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 18:18:36 crc kubenswrapper[4859]: I1008 18:18:36.434878 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 18:18:36 crc kubenswrapper[4859]: I1008 18:18:36.434915 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 18:18:36 crc kubenswrapper[4859]: I1008 18:18:36.434937 4859 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T18:18:36Z","lastTransitionTime":"2025-10-08T18:18:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 18:18:36 crc kubenswrapper[4859]: I1008 18:18:36.539313 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 18:18:36 crc kubenswrapper[4859]: I1008 18:18:36.539482 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 18:18:36 crc kubenswrapper[4859]: I1008 18:18:36.539574 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 18:18:36 crc kubenswrapper[4859]: I1008 18:18:36.539611 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 18:18:36 crc kubenswrapper[4859]: I1008 18:18:36.539674 4859 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T18:18:36Z","lastTransitionTime":"2025-10-08T18:18:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 18:18:36 crc kubenswrapper[4859]: I1008 18:18:36.643199 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 18:18:36 crc kubenswrapper[4859]: I1008 18:18:36.643260 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 18:18:36 crc kubenswrapper[4859]: I1008 18:18:36.643282 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 18:18:36 crc kubenswrapper[4859]: I1008 18:18:36.643310 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 18:18:36 crc kubenswrapper[4859]: I1008 18:18:36.643335 4859 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T18:18:36Z","lastTransitionTime":"2025-10-08T18:18:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 18:18:36 crc kubenswrapper[4859]: I1008 18:18:36.747993 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 18:18:36 crc kubenswrapper[4859]: I1008 18:18:36.748062 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 18:18:36 crc kubenswrapper[4859]: I1008 18:18:36.748077 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 18:18:36 crc kubenswrapper[4859]: I1008 18:18:36.748104 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 18:18:36 crc kubenswrapper[4859]: I1008 18:18:36.748123 4859 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T18:18:36Z","lastTransitionTime":"2025-10-08T18:18:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 18:18:36 crc kubenswrapper[4859]: I1008 18:18:36.851191 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 18:18:36 crc kubenswrapper[4859]: I1008 18:18:36.851293 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 18:18:36 crc kubenswrapper[4859]: I1008 18:18:36.851320 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 18:18:36 crc kubenswrapper[4859]: I1008 18:18:36.851352 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 18:18:36 crc kubenswrapper[4859]: I1008 18:18:36.851373 4859 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T18:18:36Z","lastTransitionTime":"2025-10-08T18:18:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 18:18:36 crc kubenswrapper[4859]: I1008 18:18:36.954334 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 18:18:36 crc kubenswrapper[4859]: I1008 18:18:36.954375 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 18:18:36 crc kubenswrapper[4859]: I1008 18:18:36.954388 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 18:18:36 crc kubenswrapper[4859]: I1008 18:18:36.954403 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 18:18:36 crc kubenswrapper[4859]: I1008 18:18:36.954414 4859 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T18:18:36Z","lastTransitionTime":"2025-10-08T18:18:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 18:18:37 crc kubenswrapper[4859]: I1008 18:18:37.056892 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 18:18:37 crc kubenswrapper[4859]: I1008 18:18:37.056960 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 18:18:37 crc kubenswrapper[4859]: I1008 18:18:37.056978 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 18:18:37 crc kubenswrapper[4859]: I1008 18:18:37.057004 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 18:18:37 crc kubenswrapper[4859]: I1008 18:18:37.057020 4859 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T18:18:37Z","lastTransitionTime":"2025-10-08T18:18:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 18:18:37 crc kubenswrapper[4859]: I1008 18:18:37.160057 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 18:18:37 crc kubenswrapper[4859]: I1008 18:18:37.160116 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 18:18:37 crc kubenswrapper[4859]: I1008 18:18:37.160128 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 18:18:37 crc kubenswrapper[4859]: I1008 18:18:37.160148 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 18:18:37 crc kubenswrapper[4859]: I1008 18:18:37.160160 4859 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T18:18:37Z","lastTransitionTime":"2025-10-08T18:18:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 18:18:37 crc kubenswrapper[4859]: I1008 18:18:37.263021 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 18:18:37 crc kubenswrapper[4859]: I1008 18:18:37.263123 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 18:18:37 crc kubenswrapper[4859]: I1008 18:18:37.263148 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 18:18:37 crc kubenswrapper[4859]: I1008 18:18:37.263182 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 18:18:37 crc kubenswrapper[4859]: I1008 18:18:37.263208 4859 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T18:18:37Z","lastTransitionTime":"2025-10-08T18:18:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 18:18:37 crc kubenswrapper[4859]: I1008 18:18:37.366355 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 18:18:37 crc kubenswrapper[4859]: I1008 18:18:37.366874 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 18:18:37 crc kubenswrapper[4859]: I1008 18:18:37.367039 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 18:18:37 crc kubenswrapper[4859]: I1008 18:18:37.367221 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 18:18:37 crc kubenswrapper[4859]: I1008 18:18:37.367397 4859 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T18:18:37Z","lastTransitionTime":"2025-10-08T18:18:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 18:18:37 crc kubenswrapper[4859]: I1008 18:18:37.468930 4859 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 08 18:18:37 crc kubenswrapper[4859]: I1008 18:18:37.468930 4859 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 08 18:18:37 crc kubenswrapper[4859]: I1008 18:18:37.468972 4859 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-4b4sz" Oct 08 18:18:37 crc kubenswrapper[4859]: E1008 18:18:37.469180 4859 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 08 18:18:37 crc kubenswrapper[4859]: E1008 18:18:37.469481 4859 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-4b4sz" podUID="2e53cd9b-64f7-4e07-8b96-fb77847c9ac6" Oct 08 18:18:37 crc kubenswrapper[4859]: E1008 18:18:37.469644 4859 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 08 18:18:37 crc kubenswrapper[4859]: I1008 18:18:37.469867 4859 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 08 18:18:37 crc kubenswrapper[4859]: E1008 18:18:37.470030 4859 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 08 18:18:37 crc kubenswrapper[4859]: I1008 18:18:37.470461 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 18:18:37 crc kubenswrapper[4859]: I1008 18:18:37.470502 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 18:18:37 crc kubenswrapper[4859]: I1008 18:18:37.470514 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 18:18:37 crc kubenswrapper[4859]: I1008 18:18:37.470532 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 18:18:37 crc kubenswrapper[4859]: I1008 18:18:37.470542 4859 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T18:18:37Z","lastTransitionTime":"2025-10-08T18:18:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 18:18:37 crc kubenswrapper[4859]: I1008 18:18:37.574035 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 18:18:37 crc kubenswrapper[4859]: I1008 18:18:37.574106 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 18:18:37 crc kubenswrapper[4859]: I1008 18:18:37.574128 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 18:18:37 crc kubenswrapper[4859]: I1008 18:18:37.574157 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 18:18:37 crc kubenswrapper[4859]: I1008 18:18:37.574176 4859 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T18:18:37Z","lastTransitionTime":"2025-10-08T18:18:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 18:18:37 crc kubenswrapper[4859]: I1008 18:18:37.678147 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 18:18:37 crc kubenswrapper[4859]: I1008 18:18:37.678211 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 18:18:37 crc kubenswrapper[4859]: I1008 18:18:37.678229 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 18:18:37 crc kubenswrapper[4859]: I1008 18:18:37.678256 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 18:18:37 crc kubenswrapper[4859]: I1008 18:18:37.678276 4859 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T18:18:37Z","lastTransitionTime":"2025-10-08T18:18:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 18:18:37 crc kubenswrapper[4859]: I1008 18:18:37.781042 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 18:18:37 crc kubenswrapper[4859]: I1008 18:18:37.781097 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 18:18:37 crc kubenswrapper[4859]: I1008 18:18:37.781113 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 18:18:37 crc kubenswrapper[4859]: I1008 18:18:37.781134 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 18:18:37 crc kubenswrapper[4859]: I1008 18:18:37.781160 4859 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T18:18:37Z","lastTransitionTime":"2025-10-08T18:18:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 18:18:37 crc kubenswrapper[4859]: I1008 18:18:37.883882 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 18:18:37 crc kubenswrapper[4859]: I1008 18:18:37.883931 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 18:18:37 crc kubenswrapper[4859]: I1008 18:18:37.883941 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 18:18:37 crc kubenswrapper[4859]: I1008 18:18:37.883957 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 18:18:37 crc kubenswrapper[4859]: I1008 18:18:37.883968 4859 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T18:18:37Z","lastTransitionTime":"2025-10-08T18:18:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 18:18:37 crc kubenswrapper[4859]: I1008 18:18:37.986519 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 18:18:37 crc kubenswrapper[4859]: I1008 18:18:37.986579 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 18:18:37 crc kubenswrapper[4859]: I1008 18:18:37.986590 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 18:18:37 crc kubenswrapper[4859]: I1008 18:18:37.986611 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 18:18:37 crc kubenswrapper[4859]: I1008 18:18:37.986624 4859 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T18:18:37Z","lastTransitionTime":"2025-10-08T18:18:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 18:18:38 crc kubenswrapper[4859]: I1008 18:18:38.089096 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 18:18:38 crc kubenswrapper[4859]: I1008 18:18:38.089146 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 18:18:38 crc kubenswrapper[4859]: I1008 18:18:38.089156 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 18:18:38 crc kubenswrapper[4859]: I1008 18:18:38.089173 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 18:18:38 crc kubenswrapper[4859]: I1008 18:18:38.089193 4859 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T18:18:38Z","lastTransitionTime":"2025-10-08T18:18:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 18:18:38 crc kubenswrapper[4859]: I1008 18:18:38.192108 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 18:18:38 crc kubenswrapper[4859]: I1008 18:18:38.192162 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 18:18:38 crc kubenswrapper[4859]: I1008 18:18:38.192171 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 18:18:38 crc kubenswrapper[4859]: I1008 18:18:38.192188 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 18:18:38 crc kubenswrapper[4859]: I1008 18:18:38.192199 4859 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T18:18:38Z","lastTransitionTime":"2025-10-08T18:18:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 18:18:38 crc kubenswrapper[4859]: I1008 18:18:38.296225 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 18:18:38 crc kubenswrapper[4859]: I1008 18:18:38.296276 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 18:18:38 crc kubenswrapper[4859]: I1008 18:18:38.296286 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 18:18:38 crc kubenswrapper[4859]: I1008 18:18:38.296305 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 18:18:38 crc kubenswrapper[4859]: I1008 18:18:38.296318 4859 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T18:18:38Z","lastTransitionTime":"2025-10-08T18:18:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 18:18:38 crc kubenswrapper[4859]: I1008 18:18:38.399928 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 18:18:38 crc kubenswrapper[4859]: I1008 18:18:38.399977 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 18:18:38 crc kubenswrapper[4859]: I1008 18:18:38.399990 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 18:18:38 crc kubenswrapper[4859]: I1008 18:18:38.400008 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 18:18:38 crc kubenswrapper[4859]: I1008 18:18:38.400022 4859 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T18:18:38Z","lastTransitionTime":"2025-10-08T18:18:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 18:18:38 crc kubenswrapper[4859]: I1008 18:18:38.502484 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 18:18:38 crc kubenswrapper[4859]: I1008 18:18:38.502521 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 18:18:38 crc kubenswrapper[4859]: I1008 18:18:38.502530 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 18:18:38 crc kubenswrapper[4859]: I1008 18:18:38.502544 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 18:18:38 crc kubenswrapper[4859]: I1008 18:18:38.502553 4859 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T18:18:38Z","lastTransitionTime":"2025-10-08T18:18:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 18:18:38 crc kubenswrapper[4859]: I1008 18:18:38.604877 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 18:18:38 crc kubenswrapper[4859]: I1008 18:18:38.604919 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 18:18:38 crc kubenswrapper[4859]: I1008 18:18:38.604930 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 18:18:38 crc kubenswrapper[4859]: I1008 18:18:38.604946 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 18:18:38 crc kubenswrapper[4859]: I1008 18:18:38.604956 4859 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T18:18:38Z","lastTransitionTime":"2025-10-08T18:18:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 18:18:38 crc kubenswrapper[4859]: I1008 18:18:38.707505 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 18:18:38 crc kubenswrapper[4859]: I1008 18:18:38.707581 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 18:18:38 crc kubenswrapper[4859]: I1008 18:18:38.707607 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 18:18:38 crc kubenswrapper[4859]: I1008 18:18:38.707643 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 18:18:38 crc kubenswrapper[4859]: I1008 18:18:38.707663 4859 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T18:18:38Z","lastTransitionTime":"2025-10-08T18:18:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 18:18:38 crc kubenswrapper[4859]: I1008 18:18:38.757750 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 18:18:38 crc kubenswrapper[4859]: I1008 18:18:38.757844 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 18:18:38 crc kubenswrapper[4859]: I1008 18:18:38.757879 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 18:18:38 crc kubenswrapper[4859]: I1008 18:18:38.757913 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 18:18:38 crc kubenswrapper[4859]: I1008 18:18:38.757934 4859 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T18:18:38Z","lastTransitionTime":"2025-10-08T18:18:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 18:18:38 crc kubenswrapper[4859]: E1008 18:18:38.774154 4859 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404552Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865352Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-10-08T18:18:38Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-08T18:18:38Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-08T18:18:38Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-08T18:18:38Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-08T18:18:38Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-08T18:18:38Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-08T18:18:38Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-08T18:18:38Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"edb3ee6d-476b-407c-bb7d-4480786e8777\\\",\\\"systemUUID\\\":\\\"8f73c7ed-44aa-4d16-bae1-d4e684803cfc\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T18:18:38Z is after 2025-08-24T17:21:41Z" Oct 08 18:18:38 crc kubenswrapper[4859]: I1008 18:18:38.779378 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 18:18:38 crc kubenswrapper[4859]: I1008 18:18:38.779419 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 18:18:38 crc kubenswrapper[4859]: I1008 18:18:38.779428 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 18:18:38 crc kubenswrapper[4859]: I1008 18:18:38.779446 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 18:18:38 crc kubenswrapper[4859]: I1008 18:18:38.779460 4859 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T18:18:38Z","lastTransitionTime":"2025-10-08T18:18:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 18:18:38 crc kubenswrapper[4859]: E1008 18:18:38.794779 4859 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404552Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865352Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-10-08T18:18:38Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-08T18:18:38Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-08T18:18:38Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-08T18:18:38Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-08T18:18:38Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-08T18:18:38Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-08T18:18:38Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-08T18:18:38Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"edb3ee6d-476b-407c-bb7d-4480786e8777\\\",\\\"systemUUID\\\":\\\"8f73c7ed-44aa-4d16-bae1-d4e684803cfc\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T18:18:38Z is after 2025-08-24T17:21:41Z" Oct 08 18:18:38 crc kubenswrapper[4859]: I1008 18:18:38.800425 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 18:18:38 crc kubenswrapper[4859]: I1008 18:18:38.800461 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 18:18:38 crc kubenswrapper[4859]: I1008 18:18:38.800473 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 18:18:38 crc kubenswrapper[4859]: I1008 18:18:38.800491 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 18:18:38 crc kubenswrapper[4859]: I1008 18:18:38.800504 4859 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T18:18:38Z","lastTransitionTime":"2025-10-08T18:18:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 18:18:38 crc kubenswrapper[4859]: E1008 18:18:38.815330 4859 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404552Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865352Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-10-08T18:18:38Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-08T18:18:38Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-08T18:18:38Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-08T18:18:38Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-08T18:18:38Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-08T18:18:38Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-08T18:18:38Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-08T18:18:38Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"edb3ee6d-476b-407c-bb7d-4480786e8777\\\",\\\"systemUUID\\\":\\\"8f73c7ed-44aa-4d16-bae1-d4e684803cfc\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T18:18:38Z is after 2025-08-24T17:21:41Z" Oct 08 18:18:38 crc kubenswrapper[4859]: I1008 18:18:38.819974 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 18:18:38 crc kubenswrapper[4859]: I1008 18:18:38.820018 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 18:18:38 crc kubenswrapper[4859]: I1008 18:18:38.820029 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 18:18:38 crc kubenswrapper[4859]: I1008 18:18:38.820047 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 18:18:38 crc kubenswrapper[4859]: I1008 18:18:38.820059 4859 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T18:18:38Z","lastTransitionTime":"2025-10-08T18:18:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 18:18:38 crc kubenswrapper[4859]: E1008 18:18:38.838683 4859 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404552Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865352Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-10-08T18:18:38Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-08T18:18:38Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-08T18:18:38Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-08T18:18:38Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-08T18:18:38Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-08T18:18:38Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-08T18:18:38Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-08T18:18:38Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"edb3ee6d-476b-407c-bb7d-4480786e8777\\\",\\\"systemUUID\\\":\\\"8f73c7ed-44aa-4d16-bae1-d4e684803cfc\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T18:18:38Z is after 2025-08-24T17:21:41Z" Oct 08 18:18:38 crc kubenswrapper[4859]: I1008 18:18:38.843543 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 18:18:38 crc kubenswrapper[4859]: I1008 18:18:38.843643 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 18:18:38 crc kubenswrapper[4859]: I1008 18:18:38.843666 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 18:18:38 crc kubenswrapper[4859]: I1008 18:18:38.843729 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 18:18:38 crc kubenswrapper[4859]: I1008 18:18:38.843755 4859 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T18:18:38Z","lastTransitionTime":"2025-10-08T18:18:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 18:18:38 crc kubenswrapper[4859]: E1008 18:18:38.863668 4859 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404552Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865352Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-10-08T18:18:38Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-08T18:18:38Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-08T18:18:38Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-08T18:18:38Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-08T18:18:38Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-08T18:18:38Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-08T18:18:38Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-08T18:18:38Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"edb3ee6d-476b-407c-bb7d-4480786e8777\\\",\\\"systemUUID\\\":\\\"8f73c7ed-44aa-4d16-bae1-d4e684803cfc\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T18:18:38Z is after 2025-08-24T17:21:41Z" Oct 08 18:18:38 crc kubenswrapper[4859]: E1008 18:18:38.863810 4859 kubelet_node_status.go:572] "Unable to update node status" err="update node status exceeds retry count" Oct 08 18:18:38 crc kubenswrapper[4859]: I1008 18:18:38.866308 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 18:18:38 crc kubenswrapper[4859]: I1008 18:18:38.866349 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 18:18:38 crc kubenswrapper[4859]: I1008 18:18:38.866359 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 18:18:38 crc kubenswrapper[4859]: I1008 18:18:38.866378 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 18:18:38 crc kubenswrapper[4859]: I1008 18:18:38.866391 4859 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T18:18:38Z","lastTransitionTime":"2025-10-08T18:18:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 18:18:38 crc kubenswrapper[4859]: I1008 18:18:38.969909 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 18:18:38 crc kubenswrapper[4859]: I1008 18:18:38.969955 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 18:18:38 crc kubenswrapper[4859]: I1008 18:18:38.969966 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 18:18:38 crc kubenswrapper[4859]: I1008 18:18:38.969984 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 18:18:38 crc kubenswrapper[4859]: I1008 18:18:38.969997 4859 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T18:18:38Z","lastTransitionTime":"2025-10-08T18:18:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 18:18:39 crc kubenswrapper[4859]: I1008 18:18:39.073767 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 18:18:39 crc kubenswrapper[4859]: I1008 18:18:39.073850 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 18:18:39 crc kubenswrapper[4859]: I1008 18:18:39.073875 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 18:18:39 crc kubenswrapper[4859]: I1008 18:18:39.073911 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 18:18:39 crc kubenswrapper[4859]: I1008 18:18:39.073937 4859 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T18:18:39Z","lastTransitionTime":"2025-10-08T18:18:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 18:18:39 crc kubenswrapper[4859]: I1008 18:18:39.177897 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 18:18:39 crc kubenswrapper[4859]: I1008 18:18:39.177966 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 18:18:39 crc kubenswrapper[4859]: I1008 18:18:39.177982 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 18:18:39 crc kubenswrapper[4859]: I1008 18:18:39.178010 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 18:18:39 crc kubenswrapper[4859]: I1008 18:18:39.178026 4859 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T18:18:39Z","lastTransitionTime":"2025-10-08T18:18:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 18:18:39 crc kubenswrapper[4859]: I1008 18:18:39.281052 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 18:18:39 crc kubenswrapper[4859]: I1008 18:18:39.281115 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 18:18:39 crc kubenswrapper[4859]: I1008 18:18:39.281144 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 18:18:39 crc kubenswrapper[4859]: I1008 18:18:39.281175 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 18:18:39 crc kubenswrapper[4859]: I1008 18:18:39.281196 4859 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T18:18:39Z","lastTransitionTime":"2025-10-08T18:18:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 18:18:39 crc kubenswrapper[4859]: I1008 18:18:39.384282 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 18:18:39 crc kubenswrapper[4859]: I1008 18:18:39.384357 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 18:18:39 crc kubenswrapper[4859]: I1008 18:18:39.384376 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 18:18:39 crc kubenswrapper[4859]: I1008 18:18:39.384403 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 18:18:39 crc kubenswrapper[4859]: I1008 18:18:39.384423 4859 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T18:18:39Z","lastTransitionTime":"2025-10-08T18:18:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 18:18:39 crc kubenswrapper[4859]: I1008 18:18:39.469646 4859 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 08 18:18:39 crc kubenswrapper[4859]: I1008 18:18:39.469792 4859 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 08 18:18:39 crc kubenswrapper[4859]: I1008 18:18:39.469834 4859 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-4b4sz" Oct 08 18:18:39 crc kubenswrapper[4859]: E1008 18:18:39.469892 4859 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 08 18:18:39 crc kubenswrapper[4859]: I1008 18:18:39.469907 4859 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 08 18:18:39 crc kubenswrapper[4859]: E1008 18:18:39.470038 4859 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 08 18:18:39 crc kubenswrapper[4859]: E1008 18:18:39.470121 4859 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 08 18:18:39 crc kubenswrapper[4859]: E1008 18:18:39.470203 4859 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-4b4sz" podUID="2e53cd9b-64f7-4e07-8b96-fb77847c9ac6" Oct 08 18:18:39 crc kubenswrapper[4859]: I1008 18:18:39.487483 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 18:18:39 crc kubenswrapper[4859]: I1008 18:18:39.487548 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 18:18:39 crc kubenswrapper[4859]: I1008 18:18:39.487565 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 18:18:39 crc kubenswrapper[4859]: I1008 18:18:39.487592 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 18:18:39 crc kubenswrapper[4859]: I1008 18:18:39.487614 4859 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T18:18:39Z","lastTransitionTime":"2025-10-08T18:18:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 18:18:39 crc kubenswrapper[4859]: I1008 18:18:39.591492 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 18:18:39 crc kubenswrapper[4859]: I1008 18:18:39.591548 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 18:18:39 crc kubenswrapper[4859]: I1008 18:18:39.591565 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 18:18:39 crc kubenswrapper[4859]: I1008 18:18:39.591590 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 18:18:39 crc kubenswrapper[4859]: I1008 18:18:39.591609 4859 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T18:18:39Z","lastTransitionTime":"2025-10-08T18:18:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 18:18:39 crc kubenswrapper[4859]: I1008 18:18:39.695349 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 18:18:39 crc kubenswrapper[4859]: I1008 18:18:39.695417 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 18:18:39 crc kubenswrapper[4859]: I1008 18:18:39.695435 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 18:18:39 crc kubenswrapper[4859]: I1008 18:18:39.695463 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 18:18:39 crc kubenswrapper[4859]: I1008 18:18:39.695484 4859 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T18:18:39Z","lastTransitionTime":"2025-10-08T18:18:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 18:18:39 crc kubenswrapper[4859]: I1008 18:18:39.799341 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 18:18:39 crc kubenswrapper[4859]: I1008 18:18:39.799460 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 18:18:39 crc kubenswrapper[4859]: I1008 18:18:39.799479 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 18:18:39 crc kubenswrapper[4859]: I1008 18:18:39.799504 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 18:18:39 crc kubenswrapper[4859]: I1008 18:18:39.799522 4859 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T18:18:39Z","lastTransitionTime":"2025-10-08T18:18:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 18:18:39 crc kubenswrapper[4859]: I1008 18:18:39.903895 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 18:18:39 crc kubenswrapper[4859]: I1008 18:18:39.903992 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 18:18:39 crc kubenswrapper[4859]: I1008 18:18:39.904011 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 18:18:39 crc kubenswrapper[4859]: I1008 18:18:39.904038 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 18:18:39 crc kubenswrapper[4859]: I1008 18:18:39.904056 4859 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T18:18:39Z","lastTransitionTime":"2025-10-08T18:18:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 18:18:40 crc kubenswrapper[4859]: I1008 18:18:40.006884 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 18:18:40 crc kubenswrapper[4859]: I1008 18:18:40.006945 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 18:18:40 crc kubenswrapper[4859]: I1008 18:18:40.006963 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 18:18:40 crc kubenswrapper[4859]: I1008 18:18:40.006986 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 18:18:40 crc kubenswrapper[4859]: I1008 18:18:40.007003 4859 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T18:18:40Z","lastTransitionTime":"2025-10-08T18:18:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 18:18:40 crc kubenswrapper[4859]: I1008 18:18:40.110184 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 18:18:40 crc kubenswrapper[4859]: I1008 18:18:40.110248 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 18:18:40 crc kubenswrapper[4859]: I1008 18:18:40.110267 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 18:18:40 crc kubenswrapper[4859]: I1008 18:18:40.110292 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 18:18:40 crc kubenswrapper[4859]: I1008 18:18:40.110320 4859 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T18:18:40Z","lastTransitionTime":"2025-10-08T18:18:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 18:18:40 crc kubenswrapper[4859]: I1008 18:18:40.213515 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 18:18:40 crc kubenswrapper[4859]: I1008 18:18:40.213595 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 18:18:40 crc kubenswrapper[4859]: I1008 18:18:40.213619 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 18:18:40 crc kubenswrapper[4859]: I1008 18:18:40.213649 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 18:18:40 crc kubenswrapper[4859]: I1008 18:18:40.213670 4859 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T18:18:40Z","lastTransitionTime":"2025-10-08T18:18:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 18:18:40 crc kubenswrapper[4859]: I1008 18:18:40.317532 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 18:18:40 crc kubenswrapper[4859]: I1008 18:18:40.317599 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 18:18:40 crc kubenswrapper[4859]: I1008 18:18:40.317618 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 18:18:40 crc kubenswrapper[4859]: I1008 18:18:40.317649 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 18:18:40 crc kubenswrapper[4859]: I1008 18:18:40.317669 4859 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T18:18:40Z","lastTransitionTime":"2025-10-08T18:18:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 18:18:40 crc kubenswrapper[4859]: I1008 18:18:40.422279 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 18:18:40 crc kubenswrapper[4859]: I1008 18:18:40.422384 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 18:18:40 crc kubenswrapper[4859]: I1008 18:18:40.422408 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 18:18:40 crc kubenswrapper[4859]: I1008 18:18:40.422443 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 18:18:40 crc kubenswrapper[4859]: I1008 18:18:40.422465 4859 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T18:18:40Z","lastTransitionTime":"2025-10-08T18:18:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 18:18:40 crc kubenswrapper[4859]: I1008 18:18:40.494508 4859 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4efc16fb-e679-464d-8b3c-0f46f082593e\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:17:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:17:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:17:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:17:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:17:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://aa8f8ea895272cd79a08d1c47cd7863a1ec10e21bb60d12aa10862a56c99e4bf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T18:17:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5c4e95d8a712ab219455564539a90286ae48bbc46ddc322a5bfd96c1a8268110\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T18:17:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://bae6ecfd8b446ef04d393d823e5ffe37fe749337c0ca17fde79c5f88d5e31393\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T18:17:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6411d1e7e18d5ceadaaceed5a8a6b20a0cacc55cede793ca9fb4dcad70c64fc6\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T18:17:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T18:17:10Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T18:18:40Z is after 2025-08-24T17:21:41Z" Oct 08 18:18:40 crc kubenswrapper[4859]: I1008 18:18:40.516369 4859 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-08T18:17:29Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T18:17:29Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T18:18:40Z is after 2025-08-24T17:21:41Z" Oct 08 18:18:40 crc kubenswrapper[4859]: I1008 18:18:40.526058 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 18:18:40 crc kubenswrapper[4859]: I1008 18:18:40.526118 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 18:18:40 crc kubenswrapper[4859]: I1008 18:18:40.526138 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 18:18:40 crc kubenswrapper[4859]: I1008 18:18:40.526165 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 18:18:40 crc kubenswrapper[4859]: I1008 18:18:40.526184 4859 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T18:18:40Z","lastTransitionTime":"2025-10-08T18:18:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 18:18:40 crc kubenswrapper[4859]: I1008 18:18:40.544609 4859 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-84k8x" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1c24078c-c9e8-4160-883a-f483b4b3dd23\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:17:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:17:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:17:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:17:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8661006834d0f50eaada7394369584ee95fc61ff54437eec35d755526afc5b48\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T18:17:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hk5xb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://18b291eb3344fb7c00ef5679dca9a6f7ed9d08cea684a803ff7dad2e8a18c585\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://18b291eb3344fb7c00ef5679dca9a6f7ed9d08cea684a803ff7dad2e8a18c585\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-08T18:17:32Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-08T18:17:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hk5xb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://160163f1a91d981583452db3c7a9244182d1c8ced45749a61b721d0d45c1dfe4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://160163f1a91d981583452db3c7a9244182d1c8ced45749a61b721d0d45c1dfe4\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-08T18:17:34Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-08T18:17:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hk5xb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://51e9b53f3847ac53f378ad83870debf8e71518d4edbc7fc0dc66644818e2103e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://51e9b53f3847ac53f378ad83870debf8e71518d4edbc7fc0dc66644818e2103e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-08T18:17:34Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-08T18:17:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hk5xb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1ff118919739759ac01b5cbe3240e6673db54eaa16353c609d670071ed7964cb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1ff118919739759ac01b5cbe3240e6673db54eaa16353c609d670071ed7964cb\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-08T18:17:35Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-08T18:17:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hk5xb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8522e803ff56eb57cd4d9fd036e0c3fb16240ea21d4a094c1db7aa7d50bfe2c2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8522e803ff56eb57cd4d9fd036e0c3fb16240ea21d4a094c1db7aa7d50bfe2c2\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-08T18:17:37Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-08T18:17:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hk5xb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://be61056edf515e037403ca2a0bfd5feb9b2537dd42212b58e133b23b294f327b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://be61056edf515e037403ca2a0bfd5feb9b2537dd42212b58e133b23b294f327b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-08T18:17:38Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-08T18:17:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hk5xb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T18:17:31Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-84k8x\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T18:18:40Z is after 2025-08-24T17:21:41Z" Oct 08 18:18:40 crc kubenswrapper[4859]: I1008 18:18:40.568048 4859 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-8plkg" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"92b2d7de-31cc-4a91-95a5-ed5ea964b028\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:17:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:17:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:18:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:18:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://729bd32fd01dcab495bffa64f2f5e44990fc1d78b05e7f131868009f92305831\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://fa00b8345d40b74a917ddc23cc1b2c5c684a9ca61b6119d090c92cc1439b71c4\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-10-08T18:18:19Z\\\",\\\"message\\\":\\\"2025-10-08T18:17:33+00:00 [cnibincopy] Successfully copied files in /usr/src/multus-cni/rhel9/bin/ to /host/opt/cni/bin/upgrade_888b130f-8754-48c6-b73e-9ff6772793f9\\\\n2025-10-08T18:17:33+00:00 [cnibincopy] Successfully moved files in /host/opt/cni/bin/upgrade_888b130f-8754-48c6-b73e-9ff6772793f9 to /host/opt/cni/bin/\\\\n2025-10-08T18:17:34Z [verbose] multus-daemon started\\\\n2025-10-08T18:17:34Z [verbose] Readiness Indicator file check\\\\n2025-10-08T18:18:19Z [error] have you checked that your default network is ready? still waiting for readinessindicatorfile @ /host/run/multus/cni/net.d/10-ovn-kubernetes.conf. pollimmediate error: timed out waiting for the condition\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-08T18:17:32Z\\\"}},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T18:18:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q9d7m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T18:17:31Z\\\"}}\" for pod \"openshift-multus\"/\"multus-8plkg\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T18:18:40Z is after 2025-08-24T17:21:41Z" Oct 08 18:18:40 crc kubenswrapper[4859]: I1008 18:18:40.583752 4859 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-btjmq" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b8605f7-fed7-493f-9b4f-6a6358adf907\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:17:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:17:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:17:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:17:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://14e9b6c4a4403f36adb78fcf1135a620ca6fc56ed974c9639dcdfc59e6917559\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T18:17:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xspg4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T18:17:35Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-btjmq\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T18:18:40Z is after 2025-08-24T17:21:41Z" Oct 08 18:18:40 crc kubenswrapper[4859]: I1008 18:18:40.597808 4859 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-g4t79" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3a92dbf1-21a8-4095-96bb-2d701d7f3d56\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:17:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:17:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:17:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:17:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f8772a25b6e605efe9a062dc32dbbc2cfe32b4fe164b4e84fbf3bcffab744179\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T18:17:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sv79p\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://92cb9986c09a7f9e1e552e3e3ac1f74f04128241e00ddee38882d2ba5c005a67\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T18:17:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-sv79p\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T18:17:45Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-g4t79\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T18:18:40Z is after 2025-08-24T17:21:41Z" Oct 08 18:18:40 crc kubenswrapper[4859]: I1008 18:18:40.611887 4859 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3ca6ce76-77e8-4e46-a2f6-b1ed4aa0fd4b\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:17:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:17:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:18:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:18:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:17:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f587e149d57ea53626ed3beb83e421c3b89ba1fec9e949705e8ccb5b4f7cc79a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T18:17:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://428cbe63daefbfbcb2ec76dad0fe638cdc65934bac80b791cebe0b36b2d6909a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T18:17:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a41399a37b1ad86a063001224232ecb8f7ea56aeb5686d3858bc08102a0c8f49\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T18:17:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://008757fce552591f7c58193f605eb60792761d5943927cbd63184549618d2b43\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://008757fce552591f7c58193f605eb60792761d5943927cbd63184549618d2b43\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-08T18:17:11Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-08T18:17:11Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T18:17:10Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T18:18:40Z is after 2025-08-24T17:21:41Z" Oct 08 18:18:40 crc kubenswrapper[4859]: I1008 18:18:40.629473 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 18:18:40 crc kubenswrapper[4859]: I1008 18:18:40.629516 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 18:18:40 crc kubenswrapper[4859]: I1008 18:18:40.629531 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 18:18:40 crc kubenswrapper[4859]: I1008 18:18:40.629555 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 18:18:40 crc kubenswrapper[4859]: I1008 18:18:40.629573 4859 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T18:18:40Z","lastTransitionTime":"2025-10-08T18:18:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 18:18:40 crc kubenswrapper[4859]: I1008 18:18:40.637492 4859 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-08T18:17:29Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T18:17:29Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T18:18:40Z is after 2025-08-24T17:21:41Z" Oct 08 18:18:40 crc kubenswrapper[4859]: I1008 18:18:40.652717 4859 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-btw28" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"97bc5598-3ead-4300-a938-3aebbff1e8c6\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:17:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:17:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:17:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:17:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d6d4d6ba771cb5fde637449bba05660f634df7e833a0c8dad2b2c7f3a422bb2b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T18:17:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-qcxsd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T18:17:31Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-btw28\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T18:18:40Z is after 2025-08-24T17:21:41Z" Oct 08 18:18:40 crc kubenswrapper[4859]: I1008 18:18:40.676505 4859 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-08T18:17:31Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T18:17:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T18:17:31Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c6d15905694b6bbd4999d49dc8f7a53eb6bfaa7adb59efb0d7e4973c21d1c7be\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T18:17:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T18:18:40Z is after 2025-08-24T17:21:41Z" Oct 08 18:18:40 crc kubenswrapper[4859]: I1008 18:18:40.693795 4859 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-08T18:17:33Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T18:17:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T18:17:33Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5d99e071bd083914418d9c128f38955f26295c08b1520f2dc774c99b1c843693\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T18:17:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T18:18:40Z is after 2025-08-24T17:21:41Z" Oct 08 18:18:40 crc kubenswrapper[4859]: I1008 18:18:40.708977 4859 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-08T18:17:31Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T18:17:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T18:17:31Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://99e2d51bd7757498dae60949eea6d732b5f7b481fa00f808dd5308a530a1daf1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T18:17:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e32aaba93699944665660ae41e9b5779178c1354c31e15ad8188bf8c3c8fa500\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T18:17:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T18:18:40Z is after 2025-08-24T17:21:41Z" Oct 08 18:18:40 crc kubenswrapper[4859]: I1008 18:18:40.723334 4859 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-08T18:17:29Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T18:17:29Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T18:18:40Z is after 2025-08-24T17:21:41Z" Oct 08 18:18:40 crc kubenswrapper[4859]: I1008 18:18:40.732307 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 18:18:40 crc kubenswrapper[4859]: I1008 18:18:40.732340 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 18:18:40 crc kubenswrapper[4859]: I1008 18:18:40.732350 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 18:18:40 crc kubenswrapper[4859]: I1008 18:18:40.732365 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 18:18:40 crc kubenswrapper[4859]: I1008 18:18:40.732374 4859 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T18:18:40Z","lastTransitionTime":"2025-10-08T18:18:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 18:18:40 crc kubenswrapper[4859]: I1008 18:18:40.740335 4859 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-82s52" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b23a6a6c-9d92-4e7b-840e-55cfda873a2d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:17:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:17:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:17:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:17:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://71302c766485b9f55743d85f97440ff578cd44ba1e5a6f5b1c024c0e808072d2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T18:17:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6vx22\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c23ea3580998adf90621a0155c20ee4bd590c512629812174e95c28857bcfef9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T18:17:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6vx22\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T18:17:31Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-82s52\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T18:18:40Z is after 2025-08-24T17:21:41Z" Oct 08 18:18:40 crc kubenswrapper[4859]: I1008 18:18:40.767844 4859 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-4g8gf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1dff864e-b75d-4e0b-b182-75f710eac8df\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:17:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:17:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:17:32Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:17:32Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://768ca274dedcc6847fc21d15234d3665fa9ccae371b5b6a882cfcb9b182f527d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T18:17:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s8jnn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://669f5f3c6d3c1ec0c4b4870e090b0a853adc01727c23383e595715b476c32668\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T18:17:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s8jnn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b8009a42cc4c9a816457a0678bdb2f7b7c53d0c8604e007a6fb141f66d06a516\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T18:17:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s8jnn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://676511a11ed4102084b57ebdd5f3fda4969741b2bc8443b415c4bf94f39ae4c9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T18:17:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s8jnn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4a74c1f0d21c90cdd7e776ee14e1f48757a04e2220fbd1f884449181ec6d1c3a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T18:17:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s8jnn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6ddc0e842bb8f81663406070023b0e5ed9cc99c98db0123b1c84f20da59fb5b5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T18:17:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s8jnn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f3a760c557ef21fbc6b54e78c3f6c5b325f296f93a022425cca7b1399eed8661\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f3a760c557ef21fbc6b54e78c3f6c5b325f296f93a022425cca7b1399eed8661\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-10-08T18:18:31Z\\\",\\\"message\\\":\\\"etrics-daemon-4b4sz\\\\nI1008 18:18:31.553175 6912 obj_retry.go:303] Retry object setup: *v1.Pod openshift-machine-config-operator/machine-config-daemon-82s52\\\\nI1008 18:18:31.553182 6912 obj_retry.go:365] Adding new object: *v1.Pod openshift-machine-config-operator/machine-config-daemon-82s52\\\\nI1008 18:18:31.553190 6912 ovn.go:134] Ensuring zone local for Pod openshift-machine-config-operator/machine-config-daemon-82s52 in node crc\\\\nI1008 18:18:31.553195 6912 obj_retry.go:386] Retry successful for *v1.Pod openshift-machine-config-operator/machine-config-daemon-82s52 after 0 failed attempt(s)\\\\nI1008 18:18:31.553199 6912 default_network_controller.go:776] Recording success event on pod openshift-machine-config-operator/machine-config-daemon-82s52\\\\nI1008 18:18:31.553197 6912 obj_retry.go:303] Retry object setup: *v1.Pod openshift-network-diagnostics/network-check-source-55646444c4-trplf\\\\nI1008 18:18:31.553210 6912 obj_retry.go:365] Adding new object: *v1.Pod openshift-network-diagnostics/network-check-source-55646444c4-trplf\\\\nF1008 18:18:31.553216 6912 ovnkube.go:137] failed to run ovnkube: [failed to start network controller: failed to start default network controller: unable to create admin network policy controller, err: could not add Event Handler for anpInformer during admin network policy controller initialization, handler {0x1\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-08T18:18:30Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 40s restarting failed container=ovnkube-controller pod=ovnkube-node-4g8gf_openshift-ovn-kubernetes(1dff864e-b75d-4e0b-b182-75f710eac8df)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s8jnn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2c67d2aed7126d76989e057d8e2d8e188a8902b2f3fe26ed6d1d45d3b013d897\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T18:17:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s8jnn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7a211d3da02b173867ff3a912eecd3b379483c7cc77e6b91122edd1ffd30fd53\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7a211d3da02b173867ff3a912eecd3b379483c7cc77e6b91122edd1ffd30fd53\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-08T18:17:32Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-08T18:17:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s8jnn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T18:17:32Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-4g8gf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T18:18:40Z is after 2025-08-24T17:21:41Z" Oct 08 18:18:40 crc kubenswrapper[4859]: I1008 18:18:40.784055 4859 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-4b4sz" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"2e53cd9b-64f7-4e07-8b96-fb77847c9ac6\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:17:45Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:17:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:17:45Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:17:45Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kxgbz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-kxgbz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T18:17:45Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-4b4sz\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T18:18:40Z is after 2025-08-24T17:21:41Z" Oct 08 18:18:40 crc kubenswrapper[4859]: I1008 18:18:40.797475 4859 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"262c958c-c49d-4c11-96ae-94106dd123d4\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:17:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:17:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:17:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:17:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:17:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a54a54b9a831ead1fbe6e03abbb622242dd16b4081a4482d559e03d47aebd2fe\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-crio\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T18:17:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kube\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3ab3bc5e8e2fd4941f30cbf8a90d9a9078b90fab91063c07da2cd48b63eef657\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3ab3bc5e8e2fd4941f30cbf8a90d9a9078b90fab91063c07da2cd48b63eef657\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-08T18:17:11Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-08T18:17:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T18:17:10Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"kube-rbac-proxy-crio-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T18:18:40Z is after 2025-08-24T17:21:41Z" Oct 08 18:18:40 crc kubenswrapper[4859]: I1008 18:18:40.813750 4859 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"63a623de-5608-4228-b1c2-77188c276986\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:17:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:17:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:17:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:17:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:17:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d4786a8fb8abd96fb22e60733c974f599017407d589ed815fb8541fd3b64f696\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T18:17:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://183d481502ed5108d5e19e483b714c289f355ea3f8b0891883e5335b802c83b6\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T18:17:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ec848c46fd5dde0469b00d3cab14ba4a652533a265490fadbee7a8b8843ef249\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T18:17:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4fe1828b4d63ff8106bfcd15e1dc1c1bd1ba7f77777cdb23ff1b291d0418774c\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2e700b2a20c81b414617f5b5c018a738278a3fa44e6c6d14b2170051d0725431\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-08T18:17:24Z\\\",\\\"message\\\":\\\"W1008 18:17:13.670625 1 cmd.go:257] Using insecure, self-signed certificates\\\\nI1008 18:17:13.671202 1 crypto.go:601] Generating new CA for check-endpoints-signer@1759947433 cert, and key in /tmp/serving-cert-3867218267/serving-signer.crt, /tmp/serving-cert-3867218267/serving-signer.key\\\\nI1008 18:17:13.944463 1 observer_polling.go:159] Starting file observer\\\\nW1008 18:17:13.951958 1 builder.go:272] unable to get owner reference (falling back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": dial tcp [::1]:6443: connect: connection refused\\\\nI1008 18:17:13.952088 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1008 18:17:13.956755 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3867218267/tls.crt::/tmp/serving-cert-3867218267/tls.key\\\\\\\"\\\\nF1008 18:17:24.290991 1 cmd.go:182] error initializing delegating authentication: unable to load configmap based request-header-client-ca-file: Get \\\\\\\"https://localhost:6443/api/v1/namespaces/kube-system/configmaps/extension-apiserver-authentication\\\\\\\": net/http: TLS handshake timeout\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-08T18:17:13Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T18:17:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c65b66f62d325ca10ef8d5d31f04529ccf4e691cf3303d231e7a414fb400e4d6\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T18:17:13Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://58c7bba4ace28215b062653a960e561b1dc1f890f2baf450145638a065af038b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://58c7bba4ace28215b062653a960e561b1dc1f890f2baf450145638a065af038b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-08T18:17:11Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-08T18:17:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T18:17:10Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T18:18:40Z is after 2025-08-24T17:21:41Z" Oct 08 18:18:40 crc kubenswrapper[4859]: I1008 18:18:40.834940 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 18:18:40 crc kubenswrapper[4859]: I1008 18:18:40.834979 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 18:18:40 crc kubenswrapper[4859]: I1008 18:18:40.834988 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 18:18:40 crc kubenswrapper[4859]: I1008 18:18:40.835003 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 18:18:40 crc kubenswrapper[4859]: I1008 18:18:40.835016 4859 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T18:18:40Z","lastTransitionTime":"2025-10-08T18:18:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 18:18:40 crc kubenswrapper[4859]: I1008 18:18:40.936955 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 18:18:40 crc kubenswrapper[4859]: I1008 18:18:40.937038 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 18:18:40 crc kubenswrapper[4859]: I1008 18:18:40.937056 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 18:18:40 crc kubenswrapper[4859]: I1008 18:18:40.937076 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 18:18:40 crc kubenswrapper[4859]: I1008 18:18:40.937089 4859 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T18:18:40Z","lastTransitionTime":"2025-10-08T18:18:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 18:18:41 crc kubenswrapper[4859]: I1008 18:18:41.041088 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 18:18:41 crc kubenswrapper[4859]: I1008 18:18:41.041144 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 18:18:41 crc kubenswrapper[4859]: I1008 18:18:41.041157 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 18:18:41 crc kubenswrapper[4859]: I1008 18:18:41.041180 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 18:18:41 crc kubenswrapper[4859]: I1008 18:18:41.041192 4859 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T18:18:41Z","lastTransitionTime":"2025-10-08T18:18:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 18:18:41 crc kubenswrapper[4859]: I1008 18:18:41.144005 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 18:18:41 crc kubenswrapper[4859]: I1008 18:18:41.144055 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 18:18:41 crc kubenswrapper[4859]: I1008 18:18:41.144067 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 18:18:41 crc kubenswrapper[4859]: I1008 18:18:41.144084 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 18:18:41 crc kubenswrapper[4859]: I1008 18:18:41.144096 4859 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T18:18:41Z","lastTransitionTime":"2025-10-08T18:18:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 18:18:41 crc kubenswrapper[4859]: I1008 18:18:41.247232 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 18:18:41 crc kubenswrapper[4859]: I1008 18:18:41.247269 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 18:18:41 crc kubenswrapper[4859]: I1008 18:18:41.247278 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 18:18:41 crc kubenswrapper[4859]: I1008 18:18:41.247292 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 18:18:41 crc kubenswrapper[4859]: I1008 18:18:41.247303 4859 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T18:18:41Z","lastTransitionTime":"2025-10-08T18:18:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 18:18:41 crc kubenswrapper[4859]: I1008 18:18:41.350927 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 18:18:41 crc kubenswrapper[4859]: I1008 18:18:41.350988 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 18:18:41 crc kubenswrapper[4859]: I1008 18:18:41.351005 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 18:18:41 crc kubenswrapper[4859]: I1008 18:18:41.351033 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 18:18:41 crc kubenswrapper[4859]: I1008 18:18:41.351052 4859 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T18:18:41Z","lastTransitionTime":"2025-10-08T18:18:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 18:18:41 crc kubenswrapper[4859]: I1008 18:18:41.454365 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 18:18:41 crc kubenswrapper[4859]: I1008 18:18:41.455061 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 18:18:41 crc kubenswrapper[4859]: I1008 18:18:41.455104 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 18:18:41 crc kubenswrapper[4859]: I1008 18:18:41.455140 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 18:18:41 crc kubenswrapper[4859]: I1008 18:18:41.455165 4859 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T18:18:41Z","lastTransitionTime":"2025-10-08T18:18:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 18:18:41 crc kubenswrapper[4859]: I1008 18:18:41.469293 4859 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 08 18:18:41 crc kubenswrapper[4859]: I1008 18:18:41.469355 4859 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-4b4sz" Oct 08 18:18:41 crc kubenswrapper[4859]: I1008 18:18:41.469446 4859 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 08 18:18:41 crc kubenswrapper[4859]: I1008 18:18:41.469779 4859 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 08 18:18:41 crc kubenswrapper[4859]: E1008 18:18:41.469942 4859 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 08 18:18:41 crc kubenswrapper[4859]: E1008 18:18:41.470061 4859 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 08 18:18:41 crc kubenswrapper[4859]: E1008 18:18:41.470185 4859 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 08 18:18:41 crc kubenswrapper[4859]: E1008 18:18:41.470363 4859 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-4b4sz" podUID="2e53cd9b-64f7-4e07-8b96-fb77847c9ac6" Oct 08 18:18:41 crc kubenswrapper[4859]: I1008 18:18:41.485928 4859 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-etcd/etcd-crc"] Oct 08 18:18:41 crc kubenswrapper[4859]: I1008 18:18:41.558171 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 18:18:41 crc kubenswrapper[4859]: I1008 18:18:41.558212 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 18:18:41 crc kubenswrapper[4859]: I1008 18:18:41.558220 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 18:18:41 crc kubenswrapper[4859]: I1008 18:18:41.558236 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 18:18:41 crc kubenswrapper[4859]: I1008 18:18:41.558249 4859 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T18:18:41Z","lastTransitionTime":"2025-10-08T18:18:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 18:18:41 crc kubenswrapper[4859]: I1008 18:18:41.660988 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 18:18:41 crc kubenswrapper[4859]: I1008 18:18:41.661053 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 18:18:41 crc kubenswrapper[4859]: I1008 18:18:41.661074 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 18:18:41 crc kubenswrapper[4859]: I1008 18:18:41.661103 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 18:18:41 crc kubenswrapper[4859]: I1008 18:18:41.661121 4859 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T18:18:41Z","lastTransitionTime":"2025-10-08T18:18:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 18:18:41 crc kubenswrapper[4859]: I1008 18:18:41.764424 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 18:18:41 crc kubenswrapper[4859]: I1008 18:18:41.764516 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 18:18:41 crc kubenswrapper[4859]: I1008 18:18:41.764542 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 18:18:41 crc kubenswrapper[4859]: I1008 18:18:41.764577 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 18:18:41 crc kubenswrapper[4859]: I1008 18:18:41.764597 4859 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T18:18:41Z","lastTransitionTime":"2025-10-08T18:18:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 18:18:41 crc kubenswrapper[4859]: I1008 18:18:41.868067 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 18:18:41 crc kubenswrapper[4859]: I1008 18:18:41.868136 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 18:18:41 crc kubenswrapper[4859]: I1008 18:18:41.868154 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 18:18:41 crc kubenswrapper[4859]: I1008 18:18:41.868192 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 18:18:41 crc kubenswrapper[4859]: I1008 18:18:41.868220 4859 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T18:18:41Z","lastTransitionTime":"2025-10-08T18:18:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 18:18:41 crc kubenswrapper[4859]: I1008 18:18:41.971479 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 18:18:41 crc kubenswrapper[4859]: I1008 18:18:41.971537 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 18:18:41 crc kubenswrapper[4859]: I1008 18:18:41.971551 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 18:18:41 crc kubenswrapper[4859]: I1008 18:18:41.971574 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 18:18:41 crc kubenswrapper[4859]: I1008 18:18:41.971589 4859 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T18:18:41Z","lastTransitionTime":"2025-10-08T18:18:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 18:18:42 crc kubenswrapper[4859]: I1008 18:18:42.074774 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 18:18:42 crc kubenswrapper[4859]: I1008 18:18:42.074831 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 18:18:42 crc kubenswrapper[4859]: I1008 18:18:42.074841 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 18:18:42 crc kubenswrapper[4859]: I1008 18:18:42.074863 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 18:18:42 crc kubenswrapper[4859]: I1008 18:18:42.074875 4859 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T18:18:42Z","lastTransitionTime":"2025-10-08T18:18:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 18:18:42 crc kubenswrapper[4859]: I1008 18:18:42.178161 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 18:18:42 crc kubenswrapper[4859]: I1008 18:18:42.178210 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 18:18:42 crc kubenswrapper[4859]: I1008 18:18:42.178223 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 18:18:42 crc kubenswrapper[4859]: I1008 18:18:42.178240 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 18:18:42 crc kubenswrapper[4859]: I1008 18:18:42.178252 4859 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T18:18:42Z","lastTransitionTime":"2025-10-08T18:18:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 18:18:42 crc kubenswrapper[4859]: I1008 18:18:42.281241 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 18:18:42 crc kubenswrapper[4859]: I1008 18:18:42.281300 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 18:18:42 crc kubenswrapper[4859]: I1008 18:18:42.281310 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 18:18:42 crc kubenswrapper[4859]: I1008 18:18:42.281329 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 18:18:42 crc kubenswrapper[4859]: I1008 18:18:42.281340 4859 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T18:18:42Z","lastTransitionTime":"2025-10-08T18:18:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 18:18:42 crc kubenswrapper[4859]: I1008 18:18:42.384980 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 18:18:42 crc kubenswrapper[4859]: I1008 18:18:42.385047 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 18:18:42 crc kubenswrapper[4859]: I1008 18:18:42.385068 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 18:18:42 crc kubenswrapper[4859]: I1008 18:18:42.385094 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 18:18:42 crc kubenswrapper[4859]: I1008 18:18:42.385112 4859 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T18:18:42Z","lastTransitionTime":"2025-10-08T18:18:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 18:18:42 crc kubenswrapper[4859]: I1008 18:18:42.488193 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 18:18:42 crc kubenswrapper[4859]: I1008 18:18:42.488284 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 18:18:42 crc kubenswrapper[4859]: I1008 18:18:42.488322 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 18:18:42 crc kubenswrapper[4859]: I1008 18:18:42.488353 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 18:18:42 crc kubenswrapper[4859]: I1008 18:18:42.488371 4859 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T18:18:42Z","lastTransitionTime":"2025-10-08T18:18:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 18:18:42 crc kubenswrapper[4859]: I1008 18:18:42.592118 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 18:18:42 crc kubenswrapper[4859]: I1008 18:18:42.592168 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 18:18:42 crc kubenswrapper[4859]: I1008 18:18:42.592179 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 18:18:42 crc kubenswrapper[4859]: I1008 18:18:42.592197 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 18:18:42 crc kubenswrapper[4859]: I1008 18:18:42.592207 4859 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T18:18:42Z","lastTransitionTime":"2025-10-08T18:18:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 18:18:42 crc kubenswrapper[4859]: I1008 18:18:42.695145 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 18:18:42 crc kubenswrapper[4859]: I1008 18:18:42.695220 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 18:18:42 crc kubenswrapper[4859]: I1008 18:18:42.695242 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 18:18:42 crc kubenswrapper[4859]: I1008 18:18:42.695276 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 18:18:42 crc kubenswrapper[4859]: I1008 18:18:42.695295 4859 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T18:18:42Z","lastTransitionTime":"2025-10-08T18:18:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 18:18:42 crc kubenswrapper[4859]: I1008 18:18:42.798224 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 18:18:42 crc kubenswrapper[4859]: I1008 18:18:42.798275 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 18:18:42 crc kubenswrapper[4859]: I1008 18:18:42.798285 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 18:18:42 crc kubenswrapper[4859]: I1008 18:18:42.798303 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 18:18:42 crc kubenswrapper[4859]: I1008 18:18:42.798314 4859 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T18:18:42Z","lastTransitionTime":"2025-10-08T18:18:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 18:18:42 crc kubenswrapper[4859]: I1008 18:18:42.901020 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 18:18:42 crc kubenswrapper[4859]: I1008 18:18:42.901087 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 18:18:42 crc kubenswrapper[4859]: I1008 18:18:42.901103 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 18:18:42 crc kubenswrapper[4859]: I1008 18:18:42.901131 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 18:18:42 crc kubenswrapper[4859]: I1008 18:18:42.901149 4859 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T18:18:42Z","lastTransitionTime":"2025-10-08T18:18:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 18:18:43 crc kubenswrapper[4859]: I1008 18:18:43.004863 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 18:18:43 crc kubenswrapper[4859]: I1008 18:18:43.004932 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 18:18:43 crc kubenswrapper[4859]: I1008 18:18:43.004953 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 18:18:43 crc kubenswrapper[4859]: I1008 18:18:43.004978 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 18:18:43 crc kubenswrapper[4859]: I1008 18:18:43.004998 4859 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T18:18:43Z","lastTransitionTime":"2025-10-08T18:18:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 18:18:43 crc kubenswrapper[4859]: I1008 18:18:43.107636 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 18:18:43 crc kubenswrapper[4859]: I1008 18:18:43.107728 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 18:18:43 crc kubenswrapper[4859]: I1008 18:18:43.107752 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 18:18:43 crc kubenswrapper[4859]: I1008 18:18:43.107782 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 18:18:43 crc kubenswrapper[4859]: I1008 18:18:43.107804 4859 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T18:18:43Z","lastTransitionTime":"2025-10-08T18:18:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 18:18:43 crc kubenswrapper[4859]: I1008 18:18:43.211303 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 18:18:43 crc kubenswrapper[4859]: I1008 18:18:43.211375 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 18:18:43 crc kubenswrapper[4859]: I1008 18:18:43.211394 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 18:18:43 crc kubenswrapper[4859]: I1008 18:18:43.211422 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 18:18:43 crc kubenswrapper[4859]: I1008 18:18:43.211441 4859 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T18:18:43Z","lastTransitionTime":"2025-10-08T18:18:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 18:18:43 crc kubenswrapper[4859]: I1008 18:18:43.315641 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 18:18:43 crc kubenswrapper[4859]: I1008 18:18:43.315746 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 18:18:43 crc kubenswrapper[4859]: I1008 18:18:43.315764 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 18:18:43 crc kubenswrapper[4859]: I1008 18:18:43.315791 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 18:18:43 crc kubenswrapper[4859]: I1008 18:18:43.315809 4859 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T18:18:43Z","lastTransitionTime":"2025-10-08T18:18:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 18:18:43 crc kubenswrapper[4859]: I1008 18:18:43.418533 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 18:18:43 crc kubenswrapper[4859]: I1008 18:18:43.418613 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 18:18:43 crc kubenswrapper[4859]: I1008 18:18:43.418636 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 18:18:43 crc kubenswrapper[4859]: I1008 18:18:43.418667 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 18:18:43 crc kubenswrapper[4859]: I1008 18:18:43.418715 4859 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T18:18:43Z","lastTransitionTime":"2025-10-08T18:18:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 18:18:43 crc kubenswrapper[4859]: I1008 18:18:43.469292 4859 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 08 18:18:43 crc kubenswrapper[4859]: E1008 18:18:43.469470 4859 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 08 18:18:43 crc kubenswrapper[4859]: I1008 18:18:43.469722 4859 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 08 18:18:43 crc kubenswrapper[4859]: E1008 18:18:43.469795 4859 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 08 18:18:43 crc kubenswrapper[4859]: I1008 18:18:43.469980 4859 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 08 18:18:43 crc kubenswrapper[4859]: I1008 18:18:43.469992 4859 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-4b4sz" Oct 08 18:18:43 crc kubenswrapper[4859]: E1008 18:18:43.470188 4859 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 08 18:18:43 crc kubenswrapper[4859]: E1008 18:18:43.470361 4859 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-4b4sz" podUID="2e53cd9b-64f7-4e07-8b96-fb77847c9ac6" Oct 08 18:18:43 crc kubenswrapper[4859]: I1008 18:18:43.525931 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 18:18:43 crc kubenswrapper[4859]: I1008 18:18:43.526009 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 18:18:43 crc kubenswrapper[4859]: I1008 18:18:43.526028 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 18:18:43 crc kubenswrapper[4859]: I1008 18:18:43.526048 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 18:18:43 crc kubenswrapper[4859]: I1008 18:18:43.526154 4859 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T18:18:43Z","lastTransitionTime":"2025-10-08T18:18:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 18:18:43 crc kubenswrapper[4859]: I1008 18:18:43.630360 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 18:18:43 crc kubenswrapper[4859]: I1008 18:18:43.630415 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 18:18:43 crc kubenswrapper[4859]: I1008 18:18:43.630427 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 18:18:43 crc kubenswrapper[4859]: I1008 18:18:43.630442 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 18:18:43 crc kubenswrapper[4859]: I1008 18:18:43.630452 4859 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T18:18:43Z","lastTransitionTime":"2025-10-08T18:18:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 18:18:43 crc kubenswrapper[4859]: I1008 18:18:43.733616 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 18:18:43 crc kubenswrapper[4859]: I1008 18:18:43.733666 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 18:18:43 crc kubenswrapper[4859]: I1008 18:18:43.733675 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 18:18:43 crc kubenswrapper[4859]: I1008 18:18:43.733719 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 18:18:43 crc kubenswrapper[4859]: I1008 18:18:43.733732 4859 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T18:18:43Z","lastTransitionTime":"2025-10-08T18:18:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 18:18:43 crc kubenswrapper[4859]: I1008 18:18:43.836077 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 18:18:43 crc kubenswrapper[4859]: I1008 18:18:43.836123 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 18:18:43 crc kubenswrapper[4859]: I1008 18:18:43.836135 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 18:18:43 crc kubenswrapper[4859]: I1008 18:18:43.836155 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 18:18:43 crc kubenswrapper[4859]: I1008 18:18:43.836168 4859 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T18:18:43Z","lastTransitionTime":"2025-10-08T18:18:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 18:18:43 crc kubenswrapper[4859]: I1008 18:18:43.939918 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 18:18:43 crc kubenswrapper[4859]: I1008 18:18:43.939991 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 18:18:43 crc kubenswrapper[4859]: I1008 18:18:43.940013 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 18:18:43 crc kubenswrapper[4859]: I1008 18:18:43.940037 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 18:18:43 crc kubenswrapper[4859]: I1008 18:18:43.940057 4859 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T18:18:43Z","lastTransitionTime":"2025-10-08T18:18:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 18:18:44 crc kubenswrapper[4859]: I1008 18:18:44.044560 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 18:18:44 crc kubenswrapper[4859]: I1008 18:18:44.044612 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 18:18:44 crc kubenswrapper[4859]: I1008 18:18:44.044622 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 18:18:44 crc kubenswrapper[4859]: I1008 18:18:44.044641 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 18:18:44 crc kubenswrapper[4859]: I1008 18:18:44.044654 4859 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T18:18:44Z","lastTransitionTime":"2025-10-08T18:18:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 18:18:44 crc kubenswrapper[4859]: I1008 18:18:44.147895 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 18:18:44 crc kubenswrapper[4859]: I1008 18:18:44.147976 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 18:18:44 crc kubenswrapper[4859]: I1008 18:18:44.147997 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 18:18:44 crc kubenswrapper[4859]: I1008 18:18:44.148016 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 18:18:44 crc kubenswrapper[4859]: I1008 18:18:44.148026 4859 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T18:18:44Z","lastTransitionTime":"2025-10-08T18:18:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 18:18:44 crc kubenswrapper[4859]: I1008 18:18:44.250147 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 18:18:44 crc kubenswrapper[4859]: I1008 18:18:44.250209 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 18:18:44 crc kubenswrapper[4859]: I1008 18:18:44.250219 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 18:18:44 crc kubenswrapper[4859]: I1008 18:18:44.250235 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 18:18:44 crc kubenswrapper[4859]: I1008 18:18:44.250244 4859 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T18:18:44Z","lastTransitionTime":"2025-10-08T18:18:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 18:18:44 crc kubenswrapper[4859]: I1008 18:18:44.353385 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 18:18:44 crc kubenswrapper[4859]: I1008 18:18:44.353487 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 18:18:44 crc kubenswrapper[4859]: I1008 18:18:44.353505 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 18:18:44 crc kubenswrapper[4859]: I1008 18:18:44.353532 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 18:18:44 crc kubenswrapper[4859]: I1008 18:18:44.353549 4859 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T18:18:44Z","lastTransitionTime":"2025-10-08T18:18:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 18:18:44 crc kubenswrapper[4859]: I1008 18:18:44.457152 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 18:18:44 crc kubenswrapper[4859]: I1008 18:18:44.457223 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 18:18:44 crc kubenswrapper[4859]: I1008 18:18:44.457243 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 18:18:44 crc kubenswrapper[4859]: I1008 18:18:44.457269 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 18:18:44 crc kubenswrapper[4859]: I1008 18:18:44.457291 4859 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T18:18:44Z","lastTransitionTime":"2025-10-08T18:18:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 18:18:44 crc kubenswrapper[4859]: I1008 18:18:44.471023 4859 scope.go:117] "RemoveContainer" containerID="f3a760c557ef21fbc6b54e78c3f6c5b325f296f93a022425cca7b1399eed8661" Oct 08 18:18:44 crc kubenswrapper[4859]: E1008 18:18:44.471289 4859 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ovnkube-controller\" with CrashLoopBackOff: \"back-off 40s restarting failed container=ovnkube-controller pod=ovnkube-node-4g8gf_openshift-ovn-kubernetes(1dff864e-b75d-4e0b-b182-75f710eac8df)\"" pod="openshift-ovn-kubernetes/ovnkube-node-4g8gf" podUID="1dff864e-b75d-4e0b-b182-75f710eac8df" Oct 08 18:18:44 crc kubenswrapper[4859]: I1008 18:18:44.559408 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 18:18:44 crc kubenswrapper[4859]: I1008 18:18:44.559460 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 18:18:44 crc kubenswrapper[4859]: I1008 18:18:44.559472 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 18:18:44 crc kubenswrapper[4859]: I1008 18:18:44.559493 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 18:18:44 crc kubenswrapper[4859]: I1008 18:18:44.559514 4859 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T18:18:44Z","lastTransitionTime":"2025-10-08T18:18:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 18:18:44 crc kubenswrapper[4859]: I1008 18:18:44.662411 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 18:18:44 crc kubenswrapper[4859]: I1008 18:18:44.662475 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 18:18:44 crc kubenswrapper[4859]: I1008 18:18:44.662487 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 18:18:44 crc kubenswrapper[4859]: I1008 18:18:44.662503 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 18:18:44 crc kubenswrapper[4859]: I1008 18:18:44.662516 4859 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T18:18:44Z","lastTransitionTime":"2025-10-08T18:18:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 18:18:44 crc kubenswrapper[4859]: I1008 18:18:44.765357 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 18:18:44 crc kubenswrapper[4859]: I1008 18:18:44.765426 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 18:18:44 crc kubenswrapper[4859]: I1008 18:18:44.765436 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 18:18:44 crc kubenswrapper[4859]: I1008 18:18:44.765456 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 18:18:44 crc kubenswrapper[4859]: I1008 18:18:44.765468 4859 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T18:18:44Z","lastTransitionTime":"2025-10-08T18:18:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 18:18:44 crc kubenswrapper[4859]: I1008 18:18:44.868745 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 18:18:44 crc kubenswrapper[4859]: I1008 18:18:44.868808 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 18:18:44 crc kubenswrapper[4859]: I1008 18:18:44.868817 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 18:18:44 crc kubenswrapper[4859]: I1008 18:18:44.868840 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 18:18:44 crc kubenswrapper[4859]: I1008 18:18:44.868857 4859 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T18:18:44Z","lastTransitionTime":"2025-10-08T18:18:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 18:18:44 crc kubenswrapper[4859]: I1008 18:18:44.972394 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 18:18:44 crc kubenswrapper[4859]: I1008 18:18:44.972497 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 18:18:44 crc kubenswrapper[4859]: I1008 18:18:44.972508 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 18:18:44 crc kubenswrapper[4859]: I1008 18:18:44.972526 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 18:18:44 crc kubenswrapper[4859]: I1008 18:18:44.972536 4859 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T18:18:44Z","lastTransitionTime":"2025-10-08T18:18:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 18:18:45 crc kubenswrapper[4859]: I1008 18:18:45.075802 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 18:18:45 crc kubenswrapper[4859]: I1008 18:18:45.075867 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 18:18:45 crc kubenswrapper[4859]: I1008 18:18:45.075900 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 18:18:45 crc kubenswrapper[4859]: I1008 18:18:45.075925 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 18:18:45 crc kubenswrapper[4859]: I1008 18:18:45.075943 4859 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T18:18:45Z","lastTransitionTime":"2025-10-08T18:18:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 18:18:45 crc kubenswrapper[4859]: I1008 18:18:45.179125 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 18:18:45 crc kubenswrapper[4859]: I1008 18:18:45.179189 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 18:18:45 crc kubenswrapper[4859]: I1008 18:18:45.179207 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 18:18:45 crc kubenswrapper[4859]: I1008 18:18:45.179259 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 18:18:45 crc kubenswrapper[4859]: I1008 18:18:45.179289 4859 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T18:18:45Z","lastTransitionTime":"2025-10-08T18:18:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 18:18:45 crc kubenswrapper[4859]: I1008 18:18:45.282926 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 18:18:45 crc kubenswrapper[4859]: I1008 18:18:45.282992 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 18:18:45 crc kubenswrapper[4859]: I1008 18:18:45.283010 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 18:18:45 crc kubenswrapper[4859]: I1008 18:18:45.283038 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 18:18:45 crc kubenswrapper[4859]: I1008 18:18:45.283057 4859 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T18:18:45Z","lastTransitionTime":"2025-10-08T18:18:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 18:18:45 crc kubenswrapper[4859]: I1008 18:18:45.386938 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 18:18:45 crc kubenswrapper[4859]: I1008 18:18:45.387000 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 18:18:45 crc kubenswrapper[4859]: I1008 18:18:45.387017 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 18:18:45 crc kubenswrapper[4859]: I1008 18:18:45.387045 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 18:18:45 crc kubenswrapper[4859]: I1008 18:18:45.387062 4859 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T18:18:45Z","lastTransitionTime":"2025-10-08T18:18:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 18:18:45 crc kubenswrapper[4859]: I1008 18:18:45.469000 4859 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 08 18:18:45 crc kubenswrapper[4859]: I1008 18:18:45.469132 4859 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 08 18:18:45 crc kubenswrapper[4859]: I1008 18:18:45.469184 4859 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-4b4sz" Oct 08 18:18:45 crc kubenswrapper[4859]: E1008 18:18:45.469403 4859 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 08 18:18:45 crc kubenswrapper[4859]: E1008 18:18:45.469549 4859 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 08 18:18:45 crc kubenswrapper[4859]: E1008 18:18:45.469770 4859 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-4b4sz" podUID="2e53cd9b-64f7-4e07-8b96-fb77847c9ac6" Oct 08 18:18:45 crc kubenswrapper[4859]: I1008 18:18:45.470154 4859 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 08 18:18:45 crc kubenswrapper[4859]: E1008 18:18:45.470443 4859 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 08 18:18:45 crc kubenswrapper[4859]: I1008 18:18:45.490169 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 18:18:45 crc kubenswrapper[4859]: I1008 18:18:45.490540 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 18:18:45 crc kubenswrapper[4859]: I1008 18:18:45.490783 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 18:18:45 crc kubenswrapper[4859]: I1008 18:18:45.490964 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 18:18:45 crc kubenswrapper[4859]: I1008 18:18:45.491115 4859 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T18:18:45Z","lastTransitionTime":"2025-10-08T18:18:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 18:18:45 crc kubenswrapper[4859]: I1008 18:18:45.594271 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 18:18:45 crc kubenswrapper[4859]: I1008 18:18:45.594330 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 18:18:45 crc kubenswrapper[4859]: I1008 18:18:45.594347 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 18:18:45 crc kubenswrapper[4859]: I1008 18:18:45.594371 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 18:18:45 crc kubenswrapper[4859]: I1008 18:18:45.594389 4859 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T18:18:45Z","lastTransitionTime":"2025-10-08T18:18:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 18:18:45 crc kubenswrapper[4859]: I1008 18:18:45.697508 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 18:18:45 crc kubenswrapper[4859]: I1008 18:18:45.697543 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 18:18:45 crc kubenswrapper[4859]: I1008 18:18:45.697552 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 18:18:45 crc kubenswrapper[4859]: I1008 18:18:45.697566 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 18:18:45 crc kubenswrapper[4859]: I1008 18:18:45.697575 4859 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T18:18:45Z","lastTransitionTime":"2025-10-08T18:18:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 18:18:45 crc kubenswrapper[4859]: I1008 18:18:45.799711 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 18:18:45 crc kubenswrapper[4859]: I1008 18:18:45.799758 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 18:18:45 crc kubenswrapper[4859]: I1008 18:18:45.799767 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 18:18:45 crc kubenswrapper[4859]: I1008 18:18:45.799783 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 18:18:45 crc kubenswrapper[4859]: I1008 18:18:45.799794 4859 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T18:18:45Z","lastTransitionTime":"2025-10-08T18:18:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 18:18:45 crc kubenswrapper[4859]: I1008 18:18:45.902608 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 18:18:45 crc kubenswrapper[4859]: I1008 18:18:45.902678 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 18:18:45 crc kubenswrapper[4859]: I1008 18:18:45.902748 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 18:18:45 crc kubenswrapper[4859]: I1008 18:18:45.902781 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 18:18:45 crc kubenswrapper[4859]: I1008 18:18:45.902807 4859 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T18:18:45Z","lastTransitionTime":"2025-10-08T18:18:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 18:18:46 crc kubenswrapper[4859]: I1008 18:18:46.006440 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 18:18:46 crc kubenswrapper[4859]: I1008 18:18:46.006515 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 18:18:46 crc kubenswrapper[4859]: I1008 18:18:46.006535 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 18:18:46 crc kubenswrapper[4859]: I1008 18:18:46.006558 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 18:18:46 crc kubenswrapper[4859]: I1008 18:18:46.006573 4859 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T18:18:46Z","lastTransitionTime":"2025-10-08T18:18:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 18:18:46 crc kubenswrapper[4859]: I1008 18:18:46.110967 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 18:18:46 crc kubenswrapper[4859]: I1008 18:18:46.111082 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 18:18:46 crc kubenswrapper[4859]: I1008 18:18:46.111111 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 18:18:46 crc kubenswrapper[4859]: I1008 18:18:46.111146 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 18:18:46 crc kubenswrapper[4859]: I1008 18:18:46.111179 4859 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T18:18:46Z","lastTransitionTime":"2025-10-08T18:18:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 18:18:46 crc kubenswrapper[4859]: I1008 18:18:46.214199 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 18:18:46 crc kubenswrapper[4859]: I1008 18:18:46.214260 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 18:18:46 crc kubenswrapper[4859]: I1008 18:18:46.214273 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 18:18:46 crc kubenswrapper[4859]: I1008 18:18:46.214294 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 18:18:46 crc kubenswrapper[4859]: I1008 18:18:46.214307 4859 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T18:18:46Z","lastTransitionTime":"2025-10-08T18:18:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 18:18:46 crc kubenswrapper[4859]: I1008 18:18:46.317225 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 18:18:46 crc kubenswrapper[4859]: I1008 18:18:46.317440 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 18:18:46 crc kubenswrapper[4859]: I1008 18:18:46.317467 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 18:18:46 crc kubenswrapper[4859]: I1008 18:18:46.317492 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 18:18:46 crc kubenswrapper[4859]: I1008 18:18:46.317509 4859 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T18:18:46Z","lastTransitionTime":"2025-10-08T18:18:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 18:18:46 crc kubenswrapper[4859]: I1008 18:18:46.421062 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 18:18:46 crc kubenswrapper[4859]: I1008 18:18:46.421158 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 18:18:46 crc kubenswrapper[4859]: I1008 18:18:46.421180 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 18:18:46 crc kubenswrapper[4859]: I1008 18:18:46.421205 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 18:18:46 crc kubenswrapper[4859]: I1008 18:18:46.421223 4859 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T18:18:46Z","lastTransitionTime":"2025-10-08T18:18:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 18:18:46 crc kubenswrapper[4859]: I1008 18:18:46.524550 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 18:18:46 crc kubenswrapper[4859]: I1008 18:18:46.524599 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 18:18:46 crc kubenswrapper[4859]: I1008 18:18:46.524609 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 18:18:46 crc kubenswrapper[4859]: I1008 18:18:46.524625 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 18:18:46 crc kubenswrapper[4859]: I1008 18:18:46.524637 4859 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T18:18:46Z","lastTransitionTime":"2025-10-08T18:18:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 18:18:46 crc kubenswrapper[4859]: I1008 18:18:46.627959 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 18:18:46 crc kubenswrapper[4859]: I1008 18:18:46.628032 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 18:18:46 crc kubenswrapper[4859]: I1008 18:18:46.628053 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 18:18:46 crc kubenswrapper[4859]: I1008 18:18:46.628091 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 18:18:46 crc kubenswrapper[4859]: I1008 18:18:46.628110 4859 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T18:18:46Z","lastTransitionTime":"2025-10-08T18:18:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 18:18:46 crc kubenswrapper[4859]: I1008 18:18:46.731667 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 18:18:46 crc kubenswrapper[4859]: I1008 18:18:46.731764 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 18:18:46 crc kubenswrapper[4859]: I1008 18:18:46.731784 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 18:18:46 crc kubenswrapper[4859]: I1008 18:18:46.731810 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 18:18:46 crc kubenswrapper[4859]: I1008 18:18:46.731828 4859 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T18:18:46Z","lastTransitionTime":"2025-10-08T18:18:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 18:18:46 crc kubenswrapper[4859]: I1008 18:18:46.835382 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 18:18:46 crc kubenswrapper[4859]: I1008 18:18:46.835462 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 18:18:46 crc kubenswrapper[4859]: I1008 18:18:46.835482 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 18:18:46 crc kubenswrapper[4859]: I1008 18:18:46.835510 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 18:18:46 crc kubenswrapper[4859]: I1008 18:18:46.835541 4859 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T18:18:46Z","lastTransitionTime":"2025-10-08T18:18:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 18:18:46 crc kubenswrapper[4859]: I1008 18:18:46.939059 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 18:18:46 crc kubenswrapper[4859]: I1008 18:18:46.939136 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 18:18:46 crc kubenswrapper[4859]: I1008 18:18:46.939156 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 18:18:46 crc kubenswrapper[4859]: I1008 18:18:46.939182 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 18:18:46 crc kubenswrapper[4859]: I1008 18:18:46.939204 4859 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T18:18:46Z","lastTransitionTime":"2025-10-08T18:18:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 18:18:47 crc kubenswrapper[4859]: I1008 18:18:47.043005 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 18:18:47 crc kubenswrapper[4859]: I1008 18:18:47.043070 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 18:18:47 crc kubenswrapper[4859]: I1008 18:18:47.043248 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 18:18:47 crc kubenswrapper[4859]: I1008 18:18:47.043284 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 18:18:47 crc kubenswrapper[4859]: I1008 18:18:47.043301 4859 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T18:18:47Z","lastTransitionTime":"2025-10-08T18:18:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 18:18:47 crc kubenswrapper[4859]: I1008 18:18:47.147536 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 18:18:47 crc kubenswrapper[4859]: I1008 18:18:47.147615 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 18:18:47 crc kubenswrapper[4859]: I1008 18:18:47.147633 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 18:18:47 crc kubenswrapper[4859]: I1008 18:18:47.147658 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 18:18:47 crc kubenswrapper[4859]: I1008 18:18:47.147677 4859 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T18:18:47Z","lastTransitionTime":"2025-10-08T18:18:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 18:18:47 crc kubenswrapper[4859]: I1008 18:18:47.251531 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 18:18:47 crc kubenswrapper[4859]: I1008 18:18:47.251590 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 18:18:47 crc kubenswrapper[4859]: I1008 18:18:47.251608 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 18:18:47 crc kubenswrapper[4859]: I1008 18:18:47.251638 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 18:18:47 crc kubenswrapper[4859]: I1008 18:18:47.251657 4859 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T18:18:47Z","lastTransitionTime":"2025-10-08T18:18:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 18:18:47 crc kubenswrapper[4859]: I1008 18:18:47.355204 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 18:18:47 crc kubenswrapper[4859]: I1008 18:18:47.355324 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 18:18:47 crc kubenswrapper[4859]: I1008 18:18:47.355361 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 18:18:47 crc kubenswrapper[4859]: I1008 18:18:47.355397 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 18:18:47 crc kubenswrapper[4859]: I1008 18:18:47.355421 4859 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T18:18:47Z","lastTransitionTime":"2025-10-08T18:18:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 18:18:47 crc kubenswrapper[4859]: I1008 18:18:47.458730 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 18:18:47 crc kubenswrapper[4859]: I1008 18:18:47.458819 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 18:18:47 crc kubenswrapper[4859]: I1008 18:18:47.458840 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 18:18:47 crc kubenswrapper[4859]: I1008 18:18:47.458879 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 18:18:47 crc kubenswrapper[4859]: I1008 18:18:47.458904 4859 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T18:18:47Z","lastTransitionTime":"2025-10-08T18:18:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 18:18:47 crc kubenswrapper[4859]: I1008 18:18:47.469169 4859 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 08 18:18:47 crc kubenswrapper[4859]: I1008 18:18:47.469177 4859 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 08 18:18:47 crc kubenswrapper[4859]: E1008 18:18:47.469367 4859 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 08 18:18:47 crc kubenswrapper[4859]: E1008 18:18:47.469523 4859 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 08 18:18:47 crc kubenswrapper[4859]: I1008 18:18:47.469207 4859 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-4b4sz" Oct 08 18:18:47 crc kubenswrapper[4859]: I1008 18:18:47.469832 4859 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 08 18:18:47 crc kubenswrapper[4859]: E1008 18:18:47.470114 4859 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 08 18:18:47 crc kubenswrapper[4859]: E1008 18:18:47.470128 4859 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-4b4sz" podUID="2e53cd9b-64f7-4e07-8b96-fb77847c9ac6" Oct 08 18:18:47 crc kubenswrapper[4859]: I1008 18:18:47.562503 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 18:18:47 crc kubenswrapper[4859]: I1008 18:18:47.563055 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 18:18:47 crc kubenswrapper[4859]: I1008 18:18:47.563209 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 18:18:47 crc kubenswrapper[4859]: I1008 18:18:47.563358 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 18:18:47 crc kubenswrapper[4859]: I1008 18:18:47.563525 4859 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T18:18:47Z","lastTransitionTime":"2025-10-08T18:18:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 18:18:47 crc kubenswrapper[4859]: I1008 18:18:47.667988 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 18:18:47 crc kubenswrapper[4859]: I1008 18:18:47.668070 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 18:18:47 crc kubenswrapper[4859]: I1008 18:18:47.668088 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 18:18:47 crc kubenswrapper[4859]: I1008 18:18:47.668117 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 18:18:47 crc kubenswrapper[4859]: I1008 18:18:47.668141 4859 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T18:18:47Z","lastTransitionTime":"2025-10-08T18:18:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 18:18:47 crc kubenswrapper[4859]: I1008 18:18:47.772099 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 18:18:47 crc kubenswrapper[4859]: I1008 18:18:47.772161 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 18:18:47 crc kubenswrapper[4859]: I1008 18:18:47.772175 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 18:18:47 crc kubenswrapper[4859]: I1008 18:18:47.772201 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 18:18:47 crc kubenswrapper[4859]: I1008 18:18:47.772217 4859 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T18:18:47Z","lastTransitionTime":"2025-10-08T18:18:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 18:18:47 crc kubenswrapper[4859]: I1008 18:18:47.875493 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 18:18:47 crc kubenswrapper[4859]: I1008 18:18:47.875570 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 18:18:47 crc kubenswrapper[4859]: I1008 18:18:47.875594 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 18:18:47 crc kubenswrapper[4859]: I1008 18:18:47.875744 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 18:18:47 crc kubenswrapper[4859]: I1008 18:18:47.875786 4859 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T18:18:47Z","lastTransitionTime":"2025-10-08T18:18:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 18:18:47 crc kubenswrapper[4859]: I1008 18:18:47.978737 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 18:18:47 crc kubenswrapper[4859]: I1008 18:18:47.978782 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 18:18:47 crc kubenswrapper[4859]: I1008 18:18:47.978796 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 18:18:47 crc kubenswrapper[4859]: I1008 18:18:47.978820 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 18:18:47 crc kubenswrapper[4859]: I1008 18:18:47.978839 4859 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T18:18:47Z","lastTransitionTime":"2025-10-08T18:18:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 18:18:48 crc kubenswrapper[4859]: I1008 18:18:48.083086 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 18:18:48 crc kubenswrapper[4859]: I1008 18:18:48.083150 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 18:18:48 crc kubenswrapper[4859]: I1008 18:18:48.083165 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 18:18:48 crc kubenswrapper[4859]: I1008 18:18:48.083185 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 18:18:48 crc kubenswrapper[4859]: I1008 18:18:48.083200 4859 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T18:18:48Z","lastTransitionTime":"2025-10-08T18:18:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 18:18:48 crc kubenswrapper[4859]: I1008 18:18:48.187240 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 18:18:48 crc kubenswrapper[4859]: I1008 18:18:48.187326 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 18:18:48 crc kubenswrapper[4859]: I1008 18:18:48.187351 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 18:18:48 crc kubenswrapper[4859]: I1008 18:18:48.187383 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 18:18:48 crc kubenswrapper[4859]: I1008 18:18:48.187408 4859 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T18:18:48Z","lastTransitionTime":"2025-10-08T18:18:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 18:18:48 crc kubenswrapper[4859]: I1008 18:18:48.290952 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 18:18:48 crc kubenswrapper[4859]: I1008 18:18:48.291002 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 18:18:48 crc kubenswrapper[4859]: I1008 18:18:48.291014 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 18:18:48 crc kubenswrapper[4859]: I1008 18:18:48.291031 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 18:18:48 crc kubenswrapper[4859]: I1008 18:18:48.291043 4859 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T18:18:48Z","lastTransitionTime":"2025-10-08T18:18:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 18:18:48 crc kubenswrapper[4859]: I1008 18:18:48.394870 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 18:18:48 crc kubenswrapper[4859]: I1008 18:18:48.394930 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 18:18:48 crc kubenswrapper[4859]: I1008 18:18:48.394947 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 18:18:48 crc kubenswrapper[4859]: I1008 18:18:48.394971 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 18:18:48 crc kubenswrapper[4859]: I1008 18:18:48.394989 4859 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T18:18:48Z","lastTransitionTime":"2025-10-08T18:18:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 18:18:48 crc kubenswrapper[4859]: I1008 18:18:48.497626 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 18:18:48 crc kubenswrapper[4859]: I1008 18:18:48.498017 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 18:18:48 crc kubenswrapper[4859]: I1008 18:18:48.498127 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 18:18:48 crc kubenswrapper[4859]: I1008 18:18:48.498224 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 18:18:48 crc kubenswrapper[4859]: I1008 18:18:48.498321 4859 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T18:18:48Z","lastTransitionTime":"2025-10-08T18:18:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 18:18:48 crc kubenswrapper[4859]: I1008 18:18:48.601732 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 18:18:48 crc kubenswrapper[4859]: I1008 18:18:48.602070 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 18:18:48 crc kubenswrapper[4859]: I1008 18:18:48.602142 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 18:18:48 crc kubenswrapper[4859]: I1008 18:18:48.602218 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 18:18:48 crc kubenswrapper[4859]: I1008 18:18:48.602294 4859 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T18:18:48Z","lastTransitionTime":"2025-10-08T18:18:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 18:18:48 crc kubenswrapper[4859]: I1008 18:18:48.705372 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 18:18:48 crc kubenswrapper[4859]: I1008 18:18:48.705433 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 18:18:48 crc kubenswrapper[4859]: I1008 18:18:48.705448 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 18:18:48 crc kubenswrapper[4859]: I1008 18:18:48.705470 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 18:18:48 crc kubenswrapper[4859]: I1008 18:18:48.705484 4859 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T18:18:48Z","lastTransitionTime":"2025-10-08T18:18:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 18:18:48 crc kubenswrapper[4859]: I1008 18:18:48.808499 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 18:18:48 crc kubenswrapper[4859]: I1008 18:18:48.808552 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 18:18:48 crc kubenswrapper[4859]: I1008 18:18:48.808562 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 18:18:48 crc kubenswrapper[4859]: I1008 18:18:48.808585 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 18:18:48 crc kubenswrapper[4859]: I1008 18:18:48.808598 4859 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T18:18:48Z","lastTransitionTime":"2025-10-08T18:18:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 18:18:48 crc kubenswrapper[4859]: I1008 18:18:48.911714 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 18:18:48 crc kubenswrapper[4859]: I1008 18:18:48.911813 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 18:18:48 crc kubenswrapper[4859]: I1008 18:18:48.911832 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 18:18:48 crc kubenswrapper[4859]: I1008 18:18:48.911861 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 18:18:48 crc kubenswrapper[4859]: I1008 18:18:48.911881 4859 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T18:18:48Z","lastTransitionTime":"2025-10-08T18:18:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 18:18:48 crc kubenswrapper[4859]: I1008 18:18:48.998375 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 18:18:48 crc kubenswrapper[4859]: I1008 18:18:48.998888 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 18:18:48 crc kubenswrapper[4859]: I1008 18:18:48.999036 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 18:18:48 crc kubenswrapper[4859]: I1008 18:18:48.999182 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 18:18:48 crc kubenswrapper[4859]: I1008 18:18:48.999351 4859 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T18:18:48Z","lastTransitionTime":"2025-10-08T18:18:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 18:18:49 crc kubenswrapper[4859]: E1008 18:18:49.019792 4859 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404552Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865352Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-10-08T18:18:48Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-08T18:18:48Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-08T18:18:48Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-08T18:18:48Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-08T18:18:48Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-08T18:18:48Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-08T18:18:48Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-08T18:18:48Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"edb3ee6d-476b-407c-bb7d-4480786e8777\\\",\\\"systemUUID\\\":\\\"8f73c7ed-44aa-4d16-bae1-d4e684803cfc\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T18:18:49Z is after 2025-08-24T17:21:41Z" Oct 08 18:18:49 crc kubenswrapper[4859]: I1008 18:18:49.025242 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 18:18:49 crc kubenswrapper[4859]: I1008 18:18:49.025561 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 18:18:49 crc kubenswrapper[4859]: I1008 18:18:49.025742 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 18:18:49 crc kubenswrapper[4859]: I1008 18:18:49.025882 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 18:18:49 crc kubenswrapper[4859]: I1008 18:18:49.026020 4859 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T18:18:49Z","lastTransitionTime":"2025-10-08T18:18:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 18:18:49 crc kubenswrapper[4859]: E1008 18:18:49.044043 4859 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404552Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865352Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-10-08T18:18:49Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-08T18:18:49Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-08T18:18:49Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-08T18:18:49Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-08T18:18:49Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-08T18:18:49Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-08T18:18:49Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-08T18:18:49Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"edb3ee6d-476b-407c-bb7d-4480786e8777\\\",\\\"systemUUID\\\":\\\"8f73c7ed-44aa-4d16-bae1-d4e684803cfc\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T18:18:49Z is after 2025-08-24T17:21:41Z" Oct 08 18:18:49 crc kubenswrapper[4859]: I1008 18:18:49.049678 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 18:18:49 crc kubenswrapper[4859]: I1008 18:18:49.049879 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 18:18:49 crc kubenswrapper[4859]: I1008 18:18:49.049983 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 18:18:49 crc kubenswrapper[4859]: I1008 18:18:49.050104 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 18:18:49 crc kubenswrapper[4859]: I1008 18:18:49.050222 4859 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T18:18:49Z","lastTransitionTime":"2025-10-08T18:18:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 18:18:49 crc kubenswrapper[4859]: E1008 18:18:49.064266 4859 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404552Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865352Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-10-08T18:18:49Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-08T18:18:49Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-08T18:18:49Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-08T18:18:49Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-08T18:18:49Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-08T18:18:49Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-08T18:18:49Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-08T18:18:49Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"edb3ee6d-476b-407c-bb7d-4480786e8777\\\",\\\"systemUUID\\\":\\\"8f73c7ed-44aa-4d16-bae1-d4e684803cfc\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T18:18:49Z is after 2025-08-24T17:21:41Z" Oct 08 18:18:49 crc kubenswrapper[4859]: I1008 18:18:49.070161 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 18:18:49 crc kubenswrapper[4859]: I1008 18:18:49.070368 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 18:18:49 crc kubenswrapper[4859]: I1008 18:18:49.070506 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 18:18:49 crc kubenswrapper[4859]: I1008 18:18:49.070611 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 18:18:49 crc kubenswrapper[4859]: I1008 18:18:49.070683 4859 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T18:18:49Z","lastTransitionTime":"2025-10-08T18:18:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 18:18:49 crc kubenswrapper[4859]: E1008 18:18:49.082589 4859 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404552Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865352Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-10-08T18:18:49Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-08T18:18:49Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-08T18:18:49Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-08T18:18:49Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-08T18:18:49Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-08T18:18:49Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-08T18:18:49Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-08T18:18:49Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"edb3ee6d-476b-407c-bb7d-4480786e8777\\\",\\\"systemUUID\\\":\\\"8f73c7ed-44aa-4d16-bae1-d4e684803cfc\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T18:18:49Z is after 2025-08-24T17:21:41Z" Oct 08 18:18:49 crc kubenswrapper[4859]: I1008 18:18:49.086276 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 18:18:49 crc kubenswrapper[4859]: I1008 18:18:49.086323 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 18:18:49 crc kubenswrapper[4859]: I1008 18:18:49.086333 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 18:18:49 crc kubenswrapper[4859]: I1008 18:18:49.086350 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 18:18:49 crc kubenswrapper[4859]: I1008 18:18:49.086363 4859 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T18:18:49Z","lastTransitionTime":"2025-10-08T18:18:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 18:18:49 crc kubenswrapper[4859]: E1008 18:18:49.097779 4859 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404552Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865352Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-10-08T18:18:49Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-08T18:18:49Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-08T18:18:49Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-08T18:18:49Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-08T18:18:49Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-08T18:18:49Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-08T18:18:49Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-08T18:18:49Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"edb3ee6d-476b-407c-bb7d-4480786e8777\\\",\\\"systemUUID\\\":\\\"8f73c7ed-44aa-4d16-bae1-d4e684803cfc\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T18:18:49Z is after 2025-08-24T17:21:41Z" Oct 08 18:18:49 crc kubenswrapper[4859]: E1008 18:18:49.097897 4859 kubelet_node_status.go:572] "Unable to update node status" err="update node status exceeds retry count" Oct 08 18:18:49 crc kubenswrapper[4859]: I1008 18:18:49.098984 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 18:18:49 crc kubenswrapper[4859]: I1008 18:18:49.099009 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 18:18:49 crc kubenswrapper[4859]: I1008 18:18:49.099018 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 18:18:49 crc kubenswrapper[4859]: I1008 18:18:49.099034 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 18:18:49 crc kubenswrapper[4859]: I1008 18:18:49.099046 4859 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T18:18:49Z","lastTransitionTime":"2025-10-08T18:18:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 18:18:49 crc kubenswrapper[4859]: I1008 18:18:49.201784 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 18:18:49 crc kubenswrapper[4859]: I1008 18:18:49.201830 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 18:18:49 crc kubenswrapper[4859]: I1008 18:18:49.201839 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 18:18:49 crc kubenswrapper[4859]: I1008 18:18:49.201858 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 18:18:49 crc kubenswrapper[4859]: I1008 18:18:49.201868 4859 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T18:18:49Z","lastTransitionTime":"2025-10-08T18:18:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 18:18:49 crc kubenswrapper[4859]: I1008 18:18:49.304427 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 18:18:49 crc kubenswrapper[4859]: I1008 18:18:49.304501 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 18:18:49 crc kubenswrapper[4859]: I1008 18:18:49.304530 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 18:18:49 crc kubenswrapper[4859]: I1008 18:18:49.304563 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 18:18:49 crc kubenswrapper[4859]: I1008 18:18:49.304586 4859 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T18:18:49Z","lastTransitionTime":"2025-10-08T18:18:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 18:18:49 crc kubenswrapper[4859]: I1008 18:18:49.408715 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 18:18:49 crc kubenswrapper[4859]: I1008 18:18:49.408795 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 18:18:49 crc kubenswrapper[4859]: I1008 18:18:49.408818 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 18:18:49 crc kubenswrapper[4859]: I1008 18:18:49.408851 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 18:18:49 crc kubenswrapper[4859]: I1008 18:18:49.408872 4859 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T18:18:49Z","lastTransitionTime":"2025-10-08T18:18:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 18:18:49 crc kubenswrapper[4859]: I1008 18:18:49.469587 4859 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-4b4sz" Oct 08 18:18:49 crc kubenswrapper[4859]: I1008 18:18:49.469613 4859 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 08 18:18:49 crc kubenswrapper[4859]: E1008 18:18:49.469809 4859 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-4b4sz" podUID="2e53cd9b-64f7-4e07-8b96-fb77847c9ac6" Oct 08 18:18:49 crc kubenswrapper[4859]: I1008 18:18:49.469876 4859 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 08 18:18:49 crc kubenswrapper[4859]: E1008 18:18:49.470010 4859 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 08 18:18:49 crc kubenswrapper[4859]: I1008 18:18:49.470534 4859 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 08 18:18:49 crc kubenswrapper[4859]: E1008 18:18:49.470557 4859 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 08 18:18:49 crc kubenswrapper[4859]: E1008 18:18:49.471162 4859 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 08 18:18:49 crc kubenswrapper[4859]: I1008 18:18:49.511325 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 18:18:49 crc kubenswrapper[4859]: I1008 18:18:49.511367 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 18:18:49 crc kubenswrapper[4859]: I1008 18:18:49.511379 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 18:18:49 crc kubenswrapper[4859]: I1008 18:18:49.511395 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 18:18:49 crc kubenswrapper[4859]: I1008 18:18:49.511405 4859 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T18:18:49Z","lastTransitionTime":"2025-10-08T18:18:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 18:18:49 crc kubenswrapper[4859]: I1008 18:18:49.614745 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 18:18:49 crc kubenswrapper[4859]: I1008 18:18:49.614805 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 18:18:49 crc kubenswrapper[4859]: I1008 18:18:49.614821 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 18:18:49 crc kubenswrapper[4859]: I1008 18:18:49.614851 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 18:18:49 crc kubenswrapper[4859]: I1008 18:18:49.614869 4859 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T18:18:49Z","lastTransitionTime":"2025-10-08T18:18:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 18:18:49 crc kubenswrapper[4859]: I1008 18:18:49.718638 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 18:18:49 crc kubenswrapper[4859]: I1008 18:18:49.719598 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 18:18:49 crc kubenswrapper[4859]: I1008 18:18:49.719724 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 18:18:49 crc kubenswrapper[4859]: I1008 18:18:49.719828 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 18:18:49 crc kubenswrapper[4859]: I1008 18:18:49.719962 4859 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T18:18:49Z","lastTransitionTime":"2025-10-08T18:18:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 18:18:49 crc kubenswrapper[4859]: I1008 18:18:49.822746 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 18:18:49 crc kubenswrapper[4859]: I1008 18:18:49.822815 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 18:18:49 crc kubenswrapper[4859]: I1008 18:18:49.822859 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 18:18:49 crc kubenswrapper[4859]: I1008 18:18:49.822898 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 18:18:49 crc kubenswrapper[4859]: I1008 18:18:49.822926 4859 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T18:18:49Z","lastTransitionTime":"2025-10-08T18:18:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 18:18:49 crc kubenswrapper[4859]: I1008 18:18:49.925923 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 18:18:49 crc kubenswrapper[4859]: I1008 18:18:49.926305 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 18:18:49 crc kubenswrapper[4859]: I1008 18:18:49.926507 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 18:18:49 crc kubenswrapper[4859]: I1008 18:18:49.926657 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 18:18:49 crc kubenswrapper[4859]: I1008 18:18:49.926816 4859 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T18:18:49Z","lastTransitionTime":"2025-10-08T18:18:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 18:18:50 crc kubenswrapper[4859]: I1008 18:18:50.029517 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 18:18:50 crc kubenswrapper[4859]: I1008 18:18:50.029559 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 18:18:50 crc kubenswrapper[4859]: I1008 18:18:50.029571 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 18:18:50 crc kubenswrapper[4859]: I1008 18:18:50.029589 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 18:18:50 crc kubenswrapper[4859]: I1008 18:18:50.029602 4859 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T18:18:50Z","lastTransitionTime":"2025-10-08T18:18:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 18:18:50 crc kubenswrapper[4859]: I1008 18:18:50.051507 4859 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/2e53cd9b-64f7-4e07-8b96-fb77847c9ac6-metrics-certs\") pod \"network-metrics-daemon-4b4sz\" (UID: \"2e53cd9b-64f7-4e07-8b96-fb77847c9ac6\") " pod="openshift-multus/network-metrics-daemon-4b4sz" Oct 08 18:18:50 crc kubenswrapper[4859]: E1008 18:18:50.051879 4859 secret.go:188] Couldn't get secret openshift-multus/metrics-daemon-secret: object "openshift-multus"/"metrics-daemon-secret" not registered Oct 08 18:18:50 crc kubenswrapper[4859]: E1008 18:18:50.051976 4859 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/2e53cd9b-64f7-4e07-8b96-fb77847c9ac6-metrics-certs podName:2e53cd9b-64f7-4e07-8b96-fb77847c9ac6 nodeName:}" failed. No retries permitted until 2025-10-08 18:19:54.05194974 +0000 UTC m=+164.298789159 (durationBeforeRetry 1m4s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/2e53cd9b-64f7-4e07-8b96-fb77847c9ac6-metrics-certs") pod "network-metrics-daemon-4b4sz" (UID: "2e53cd9b-64f7-4e07-8b96-fb77847c9ac6") : object "openshift-multus"/"metrics-daemon-secret" not registered Oct 08 18:18:50 crc kubenswrapper[4859]: I1008 18:18:50.132477 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 18:18:50 crc kubenswrapper[4859]: I1008 18:18:50.132849 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 18:18:50 crc kubenswrapper[4859]: I1008 18:18:50.132922 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 18:18:50 crc kubenswrapper[4859]: I1008 18:18:50.132991 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 18:18:50 crc kubenswrapper[4859]: I1008 18:18:50.133084 4859 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T18:18:50Z","lastTransitionTime":"2025-10-08T18:18:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 18:18:50 crc kubenswrapper[4859]: I1008 18:18:50.235382 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 18:18:50 crc kubenswrapper[4859]: I1008 18:18:50.235478 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 18:18:50 crc kubenswrapper[4859]: I1008 18:18:50.235501 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 18:18:50 crc kubenswrapper[4859]: I1008 18:18:50.235531 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 18:18:50 crc kubenswrapper[4859]: I1008 18:18:50.235553 4859 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T18:18:50Z","lastTransitionTime":"2025-10-08T18:18:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 18:18:50 crc kubenswrapper[4859]: I1008 18:18:50.338654 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 18:18:50 crc kubenswrapper[4859]: I1008 18:18:50.338727 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 18:18:50 crc kubenswrapper[4859]: I1008 18:18:50.338740 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 18:18:50 crc kubenswrapper[4859]: I1008 18:18:50.338759 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 18:18:50 crc kubenswrapper[4859]: I1008 18:18:50.338772 4859 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T18:18:50Z","lastTransitionTime":"2025-10-08T18:18:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 18:18:50 crc kubenswrapper[4859]: I1008 18:18:50.441502 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 18:18:50 crc kubenswrapper[4859]: I1008 18:18:50.441586 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 18:18:50 crc kubenswrapper[4859]: I1008 18:18:50.441611 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 18:18:50 crc kubenswrapper[4859]: I1008 18:18:50.441639 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 18:18:50 crc kubenswrapper[4859]: I1008 18:18:50.441658 4859 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T18:18:50Z","lastTransitionTime":"2025-10-08T18:18:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 18:18:50 crc kubenswrapper[4859]: I1008 18:18:50.488274 4859 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4efc16fb-e679-464d-8b3c-0f46f082593e\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:17:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:17:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:17:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:17:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:17:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://aa8f8ea895272cd79a08d1c47cd7863a1ec10e21bb60d12aa10862a56c99e4bf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T18:17:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5c4e95d8a712ab219455564539a90286ae48bbc46ddc322a5bfd96c1a8268110\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T18:17:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://bae6ecfd8b446ef04d393d823e5ffe37fe749337c0ca17fde79c5f88d5e31393\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T18:17:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6411d1e7e18d5ceadaaceed5a8a6b20a0cacc55cede793ca9fb4dcad70c64fc6\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T18:17:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T18:17:10Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T18:18:50Z is after 2025-08-24T17:21:41Z" Oct 08 18:18:50 crc kubenswrapper[4859]: I1008 18:18:50.501462 4859 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-08T18:17:29Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-08T18:17:29Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T18:18:50Z is after 2025-08-24T17:21:41Z" Oct 08 18:18:50 crc kubenswrapper[4859]: I1008 18:18:50.518999 4859 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-84k8x" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1c24078c-c9e8-4160-883a-f483b4b3dd23\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:17:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:17:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:17:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:17:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8661006834d0f50eaada7394369584ee95fc61ff54437eec35d755526afc5b48\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T18:17:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hk5xb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://18b291eb3344fb7c00ef5679dca9a6f7ed9d08cea684a803ff7dad2e8a18c585\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://18b291eb3344fb7c00ef5679dca9a6f7ed9d08cea684a803ff7dad2e8a18c585\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-08T18:17:32Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-08T18:17:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hk5xb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://160163f1a91d981583452db3c7a9244182d1c8ced45749a61b721d0d45c1dfe4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://160163f1a91d981583452db3c7a9244182d1c8ced45749a61b721d0d45c1dfe4\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-08T18:17:34Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-08T18:17:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hk5xb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://51e9b53f3847ac53f378ad83870debf8e71518d4edbc7fc0dc66644818e2103e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://51e9b53f3847ac53f378ad83870debf8e71518d4edbc7fc0dc66644818e2103e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-08T18:17:34Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-08T18:17:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hk5xb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1ff118919739759ac01b5cbe3240e6673db54eaa16353c609d670071ed7964cb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1ff118919739759ac01b5cbe3240e6673db54eaa16353c609d670071ed7964cb\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-08T18:17:35Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-08T18:17:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hk5xb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8522e803ff56eb57cd4d9fd036e0c3fb16240ea21d4a094c1db7aa7d50bfe2c2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8522e803ff56eb57cd4d9fd036e0c3fb16240ea21d4a094c1db7aa7d50bfe2c2\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-08T18:17:37Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-08T18:17:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hk5xb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://be61056edf515e037403ca2a0bfd5feb9b2537dd42212b58e133b23b294f327b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://be61056edf515e037403ca2a0bfd5feb9b2537dd42212b58e133b23b294f327b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-08T18:17:38Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-08T18:17:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hk5xb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T18:17:31Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-84k8x\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T18:18:50Z is after 2025-08-24T17:21:41Z" Oct 08 18:18:50 crc kubenswrapper[4859]: I1008 18:18:50.537216 4859 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-8plkg" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"92b2d7de-31cc-4a91-95a5-ed5ea964b028\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:17:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:17:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:18:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:18:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://729bd32fd01dcab495bffa64f2f5e44990fc1d78b05e7f131868009f92305831\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://fa00b8345d40b74a917ddc23cc1b2c5c684a9ca61b6119d090c92cc1439b71c4\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-10-08T18:18:19Z\\\",\\\"message\\\":\\\"2025-10-08T18:17:33+00:00 [cnibincopy] Successfully copied files in /usr/src/multus-cni/rhel9/bin/ to /host/opt/cni/bin/upgrade_888b130f-8754-48c6-b73e-9ff6772793f9\\\\n2025-10-08T18:17:33+00:00 [cnibincopy] Successfully moved files in /host/opt/cni/bin/upgrade_888b130f-8754-48c6-b73e-9ff6772793f9 to /host/opt/cni/bin/\\\\n2025-10-08T18:17:34Z [verbose] multus-daemon started\\\\n2025-10-08T18:17:34Z [verbose] Readiness Indicator file check\\\\n2025-10-08T18:18:19Z [error] have you checked that your default network is ready? still waiting for readinessindicatorfile @ /host/run/multus/cni/net.d/10-ovn-kubernetes.conf. pollimmediate error: timed out waiting for the condition\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-08T18:17:32Z\\\"}},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T18:18:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q9d7m\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T18:17:31Z\\\"}}\" for pod \"openshift-multus\"/\"multus-8plkg\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T18:18:50Z is after 2025-08-24T17:21:41Z" Oct 08 18:18:50 crc kubenswrapper[4859]: I1008 18:18:50.544381 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 18:18:50 crc kubenswrapper[4859]: I1008 18:18:50.544416 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 18:18:50 crc kubenswrapper[4859]: I1008 18:18:50.544425 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 18:18:50 crc kubenswrapper[4859]: I1008 18:18:50.544439 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 18:18:50 crc kubenswrapper[4859]: I1008 18:18:50.544449 4859 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T18:18:50Z","lastTransitionTime":"2025-10-08T18:18:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 18:18:50 crc kubenswrapper[4859]: I1008 18:18:50.554957 4859 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-btjmq" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b8605f7-fed7-493f-9b4f-6a6358adf907\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:17:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:17:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:17:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-08T18:17:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://14e9b6c4a4403f36adb78fcf1135a620ca6fc56ed974c9639dcdfc59e6917559\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-08T18:17:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xspg4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-08T18:17:35Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-btjmq\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-08T18:18:50Z is after 2025-08-24T17:21:41Z" Oct 08 18:18:50 crc kubenswrapper[4859]: I1008 18:18:50.599451 4859 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-g4t79" podStartSLOduration=79.599416673 podStartE2EDuration="1m19.599416673s" podCreationTimestamp="2025-10-08 18:17:31 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-08 18:18:50.584347672 +0000 UTC m=+100.831187071" watchObservedRunningTime="2025-10-08 18:18:50.599416673 +0000 UTC m=+100.846256092" Oct 08 18:18:50 crc kubenswrapper[4859]: I1008 18:18:50.616844 4859 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" podStartSLOduration=50.61680528 podStartE2EDuration="50.61680528s" podCreationTimestamp="2025-10-08 18:18:00 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-08 18:18:50.60037887 +0000 UTC m=+100.847218319" watchObservedRunningTime="2025-10-08 18:18:50.61680528 +0000 UTC m=+100.863644699" Oct 08 18:18:50 crc kubenswrapper[4859]: I1008 18:18:50.630164 4859 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-dns/node-resolver-btw28" podStartSLOduration=80.630141611 podStartE2EDuration="1m20.630141611s" podCreationTimestamp="2025-10-08 18:17:30 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-08 18:18:50.629949515 +0000 UTC m=+100.876788904" watchObservedRunningTime="2025-10-08 18:18:50.630141611 +0000 UTC m=+100.876980990" Oct 08 18:18:50 crc kubenswrapper[4859]: I1008 18:18:50.647187 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 18:18:50 crc kubenswrapper[4859]: I1008 18:18:50.647231 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 18:18:50 crc kubenswrapper[4859]: I1008 18:18:50.647241 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 18:18:50 crc kubenswrapper[4859]: I1008 18:18:50.647253 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 18:18:50 crc kubenswrapper[4859]: I1008 18:18:50.647264 4859 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T18:18:50Z","lastTransitionTime":"2025-10-08T18:18:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 18:18:50 crc kubenswrapper[4859]: I1008 18:18:50.708149 4859 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-machine-config-operator/machine-config-daemon-82s52" podStartSLOduration=80.708120329 podStartE2EDuration="1m20.708120329s" podCreationTimestamp="2025-10-08 18:17:30 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-08 18:18:50.708017816 +0000 UTC m=+100.954857195" watchObservedRunningTime="2025-10-08 18:18:50.708120329 +0000 UTC m=+100.954959708" Oct 08 18:18:50 crc kubenswrapper[4859]: I1008 18:18:50.749296 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 18:18:50 crc kubenswrapper[4859]: I1008 18:18:50.749336 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 18:18:50 crc kubenswrapper[4859]: I1008 18:18:50.749343 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 18:18:50 crc kubenswrapper[4859]: I1008 18:18:50.749359 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 18:18:50 crc kubenswrapper[4859]: I1008 18:18:50.749368 4859 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T18:18:50Z","lastTransitionTime":"2025-10-08T18:18:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 18:18:50 crc kubenswrapper[4859]: I1008 18:18:50.754304 4859 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" podStartSLOduration=30.754282478 podStartE2EDuration="30.754282478s" podCreationTimestamp="2025-10-08 18:18:20 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-08 18:18:50.753729832 +0000 UTC m=+101.000569231" watchObservedRunningTime="2025-10-08 18:18:50.754282478 +0000 UTC m=+101.001121857" Oct 08 18:18:50 crc kubenswrapper[4859]: I1008 18:18:50.778232 4859 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-etcd/etcd-crc" podStartSLOduration=9.778207751 podStartE2EDuration="9.778207751s" podCreationTimestamp="2025-10-08 18:18:41 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-08 18:18:50.777535602 +0000 UTC m=+101.024375011" watchObservedRunningTime="2025-10-08 18:18:50.778207751 +0000 UTC m=+101.025047130" Oct 08 18:18:50 crc kubenswrapper[4859]: I1008 18:18:50.855410 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 18:18:50 crc kubenswrapper[4859]: I1008 18:18:50.855474 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 18:18:50 crc kubenswrapper[4859]: I1008 18:18:50.855490 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 18:18:50 crc kubenswrapper[4859]: I1008 18:18:50.855513 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 18:18:50 crc kubenswrapper[4859]: I1008 18:18:50.855530 4859 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T18:18:50Z","lastTransitionTime":"2025-10-08T18:18:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 18:18:50 crc kubenswrapper[4859]: I1008 18:18:50.958394 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 18:18:50 crc kubenswrapper[4859]: I1008 18:18:50.958476 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 18:18:50 crc kubenswrapper[4859]: I1008 18:18:50.958496 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 18:18:50 crc kubenswrapper[4859]: I1008 18:18:50.958529 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 18:18:50 crc kubenswrapper[4859]: I1008 18:18:50.958553 4859 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T18:18:50Z","lastTransitionTime":"2025-10-08T18:18:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 18:18:51 crc kubenswrapper[4859]: I1008 18:18:51.061392 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 18:18:51 crc kubenswrapper[4859]: I1008 18:18:51.061436 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 18:18:51 crc kubenswrapper[4859]: I1008 18:18:51.061447 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 18:18:51 crc kubenswrapper[4859]: I1008 18:18:51.061465 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 18:18:51 crc kubenswrapper[4859]: I1008 18:18:51.061477 4859 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T18:18:51Z","lastTransitionTime":"2025-10-08T18:18:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 18:18:51 crc kubenswrapper[4859]: I1008 18:18:51.165795 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 18:18:51 crc kubenswrapper[4859]: I1008 18:18:51.165863 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 18:18:51 crc kubenswrapper[4859]: I1008 18:18:51.165881 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 18:18:51 crc kubenswrapper[4859]: I1008 18:18:51.165907 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 18:18:51 crc kubenswrapper[4859]: I1008 18:18:51.165926 4859 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T18:18:51Z","lastTransitionTime":"2025-10-08T18:18:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 18:18:51 crc kubenswrapper[4859]: I1008 18:18:51.268807 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 18:18:51 crc kubenswrapper[4859]: I1008 18:18:51.268858 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 18:18:51 crc kubenswrapper[4859]: I1008 18:18:51.268871 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 18:18:51 crc kubenswrapper[4859]: I1008 18:18:51.268894 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 18:18:51 crc kubenswrapper[4859]: I1008 18:18:51.268909 4859 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T18:18:51Z","lastTransitionTime":"2025-10-08T18:18:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 18:18:51 crc kubenswrapper[4859]: I1008 18:18:51.372475 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 18:18:51 crc kubenswrapper[4859]: I1008 18:18:51.372556 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 18:18:51 crc kubenswrapper[4859]: I1008 18:18:51.372574 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 18:18:51 crc kubenswrapper[4859]: I1008 18:18:51.372602 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 18:18:51 crc kubenswrapper[4859]: I1008 18:18:51.372623 4859 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T18:18:51Z","lastTransitionTime":"2025-10-08T18:18:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 18:18:51 crc kubenswrapper[4859]: I1008 18:18:51.469446 4859 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 08 18:18:51 crc kubenswrapper[4859]: E1008 18:18:51.469660 4859 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 08 18:18:51 crc kubenswrapper[4859]: I1008 18:18:51.469747 4859 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 08 18:18:51 crc kubenswrapper[4859]: I1008 18:18:51.469801 4859 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 08 18:18:51 crc kubenswrapper[4859]: E1008 18:18:51.469842 4859 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 08 18:18:51 crc kubenswrapper[4859]: E1008 18:18:51.470087 4859 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 08 18:18:51 crc kubenswrapper[4859]: I1008 18:18:51.470508 4859 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-4b4sz" Oct 08 18:18:51 crc kubenswrapper[4859]: E1008 18:18:51.470918 4859 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-4b4sz" podUID="2e53cd9b-64f7-4e07-8b96-fb77847c9ac6" Oct 08 18:18:51 crc kubenswrapper[4859]: I1008 18:18:51.475785 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 18:18:51 crc kubenswrapper[4859]: I1008 18:18:51.475837 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 18:18:51 crc kubenswrapper[4859]: I1008 18:18:51.475855 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 18:18:51 crc kubenswrapper[4859]: I1008 18:18:51.475880 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 18:18:51 crc kubenswrapper[4859]: I1008 18:18:51.475899 4859 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T18:18:51Z","lastTransitionTime":"2025-10-08T18:18:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 18:18:51 crc kubenswrapper[4859]: I1008 18:18:51.579391 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 18:18:51 crc kubenswrapper[4859]: I1008 18:18:51.579937 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 18:18:51 crc kubenswrapper[4859]: I1008 18:18:51.579977 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 18:18:51 crc kubenswrapper[4859]: I1008 18:18:51.579999 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 18:18:51 crc kubenswrapper[4859]: I1008 18:18:51.580014 4859 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T18:18:51Z","lastTransitionTime":"2025-10-08T18:18:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 18:18:51 crc kubenswrapper[4859]: I1008 18:18:51.684868 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 18:18:51 crc kubenswrapper[4859]: I1008 18:18:51.684918 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 18:18:51 crc kubenswrapper[4859]: I1008 18:18:51.684933 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 18:18:51 crc kubenswrapper[4859]: I1008 18:18:51.684952 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 18:18:51 crc kubenswrapper[4859]: I1008 18:18:51.684972 4859 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T18:18:51Z","lastTransitionTime":"2025-10-08T18:18:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 18:18:51 crc kubenswrapper[4859]: I1008 18:18:51.788001 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 18:18:51 crc kubenswrapper[4859]: I1008 18:18:51.788073 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 18:18:51 crc kubenswrapper[4859]: I1008 18:18:51.788090 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 18:18:51 crc kubenswrapper[4859]: I1008 18:18:51.788123 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 18:18:51 crc kubenswrapper[4859]: I1008 18:18:51.788143 4859 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T18:18:51Z","lastTransitionTime":"2025-10-08T18:18:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 18:18:51 crc kubenswrapper[4859]: I1008 18:18:51.890864 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 18:18:51 crc kubenswrapper[4859]: I1008 18:18:51.890945 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 18:18:51 crc kubenswrapper[4859]: I1008 18:18:51.890962 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 18:18:51 crc kubenswrapper[4859]: I1008 18:18:51.890992 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 18:18:51 crc kubenswrapper[4859]: I1008 18:18:51.891009 4859 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T18:18:51Z","lastTransitionTime":"2025-10-08T18:18:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 18:18:51 crc kubenswrapper[4859]: I1008 18:18:51.994676 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 18:18:51 crc kubenswrapper[4859]: I1008 18:18:51.994761 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 18:18:51 crc kubenswrapper[4859]: I1008 18:18:51.994774 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 18:18:51 crc kubenswrapper[4859]: I1008 18:18:51.994794 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 18:18:51 crc kubenswrapper[4859]: I1008 18:18:51.994808 4859 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T18:18:51Z","lastTransitionTime":"2025-10-08T18:18:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 18:18:52 crc kubenswrapper[4859]: I1008 18:18:52.097511 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 18:18:52 crc kubenswrapper[4859]: I1008 18:18:52.097561 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 18:18:52 crc kubenswrapper[4859]: I1008 18:18:52.097574 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 18:18:52 crc kubenswrapper[4859]: I1008 18:18:52.097594 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 18:18:52 crc kubenswrapper[4859]: I1008 18:18:52.097609 4859 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T18:18:52Z","lastTransitionTime":"2025-10-08T18:18:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 18:18:52 crc kubenswrapper[4859]: I1008 18:18:52.200797 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 18:18:52 crc kubenswrapper[4859]: I1008 18:18:52.200848 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 18:18:52 crc kubenswrapper[4859]: I1008 18:18:52.200862 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 18:18:52 crc kubenswrapper[4859]: I1008 18:18:52.200882 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 18:18:52 crc kubenswrapper[4859]: I1008 18:18:52.200893 4859 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T18:18:52Z","lastTransitionTime":"2025-10-08T18:18:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 18:18:52 crc kubenswrapper[4859]: I1008 18:18:52.304235 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 18:18:52 crc kubenswrapper[4859]: I1008 18:18:52.304306 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 18:18:52 crc kubenswrapper[4859]: I1008 18:18:52.304330 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 18:18:52 crc kubenswrapper[4859]: I1008 18:18:52.304362 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 18:18:52 crc kubenswrapper[4859]: I1008 18:18:52.304385 4859 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T18:18:52Z","lastTransitionTime":"2025-10-08T18:18:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 18:18:52 crc kubenswrapper[4859]: I1008 18:18:52.407351 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 18:18:52 crc kubenswrapper[4859]: I1008 18:18:52.407414 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 18:18:52 crc kubenswrapper[4859]: I1008 18:18:52.407433 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 18:18:52 crc kubenswrapper[4859]: I1008 18:18:52.407457 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 18:18:52 crc kubenswrapper[4859]: I1008 18:18:52.407474 4859 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T18:18:52Z","lastTransitionTime":"2025-10-08T18:18:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 18:18:52 crc kubenswrapper[4859]: I1008 18:18:52.510633 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 18:18:52 crc kubenswrapper[4859]: I1008 18:18:52.510794 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 18:18:52 crc kubenswrapper[4859]: I1008 18:18:52.510825 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 18:18:52 crc kubenswrapper[4859]: I1008 18:18:52.510910 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 18:18:52 crc kubenswrapper[4859]: I1008 18:18:52.510935 4859 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T18:18:52Z","lastTransitionTime":"2025-10-08T18:18:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 18:18:52 crc kubenswrapper[4859]: I1008 18:18:52.614280 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 18:18:52 crc kubenswrapper[4859]: I1008 18:18:52.614334 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 18:18:52 crc kubenswrapper[4859]: I1008 18:18:52.614349 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 18:18:52 crc kubenswrapper[4859]: I1008 18:18:52.614368 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 18:18:52 crc kubenswrapper[4859]: I1008 18:18:52.614381 4859 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T18:18:52Z","lastTransitionTime":"2025-10-08T18:18:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 18:18:52 crc kubenswrapper[4859]: I1008 18:18:52.718419 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 18:18:52 crc kubenswrapper[4859]: I1008 18:18:52.718484 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 18:18:52 crc kubenswrapper[4859]: I1008 18:18:52.718503 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 18:18:52 crc kubenswrapper[4859]: I1008 18:18:52.718528 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 18:18:52 crc kubenswrapper[4859]: I1008 18:18:52.718550 4859 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T18:18:52Z","lastTransitionTime":"2025-10-08T18:18:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 18:18:52 crc kubenswrapper[4859]: I1008 18:18:52.820956 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 18:18:52 crc kubenswrapper[4859]: I1008 18:18:52.821026 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 18:18:52 crc kubenswrapper[4859]: I1008 18:18:52.821049 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 18:18:52 crc kubenswrapper[4859]: I1008 18:18:52.821082 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 18:18:52 crc kubenswrapper[4859]: I1008 18:18:52.821104 4859 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T18:18:52Z","lastTransitionTime":"2025-10-08T18:18:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 18:18:52 crc kubenswrapper[4859]: I1008 18:18:52.924724 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 18:18:52 crc kubenswrapper[4859]: I1008 18:18:52.924787 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 18:18:52 crc kubenswrapper[4859]: I1008 18:18:52.924843 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 18:18:52 crc kubenswrapper[4859]: I1008 18:18:52.924876 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 18:18:52 crc kubenswrapper[4859]: I1008 18:18:52.924898 4859 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T18:18:52Z","lastTransitionTime":"2025-10-08T18:18:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 18:18:53 crc kubenswrapper[4859]: I1008 18:18:53.028399 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 18:18:53 crc kubenswrapper[4859]: I1008 18:18:53.028463 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 18:18:53 crc kubenswrapper[4859]: I1008 18:18:53.028485 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 18:18:53 crc kubenswrapper[4859]: I1008 18:18:53.028519 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 18:18:53 crc kubenswrapper[4859]: I1008 18:18:53.028539 4859 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T18:18:53Z","lastTransitionTime":"2025-10-08T18:18:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 18:18:53 crc kubenswrapper[4859]: I1008 18:18:53.131558 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 18:18:53 crc kubenswrapper[4859]: I1008 18:18:53.131992 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 18:18:53 crc kubenswrapper[4859]: I1008 18:18:53.132247 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 18:18:53 crc kubenswrapper[4859]: I1008 18:18:53.132450 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 18:18:53 crc kubenswrapper[4859]: I1008 18:18:53.132592 4859 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T18:18:53Z","lastTransitionTime":"2025-10-08T18:18:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 18:18:53 crc kubenswrapper[4859]: I1008 18:18:53.236145 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 18:18:53 crc kubenswrapper[4859]: I1008 18:18:53.236220 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 18:18:53 crc kubenswrapper[4859]: I1008 18:18:53.236230 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 18:18:53 crc kubenswrapper[4859]: I1008 18:18:53.236256 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 18:18:53 crc kubenswrapper[4859]: I1008 18:18:53.236268 4859 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T18:18:53Z","lastTransitionTime":"2025-10-08T18:18:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 18:18:53 crc kubenswrapper[4859]: I1008 18:18:53.339896 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 18:18:53 crc kubenswrapper[4859]: I1008 18:18:53.340393 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 18:18:53 crc kubenswrapper[4859]: I1008 18:18:53.340588 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 18:18:53 crc kubenswrapper[4859]: I1008 18:18:53.340773 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 18:18:53 crc kubenswrapper[4859]: I1008 18:18:53.341078 4859 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T18:18:53Z","lastTransitionTime":"2025-10-08T18:18:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 18:18:53 crc kubenswrapper[4859]: I1008 18:18:53.444606 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 18:18:53 crc kubenswrapper[4859]: I1008 18:18:53.444661 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 18:18:53 crc kubenswrapper[4859]: I1008 18:18:53.444680 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 18:18:53 crc kubenswrapper[4859]: I1008 18:18:53.444732 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 18:18:53 crc kubenswrapper[4859]: I1008 18:18:53.444748 4859 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T18:18:53Z","lastTransitionTime":"2025-10-08T18:18:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 18:18:53 crc kubenswrapper[4859]: I1008 18:18:53.469169 4859 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 08 18:18:53 crc kubenswrapper[4859]: E1008 18:18:53.469814 4859 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 08 18:18:53 crc kubenswrapper[4859]: I1008 18:18:53.470067 4859 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-4b4sz" Oct 08 18:18:53 crc kubenswrapper[4859]: E1008 18:18:53.470318 4859 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-4b4sz" podUID="2e53cd9b-64f7-4e07-8b96-fb77847c9ac6" Oct 08 18:18:53 crc kubenswrapper[4859]: I1008 18:18:53.473595 4859 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 08 18:18:53 crc kubenswrapper[4859]: E1008 18:18:53.474073 4859 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 08 18:18:53 crc kubenswrapper[4859]: I1008 18:18:53.476078 4859 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 08 18:18:53 crc kubenswrapper[4859]: E1008 18:18:53.476277 4859 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 08 18:18:53 crc kubenswrapper[4859]: I1008 18:18:53.547559 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 18:18:53 crc kubenswrapper[4859]: I1008 18:18:53.547611 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 18:18:53 crc kubenswrapper[4859]: I1008 18:18:53.547621 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 18:18:53 crc kubenswrapper[4859]: I1008 18:18:53.547642 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 18:18:53 crc kubenswrapper[4859]: I1008 18:18:53.547654 4859 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T18:18:53Z","lastTransitionTime":"2025-10-08T18:18:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 18:18:53 crc kubenswrapper[4859]: I1008 18:18:53.651264 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 18:18:53 crc kubenswrapper[4859]: I1008 18:18:53.651335 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 18:18:53 crc kubenswrapper[4859]: I1008 18:18:53.651360 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 18:18:53 crc kubenswrapper[4859]: I1008 18:18:53.651392 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 18:18:53 crc kubenswrapper[4859]: I1008 18:18:53.651422 4859 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T18:18:53Z","lastTransitionTime":"2025-10-08T18:18:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 18:18:53 crc kubenswrapper[4859]: I1008 18:18:53.754468 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 18:18:53 crc kubenswrapper[4859]: I1008 18:18:53.754516 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 18:18:53 crc kubenswrapper[4859]: I1008 18:18:53.754530 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 18:18:53 crc kubenswrapper[4859]: I1008 18:18:53.754549 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 18:18:53 crc kubenswrapper[4859]: I1008 18:18:53.754561 4859 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T18:18:53Z","lastTransitionTime":"2025-10-08T18:18:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 18:18:53 crc kubenswrapper[4859]: I1008 18:18:53.857580 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 18:18:53 crc kubenswrapper[4859]: I1008 18:18:53.857649 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 18:18:53 crc kubenswrapper[4859]: I1008 18:18:53.857666 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 18:18:53 crc kubenswrapper[4859]: I1008 18:18:53.857727 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 18:18:53 crc kubenswrapper[4859]: I1008 18:18:53.857748 4859 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T18:18:53Z","lastTransitionTime":"2025-10-08T18:18:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 18:18:53 crc kubenswrapper[4859]: I1008 18:18:53.960853 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 18:18:53 crc kubenswrapper[4859]: I1008 18:18:53.960934 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 18:18:53 crc kubenswrapper[4859]: I1008 18:18:53.960957 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 18:18:53 crc kubenswrapper[4859]: I1008 18:18:53.960985 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 18:18:53 crc kubenswrapper[4859]: I1008 18:18:53.961004 4859 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T18:18:53Z","lastTransitionTime":"2025-10-08T18:18:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 18:18:54 crc kubenswrapper[4859]: I1008 18:18:54.064143 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 18:18:54 crc kubenswrapper[4859]: I1008 18:18:54.064207 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 18:18:54 crc kubenswrapper[4859]: I1008 18:18:54.064224 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 18:18:54 crc kubenswrapper[4859]: I1008 18:18:54.064249 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 18:18:54 crc kubenswrapper[4859]: I1008 18:18:54.064268 4859 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T18:18:54Z","lastTransitionTime":"2025-10-08T18:18:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 18:18:54 crc kubenswrapper[4859]: I1008 18:18:54.167196 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 18:18:54 crc kubenswrapper[4859]: I1008 18:18:54.167256 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 18:18:54 crc kubenswrapper[4859]: I1008 18:18:54.167270 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 18:18:54 crc kubenswrapper[4859]: I1008 18:18:54.167296 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 18:18:54 crc kubenswrapper[4859]: I1008 18:18:54.167313 4859 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T18:18:54Z","lastTransitionTime":"2025-10-08T18:18:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 18:18:54 crc kubenswrapper[4859]: I1008 18:18:54.269949 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 18:18:54 crc kubenswrapper[4859]: I1008 18:18:54.270006 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 18:18:54 crc kubenswrapper[4859]: I1008 18:18:54.270018 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 18:18:54 crc kubenswrapper[4859]: I1008 18:18:54.270036 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 18:18:54 crc kubenswrapper[4859]: I1008 18:18:54.270050 4859 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T18:18:54Z","lastTransitionTime":"2025-10-08T18:18:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 18:18:54 crc kubenswrapper[4859]: I1008 18:18:54.373459 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 18:18:54 crc kubenswrapper[4859]: I1008 18:18:54.373509 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 18:18:54 crc kubenswrapper[4859]: I1008 18:18:54.373527 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 18:18:54 crc kubenswrapper[4859]: I1008 18:18:54.373547 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 18:18:54 crc kubenswrapper[4859]: I1008 18:18:54.373563 4859 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T18:18:54Z","lastTransitionTime":"2025-10-08T18:18:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 18:18:54 crc kubenswrapper[4859]: I1008 18:18:54.476106 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 18:18:54 crc kubenswrapper[4859]: I1008 18:18:54.476437 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 18:18:54 crc kubenswrapper[4859]: I1008 18:18:54.476535 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 18:18:54 crc kubenswrapper[4859]: I1008 18:18:54.476620 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 18:18:54 crc kubenswrapper[4859]: I1008 18:18:54.476748 4859 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T18:18:54Z","lastTransitionTime":"2025-10-08T18:18:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 18:18:54 crc kubenswrapper[4859]: I1008 18:18:54.580052 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 18:18:54 crc kubenswrapper[4859]: I1008 18:18:54.580121 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 18:18:54 crc kubenswrapper[4859]: I1008 18:18:54.580136 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 18:18:54 crc kubenswrapper[4859]: I1008 18:18:54.580156 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 18:18:54 crc kubenswrapper[4859]: I1008 18:18:54.580169 4859 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T18:18:54Z","lastTransitionTime":"2025-10-08T18:18:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 18:18:54 crc kubenswrapper[4859]: I1008 18:18:54.683959 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 18:18:54 crc kubenswrapper[4859]: I1008 18:18:54.684032 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 18:18:54 crc kubenswrapper[4859]: I1008 18:18:54.684054 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 18:18:54 crc kubenswrapper[4859]: I1008 18:18:54.684086 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 18:18:54 crc kubenswrapper[4859]: I1008 18:18:54.684112 4859 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T18:18:54Z","lastTransitionTime":"2025-10-08T18:18:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 18:18:54 crc kubenswrapper[4859]: I1008 18:18:54.788360 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 18:18:54 crc kubenswrapper[4859]: I1008 18:18:54.788447 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 18:18:54 crc kubenswrapper[4859]: I1008 18:18:54.788476 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 18:18:54 crc kubenswrapper[4859]: I1008 18:18:54.788511 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 18:18:54 crc kubenswrapper[4859]: I1008 18:18:54.788531 4859 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T18:18:54Z","lastTransitionTime":"2025-10-08T18:18:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 18:18:54 crc kubenswrapper[4859]: I1008 18:18:54.892188 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 18:18:54 crc kubenswrapper[4859]: I1008 18:18:54.892285 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 18:18:54 crc kubenswrapper[4859]: I1008 18:18:54.892304 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 18:18:54 crc kubenswrapper[4859]: I1008 18:18:54.892333 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 18:18:54 crc kubenswrapper[4859]: I1008 18:18:54.892360 4859 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T18:18:54Z","lastTransitionTime":"2025-10-08T18:18:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 18:18:54 crc kubenswrapper[4859]: I1008 18:18:54.995518 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 18:18:54 crc kubenswrapper[4859]: I1008 18:18:54.995574 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 18:18:54 crc kubenswrapper[4859]: I1008 18:18:54.995583 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 18:18:54 crc kubenswrapper[4859]: I1008 18:18:54.995602 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 18:18:54 crc kubenswrapper[4859]: I1008 18:18:54.995616 4859 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T18:18:54Z","lastTransitionTime":"2025-10-08T18:18:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 18:18:55 crc kubenswrapper[4859]: I1008 18:18:55.099874 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 18:18:55 crc kubenswrapper[4859]: I1008 18:18:55.099940 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 18:18:55 crc kubenswrapper[4859]: I1008 18:18:55.099958 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 18:18:55 crc kubenswrapper[4859]: I1008 18:18:55.099985 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 18:18:55 crc kubenswrapper[4859]: I1008 18:18:55.100004 4859 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T18:18:55Z","lastTransitionTime":"2025-10-08T18:18:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 18:18:55 crc kubenswrapper[4859]: I1008 18:18:55.204649 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 18:18:55 crc kubenswrapper[4859]: I1008 18:18:55.204755 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 18:18:55 crc kubenswrapper[4859]: I1008 18:18:55.204779 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 18:18:55 crc kubenswrapper[4859]: I1008 18:18:55.204809 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 18:18:55 crc kubenswrapper[4859]: I1008 18:18:55.204835 4859 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T18:18:55Z","lastTransitionTime":"2025-10-08T18:18:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 18:18:55 crc kubenswrapper[4859]: I1008 18:18:55.308475 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 18:18:55 crc kubenswrapper[4859]: I1008 18:18:55.308549 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 18:18:55 crc kubenswrapper[4859]: I1008 18:18:55.308572 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 18:18:55 crc kubenswrapper[4859]: I1008 18:18:55.308607 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 18:18:55 crc kubenswrapper[4859]: I1008 18:18:55.308630 4859 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T18:18:55Z","lastTransitionTime":"2025-10-08T18:18:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 18:18:55 crc kubenswrapper[4859]: I1008 18:18:55.411561 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 18:18:55 crc kubenswrapper[4859]: I1008 18:18:55.411618 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 18:18:55 crc kubenswrapper[4859]: I1008 18:18:55.411643 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 18:18:55 crc kubenswrapper[4859]: I1008 18:18:55.411661 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 18:18:55 crc kubenswrapper[4859]: I1008 18:18:55.411673 4859 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T18:18:55Z","lastTransitionTime":"2025-10-08T18:18:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 18:18:55 crc kubenswrapper[4859]: I1008 18:18:55.469462 4859 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 08 18:18:55 crc kubenswrapper[4859]: I1008 18:18:55.469505 4859 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 08 18:18:55 crc kubenswrapper[4859]: I1008 18:18:55.469505 4859 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 08 18:18:55 crc kubenswrapper[4859]: I1008 18:18:55.469462 4859 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-4b4sz" Oct 08 18:18:55 crc kubenswrapper[4859]: E1008 18:18:55.469619 4859 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 08 18:18:55 crc kubenswrapper[4859]: E1008 18:18:55.469809 4859 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-4b4sz" podUID="2e53cd9b-64f7-4e07-8b96-fb77847c9ac6" Oct 08 18:18:55 crc kubenswrapper[4859]: E1008 18:18:55.470089 4859 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 08 18:18:55 crc kubenswrapper[4859]: E1008 18:18:55.470197 4859 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 08 18:18:55 crc kubenswrapper[4859]: I1008 18:18:55.515324 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 18:18:55 crc kubenswrapper[4859]: I1008 18:18:55.515384 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 18:18:55 crc kubenswrapper[4859]: I1008 18:18:55.515394 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 18:18:55 crc kubenswrapper[4859]: I1008 18:18:55.515411 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 18:18:55 crc kubenswrapper[4859]: I1008 18:18:55.515425 4859 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T18:18:55Z","lastTransitionTime":"2025-10-08T18:18:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 18:18:55 crc kubenswrapper[4859]: I1008 18:18:55.617965 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 18:18:55 crc kubenswrapper[4859]: I1008 18:18:55.618025 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 18:18:55 crc kubenswrapper[4859]: I1008 18:18:55.618036 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 18:18:55 crc kubenswrapper[4859]: I1008 18:18:55.618054 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 18:18:55 crc kubenswrapper[4859]: I1008 18:18:55.618142 4859 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T18:18:55Z","lastTransitionTime":"2025-10-08T18:18:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 18:18:55 crc kubenswrapper[4859]: I1008 18:18:55.720798 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 18:18:55 crc kubenswrapper[4859]: I1008 18:18:55.720847 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 18:18:55 crc kubenswrapper[4859]: I1008 18:18:55.720858 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 18:18:55 crc kubenswrapper[4859]: I1008 18:18:55.720872 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 18:18:55 crc kubenswrapper[4859]: I1008 18:18:55.720884 4859 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T18:18:55Z","lastTransitionTime":"2025-10-08T18:18:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 18:18:55 crc kubenswrapper[4859]: I1008 18:18:55.823866 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 18:18:55 crc kubenswrapper[4859]: I1008 18:18:55.823925 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 18:18:55 crc kubenswrapper[4859]: I1008 18:18:55.823937 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 18:18:55 crc kubenswrapper[4859]: I1008 18:18:55.823956 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 18:18:55 crc kubenswrapper[4859]: I1008 18:18:55.823970 4859 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T18:18:55Z","lastTransitionTime":"2025-10-08T18:18:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 18:18:55 crc kubenswrapper[4859]: I1008 18:18:55.928245 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 18:18:55 crc kubenswrapper[4859]: I1008 18:18:55.928295 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 18:18:55 crc kubenswrapper[4859]: I1008 18:18:55.928304 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 18:18:55 crc kubenswrapper[4859]: I1008 18:18:55.928322 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 18:18:55 crc kubenswrapper[4859]: I1008 18:18:55.928332 4859 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T18:18:55Z","lastTransitionTime":"2025-10-08T18:18:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 18:18:56 crc kubenswrapper[4859]: I1008 18:18:56.030699 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 18:18:56 crc kubenswrapper[4859]: I1008 18:18:56.030753 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 18:18:56 crc kubenswrapper[4859]: I1008 18:18:56.030770 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 18:18:56 crc kubenswrapper[4859]: I1008 18:18:56.030792 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 18:18:56 crc kubenswrapper[4859]: I1008 18:18:56.030806 4859 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T18:18:56Z","lastTransitionTime":"2025-10-08T18:18:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 18:18:56 crc kubenswrapper[4859]: I1008 18:18:56.133469 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 18:18:56 crc kubenswrapper[4859]: I1008 18:18:56.133526 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 18:18:56 crc kubenswrapper[4859]: I1008 18:18:56.133536 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 18:18:56 crc kubenswrapper[4859]: I1008 18:18:56.133560 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 18:18:56 crc kubenswrapper[4859]: I1008 18:18:56.133576 4859 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T18:18:56Z","lastTransitionTime":"2025-10-08T18:18:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 18:18:56 crc kubenswrapper[4859]: I1008 18:18:56.237180 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 18:18:56 crc kubenswrapper[4859]: I1008 18:18:56.237529 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 18:18:56 crc kubenswrapper[4859]: I1008 18:18:56.237611 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 18:18:56 crc kubenswrapper[4859]: I1008 18:18:56.237780 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 18:18:56 crc kubenswrapper[4859]: I1008 18:18:56.237885 4859 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T18:18:56Z","lastTransitionTime":"2025-10-08T18:18:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 18:18:56 crc kubenswrapper[4859]: I1008 18:18:56.341106 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 18:18:56 crc kubenswrapper[4859]: I1008 18:18:56.341147 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 18:18:56 crc kubenswrapper[4859]: I1008 18:18:56.341157 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 18:18:56 crc kubenswrapper[4859]: I1008 18:18:56.341174 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 18:18:56 crc kubenswrapper[4859]: I1008 18:18:56.341185 4859 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T18:18:56Z","lastTransitionTime":"2025-10-08T18:18:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 18:18:56 crc kubenswrapper[4859]: I1008 18:18:56.444169 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 18:18:56 crc kubenswrapper[4859]: I1008 18:18:56.444217 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 18:18:56 crc kubenswrapper[4859]: I1008 18:18:56.444228 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 18:18:56 crc kubenswrapper[4859]: I1008 18:18:56.444245 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 18:18:56 crc kubenswrapper[4859]: I1008 18:18:56.444254 4859 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T18:18:56Z","lastTransitionTime":"2025-10-08T18:18:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 18:18:56 crc kubenswrapper[4859]: I1008 18:18:56.469797 4859 scope.go:117] "RemoveContainer" containerID="f3a760c557ef21fbc6b54e78c3f6c5b325f296f93a022425cca7b1399eed8661" Oct 08 18:18:56 crc kubenswrapper[4859]: E1008 18:18:56.469991 4859 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ovnkube-controller\" with CrashLoopBackOff: \"back-off 40s restarting failed container=ovnkube-controller pod=ovnkube-node-4g8gf_openshift-ovn-kubernetes(1dff864e-b75d-4e0b-b182-75f710eac8df)\"" pod="openshift-ovn-kubernetes/ovnkube-node-4g8gf" podUID="1dff864e-b75d-4e0b-b182-75f710eac8df" Oct 08 18:18:56 crc kubenswrapper[4859]: I1008 18:18:56.546758 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 18:18:56 crc kubenswrapper[4859]: I1008 18:18:56.546806 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 18:18:56 crc kubenswrapper[4859]: I1008 18:18:56.546819 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 18:18:56 crc kubenswrapper[4859]: I1008 18:18:56.546838 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 18:18:56 crc kubenswrapper[4859]: I1008 18:18:56.546853 4859 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T18:18:56Z","lastTransitionTime":"2025-10-08T18:18:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 18:18:56 crc kubenswrapper[4859]: I1008 18:18:56.650414 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 18:18:56 crc kubenswrapper[4859]: I1008 18:18:56.650469 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 18:18:56 crc kubenswrapper[4859]: I1008 18:18:56.650479 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 18:18:56 crc kubenswrapper[4859]: I1008 18:18:56.650500 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 18:18:56 crc kubenswrapper[4859]: I1008 18:18:56.650512 4859 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T18:18:56Z","lastTransitionTime":"2025-10-08T18:18:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 18:18:56 crc kubenswrapper[4859]: I1008 18:18:56.753013 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 18:18:56 crc kubenswrapper[4859]: I1008 18:18:56.753054 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 18:18:56 crc kubenswrapper[4859]: I1008 18:18:56.753067 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 18:18:56 crc kubenswrapper[4859]: I1008 18:18:56.753084 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 18:18:56 crc kubenswrapper[4859]: I1008 18:18:56.753096 4859 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T18:18:56Z","lastTransitionTime":"2025-10-08T18:18:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 18:18:56 crc kubenswrapper[4859]: I1008 18:18:56.855903 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 18:18:56 crc kubenswrapper[4859]: I1008 18:18:56.855968 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 18:18:56 crc kubenswrapper[4859]: I1008 18:18:56.855979 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 18:18:56 crc kubenswrapper[4859]: I1008 18:18:56.856008 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 18:18:56 crc kubenswrapper[4859]: I1008 18:18:56.856023 4859 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T18:18:56Z","lastTransitionTime":"2025-10-08T18:18:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 18:18:56 crc kubenswrapper[4859]: I1008 18:18:56.958796 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 18:18:56 crc kubenswrapper[4859]: I1008 18:18:56.958839 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 18:18:56 crc kubenswrapper[4859]: I1008 18:18:56.958849 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 18:18:56 crc kubenswrapper[4859]: I1008 18:18:56.958866 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 18:18:56 crc kubenswrapper[4859]: I1008 18:18:56.958877 4859 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T18:18:56Z","lastTransitionTime":"2025-10-08T18:18:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 18:18:57 crc kubenswrapper[4859]: I1008 18:18:57.062220 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 18:18:57 crc kubenswrapper[4859]: I1008 18:18:57.062637 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 18:18:57 crc kubenswrapper[4859]: I1008 18:18:57.062907 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 18:18:57 crc kubenswrapper[4859]: I1008 18:18:57.063110 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 18:18:57 crc kubenswrapper[4859]: I1008 18:18:57.063279 4859 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T18:18:57Z","lastTransitionTime":"2025-10-08T18:18:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 18:18:57 crc kubenswrapper[4859]: I1008 18:18:57.166725 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 18:18:57 crc kubenswrapper[4859]: I1008 18:18:57.167089 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 18:18:57 crc kubenswrapper[4859]: I1008 18:18:57.167248 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 18:18:57 crc kubenswrapper[4859]: I1008 18:18:57.167391 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 18:18:57 crc kubenswrapper[4859]: I1008 18:18:57.167532 4859 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T18:18:57Z","lastTransitionTime":"2025-10-08T18:18:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 18:18:57 crc kubenswrapper[4859]: I1008 18:18:57.271533 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 18:18:57 crc kubenswrapper[4859]: I1008 18:18:57.272090 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 18:18:57 crc kubenswrapper[4859]: I1008 18:18:57.272285 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 18:18:57 crc kubenswrapper[4859]: I1008 18:18:57.272476 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 18:18:57 crc kubenswrapper[4859]: I1008 18:18:57.272657 4859 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T18:18:57Z","lastTransitionTime":"2025-10-08T18:18:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 18:18:57 crc kubenswrapper[4859]: I1008 18:18:57.375484 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 18:18:57 crc kubenswrapper[4859]: I1008 18:18:57.375998 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 18:18:57 crc kubenswrapper[4859]: I1008 18:18:57.376159 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 18:18:57 crc kubenswrapper[4859]: I1008 18:18:57.376311 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 18:18:57 crc kubenswrapper[4859]: I1008 18:18:57.376465 4859 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T18:18:57Z","lastTransitionTime":"2025-10-08T18:18:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 18:18:57 crc kubenswrapper[4859]: I1008 18:18:57.469537 4859 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-4b4sz" Oct 08 18:18:57 crc kubenswrapper[4859]: I1008 18:18:57.469721 4859 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 08 18:18:57 crc kubenswrapper[4859]: I1008 18:18:57.469744 4859 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 08 18:18:57 crc kubenswrapper[4859]: I1008 18:18:57.469847 4859 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 08 18:18:57 crc kubenswrapper[4859]: E1008 18:18:57.469855 4859 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-4b4sz" podUID="2e53cd9b-64f7-4e07-8b96-fb77847c9ac6" Oct 08 18:18:57 crc kubenswrapper[4859]: E1008 18:18:57.470083 4859 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 08 18:18:57 crc kubenswrapper[4859]: E1008 18:18:57.470200 4859 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 08 18:18:57 crc kubenswrapper[4859]: E1008 18:18:57.470242 4859 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 08 18:18:57 crc kubenswrapper[4859]: I1008 18:18:57.479931 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 18:18:57 crc kubenswrapper[4859]: I1008 18:18:57.479994 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 18:18:57 crc kubenswrapper[4859]: I1008 18:18:57.480013 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 18:18:57 crc kubenswrapper[4859]: I1008 18:18:57.480039 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 18:18:57 crc kubenswrapper[4859]: I1008 18:18:57.480061 4859 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T18:18:57Z","lastTransitionTime":"2025-10-08T18:18:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 18:18:57 crc kubenswrapper[4859]: I1008 18:18:57.583234 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 18:18:57 crc kubenswrapper[4859]: I1008 18:18:57.583305 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 18:18:57 crc kubenswrapper[4859]: I1008 18:18:57.583319 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 18:18:57 crc kubenswrapper[4859]: I1008 18:18:57.583337 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 18:18:57 crc kubenswrapper[4859]: I1008 18:18:57.583350 4859 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T18:18:57Z","lastTransitionTime":"2025-10-08T18:18:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 18:18:57 crc kubenswrapper[4859]: I1008 18:18:57.685999 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 18:18:57 crc kubenswrapper[4859]: I1008 18:18:57.686054 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 18:18:57 crc kubenswrapper[4859]: I1008 18:18:57.686069 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 18:18:57 crc kubenswrapper[4859]: I1008 18:18:57.686089 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 18:18:57 crc kubenswrapper[4859]: I1008 18:18:57.686102 4859 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T18:18:57Z","lastTransitionTime":"2025-10-08T18:18:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 18:18:57 crc kubenswrapper[4859]: I1008 18:18:57.788601 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 18:18:57 crc kubenswrapper[4859]: I1008 18:18:57.788661 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 18:18:57 crc kubenswrapper[4859]: I1008 18:18:57.788673 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 18:18:57 crc kubenswrapper[4859]: I1008 18:18:57.788725 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 18:18:57 crc kubenswrapper[4859]: I1008 18:18:57.788744 4859 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T18:18:57Z","lastTransitionTime":"2025-10-08T18:18:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 18:18:57 crc kubenswrapper[4859]: I1008 18:18:57.892424 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 18:18:57 crc kubenswrapper[4859]: I1008 18:18:57.892479 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 18:18:57 crc kubenswrapper[4859]: I1008 18:18:57.892495 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 18:18:57 crc kubenswrapper[4859]: I1008 18:18:57.892519 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 18:18:57 crc kubenswrapper[4859]: I1008 18:18:57.892534 4859 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T18:18:57Z","lastTransitionTime":"2025-10-08T18:18:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 18:18:57 crc kubenswrapper[4859]: I1008 18:18:57.996039 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 18:18:57 crc kubenswrapper[4859]: I1008 18:18:57.996116 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 18:18:57 crc kubenswrapper[4859]: I1008 18:18:57.996126 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 18:18:57 crc kubenswrapper[4859]: I1008 18:18:57.996144 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 18:18:57 crc kubenswrapper[4859]: I1008 18:18:57.996156 4859 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T18:18:57Z","lastTransitionTime":"2025-10-08T18:18:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 18:18:58 crc kubenswrapper[4859]: I1008 18:18:58.099531 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 18:18:58 crc kubenswrapper[4859]: I1008 18:18:58.100584 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 18:18:58 crc kubenswrapper[4859]: I1008 18:18:58.100612 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 18:18:58 crc kubenswrapper[4859]: I1008 18:18:58.100651 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 18:18:58 crc kubenswrapper[4859]: I1008 18:18:58.100674 4859 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T18:18:58Z","lastTransitionTime":"2025-10-08T18:18:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 18:18:58 crc kubenswrapper[4859]: I1008 18:18:58.203552 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 18:18:58 crc kubenswrapper[4859]: I1008 18:18:58.203675 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 18:18:58 crc kubenswrapper[4859]: I1008 18:18:58.203730 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 18:18:58 crc kubenswrapper[4859]: I1008 18:18:58.203757 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 18:18:58 crc kubenswrapper[4859]: I1008 18:18:58.203777 4859 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T18:18:58Z","lastTransitionTime":"2025-10-08T18:18:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 18:18:58 crc kubenswrapper[4859]: I1008 18:18:58.306385 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 18:18:58 crc kubenswrapper[4859]: I1008 18:18:58.306430 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 18:18:58 crc kubenswrapper[4859]: I1008 18:18:58.306439 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 18:18:58 crc kubenswrapper[4859]: I1008 18:18:58.306456 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 18:18:58 crc kubenswrapper[4859]: I1008 18:18:58.306469 4859 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T18:18:58Z","lastTransitionTime":"2025-10-08T18:18:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 18:18:58 crc kubenswrapper[4859]: I1008 18:18:58.410585 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 18:18:58 crc kubenswrapper[4859]: I1008 18:18:58.410673 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 18:18:58 crc kubenswrapper[4859]: I1008 18:18:58.410737 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 18:18:58 crc kubenswrapper[4859]: I1008 18:18:58.410772 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 18:18:58 crc kubenswrapper[4859]: I1008 18:18:58.410803 4859 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T18:18:58Z","lastTransitionTime":"2025-10-08T18:18:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 18:18:58 crc kubenswrapper[4859]: I1008 18:18:58.513962 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 18:18:58 crc kubenswrapper[4859]: I1008 18:18:58.514027 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 18:18:58 crc kubenswrapper[4859]: I1008 18:18:58.514039 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 18:18:58 crc kubenswrapper[4859]: I1008 18:18:58.514059 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 18:18:58 crc kubenswrapper[4859]: I1008 18:18:58.514071 4859 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T18:18:58Z","lastTransitionTime":"2025-10-08T18:18:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 18:18:58 crc kubenswrapper[4859]: I1008 18:18:58.616640 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 18:18:58 crc kubenswrapper[4859]: I1008 18:18:58.616762 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 18:18:58 crc kubenswrapper[4859]: I1008 18:18:58.616789 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 18:18:58 crc kubenswrapper[4859]: I1008 18:18:58.616813 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 18:18:58 crc kubenswrapper[4859]: I1008 18:18:58.616837 4859 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T18:18:58Z","lastTransitionTime":"2025-10-08T18:18:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 18:18:58 crc kubenswrapper[4859]: I1008 18:18:58.720506 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 18:18:58 crc kubenswrapper[4859]: I1008 18:18:58.720559 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 18:18:58 crc kubenswrapper[4859]: I1008 18:18:58.720579 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 18:18:58 crc kubenswrapper[4859]: I1008 18:18:58.720600 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 18:18:58 crc kubenswrapper[4859]: I1008 18:18:58.720623 4859 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T18:18:58Z","lastTransitionTime":"2025-10-08T18:18:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 18:18:58 crc kubenswrapper[4859]: I1008 18:18:58.823563 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 18:18:58 crc kubenswrapper[4859]: I1008 18:18:58.823605 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 18:18:58 crc kubenswrapper[4859]: I1008 18:18:58.823616 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 18:18:58 crc kubenswrapper[4859]: I1008 18:18:58.823634 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 18:18:58 crc kubenswrapper[4859]: I1008 18:18:58.823644 4859 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T18:18:58Z","lastTransitionTime":"2025-10-08T18:18:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 18:18:58 crc kubenswrapper[4859]: I1008 18:18:58.926963 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 18:18:58 crc kubenswrapper[4859]: I1008 18:18:58.927034 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 18:18:58 crc kubenswrapper[4859]: I1008 18:18:58.927048 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 18:18:58 crc kubenswrapper[4859]: I1008 18:18:58.927071 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 18:18:58 crc kubenswrapper[4859]: I1008 18:18:58.927086 4859 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T18:18:58Z","lastTransitionTime":"2025-10-08T18:18:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 18:18:59 crc kubenswrapper[4859]: I1008 18:18:59.029813 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 18:18:59 crc kubenswrapper[4859]: I1008 18:18:59.029904 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 18:18:59 crc kubenswrapper[4859]: I1008 18:18:59.029920 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 18:18:59 crc kubenswrapper[4859]: I1008 18:18:59.029945 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 18:18:59 crc kubenswrapper[4859]: I1008 18:18:59.029960 4859 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T18:18:59Z","lastTransitionTime":"2025-10-08T18:18:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 18:18:59 crc kubenswrapper[4859]: I1008 18:18:59.132767 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 18:18:59 crc kubenswrapper[4859]: I1008 18:18:59.132868 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 18:18:59 crc kubenswrapper[4859]: I1008 18:18:59.132881 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 18:18:59 crc kubenswrapper[4859]: I1008 18:18:59.132900 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 18:18:59 crc kubenswrapper[4859]: I1008 18:18:59.132910 4859 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T18:18:59Z","lastTransitionTime":"2025-10-08T18:18:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 18:18:59 crc kubenswrapper[4859]: I1008 18:18:59.236906 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 18:18:59 crc kubenswrapper[4859]: I1008 18:18:59.236980 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 18:18:59 crc kubenswrapper[4859]: I1008 18:18:59.236997 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 18:18:59 crc kubenswrapper[4859]: I1008 18:18:59.237023 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 18:18:59 crc kubenswrapper[4859]: I1008 18:18:59.237043 4859 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T18:18:59Z","lastTransitionTime":"2025-10-08T18:18:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 18:18:59 crc kubenswrapper[4859]: I1008 18:18:59.340797 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 18:18:59 crc kubenswrapper[4859]: I1008 18:18:59.340853 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 18:18:59 crc kubenswrapper[4859]: I1008 18:18:59.340865 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 18:18:59 crc kubenswrapper[4859]: I1008 18:18:59.340885 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 18:18:59 crc kubenswrapper[4859]: I1008 18:18:59.340897 4859 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T18:18:59Z","lastTransitionTime":"2025-10-08T18:18:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 18:18:59 crc kubenswrapper[4859]: I1008 18:18:59.443892 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 18:18:59 crc kubenswrapper[4859]: I1008 18:18:59.443960 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 18:18:59 crc kubenswrapper[4859]: I1008 18:18:59.443972 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 18:18:59 crc kubenswrapper[4859]: I1008 18:18:59.443991 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 18:18:59 crc kubenswrapper[4859]: I1008 18:18:59.444006 4859 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T18:18:59Z","lastTransitionTime":"2025-10-08T18:18:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 18:18:59 crc kubenswrapper[4859]: I1008 18:18:59.469701 4859 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-4b4sz" Oct 08 18:18:59 crc kubenswrapper[4859]: I1008 18:18:59.469812 4859 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 08 18:18:59 crc kubenswrapper[4859]: I1008 18:18:59.469896 4859 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 08 18:18:59 crc kubenswrapper[4859]: E1008 18:18:59.469903 4859 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-4b4sz" podUID="2e53cd9b-64f7-4e07-8b96-fb77847c9ac6" Oct 08 18:18:59 crc kubenswrapper[4859]: I1008 18:18:59.469730 4859 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 08 18:18:59 crc kubenswrapper[4859]: E1008 18:18:59.470143 4859 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 08 18:18:59 crc kubenswrapper[4859]: E1008 18:18:59.470191 4859 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 08 18:18:59 crc kubenswrapper[4859]: E1008 18:18:59.470293 4859 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 08 18:18:59 crc kubenswrapper[4859]: I1008 18:18:59.494932 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 08 18:18:59 crc kubenswrapper[4859]: I1008 18:18:59.495186 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 08 18:18:59 crc kubenswrapper[4859]: I1008 18:18:59.495313 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 08 18:18:59 crc kubenswrapper[4859]: I1008 18:18:59.495413 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 08 18:18:59 crc kubenswrapper[4859]: I1008 18:18:59.495501 4859 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-08T18:18:59Z","lastTransitionTime":"2025-10-08T18:18:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 08 18:18:59 crc kubenswrapper[4859]: I1008 18:18:59.571516 4859 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-apiserver/kube-apiserver-crc" podStartSLOduration=90.571452395 podStartE2EDuration="1m30.571452395s" podCreationTimestamp="2025-10-08 18:17:29 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-08 18:18:50.799242582 +0000 UTC m=+101.046081971" watchObservedRunningTime="2025-10-08 18:18:59.571452395 +0000 UTC m=+109.818291814" Oct 08 18:18:59 crc kubenswrapper[4859]: I1008 18:18:59.573378 4859 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-cluster-version/cluster-version-operator-5c965bbfc6-vzwzx"] Oct 08 18:18:59 crc kubenswrapper[4859]: I1008 18:18:59.574213 4859 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-vzwzx" Oct 08 18:18:59 crc kubenswrapper[4859]: I1008 18:18:59.576857 4859 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-version"/"kube-root-ca.crt" Oct 08 18:18:59 crc kubenswrapper[4859]: I1008 18:18:59.578393 4859 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-version"/"cluster-version-operator-serving-cert" Oct 08 18:18:59 crc kubenswrapper[4859]: I1008 18:18:59.578460 4859 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-version"/"openshift-service-ca.crt" Oct 08 18:18:59 crc kubenswrapper[4859]: I1008 18:18:59.578573 4859 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-version"/"default-dockercfg-gxtc4" Oct 08 18:18:59 crc kubenswrapper[4859]: I1008 18:18:59.603952 4859 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-multus/multus-8plkg" podStartSLOduration=89.603919073 podStartE2EDuration="1m29.603919073s" podCreationTimestamp="2025-10-08 18:17:30 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-08 18:18:59.603426719 +0000 UTC m=+109.850266128" watchObservedRunningTime="2025-10-08 18:18:59.603919073 +0000 UTC m=+109.850758482" Oct 08 18:18:59 crc kubenswrapper[4859]: I1008 18:18:59.625997 4859 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-image-registry/node-ca-btjmq" podStartSLOduration=89.625964533 podStartE2EDuration="1m29.625964533s" podCreationTimestamp="2025-10-08 18:17:30 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-08 18:18:59.62341726 +0000 UTC m=+109.870256669" watchObservedRunningTime="2025-10-08 18:18:59.625964533 +0000 UTC m=+109.872803962" Oct 08 18:18:59 crc kubenswrapper[4859]: I1008 18:18:59.663667 4859 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-cvo-updatepayloads\" (UniqueName: \"kubernetes.io/host-path/cb337667-30d7-4bcf-bac7-392c18577227-etc-cvo-updatepayloads\") pod \"cluster-version-operator-5c965bbfc6-vzwzx\" (UID: \"cb337667-30d7-4bcf-bac7-392c18577227\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-vzwzx" Oct 08 18:18:59 crc kubenswrapper[4859]: I1008 18:18:59.663755 4859 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/cb337667-30d7-4bcf-bac7-392c18577227-serving-cert\") pod \"cluster-version-operator-5c965bbfc6-vzwzx\" (UID: \"cb337667-30d7-4bcf-bac7-392c18577227\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-vzwzx" Oct 08 18:18:59 crc kubenswrapper[4859]: I1008 18:18:59.663828 4859 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-ssl-certs\" (UniqueName: \"kubernetes.io/host-path/cb337667-30d7-4bcf-bac7-392c18577227-etc-ssl-certs\") pod \"cluster-version-operator-5c965bbfc6-vzwzx\" (UID: \"cb337667-30d7-4bcf-bac7-392c18577227\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-vzwzx" Oct 08 18:18:59 crc kubenswrapper[4859]: I1008 18:18:59.663880 4859 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/cb337667-30d7-4bcf-bac7-392c18577227-kube-api-access\") pod \"cluster-version-operator-5c965bbfc6-vzwzx\" (UID: \"cb337667-30d7-4bcf-bac7-392c18577227\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-vzwzx" Oct 08 18:18:59 crc kubenswrapper[4859]: I1008 18:18:59.663838 4859 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-controller-manager/kube-controller-manager-crc" podStartSLOduration=85.663805624 podStartE2EDuration="1m25.663805624s" podCreationTimestamp="2025-10-08 18:17:34 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-08 18:18:59.663382852 +0000 UTC m=+109.910222281" watchObservedRunningTime="2025-10-08 18:18:59.663805624 +0000 UTC m=+109.910645043" Oct 08 18:18:59 crc kubenswrapper[4859]: I1008 18:18:59.664030 4859 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/cb337667-30d7-4bcf-bac7-392c18577227-service-ca\") pod \"cluster-version-operator-5c965bbfc6-vzwzx\" (UID: \"cb337667-30d7-4bcf-bac7-392c18577227\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-vzwzx" Oct 08 18:18:59 crc kubenswrapper[4859]: I1008 18:18:59.704432 4859 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-multus/multus-additional-cni-plugins-84k8x" podStartSLOduration=89.704376863 podStartE2EDuration="1m29.704376863s" podCreationTimestamp="2025-10-08 18:17:30 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-08 18:18:59.703451907 +0000 UTC m=+109.950291336" watchObservedRunningTime="2025-10-08 18:18:59.704376863 +0000 UTC m=+109.951216262" Oct 08 18:18:59 crc kubenswrapper[4859]: I1008 18:18:59.764890 4859 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/cb337667-30d7-4bcf-bac7-392c18577227-service-ca\") pod \"cluster-version-operator-5c965bbfc6-vzwzx\" (UID: \"cb337667-30d7-4bcf-bac7-392c18577227\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-vzwzx" Oct 08 18:18:59 crc kubenswrapper[4859]: I1008 18:18:59.765221 4859 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/cb337667-30d7-4bcf-bac7-392c18577227-kube-api-access\") pod \"cluster-version-operator-5c965bbfc6-vzwzx\" (UID: \"cb337667-30d7-4bcf-bac7-392c18577227\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-vzwzx" Oct 08 18:18:59 crc kubenswrapper[4859]: I1008 18:18:59.765343 4859 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-cvo-updatepayloads\" (UniqueName: \"kubernetes.io/host-path/cb337667-30d7-4bcf-bac7-392c18577227-etc-cvo-updatepayloads\") pod \"cluster-version-operator-5c965bbfc6-vzwzx\" (UID: \"cb337667-30d7-4bcf-bac7-392c18577227\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-vzwzx" Oct 08 18:18:59 crc kubenswrapper[4859]: I1008 18:18:59.765428 4859 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/cb337667-30d7-4bcf-bac7-392c18577227-serving-cert\") pod \"cluster-version-operator-5c965bbfc6-vzwzx\" (UID: \"cb337667-30d7-4bcf-bac7-392c18577227\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-vzwzx" Oct 08 18:18:59 crc kubenswrapper[4859]: I1008 18:18:59.765628 4859 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-ssl-certs\" (UniqueName: \"kubernetes.io/host-path/cb337667-30d7-4bcf-bac7-392c18577227-etc-ssl-certs\") pod \"cluster-version-operator-5c965bbfc6-vzwzx\" (UID: \"cb337667-30d7-4bcf-bac7-392c18577227\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-vzwzx" Oct 08 18:18:59 crc kubenswrapper[4859]: I1008 18:18:59.765429 4859 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-cvo-updatepayloads\" (UniqueName: \"kubernetes.io/host-path/cb337667-30d7-4bcf-bac7-392c18577227-etc-cvo-updatepayloads\") pod \"cluster-version-operator-5c965bbfc6-vzwzx\" (UID: \"cb337667-30d7-4bcf-bac7-392c18577227\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-vzwzx" Oct 08 18:18:59 crc kubenswrapper[4859]: I1008 18:18:59.765757 4859 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-ssl-certs\" (UniqueName: \"kubernetes.io/host-path/cb337667-30d7-4bcf-bac7-392c18577227-etc-ssl-certs\") pod \"cluster-version-operator-5c965bbfc6-vzwzx\" (UID: \"cb337667-30d7-4bcf-bac7-392c18577227\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-vzwzx" Oct 08 18:18:59 crc kubenswrapper[4859]: I1008 18:18:59.766075 4859 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/cb337667-30d7-4bcf-bac7-392c18577227-service-ca\") pod \"cluster-version-operator-5c965bbfc6-vzwzx\" (UID: \"cb337667-30d7-4bcf-bac7-392c18577227\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-vzwzx" Oct 08 18:18:59 crc kubenswrapper[4859]: I1008 18:18:59.774125 4859 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/cb337667-30d7-4bcf-bac7-392c18577227-serving-cert\") pod \"cluster-version-operator-5c965bbfc6-vzwzx\" (UID: \"cb337667-30d7-4bcf-bac7-392c18577227\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-vzwzx" Oct 08 18:18:59 crc kubenswrapper[4859]: I1008 18:18:59.789065 4859 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/cb337667-30d7-4bcf-bac7-392c18577227-kube-api-access\") pod \"cluster-version-operator-5c965bbfc6-vzwzx\" (UID: \"cb337667-30d7-4bcf-bac7-392c18577227\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-vzwzx" Oct 08 18:18:59 crc kubenswrapper[4859]: I1008 18:18:59.904490 4859 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-vzwzx" Oct 08 18:19:00 crc kubenswrapper[4859]: I1008 18:19:00.141504 4859 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-vzwzx" event={"ID":"cb337667-30d7-4bcf-bac7-392c18577227","Type":"ContainerStarted","Data":"79f113854c66dcb10072d8763a8a41b51709947eaec157441b9b648ed000f07a"} Oct 08 18:19:01 crc kubenswrapper[4859]: I1008 18:19:01.147301 4859 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-vzwzx" event={"ID":"cb337667-30d7-4bcf-bac7-392c18577227","Type":"ContainerStarted","Data":"5fdf590fbab0b8a2b44e65733797ca6b82dfa3a69931afdde77571d8d7473392"} Oct 08 18:19:01 crc kubenswrapper[4859]: I1008 18:19:01.167918 4859 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-vzwzx" podStartSLOduration=91.167895419 podStartE2EDuration="1m31.167895419s" podCreationTimestamp="2025-10-08 18:17:30 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-08 18:19:01.167146898 +0000 UTC m=+111.413986297" watchObservedRunningTime="2025-10-08 18:19:01.167895419 +0000 UTC m=+111.414734808" Oct 08 18:19:01 crc kubenswrapper[4859]: I1008 18:19:01.468949 4859 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 08 18:19:01 crc kubenswrapper[4859]: I1008 18:19:01.469006 4859 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-4b4sz" Oct 08 18:19:01 crc kubenswrapper[4859]: E1008 18:19:01.469112 4859 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 08 18:19:01 crc kubenswrapper[4859]: I1008 18:19:01.469116 4859 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 08 18:19:01 crc kubenswrapper[4859]: I1008 18:19:01.469183 4859 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 08 18:19:01 crc kubenswrapper[4859]: E1008 18:19:01.469296 4859 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-4b4sz" podUID="2e53cd9b-64f7-4e07-8b96-fb77847c9ac6" Oct 08 18:19:01 crc kubenswrapper[4859]: E1008 18:19:01.469538 4859 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 08 18:19:01 crc kubenswrapper[4859]: E1008 18:19:01.469732 4859 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 08 18:19:03 crc kubenswrapper[4859]: I1008 18:19:03.470011 4859 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 08 18:19:03 crc kubenswrapper[4859]: I1008 18:19:03.470108 4859 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-4b4sz" Oct 08 18:19:03 crc kubenswrapper[4859]: I1008 18:19:03.470149 4859 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 08 18:19:03 crc kubenswrapper[4859]: I1008 18:19:03.470208 4859 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 08 18:19:03 crc kubenswrapper[4859]: E1008 18:19:03.470463 4859 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 08 18:19:03 crc kubenswrapper[4859]: E1008 18:19:03.470661 4859 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-4b4sz" podUID="2e53cd9b-64f7-4e07-8b96-fb77847c9ac6" Oct 08 18:19:03 crc kubenswrapper[4859]: E1008 18:19:03.470889 4859 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 08 18:19:03 crc kubenswrapper[4859]: E1008 18:19:03.471087 4859 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 08 18:19:05 crc kubenswrapper[4859]: I1008 18:19:05.469498 4859 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 08 18:19:05 crc kubenswrapper[4859]: I1008 18:19:05.469610 4859 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-4b4sz" Oct 08 18:19:05 crc kubenswrapper[4859]: E1008 18:19:05.469750 4859 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 08 18:19:05 crc kubenswrapper[4859]: I1008 18:19:05.469639 4859 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 08 18:19:05 crc kubenswrapper[4859]: E1008 18:19:05.469839 4859 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-4b4sz" podUID="2e53cd9b-64f7-4e07-8b96-fb77847c9ac6" Oct 08 18:19:05 crc kubenswrapper[4859]: E1008 18:19:05.469860 4859 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 08 18:19:05 crc kubenswrapper[4859]: I1008 18:19:05.469610 4859 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 08 18:19:05 crc kubenswrapper[4859]: E1008 18:19:05.469984 4859 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 08 18:19:06 crc kubenswrapper[4859]: I1008 18:19:06.176095 4859 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-8plkg_92b2d7de-31cc-4a91-95a5-ed5ea964b028/kube-multus/1.log" Oct 08 18:19:06 crc kubenswrapper[4859]: I1008 18:19:06.176782 4859 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-8plkg_92b2d7de-31cc-4a91-95a5-ed5ea964b028/kube-multus/0.log" Oct 08 18:19:06 crc kubenswrapper[4859]: I1008 18:19:06.176867 4859 generic.go:334] "Generic (PLEG): container finished" podID="92b2d7de-31cc-4a91-95a5-ed5ea964b028" containerID="729bd32fd01dcab495bffa64f2f5e44990fc1d78b05e7f131868009f92305831" exitCode=1 Oct 08 18:19:06 crc kubenswrapper[4859]: I1008 18:19:06.176922 4859 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-8plkg" event={"ID":"92b2d7de-31cc-4a91-95a5-ed5ea964b028","Type":"ContainerDied","Data":"729bd32fd01dcab495bffa64f2f5e44990fc1d78b05e7f131868009f92305831"} Oct 08 18:19:06 crc kubenswrapper[4859]: I1008 18:19:06.176988 4859 scope.go:117] "RemoveContainer" containerID="fa00b8345d40b74a917ddc23cc1b2c5c684a9ca61b6119d090c92cc1439b71c4" Oct 08 18:19:06 crc kubenswrapper[4859]: I1008 18:19:06.177743 4859 scope.go:117] "RemoveContainer" containerID="729bd32fd01dcab495bffa64f2f5e44990fc1d78b05e7f131868009f92305831" Oct 08 18:19:06 crc kubenswrapper[4859]: E1008 18:19:06.178122 4859 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"kube-multus\" with CrashLoopBackOff: \"back-off 10s restarting failed container=kube-multus pod=multus-8plkg_openshift-multus(92b2d7de-31cc-4a91-95a5-ed5ea964b028)\"" pod="openshift-multus/multus-8plkg" podUID="92b2d7de-31cc-4a91-95a5-ed5ea964b028" Oct 08 18:19:07 crc kubenswrapper[4859]: I1008 18:19:07.184144 4859 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-8plkg_92b2d7de-31cc-4a91-95a5-ed5ea964b028/kube-multus/1.log" Oct 08 18:19:07 crc kubenswrapper[4859]: I1008 18:19:07.469224 4859 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 08 18:19:07 crc kubenswrapper[4859]: I1008 18:19:07.469256 4859 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-4b4sz" Oct 08 18:19:07 crc kubenswrapper[4859]: I1008 18:19:07.469306 4859 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 08 18:19:07 crc kubenswrapper[4859]: E1008 18:19:07.469400 4859 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 08 18:19:07 crc kubenswrapper[4859]: I1008 18:19:07.469413 4859 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 08 18:19:07 crc kubenswrapper[4859]: E1008 18:19:07.469506 4859 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 08 18:19:07 crc kubenswrapper[4859]: E1008 18:19:07.469585 4859 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 08 18:19:07 crc kubenswrapper[4859]: E1008 18:19:07.469807 4859 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-4b4sz" podUID="2e53cd9b-64f7-4e07-8b96-fb77847c9ac6" Oct 08 18:19:09 crc kubenswrapper[4859]: I1008 18:19:09.469766 4859 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 08 18:19:09 crc kubenswrapper[4859]: I1008 18:19:09.469785 4859 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 08 18:19:09 crc kubenswrapper[4859]: I1008 18:19:09.469797 4859 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 08 18:19:09 crc kubenswrapper[4859]: I1008 18:19:09.469766 4859 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-4b4sz" Oct 08 18:19:09 crc kubenswrapper[4859]: E1008 18:19:09.469993 4859 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 08 18:19:09 crc kubenswrapper[4859]: E1008 18:19:09.470599 4859 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 08 18:19:09 crc kubenswrapper[4859]: E1008 18:19:09.470763 4859 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 08 18:19:09 crc kubenswrapper[4859]: E1008 18:19:09.470859 4859 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-4b4sz" podUID="2e53cd9b-64f7-4e07-8b96-fb77847c9ac6" Oct 08 18:19:09 crc kubenswrapper[4859]: I1008 18:19:09.471108 4859 scope.go:117] "RemoveContainer" containerID="f3a760c557ef21fbc6b54e78c3f6c5b325f296f93a022425cca7b1399eed8661" Oct 08 18:19:09 crc kubenswrapper[4859]: E1008 18:19:09.471461 4859 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ovnkube-controller\" with CrashLoopBackOff: \"back-off 40s restarting failed container=ovnkube-controller pod=ovnkube-node-4g8gf_openshift-ovn-kubernetes(1dff864e-b75d-4e0b-b182-75f710eac8df)\"" pod="openshift-ovn-kubernetes/ovnkube-node-4g8gf" podUID="1dff864e-b75d-4e0b-b182-75f710eac8df" Oct 08 18:19:10 crc kubenswrapper[4859]: E1008 18:19:10.401105 4859 kubelet_node_status.go:497] "Node not becoming ready in time after startup" Oct 08 18:19:10 crc kubenswrapper[4859]: E1008 18:19:10.583588 4859 kubelet.go:2916] "Container runtime network not ready" networkReady="NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" Oct 08 18:19:11 crc kubenswrapper[4859]: I1008 18:19:11.469819 4859 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 08 18:19:11 crc kubenswrapper[4859]: I1008 18:19:11.469866 4859 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 08 18:19:11 crc kubenswrapper[4859]: I1008 18:19:11.469899 4859 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-4b4sz" Oct 08 18:19:11 crc kubenswrapper[4859]: I1008 18:19:11.469817 4859 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 08 18:19:11 crc kubenswrapper[4859]: E1008 18:19:11.470021 4859 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 08 18:19:11 crc kubenswrapper[4859]: E1008 18:19:11.470187 4859 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 08 18:19:11 crc kubenswrapper[4859]: E1008 18:19:11.470274 4859 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 08 18:19:11 crc kubenswrapper[4859]: E1008 18:19:11.470351 4859 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-4b4sz" podUID="2e53cd9b-64f7-4e07-8b96-fb77847c9ac6" Oct 08 18:19:13 crc kubenswrapper[4859]: I1008 18:19:13.469230 4859 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 08 18:19:13 crc kubenswrapper[4859]: I1008 18:19:13.469252 4859 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 08 18:19:13 crc kubenswrapper[4859]: I1008 18:19:13.469252 4859 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 08 18:19:13 crc kubenswrapper[4859]: E1008 18:19:13.469466 4859 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 08 18:19:13 crc kubenswrapper[4859]: E1008 18:19:13.469609 4859 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 08 18:19:13 crc kubenswrapper[4859]: I1008 18:19:13.469720 4859 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-4b4sz" Oct 08 18:19:13 crc kubenswrapper[4859]: E1008 18:19:13.469809 4859 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 08 18:19:13 crc kubenswrapper[4859]: E1008 18:19:13.469897 4859 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-4b4sz" podUID="2e53cd9b-64f7-4e07-8b96-fb77847c9ac6" Oct 08 18:19:15 crc kubenswrapper[4859]: I1008 18:19:15.469953 4859 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 08 18:19:15 crc kubenswrapper[4859]: I1008 18:19:15.470041 4859 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-4b4sz" Oct 08 18:19:15 crc kubenswrapper[4859]: I1008 18:19:15.469977 4859 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 08 18:19:15 crc kubenswrapper[4859]: I1008 18:19:15.470069 4859 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 08 18:19:15 crc kubenswrapper[4859]: E1008 18:19:15.470353 4859 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 08 18:19:15 crc kubenswrapper[4859]: E1008 18:19:15.470494 4859 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-4b4sz" podUID="2e53cd9b-64f7-4e07-8b96-fb77847c9ac6" Oct 08 18:19:15 crc kubenswrapper[4859]: E1008 18:19:15.470621 4859 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 08 18:19:15 crc kubenswrapper[4859]: E1008 18:19:15.470770 4859 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 08 18:19:15 crc kubenswrapper[4859]: E1008 18:19:15.585493 4859 kubelet.go:2916] "Container runtime network not ready" networkReady="NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" Oct 08 18:19:17 crc kubenswrapper[4859]: I1008 18:19:17.469175 4859 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 08 18:19:17 crc kubenswrapper[4859]: I1008 18:19:17.469171 4859 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 08 18:19:17 crc kubenswrapper[4859]: I1008 18:19:17.469193 4859 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 08 18:19:17 crc kubenswrapper[4859]: I1008 18:19:17.469475 4859 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-4b4sz" Oct 08 18:19:17 crc kubenswrapper[4859]: E1008 18:19:17.469821 4859 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 08 18:19:17 crc kubenswrapper[4859]: E1008 18:19:17.469958 4859 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 08 18:19:17 crc kubenswrapper[4859]: E1008 18:19:17.470247 4859 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 08 18:19:17 crc kubenswrapper[4859]: E1008 18:19:17.470562 4859 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-4b4sz" podUID="2e53cd9b-64f7-4e07-8b96-fb77847c9ac6" Oct 08 18:19:18 crc kubenswrapper[4859]: I1008 18:19:18.470546 4859 scope.go:117] "RemoveContainer" containerID="729bd32fd01dcab495bffa64f2f5e44990fc1d78b05e7f131868009f92305831" Oct 08 18:19:19 crc kubenswrapper[4859]: I1008 18:19:19.234256 4859 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-8plkg_92b2d7de-31cc-4a91-95a5-ed5ea964b028/kube-multus/1.log" Oct 08 18:19:19 crc kubenswrapper[4859]: I1008 18:19:19.234335 4859 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-8plkg" event={"ID":"92b2d7de-31cc-4a91-95a5-ed5ea964b028","Type":"ContainerStarted","Data":"d18a8ef0ab4fc3ed687ccf7e94753da8a4a29dae40a39aae2b64a854761f90d0"} Oct 08 18:19:19 crc kubenswrapper[4859]: I1008 18:19:19.469386 4859 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 08 18:19:19 crc kubenswrapper[4859]: E1008 18:19:19.469625 4859 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 08 18:19:19 crc kubenswrapper[4859]: I1008 18:19:19.469909 4859 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-4b4sz" Oct 08 18:19:19 crc kubenswrapper[4859]: I1008 18:19:19.469871 4859 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 08 18:19:19 crc kubenswrapper[4859]: E1008 18:19:19.470053 4859 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-4b4sz" podUID="2e53cd9b-64f7-4e07-8b96-fb77847c9ac6" Oct 08 18:19:19 crc kubenswrapper[4859]: I1008 18:19:19.469968 4859 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 08 18:19:19 crc kubenswrapper[4859]: E1008 18:19:19.470236 4859 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 08 18:19:19 crc kubenswrapper[4859]: E1008 18:19:19.470343 4859 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 08 18:19:20 crc kubenswrapper[4859]: I1008 18:19:20.476547 4859 scope.go:117] "RemoveContainer" containerID="f3a760c557ef21fbc6b54e78c3f6c5b325f296f93a022425cca7b1399eed8661" Oct 08 18:19:20 crc kubenswrapper[4859]: E1008 18:19:20.586967 4859 kubelet.go:2916] "Container runtime network not ready" networkReady="NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" Oct 08 18:19:21 crc kubenswrapper[4859]: I1008 18:19:21.245023 4859 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-4g8gf_1dff864e-b75d-4e0b-b182-75f710eac8df/ovnkube-controller/3.log" Oct 08 18:19:21 crc kubenswrapper[4859]: I1008 18:19:21.247089 4859 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-4g8gf" event={"ID":"1dff864e-b75d-4e0b-b182-75f710eac8df","Type":"ContainerStarted","Data":"b737cb013e2877d970d7e301048f95e0ba0ade0b8db42b2607a78fdf2580659f"} Oct 08 18:19:21 crc kubenswrapper[4859]: I1008 18:19:21.247540 4859 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ovn-kubernetes/ovnkube-node-4g8gf" Oct 08 18:19:21 crc kubenswrapper[4859]: I1008 18:19:21.274907 4859 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-ovn-kubernetes/ovnkube-node-4g8gf" podStartSLOduration=110.274876482 podStartE2EDuration="1m50.274876482s" podCreationTimestamp="2025-10-08 18:17:31 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-08 18:19:21.274841921 +0000 UTC m=+131.521681320" watchObservedRunningTime="2025-10-08 18:19:21.274876482 +0000 UTC m=+131.521715861" Oct 08 18:19:21 crc kubenswrapper[4859]: I1008 18:19:21.382138 4859 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-multus/network-metrics-daemon-4b4sz"] Oct 08 18:19:21 crc kubenswrapper[4859]: I1008 18:19:21.382270 4859 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-4b4sz" Oct 08 18:19:21 crc kubenswrapper[4859]: E1008 18:19:21.382351 4859 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-4b4sz" podUID="2e53cd9b-64f7-4e07-8b96-fb77847c9ac6" Oct 08 18:19:21 crc kubenswrapper[4859]: I1008 18:19:21.469153 4859 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 08 18:19:21 crc kubenswrapper[4859]: E1008 18:19:21.469347 4859 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 08 18:19:21 crc kubenswrapper[4859]: I1008 18:19:21.469442 4859 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 08 18:19:21 crc kubenswrapper[4859]: E1008 18:19:21.469551 4859 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 08 18:19:21 crc kubenswrapper[4859]: I1008 18:19:21.469637 4859 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 08 18:19:21 crc kubenswrapper[4859]: E1008 18:19:21.469792 4859 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 08 18:19:22 crc kubenswrapper[4859]: I1008 18:19:22.469061 4859 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-4b4sz" Oct 08 18:19:22 crc kubenswrapper[4859]: E1008 18:19:22.469349 4859 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-4b4sz" podUID="2e53cd9b-64f7-4e07-8b96-fb77847c9ac6" Oct 08 18:19:23 crc kubenswrapper[4859]: I1008 18:19:23.468986 4859 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 08 18:19:23 crc kubenswrapper[4859]: I1008 18:19:23.468986 4859 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 08 18:19:23 crc kubenswrapper[4859]: E1008 18:19:23.469238 4859 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 08 18:19:23 crc kubenswrapper[4859]: I1008 18:19:23.469038 4859 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 08 18:19:23 crc kubenswrapper[4859]: E1008 18:19:23.469492 4859 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 08 18:19:23 crc kubenswrapper[4859]: E1008 18:19:23.469611 4859 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 08 18:19:24 crc kubenswrapper[4859]: I1008 18:19:24.469212 4859 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-4b4sz" Oct 08 18:19:24 crc kubenswrapper[4859]: E1008 18:19:24.469445 4859 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-4b4sz" podUID="2e53cd9b-64f7-4e07-8b96-fb77847c9ac6" Oct 08 18:19:25 crc kubenswrapper[4859]: I1008 18:19:25.469490 4859 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 08 18:19:25 crc kubenswrapper[4859]: I1008 18:19:25.469490 4859 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 08 18:19:25 crc kubenswrapper[4859]: I1008 18:19:25.469612 4859 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 08 18:19:25 crc kubenswrapper[4859]: E1008 18:19:25.469647 4859 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 08 18:19:25 crc kubenswrapper[4859]: E1008 18:19:25.469751 4859 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 08 18:19:25 crc kubenswrapper[4859]: E1008 18:19:25.469845 4859 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 08 18:19:26 crc kubenswrapper[4859]: I1008 18:19:26.469474 4859 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-4b4sz" Oct 08 18:19:26 crc kubenswrapper[4859]: I1008 18:19:26.473337 4859 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"metrics-daemon-sa-dockercfg-d427c" Oct 08 18:19:26 crc kubenswrapper[4859]: I1008 18:19:26.473508 4859 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"metrics-daemon-secret" Oct 08 18:19:27 crc kubenswrapper[4859]: I1008 18:19:27.468928 4859 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 08 18:19:27 crc kubenswrapper[4859]: I1008 18:19:27.468985 4859 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 08 18:19:27 crc kubenswrapper[4859]: I1008 18:19:27.469007 4859 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 08 18:19:27 crc kubenswrapper[4859]: I1008 18:19:27.472165 4859 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-console"/"networking-console-plugin" Oct 08 18:19:27 crc kubenswrapper[4859]: I1008 18:19:27.472449 4859 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-diagnostics"/"kube-root-ca.crt" Oct 08 18:19:27 crc kubenswrapper[4859]: I1008 18:19:27.474430 4859 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-diagnostics"/"openshift-service-ca.crt" Oct 08 18:19:27 crc kubenswrapper[4859]: I1008 18:19:27.475022 4859 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-network-console"/"networking-console-plugin-cert" Oct 08 18:19:30 crc kubenswrapper[4859]: I1008 18:19:30.093315 4859 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeReady" Oct 08 18:19:30 crc kubenswrapper[4859]: I1008 18:19:30.127869 4859 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-oauth-apiserver/apiserver-7bbb656c7d-6j5rk"] Oct 08 18:19:30 crc kubenswrapper[4859]: I1008 18:19:30.128344 4859 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-6j5rk" Oct 08 18:19:30 crc kubenswrapper[4859]: I1008 18:19:30.129918 4859 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-authentication/oauth-openshift-558db77b4-j8gh2"] Oct 08 18:19:30 crc kubenswrapper[4859]: I1008 18:19:30.130132 4859 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-authentication/oauth-openshift-558db77b4-j8gh2" Oct 08 18:19:30 crc kubenswrapper[4859]: I1008 18:19:30.131838 4859 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-route-controller-manager/route-controller-manager-6576b87f9c-xlfdd"] Oct 08 18:19:30 crc kubenswrapper[4859]: I1008 18:19:30.132312 4859 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-xlfdd" Oct 08 18:19:30 crc kubenswrapper[4859]: I1008 18:19:30.132519 4859 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-config-operator/openshift-config-operator-7777fb866f-ncj6n"] Oct 08 18:19:30 crc kubenswrapper[4859]: I1008 18:19:30.132876 4859 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-config-operator/openshift-config-operator-7777fb866f-ncj6n" Oct 08 18:19:30 crc kubenswrapper[4859]: I1008 18:19:30.134579 4859 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-controller-manager/controller-manager-879f6c89f-9gpzj"] Oct 08 18:19:30 crc kubenswrapper[4859]: I1008 18:19:30.135273 4859 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-authentication-operator/authentication-operator-69f744f599-jd5kh"] Oct 08 18:19:30 crc kubenswrapper[4859]: I1008 18:19:30.135362 4859 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-879f6c89f-9gpzj" Oct 08 18:19:30 crc kubenswrapper[4859]: I1008 18:19:30.135622 4859 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-authentication-operator/authentication-operator-69f744f599-jd5kh" Oct 08 18:19:30 crc kubenswrapper[4859]: I1008 18:19:30.136824 4859 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-cluster-machine-approver/machine-approver-56656f9798-95s6m"] Oct 08 18:19:30 crc kubenswrapper[4859]: I1008 18:19:30.137242 4859 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-cluster-machine-approver/machine-approver-56656f9798-95s6m" Oct 08 18:19:30 crc kubenswrapper[4859]: W1008 18:19:30.137863 4859 reflector.go:561] object-"openshift-oauth-apiserver"/"oauth-apiserver-sa-dockercfg-6r2bq": failed to list *v1.Secret: secrets "oauth-apiserver-sa-dockercfg-6r2bq" is forbidden: User "system:node:crc" cannot list resource "secrets" in API group "" in the namespace "openshift-oauth-apiserver": no relationship found between node 'crc' and this object Oct 08 18:19:30 crc kubenswrapper[4859]: E1008 18:19:30.137920 4859 reflector.go:158] "Unhandled Error" err="object-\"openshift-oauth-apiserver\"/\"oauth-apiserver-sa-dockercfg-6r2bq\": Failed to watch *v1.Secret: failed to list *v1.Secret: secrets \"oauth-apiserver-sa-dockercfg-6r2bq\" is forbidden: User \"system:node:crc\" cannot list resource \"secrets\" in API group \"\" in the namespace \"openshift-oauth-apiserver\": no relationship found between node 'crc' and this object" logger="UnhandledError" Oct 08 18:19:30 crc kubenswrapper[4859]: W1008 18:19:30.138073 4859 reflector.go:561] object-"openshift-oauth-apiserver"/"etcd-client": failed to list *v1.Secret: secrets "etcd-client" is forbidden: User "system:node:crc" cannot list resource "secrets" in API group "" in the namespace "openshift-oauth-apiserver": no relationship found between node 'crc' and this object Oct 08 18:19:30 crc kubenswrapper[4859]: W1008 18:19:30.138120 4859 reflector.go:561] object-"openshift-oauth-apiserver"/"audit-1": failed to list *v1.ConfigMap: configmaps "audit-1" is forbidden: User "system:node:crc" cannot list resource "configmaps" in API group "" in the namespace "openshift-oauth-apiserver": no relationship found between node 'crc' and this object Oct 08 18:19:30 crc kubenswrapper[4859]: E1008 18:19:30.138149 4859 reflector.go:158] "Unhandled Error" err="object-\"openshift-oauth-apiserver\"/\"audit-1\": Failed to watch *v1.ConfigMap: failed to list *v1.ConfigMap: configmaps \"audit-1\" is forbidden: User \"system:node:crc\" cannot list resource \"configmaps\" in API group \"\" in the namespace \"openshift-oauth-apiserver\": no relationship found between node 'crc' and this object" logger="UnhandledError" Oct 08 18:19:30 crc kubenswrapper[4859]: E1008 18:19:30.138115 4859 reflector.go:158] "Unhandled Error" err="object-\"openshift-oauth-apiserver\"/\"etcd-client\": Failed to watch *v1.Secret: failed to list *v1.Secret: secrets \"etcd-client\" is forbidden: User \"system:node:crc\" cannot list resource \"secrets\" in API group \"\" in the namespace \"openshift-oauth-apiserver\": no relationship found between node 'crc' and this object" logger="UnhandledError" Oct 08 18:19:30 crc kubenswrapper[4859]: W1008 18:19:30.138239 4859 reflector.go:561] object-"openshift-authentication"/"v4-0-config-system-session": failed to list *v1.Secret: secrets "v4-0-config-system-session" is forbidden: User "system:node:crc" cannot list resource "secrets" in API group "" in the namespace "openshift-authentication": no relationship found between node 'crc' and this object Oct 08 18:19:30 crc kubenswrapper[4859]: E1008 18:19:30.138255 4859 reflector.go:158] "Unhandled Error" err="object-\"openshift-authentication\"/\"v4-0-config-system-session\": Failed to watch *v1.Secret: failed to list *v1.Secret: secrets \"v4-0-config-system-session\" is forbidden: User \"system:node:crc\" cannot list resource \"secrets\" in API group \"\" in the namespace \"openshift-authentication\": no relationship found between node 'crc' and this object" logger="UnhandledError" Oct 08 18:19:30 crc kubenswrapper[4859]: I1008 18:19:30.139472 4859 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-console/console-f9d7485db-m2xmp"] Oct 08 18:19:30 crc kubenswrapper[4859]: W1008 18:19:30.139546 4859 reflector.go:561] object-"openshift-oauth-apiserver"/"kube-root-ca.crt": failed to list *v1.ConfigMap: configmaps "kube-root-ca.crt" is forbidden: User "system:node:crc" cannot list resource "configmaps" in API group "" in the namespace "openshift-oauth-apiserver": no relationship found between node 'crc' and this object Oct 08 18:19:30 crc kubenswrapper[4859]: E1008 18:19:30.139607 4859 reflector.go:158] "Unhandled Error" err="object-\"openshift-oauth-apiserver\"/\"kube-root-ca.crt\": Failed to watch *v1.ConfigMap: failed to list *v1.ConfigMap: configmaps \"kube-root-ca.crt\" is forbidden: User \"system:node:crc\" cannot list resource \"configmaps\" in API group \"\" in the namespace \"openshift-oauth-apiserver\": no relationship found between node 'crc' and this object" logger="UnhandledError" Oct 08 18:19:30 crc kubenswrapper[4859]: I1008 18:19:30.139947 4859 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-console/console-f9d7485db-m2xmp" Oct 08 18:19:30 crc kubenswrapper[4859]: I1008 18:19:30.142144 4859 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-user-template-error" Oct 08 18:19:30 crc kubenswrapper[4859]: I1008 18:19:30.142155 4859 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-c5p45"] Oct 08 18:19:30 crc kubenswrapper[4859]: I1008 18:19:30.143939 4859 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"v4-0-config-system-service-ca" Oct 08 18:19:30 crc kubenswrapper[4859]: W1008 18:19:30.144078 4859 reflector.go:561] object-"openshift-oauth-apiserver"/"etcd-serving-ca": failed to list *v1.ConfigMap: configmaps "etcd-serving-ca" is forbidden: User "system:node:crc" cannot list resource "configmaps" in API group "" in the namespace "openshift-oauth-apiserver": no relationship found between node 'crc' and this object Oct 08 18:19:30 crc kubenswrapper[4859]: E1008 18:19:30.144106 4859 reflector.go:158] "Unhandled Error" err="object-\"openshift-oauth-apiserver\"/\"etcd-serving-ca\": Failed to watch *v1.ConfigMap: failed to list *v1.ConfigMap: configmaps \"etcd-serving-ca\" is forbidden: User \"system:node:crc\" cannot list resource \"configmaps\" in API group \"\" in the namespace \"openshift-oauth-apiserver\": no relationship found between node 'crc' and this object" logger="UnhandledError" Oct 08 18:19:30 crc kubenswrapper[4859]: W1008 18:19:30.144151 4859 reflector.go:561] object-"openshift-oauth-apiserver"/"trusted-ca-bundle": failed to list *v1.ConfigMap: configmaps "trusted-ca-bundle" is forbidden: User "system:node:crc" cannot list resource "configmaps" in API group "" in the namespace "openshift-oauth-apiserver": no relationship found between node 'crc' and this object Oct 08 18:19:30 crc kubenswrapper[4859]: E1008 18:19:30.144162 4859 reflector.go:158] "Unhandled Error" err="object-\"openshift-oauth-apiserver\"/\"trusted-ca-bundle\": Failed to watch *v1.ConfigMap: failed to list *v1.ConfigMap: configmaps \"trusted-ca-bundle\" is forbidden: User \"system:node:crc\" cannot list resource \"configmaps\" in API group \"\" in the namespace \"openshift-oauth-apiserver\": no relationship found between node 'crc' and this object" logger="UnhandledError" Oct 08 18:19:30 crc kubenswrapper[4859]: W1008 18:19:30.144201 4859 reflector.go:561] object-"openshift-authentication"/"v4-0-config-user-template-provider-selection": failed to list *v1.Secret: secrets "v4-0-config-user-template-provider-selection" is forbidden: User "system:node:crc" cannot list resource "secrets" in API group "" in the namespace "openshift-authentication": no relationship found between node 'crc' and this object Oct 08 18:19:30 crc kubenswrapper[4859]: E1008 18:19:30.144213 4859 reflector.go:158] "Unhandled Error" err="object-\"openshift-authentication\"/\"v4-0-config-user-template-provider-selection\": Failed to watch *v1.Secret: failed to list *v1.Secret: secrets \"v4-0-config-user-template-provider-selection\" is forbidden: User \"system:node:crc\" cannot list resource \"secrets\" in API group \"\" in the namespace \"openshift-authentication\": no relationship found between node 'crc' and this object" logger="UnhandledError" Oct 08 18:19:30 crc kubenswrapper[4859]: W1008 18:19:30.144250 4859 reflector.go:561] object-"openshift-oauth-apiserver"/"openshift-service-ca.crt": failed to list *v1.ConfigMap: configmaps "openshift-service-ca.crt" is forbidden: User "system:node:crc" cannot list resource "configmaps" in API group "" in the namespace "openshift-oauth-apiserver": no relationship found between node 'crc' and this object Oct 08 18:19:30 crc kubenswrapper[4859]: E1008 18:19:30.144260 4859 reflector.go:158] "Unhandled Error" err="object-\"openshift-oauth-apiserver\"/\"openshift-service-ca.crt\": Failed to watch *v1.ConfigMap: failed to list *v1.ConfigMap: configmaps \"openshift-service-ca.crt\" is forbidden: User \"system:node:crc\" cannot list resource \"configmaps\" in API group \"\" in the namespace \"openshift-oauth-apiserver\": no relationship found between node 'crc' and this object" logger="UnhandledError" Oct 08 18:19:30 crc kubenswrapper[4859]: I1008 18:19:30.145762 4859 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"openshift-service-ca.crt" Oct 08 18:19:30 crc kubenswrapper[4859]: W1008 18:19:30.146101 4859 reflector.go:561] object-"openshift-oauth-apiserver"/"serving-cert": failed to list *v1.Secret: secrets "serving-cert" is forbidden: User "system:node:crc" cannot list resource "secrets" in API group "" in the namespace "openshift-oauth-apiserver": no relationship found between node 'crc' and this object Oct 08 18:19:30 crc kubenswrapper[4859]: E1008 18:19:30.146123 4859 reflector.go:158] "Unhandled Error" err="object-\"openshift-oauth-apiserver\"/\"serving-cert\": Failed to watch *v1.Secret: failed to list *v1.Secret: secrets \"serving-cert\" is forbidden: User \"system:node:crc\" cannot list resource \"secrets\" in API group \"\" in the namespace \"openshift-oauth-apiserver\": no relationship found between node 'crc' and this object" logger="UnhandledError" Oct 08 18:19:30 crc kubenswrapper[4859]: W1008 18:19:30.146162 4859 reflector.go:561] object-"openshift-oauth-apiserver"/"encryption-config-1": failed to list *v1.Secret: secrets "encryption-config-1" is forbidden: User "system:node:crc" cannot list resource "secrets" in API group "" in the namespace "openshift-oauth-apiserver": no relationship found between node 'crc' and this object Oct 08 18:19:30 crc kubenswrapper[4859]: E1008 18:19:30.146172 4859 reflector.go:158] "Unhandled Error" err="object-\"openshift-oauth-apiserver\"/\"encryption-config-1\": Failed to watch *v1.Secret: failed to list *v1.Secret: secrets \"encryption-config-1\" is forbidden: User \"system:node:crc\" cannot list resource \"secrets\" in API group \"\" in the namespace \"openshift-oauth-apiserver\": no relationship found between node 'crc' and this object" logger="UnhandledError" Oct 08 18:19:30 crc kubenswrapper[4859]: W1008 18:19:30.146208 4859 reflector.go:561] object-"openshift-authentication"/"audit": failed to list *v1.ConfigMap: configmaps "audit" is forbidden: User "system:node:crc" cannot list resource "configmaps" in API group "" in the namespace "openshift-authentication": no relationship found between node 'crc' and this object Oct 08 18:19:30 crc kubenswrapper[4859]: E1008 18:19:30.146217 4859 reflector.go:158] "Unhandled Error" err="object-\"openshift-authentication\"/\"audit\": Failed to watch *v1.ConfigMap: failed to list *v1.ConfigMap: configmaps \"audit\" is forbidden: User \"system:node:crc\" cannot list resource \"configmaps\" in API group \"\" in the namespace \"openshift-authentication\": no relationship found between node 'crc' and this object" logger="UnhandledError" Oct 08 18:19:30 crc kubenswrapper[4859]: I1008 18:19:30.146678 4859 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-machine-api/machine-api-operator-5694c8668f-gmbk9"] Oct 08 18:19:30 crc kubenswrapper[4859]: I1008 18:19:30.147126 4859 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-c5p45" Oct 08 18:19:30 crc kubenswrapper[4859]: I1008 18:19:30.147488 4859 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-console-operator/console-operator-58897d9998-8sflk"] Oct 08 18:19:30 crc kubenswrapper[4859]: I1008 18:19:30.148214 4859 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-api/machine-api-operator-5694c8668f-gmbk9" Oct 08 18:19:30 crc kubenswrapper[4859]: I1008 18:19:30.150391 4859 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-apiserver/apiserver-76f77b778f-fv9bp"] Oct 08 18:19:30 crc kubenswrapper[4859]: I1008 18:19:30.150427 4859 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-console-operator/console-operator-58897d9998-8sflk" Oct 08 18:19:30 crc kubenswrapper[4859]: I1008 18:19:30.151036 4859 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-apiserver/apiserver-76f77b778f-fv9bp" Oct 08 18:19:30 crc kubenswrapper[4859]: I1008 18:19:30.156791 4859 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-system-serving-cert" Oct 08 18:19:30 crc kubenswrapper[4859]: I1008 18:19:30.157242 4859 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"kube-root-ca.crt" Oct 08 18:19:30 crc kubenswrapper[4859]: I1008 18:19:30.157942 4859 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"config" Oct 08 18:19:30 crc kubenswrapper[4859]: I1008 18:19:30.158027 4859 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-config-operator"/"openshift-service-ca.crt" Oct 08 18:19:30 crc kubenswrapper[4859]: I1008 18:19:30.158219 4859 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"openshift-service-ca.crt" Oct 08 18:19:30 crc kubenswrapper[4859]: I1008 18:19:30.158399 4859 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-route-controller-manager"/"serving-cert" Oct 08 18:19:30 crc kubenswrapper[4859]: I1008 18:19:30.158471 4859 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"openshift-service-ca.crt" Oct 08 18:19:30 crc kubenswrapper[4859]: I1008 18:19:30.158572 4859 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"config" Oct 08 18:19:30 crc kubenswrapper[4859]: I1008 18:19:30.166135 4859 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-machine-approver"/"machine-approver-config" Oct 08 18:19:30 crc kubenswrapper[4859]: I1008 18:19:30.167509 4859 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-machine-approver"/"kube-rbac-proxy" Oct 08 18:19:30 crc kubenswrapper[4859]: I1008 18:19:30.168897 4859 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager"/"openshift-controller-manager-sa-dockercfg-msq4c" Oct 08 18:19:30 crc kubenswrapper[4859]: I1008 18:19:30.169238 4859 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager"/"serving-cert" Oct 08 18:19:30 crc kubenswrapper[4859]: I1008 18:19:30.169414 4859 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"kube-root-ca.crt" Oct 08 18:19:30 crc kubenswrapper[4859]: I1008 18:19:30.169667 4859 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"kube-root-ca.crt" Oct 08 18:19:30 crc kubenswrapper[4859]: I1008 18:19:30.169778 4859 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"v4-0-config-system-cliconfig" Oct 08 18:19:30 crc kubenswrapper[4859]: I1008 18:19:30.169947 4859 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication-operator"/"authentication-operator-dockercfg-mz9bj" Oct 08 18:19:30 crc kubenswrapper[4859]: I1008 18:19:30.169981 4859 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"oauth-serving-cert" Oct 08 18:19:30 crc kubenswrapper[4859]: I1008 18:19:30.170076 4859 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-config-operator"/"openshift-config-operator-dockercfg-7pc5z" Oct 08 18:19:30 crc kubenswrapper[4859]: I1008 18:19:30.170103 4859 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"client-ca" Oct 08 18:19:30 crc kubenswrapper[4859]: I1008 18:19:30.170265 4859 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication-operator"/"serving-cert" Oct 08 18:19:30 crc kubenswrapper[4859]: I1008 18:19:30.170379 4859 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"oauth-openshift-dockercfg-znhcc" Oct 08 18:19:30 crc kubenswrapper[4859]: I1008 18:19:30.170419 4859 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication-operator"/"authentication-operator-config" Oct 08 18:19:30 crc kubenswrapper[4859]: I1008 18:19:30.170529 4859 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"client-ca" Oct 08 18:19:30 crc kubenswrapper[4859]: I1008 18:19:30.170597 4859 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-machine-approver"/"machine-approver-sa-dockercfg-nl2j4" Oct 08 18:19:30 crc kubenswrapper[4859]: I1008 18:19:30.171199 4859 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-machine-approver"/"openshift-service-ca.crt" Oct 08 18:19:30 crc kubenswrapper[4859]: I1008 18:19:30.171371 4859 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-system-router-certs" Oct 08 18:19:30 crc kubenswrapper[4859]: I1008 18:19:30.171784 4859 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication-operator"/"service-ca-bundle" Oct 08 18:19:30 crc kubenswrapper[4859]: I1008 18:19:30.171971 4859 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-user-idp-0-file-data" Oct 08 18:19:30 crc kubenswrapper[4859]: I1008 18:19:30.183168 4859 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-kq7mx"] Oct 08 18:19:30 crc kubenswrapper[4859]: I1008 18:19:30.183360 4859 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication-operator"/"openshift-service-ca.crt" Oct 08 18:19:30 crc kubenswrapper[4859]: I1008 18:19:30.183536 4859 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-machine-approver"/"machine-approver-tls" Oct 08 18:19:30 crc kubenswrapper[4859]: I1008 18:19:30.183638 4859 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-machine-approver"/"kube-root-ca.crt" Oct 08 18:19:30 crc kubenswrapper[4859]: I1008 18:19:30.183839 4859 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"console-config" Oct 08 18:19:30 crc kubenswrapper[4859]: I1008 18:19:30.183908 4859 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-kq7mx" Oct 08 18:19:30 crc kubenswrapper[4859]: I1008 18:19:30.183949 4859 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console"/"console-dockercfg-f62pw" Oct 08 18:19:30 crc kubenswrapper[4859]: I1008 18:19:30.184052 4859 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-api"/"machine-api-operator-dockercfg-mfbb7" Oct 08 18:19:30 crc kubenswrapper[4859]: I1008 18:19:30.184130 4859 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console"/"console-oauth-config" Oct 08 18:19:30 crc kubenswrapper[4859]: I1008 18:19:30.184212 4859 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console-operator"/"console-operator-dockercfg-4xjcr" Oct 08 18:19:30 crc kubenswrapper[4859]: I1008 18:19:30.184397 4859 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console-operator"/"console-operator-config" Oct 08 18:19:30 crc kubenswrapper[4859]: I1008 18:19:30.184382 4859 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-oauth-apiserver/apiserver-7bbb656c7d-6j5rk"] Oct 08 18:19:30 crc kubenswrapper[4859]: I1008 18:19:30.184502 4859 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-api"/"openshift-service-ca.crt" Oct 08 18:19:30 crc kubenswrapper[4859]: I1008 18:19:30.184610 4859 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"service-ca" Oct 08 18:19:30 crc kubenswrapper[4859]: I1008 18:19:30.184671 4859 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-api"/"kube-rbac-proxy" Oct 08 18:19:30 crc kubenswrapper[4859]: I1008 18:19:30.185017 4859 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-config-operator"/"kube-root-ca.crt" Oct 08 18:19:30 crc kubenswrapper[4859]: I1008 18:19:30.185177 4859 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver"/"encryption-config-1" Oct 08 18:19:30 crc kubenswrapper[4859]: I1008 18:19:30.185299 4859 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-config-operator"/"config-operator-serving-cert" Oct 08 18:19:30 crc kubenswrapper[4859]: I1008 18:19:30.185342 4859 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-route-controller-manager"/"route-controller-manager-sa-dockercfg-h2zr2" Oct 08 18:19:30 crc kubenswrapper[4859]: I1008 18:19:30.185499 4859 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"openshift-service-ca.crt" Oct 08 18:19:30 crc kubenswrapper[4859]: I1008 18:19:30.185593 4859 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console"/"console-serving-cert" Oct 08 18:19:30 crc kubenswrapper[4859]: I1008 18:19:30.185761 4859 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console-operator"/"kube-root-ca.crt" Oct 08 18:19:30 crc kubenswrapper[4859]: I1008 18:19:30.186127 4859 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"audit-1" Oct 08 18:19:30 crc kubenswrapper[4859]: I1008 18:19:30.186154 4859 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-api"/"machine-api-operator-tls" Oct 08 18:19:30 crc kubenswrapper[4859]: I1008 18:19:30.186305 4859 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console-operator"/"openshift-service-ca.crt" Oct 08 18:19:30 crc kubenswrapper[4859]: I1008 18:19:30.186383 4859 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"etcd-serving-ca" Oct 08 18:19:30 crc kubenswrapper[4859]: I1008 18:19:30.186427 4859 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-samples-operator"/"kube-root-ca.crt" Oct 08 18:19:30 crc kubenswrapper[4859]: I1008 18:19:30.186476 4859 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver"/"etcd-client" Oct 08 18:19:30 crc kubenswrapper[4859]: I1008 18:19:30.186589 4859 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-samples-operator"/"cluster-samples-operator-dockercfg-xpp9w" Oct 08 18:19:30 crc kubenswrapper[4859]: I1008 18:19:30.186747 4859 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console-operator"/"serving-cert" Oct 08 18:19:30 crc kubenswrapper[4859]: I1008 18:19:30.186967 4859 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"openshift-service-ca.crt" Oct 08 18:19:30 crc kubenswrapper[4859]: I1008 18:19:30.187054 4859 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication-operator"/"kube-root-ca.crt" Oct 08 18:19:30 crc kubenswrapper[4859]: I1008 18:19:30.187125 4859 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver"/"openshift-apiserver-sa-dockercfg-djjff" Oct 08 18:19:30 crc kubenswrapper[4859]: I1008 18:19:30.187170 4859 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"kube-root-ca.crt" Oct 08 18:19:30 crc kubenswrapper[4859]: I1008 18:19:30.187220 4859 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-api"/"kube-root-ca.crt" Oct 08 18:19:30 crc kubenswrapper[4859]: I1008 18:19:30.187322 4859 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-samples-operator"/"samples-operator-tls" Oct 08 18:19:30 crc kubenswrapper[4859]: I1008 18:19:30.187133 4859 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-api"/"machine-api-operator-images" Oct 08 18:19:30 crc kubenswrapper[4859]: I1008 18:19:30.187538 4859 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"kube-root-ca.crt" Oct 08 18:19:30 crc kubenswrapper[4859]: I1008 18:19:30.187730 4859 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"config" Oct 08 18:19:30 crc kubenswrapper[4859]: I1008 18:19:30.189206 4859 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver"/"serving-cert" Oct 08 18:19:30 crc kubenswrapper[4859]: I1008 18:19:30.189328 4859 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"image-import-ca" Oct 08 18:19:30 crc kubenswrapper[4859]: I1008 18:19:30.189435 4859 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-samples-operator"/"openshift-service-ca.crt" Oct 08 18:19:30 crc kubenswrapper[4859]: I1008 18:19:30.194701 4859 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-controller-manager/controller-manager-879f6c89f-9gpzj"] Oct 08 18:19:30 crc kubenswrapper[4859]: I1008 18:19:30.197764 4859 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-authentication-operator/authentication-operator-69f744f599-jd5kh"] Oct 08 18:19:30 crc kubenswrapper[4859]: I1008 18:19:30.197831 4859 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-console/downloads-7954f5f757-8q5jz"] Oct 08 18:19:30 crc kubenswrapper[4859]: I1008 18:19:30.198408 4859 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-console/downloads-7954f5f757-8q5jz" Oct 08 18:19:30 crc kubenswrapper[4859]: I1008 18:19:30.200956 4859 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-czj5g"] Oct 08 18:19:30 crc kubenswrapper[4859]: I1008 18:19:30.201652 4859 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-82pcg"] Oct 08 18:19:30 crc kubenswrapper[4859]: I1008 18:19:30.202002 4859 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-etcd-operator/etcd-operator-b45778765-tnn6l"] Oct 08 18:19:30 crc kubenswrapper[4859]: I1008 18:19:30.202348 4859 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-etcd-operator/etcd-operator-b45778765-tnn6l" Oct 08 18:19:30 crc kubenswrapper[4859]: I1008 18:19:30.202739 4859 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-czj5g" Oct 08 18:19:30 crc kubenswrapper[4859]: I1008 18:19:30.202953 4859 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-82pcg" Oct 08 18:19:30 crc kubenswrapper[4859]: I1008 18:19:30.209723 4859 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-etcd-operator"/"etcd-client" Oct 08 18:19:30 crc kubenswrapper[4859]: I1008 18:19:30.211619 4859 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-kc42p"] Oct 08 18:19:30 crc kubenswrapper[4859]: I1008 18:19:30.211849 4859 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"cluster-image-registry-operator-dockercfg-m4qtx" Oct 08 18:19:30 crc kubenswrapper[4859]: I1008 18:19:30.212166 4859 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-kc42p" Oct 08 18:19:30 crc kubenswrapper[4859]: I1008 18:19:30.213369 4859 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-image-registry/image-registry-697d97f7c8-kg44m"] Oct 08 18:19:30 crc kubenswrapper[4859]: I1008 18:19:30.214188 4859 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-dns-operator/dns-operator-744455d44c-q222b"] Oct 08 18:19:30 crc kubenswrapper[4859]: I1008 18:19:30.214312 4859 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/image-registry-697d97f7c8-kg44m" Oct 08 18:19:30 crc kubenswrapper[4859]: I1008 18:19:30.214607 4859 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-dns-operator/dns-operator-744455d44c-q222b" Oct 08 18:19:30 crc kubenswrapper[4859]: I1008 18:19:30.215436 4859 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-ingress-operator/ingress-operator-5b745b69d9-pf7j6"] Oct 08 18:19:30 crc kubenswrapper[4859]: I1008 18:19:30.216574 4859 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ingress-operator/ingress-operator-5b745b69d9-pf7j6" Oct 08 18:19:30 crc kubenswrapper[4859]: I1008 18:19:30.220358 4859 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver-operator"/"openshift-service-ca.crt" Oct 08 18:19:30 crc kubenswrapper[4859]: I1008 18:19:30.220468 4859 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager-operator"/"openshift-service-ca.crt" Oct 08 18:19:30 crc kubenswrapper[4859]: I1008 18:19:30.222002 4859 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"image-registry-operator-tls" Oct 08 18:19:30 crc kubenswrapper[4859]: I1008 18:19:30.223104 4859 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"openshift-global-ca" Oct 08 18:19:30 crc kubenswrapper[4859]: I1008 18:19:30.224036 4859 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-user-template-login" Oct 08 18:19:30 crc kubenswrapper[4859]: I1008 18:19:30.224344 4859 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-etcd-operator"/"openshift-service-ca.crt" Oct 08 18:19:30 crc kubenswrapper[4859]: I1008 18:19:30.224478 4859 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-etcd-operator"/"etcd-ca-bundle" Oct 08 18:19:30 crc kubenswrapper[4859]: I1008 18:19:30.225267 4859 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-etcd-operator"/"etcd-operator-config" Oct 08 18:19:30 crc kubenswrapper[4859]: I1008 18:19:30.225965 4859 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"trusted-ca-bundle" Oct 08 18:19:30 crc kubenswrapper[4859]: I1008 18:19:30.226707 4859 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-etcd-operator"/"etcd-operator-dockercfg-r9srn" Oct 08 18:19:30 crc kubenswrapper[4859]: I1008 18:19:30.226896 4859 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console"/"default-dockercfg-chnjx" Oct 08 18:19:30 crc kubenswrapper[4859]: I1008 18:19:30.226941 4859 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-etcd-operator"/"etcd-operator-serving-cert" Oct 08 18:19:30 crc kubenswrapper[4859]: I1008 18:19:30.240179 4859 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"v4-0-config-system-trusted-ca-bundle" Oct 08 18:19:30 crc kubenswrapper[4859]: I1008 18:19:30.241288 4859 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-machine-config-operator/machine-config-operator-74547568cd-xsc6l"] Oct 08 18:19:30 crc kubenswrapper[4859]: I1008 18:19:30.248659 4859 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-5f4q5"] Oct 08 18:19:30 crc kubenswrapper[4859]: I1008 18:19:30.249069 4859 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/machine-config-operator-74547568cd-xsc6l" Oct 08 18:19:30 crc kubenswrapper[4859]: I1008 18:19:30.264339 4859 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-5f4q5" Oct 08 18:19:30 crc kubenswrapper[4859]: I1008 18:19:30.264540 4859 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-4s5w9"] Oct 08 18:19:30 crc kubenswrapper[4859]: I1008 18:19:30.265160 4859 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-4s5w9" Oct 08 18:19:30 crc kubenswrapper[4859]: I1008 18:19:30.269178 4859 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-ingress/router-default-5444994796-r5gs2"] Oct 08 18:19:30 crc kubenswrapper[4859]: I1008 18:19:30.269479 4859 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-system-ocp-branding-template" Oct 08 18:19:30 crc kubenswrapper[4859]: I1008 18:19:30.269729 4859 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-machine-config-operator/machine-config-controller-84d6567774-m7m5n"] Oct 08 18:19:30 crc kubenswrapper[4859]: I1008 18:19:30.270106 4859 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-storage-version-migrator/migrator-59844c95c7-pmngm"] Oct 08 18:19:30 crc kubenswrapper[4859]: I1008 18:19:30.270344 4859 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/machine-config-controller-84d6567774-m7m5n" Oct 08 18:19:30 crc kubenswrapper[4859]: I1008 18:19:30.270563 4859 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ingress/router-default-5444994796-r5gs2" Oct 08 18:19:30 crc kubenswrapper[4859]: I1008 18:19:30.270869 4859 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication-operator"/"trusted-ca-bundle" Oct 08 18:19:30 crc kubenswrapper[4859]: I1008 18:19:30.271345 4859 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-storage-version-migrator/migrator-59844c95c7-pmngm" Oct 08 18:19:30 crc kubenswrapper[4859]: I1008 18:19:30.272445 4859 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-image-registry"/"trusted-ca" Oct 08 18:19:30 crc kubenswrapper[4859]: I1008 18:19:30.277148 4859 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"trusted-ca-bundle" Oct 08 18:19:30 crc kubenswrapper[4859]: I1008 18:19:30.277463 4859 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver-operator"/"openshift-apiserver-operator-dockercfg-xtcjv" Oct 08 18:19:30 crc kubenswrapper[4859]: I1008 18:19:30.277758 4859 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-etcd-operator"/"etcd-service-ca-bundle" Oct 08 18:19:30 crc kubenswrapper[4859]: I1008 18:19:30.277787 4859 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-machine-api/machine-api-operator-5694c8668f-gmbk9"] Oct 08 18:19:30 crc kubenswrapper[4859]: I1008 18:19:30.278005 4859 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-etcd-operator"/"kube-root-ca.crt" Oct 08 18:19:30 crc kubenswrapper[4859]: I1008 18:19:30.278731 4859 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console-operator"/"trusted-ca" Oct 08 18:19:30 crc kubenswrapper[4859]: I1008 18:19:30.278930 4859 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-authentication/oauth-openshift-558db77b4-j8gh2"] Oct 08 18:19:30 crc kubenswrapper[4859]: I1008 18:19:30.281922 4859 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-service-ca/service-ca-9c57cc56f-cs77r"] Oct 08 18:19:30 crc kubenswrapper[4859]: I1008 18:19:30.282628 4859 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-service-ca/service-ca-9c57cc56f-cs77r" Oct 08 18:19:30 crc kubenswrapper[4859]: I1008 18:19:30.283033 4859 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-vc7bh"] Oct 08 18:19:30 crc kubenswrapper[4859]: I1008 18:19:30.283560 4859 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-vc7bh" Oct 08 18:19:30 crc kubenswrapper[4859]: I1008 18:19:30.284270 4859 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-kq7mx"] Oct 08 18:19:30 crc kubenswrapper[4859]: I1008 18:19:30.286041 4859 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-5n49w"] Oct 08 18:19:30 crc kubenswrapper[4859]: I1008 18:19:30.286418 4859 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-multus/multus-admission-controller-857f4d67dd-lrqkp"] Oct 08 18:19:30 crc kubenswrapper[4859]: I1008 18:19:30.286918 4859 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29332455-w9wt8"] Oct 08 18:19:30 crc kubenswrapper[4859]: I1008 18:19:30.287269 4859 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29332455-w9wt8" Oct 08 18:19:30 crc kubenswrapper[4859]: I1008 18:19:30.287433 4859 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/multus-admission-controller-857f4d67dd-lrqkp" Oct 08 18:19:30 crc kubenswrapper[4859]: I1008 18:19:30.287521 4859 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-5n49w" Oct 08 18:19:30 crc kubenswrapper[4859]: I1008 18:19:30.294244 4859 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/marketplace-operator-79b997595-bz6v8"] Oct 08 18:19:30 crc kubenswrapper[4859]: I1008 18:19:30.294538 4859 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager-operator"/"openshift-controller-manager-operator-dockercfg-vw8fw" Oct 08 18:19:30 crc kubenswrapper[4859]: I1008 18:19:30.296214 4859 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/catalog-operator-68c6474976-xpnsv"] Oct 08 18:19:30 crc kubenswrapper[4859]: I1008 18:19:30.296513 4859 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/marketplace-operator-79b997595-bz6v8" Oct 08 18:19:30 crc kubenswrapper[4859]: I1008 18:19:30.296846 4859 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-xpnsv" Oct 08 18:19:30 crc kubenswrapper[4859]: I1008 18:19:30.298467 4859 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-6576b87f9c-xlfdd"] Oct 08 18:19:30 crc kubenswrapper[4859]: I1008 18:19:30.298542 4859 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-console-operator/console-operator-58897d9998-8sflk"] Oct 08 18:19:30 crc kubenswrapper[4859]: I1008 18:19:30.299241 4859 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-hf2fp"] Oct 08 18:19:30 crc kubenswrapper[4859]: I1008 18:19:30.300294 4859 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-hf2fp" Oct 08 18:19:30 crc kubenswrapper[4859]: I1008 18:19:30.302761 4859 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-dns/dns-default-x7pp9"] Oct 08 18:19:30 crc kubenswrapper[4859]: I1008 18:19:30.303446 4859 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-service-ca-operator/service-ca-operator-777779d784-847r8"] Oct 08 18:19:30 crc kubenswrapper[4859]: I1008 18:19:30.304514 4859 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-dns/dns-default-x7pp9" Oct 08 18:19:30 crc kubenswrapper[4859]: I1008 18:19:30.307263 4859 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-6dlgn"] Oct 08 18:19:30 crc kubenswrapper[4859]: I1008 18:19:30.307795 4859 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-6dlgn" Oct 08 18:19:30 crc kubenswrapper[4859]: I1008 18:19:30.307797 4859 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-service-ca-operator/service-ca-operator-777779d784-847r8" Oct 08 18:19:30 crc kubenswrapper[4859]: I1008 18:19:30.313088 4859 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/d407fd47-096d-4f2a-ace1-dab4ddf7192a-audit-dir\") pod \"apiserver-7bbb656c7d-6j5rk\" (UID: \"d407fd47-096d-4f2a-ace1-dab4ddf7192a\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-6j5rk" Oct 08 18:19:30 crc kubenswrapper[4859]: I1008 18:19:30.313126 4859 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/0902ccb2-78ab-48d0-be02-4c42f03e12fb-v4-0-config-system-service-ca\") pod \"oauth-openshift-558db77b4-j8gh2\" (UID: \"0902ccb2-78ab-48d0-be02-4c42f03e12fb\") " pod="openshift-authentication/oauth-openshift-558db77b4-j8gh2" Oct 08 18:19:30 crc kubenswrapper[4859]: I1008 18:19:30.313165 4859 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-g65gr\" (UniqueName: \"kubernetes.io/projected/9b239e22-0030-4c68-b0f1-66fbaedd5f9b-kube-api-access-g65gr\") pod \"cluster-samples-operator-665b6dd947-c5p45\" (UID: \"9b239e22-0030-4c68-b0f1-66fbaedd5f9b\") " pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-c5p45" Oct 08 18:19:30 crc kubenswrapper[4859]: I1008 18:19:30.313190 4859 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"node-pullsecrets\" (UniqueName: \"kubernetes.io/host-path/70f88ce6-b9fe-422e-827e-af2b6840a783-node-pullsecrets\") pod \"apiserver-76f77b778f-fv9bp\" (UID: \"70f88ce6-b9fe-422e-827e-af2b6840a783\") " pod="openshift-apiserver/apiserver-76f77b778f-fv9bp" Oct 08 18:19:30 crc kubenswrapper[4859]: I1008 18:19:30.313213 4859 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit\" (UniqueName: \"kubernetes.io/configmap/70f88ce6-b9fe-422e-827e-af2b6840a783-audit\") pod \"apiserver-76f77b778f-fv9bp\" (UID: \"70f88ce6-b9fe-422e-827e-af2b6840a783\") " pod="openshift-apiserver/apiserver-76f77b778f-fv9bp" Oct 08 18:19:30 crc kubenswrapper[4859]: I1008 18:19:30.313239 4859 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/2736d658-3609-4c30-a010-7b00c6300e2e-serving-cert\") pod \"console-operator-58897d9998-8sflk\" (UID: \"2736d658-3609-4c30-a010-7b00c6300e2e\") " pod="openshift-console-operator/console-operator-58897d9998-8sflk" Oct 08 18:19:30 crc kubenswrapper[4859]: I1008 18:19:30.313258 4859 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/70f88ce6-b9fe-422e-827e-af2b6840a783-encryption-config\") pod \"apiserver-76f77b778f-fv9bp\" (UID: \"70f88ce6-b9fe-422e-827e-af2b6840a783\") " pod="openshift-apiserver/apiserver-76f77b778f-fv9bp" Oct 08 18:19:30 crc kubenswrapper[4859]: I1008 18:19:30.313279 4859 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-hf7dh\" (UniqueName: \"kubernetes.io/projected/b8d9f66c-e7a4-467f-bcf5-701cc0b68d52-kube-api-access-hf7dh\") pod \"authentication-operator-69f744f599-jd5kh\" (UID: \"b8d9f66c-e7a4-467f-bcf5-701cc0b68d52\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-jd5kh" Oct 08 18:19:30 crc kubenswrapper[4859]: I1008 18:19:30.313296 4859 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/2736d658-3609-4c30-a010-7b00c6300e2e-trusted-ca\") pod \"console-operator-58897d9998-8sflk\" (UID: \"2736d658-3609-4c30-a010-7b00c6300e2e\") " pod="openshift-console-operator/console-operator-58897d9998-8sflk" Oct 08 18:19:30 crc kubenswrapper[4859]: I1008 18:19:30.313317 4859 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/7d9d11d4-9f7b-4f33-a9d2-23db86bb2343-serving-cert\") pod \"openshift-apiserver-operator-796bbdcf4f-czj5g\" (UID: \"7d9d11d4-9f7b-4f33-a9d2-23db86bb2343\") " pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-czj5g" Oct 08 18:19:30 crc kubenswrapper[4859]: I1008 18:19:30.313337 4859 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-8vzp2\" (UniqueName: \"kubernetes.io/projected/d407fd47-096d-4f2a-ace1-dab4ddf7192a-kube-api-access-8vzp2\") pod \"apiserver-7bbb656c7d-6j5rk\" (UID: \"d407fd47-096d-4f2a-ace1-dab4ddf7192a\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-6j5rk" Oct 08 18:19:30 crc kubenswrapper[4859]: I1008 18:19:30.313364 4859 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/460ad6e7-859d-4111-a555-b55c5b7228e0-config\") pod \"controller-manager-879f6c89f-9gpzj\" (UID: \"460ad6e7-859d-4111-a555-b55c5b7228e0\") " pod="openshift-controller-manager/controller-manager-879f6c89f-9gpzj" Oct 08 18:19:30 crc kubenswrapper[4859]: I1008 18:19:30.313388 4859 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/9a829106-59b1-4389-95bf-2ee4d56e317c-console-serving-cert\") pod \"console-f9d7485db-m2xmp\" (UID: \"9a829106-59b1-4389-95bf-2ee4d56e317c\") " pod="openshift-console/console-f9d7485db-m2xmp" Oct 08 18:19:30 crc kubenswrapper[4859]: I1008 18:19:30.313410 4859 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-df4rr\" (UniqueName: \"kubernetes.io/projected/9a829106-59b1-4389-95bf-2ee4d56e317c-kube-api-access-df4rr\") pod \"console-f9d7485db-m2xmp\" (UID: \"9a829106-59b1-4389-95bf-2ee4d56e317c\") " pod="openshift-console/console-f9d7485db-m2xmp" Oct 08 18:19:30 crc kubenswrapper[4859]: I1008 18:19:30.313432 4859 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/d80051ae-5521-4f01-b34d-34d06a661177-serving-cert\") pod \"openshift-controller-manager-operator-756b6f6bc6-82pcg\" (UID: \"d80051ae-5521-4f01-b34d-34d06a661177\") " pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-82pcg" Oct 08 18:19:30 crc kubenswrapper[4859]: I1008 18:19:30.313476 4859 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-w5l7z\" (UniqueName: \"kubernetes.io/projected/39b30c11-d0f6-4302-9ef8-e7fcf1b2170e-kube-api-access-w5l7z\") pod \"machine-approver-56656f9798-95s6m\" (UID: \"39b30c11-d0f6-4302-9ef8-e7fcf1b2170e\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-95s6m" Oct 08 18:19:30 crc kubenswrapper[4859]: I1008 18:19:30.313497 4859 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/70f88ce6-b9fe-422e-827e-af2b6840a783-config\") pod \"apiserver-76f77b778f-fv9bp\" (UID: \"70f88ce6-b9fe-422e-827e-af2b6840a783\") " pod="openshift-apiserver/apiserver-76f77b778f-fv9bp" Oct 08 18:19:30 crc kubenswrapper[4859]: I1008 18:19:30.313516 4859 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/c79c3857-6ecb-4e3c-b59f-6a58ba0f94e9-trusted-ca\") pod \"cluster-image-registry-operator-dc59b4c8b-kq7mx\" (UID: \"c79c3857-6ecb-4e3c-b59f-6a58ba0f94e9\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-kq7mx" Oct 08 18:19:30 crc kubenswrapper[4859]: I1008 18:19:30.313541 4859 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/70f88ce6-b9fe-422e-827e-af2b6840a783-etcd-serving-ca\") pod \"apiserver-76f77b778f-fv9bp\" (UID: \"70f88ce6-b9fe-422e-827e-af2b6840a783\") " pod="openshift-apiserver/apiserver-76f77b778f-fv9bp" Oct 08 18:19:30 crc kubenswrapper[4859]: I1008 18:19:30.313561 4859 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/2736d658-3609-4c30-a010-7b00c6300e2e-config\") pod \"console-operator-58897d9998-8sflk\" (UID: \"2736d658-3609-4c30-a010-7b00c6300e2e\") " pod="openshift-console-operator/console-operator-58897d9998-8sflk" Oct 08 18:19:30 crc kubenswrapper[4859]: I1008 18:19:30.313582 4859 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/460ad6e7-859d-4111-a555-b55c5b7228e0-client-ca\") pod \"controller-manager-879f6c89f-9gpzj\" (UID: \"460ad6e7-859d-4111-a555-b55c5b7228e0\") " pod="openshift-controller-manager/controller-manager-879f6c89f-9gpzj" Oct 08 18:19:30 crc kubenswrapper[4859]: I1008 18:19:30.313605 4859 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/0902ccb2-78ab-48d0-be02-4c42f03e12fb-v4-0-config-system-cliconfig\") pod \"oauth-openshift-558db77b4-j8gh2\" (UID: \"0902ccb2-78ab-48d0-be02-4c42f03e12fb\") " pod="openshift-authentication/oauth-openshift-558db77b4-j8gh2" Oct 08 18:19:30 crc kubenswrapper[4859]: I1008 18:19:30.313631 4859 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/e4b12149-79e9-4b2e-8991-519ad5daa14b-etcd-client\") pod \"etcd-operator-b45778765-tnn6l\" (UID: \"e4b12149-79e9-4b2e-8991-519ad5daa14b\") " pod="openshift-etcd-operator/etcd-operator-b45778765-tnn6l" Oct 08 18:19:30 crc kubenswrapper[4859]: I1008 18:19:30.313652 4859 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-tlm28\" (UniqueName: \"kubernetes.io/projected/e4b12149-79e9-4b2e-8991-519ad5daa14b-kube-api-access-tlm28\") pod \"etcd-operator-b45778765-tnn6l\" (UID: \"e4b12149-79e9-4b2e-8991-519ad5daa14b\") " pod="openshift-etcd-operator/etcd-operator-b45778765-tnn6l" Oct 08 18:19:30 crc kubenswrapper[4859]: I1008 18:19:30.313673 4859 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/bb6dc185-94e3-4cb7-8076-1011c1c60fab-kube-api-access\") pod \"kube-controller-manager-operator-78b949d7b-kc42p\" (UID: \"bb6dc185-94e3-4cb7-8076-1011c1c60fab\") " pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-kc42p" Oct 08 18:19:30 crc kubenswrapper[4859]: I1008 18:19:30.313721 4859 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/c79c3857-6ecb-4e3c-b59f-6a58ba0f94e9-bound-sa-token\") pod \"cluster-image-registry-operator-dc59b4c8b-kq7mx\" (UID: \"c79c3857-6ecb-4e3c-b59f-6a58ba0f94e9\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-kq7mx" Oct 08 18:19:30 crc kubenswrapper[4859]: I1008 18:19:30.313750 4859 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/d407fd47-096d-4f2a-ace1-dab4ddf7192a-etcd-client\") pod \"apiserver-7bbb656c7d-6j5rk\" (UID: \"d407fd47-096d-4f2a-ace1-dab4ddf7192a\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-6j5rk" Oct 08 18:19:30 crc kubenswrapper[4859]: I1008 18:19:30.313778 4859 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/70f88ce6-b9fe-422e-827e-af2b6840a783-audit-dir\") pod \"apiserver-76f77b778f-fv9bp\" (UID: \"70f88ce6-b9fe-422e-827e-af2b6840a783\") " pod="openshift-apiserver/apiserver-76f77b778f-fv9bp" Oct 08 18:19:30 crc kubenswrapper[4859]: I1008 18:19:30.313805 4859 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/9a829106-59b1-4389-95bf-2ee4d56e317c-trusted-ca-bundle\") pod \"console-f9d7485db-m2xmp\" (UID: \"9a829106-59b1-4389-95bf-2ee4d56e317c\") " pod="openshift-console/console-f9d7485db-m2xmp" Oct 08 18:19:30 crc kubenswrapper[4859]: I1008 18:19:30.313834 4859 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"image-registry-operator-tls\" (UniqueName: \"kubernetes.io/secret/c79c3857-6ecb-4e3c-b59f-6a58ba0f94e9-image-registry-operator-tls\") pod \"cluster-image-registry-operator-dc59b4c8b-kq7mx\" (UID: \"c79c3857-6ecb-4e3c-b59f-6a58ba0f94e9\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-kq7mx" Oct 08 18:19:30 crc kubenswrapper[4859]: I1008 18:19:30.313860 4859 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/d407fd47-096d-4f2a-ace1-dab4ddf7192a-trusted-ca-bundle\") pod \"apiserver-7bbb656c7d-6j5rk\" (UID: \"d407fd47-096d-4f2a-ace1-dab4ddf7192a\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-6j5rk" Oct 08 18:19:30 crc kubenswrapper[4859]: I1008 18:19:30.313886 4859 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/39b30c11-d0f6-4302-9ef8-e7fcf1b2170e-config\") pod \"machine-approver-56656f9798-95s6m\" (UID: \"39b30c11-d0f6-4302-9ef8-e7fcf1b2170e\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-95s6m" Oct 08 18:19:30 crc kubenswrapper[4859]: I1008 18:19:30.313913 4859 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-w2wf9\" (UniqueName: \"kubernetes.io/projected/9dcb8f76-0fff-436c-9191-fe3f787b7ac0-kube-api-access-w2wf9\") pod \"downloads-7954f5f757-8q5jz\" (UID: \"9dcb8f76-0fff-436c-9191-fe3f787b7ac0\") " pod="openshift-console/downloads-7954f5f757-8q5jz" Oct 08 18:19:30 crc kubenswrapper[4859]: I1008 18:19:30.313935 4859 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/70f88ce6-b9fe-422e-827e-af2b6840a783-etcd-client\") pod \"apiserver-76f77b778f-fv9bp\" (UID: \"70f88ce6-b9fe-422e-827e-af2b6840a783\") " pod="openshift-apiserver/apiserver-76f77b778f-fv9bp" Oct 08 18:19:30 crc kubenswrapper[4859]: I1008 18:19:30.313957 4859 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/d407fd47-096d-4f2a-ace1-dab4ddf7192a-serving-cert\") pod \"apiserver-7bbb656c7d-6j5rk\" (UID: \"d407fd47-096d-4f2a-ace1-dab4ddf7192a\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-6j5rk" Oct 08 18:19:30 crc kubenswrapper[4859]: I1008 18:19:30.313983 4859 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-machine-config-operator/machine-config-server-7tjh2"] Oct 08 18:19:30 crc kubenswrapper[4859]: I1008 18:19:30.313997 4859 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/b8d9f66c-e7a4-467f-bcf5-701cc0b68d52-trusted-ca-bundle\") pod \"authentication-operator-69f744f599-jd5kh\" (UID: \"b8d9f66c-e7a4-467f-bcf5-701cc0b68d52\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-jd5kh" Oct 08 18:19:30 crc kubenswrapper[4859]: I1008 18:19:30.314022 4859 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-9vc6p\" (UniqueName: \"kubernetes.io/projected/2736d658-3609-4c30-a010-7b00c6300e2e-kube-api-access-9vc6p\") pod \"console-operator-58897d9998-8sflk\" (UID: \"2736d658-3609-4c30-a010-7b00c6300e2e\") " pod="openshift-console-operator/console-operator-58897d9998-8sflk" Oct 08 18:19:30 crc kubenswrapper[4859]: I1008 18:19:30.314043 4859 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/9a829106-59b1-4389-95bf-2ee4d56e317c-console-oauth-config\") pod \"console-f9d7485db-m2xmp\" (UID: \"9a829106-59b1-4389-95bf-2ee4d56e317c\") " pod="openshift-console/console-f9d7485db-m2xmp" Oct 08 18:19:30 crc kubenswrapper[4859]: I1008 18:19:30.314062 4859 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/9a829106-59b1-4389-95bf-2ee4d56e317c-service-ca\") pod \"console-f9d7485db-m2xmp\" (UID: \"9a829106-59b1-4389-95bf-2ee4d56e317c\") " pod="openshift-console/console-f9d7485db-m2xmp" Oct 08 18:19:30 crc kubenswrapper[4859]: I1008 18:19:30.314083 4859 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/d80051ae-5521-4f01-b34d-34d06a661177-config\") pod \"openshift-controller-manager-operator-756b6f6bc6-82pcg\" (UID: \"d80051ae-5521-4f01-b34d-34d06a661177\") " pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-82pcg" Oct 08 18:19:30 crc kubenswrapper[4859]: I1008 18:19:30.314103 4859 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etcd-ca\" (UniqueName: \"kubernetes.io/configmap/e4b12149-79e9-4b2e-8991-519ad5daa14b-etcd-ca\") pod \"etcd-operator-b45778765-tnn6l\" (UID: \"e4b12149-79e9-4b2e-8991-519ad5daa14b\") " pod="openshift-etcd-operator/etcd-operator-b45778765-tnn6l" Oct 08 18:19:30 crc kubenswrapper[4859]: I1008 18:19:30.314125 4859 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/bb6dc185-94e3-4cb7-8076-1011c1c60fab-serving-cert\") pod \"kube-controller-manager-operator-78b949d7b-kc42p\" (UID: \"bb6dc185-94e3-4cb7-8076-1011c1c60fab\") " pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-kc42p" Oct 08 18:19:30 crc kubenswrapper[4859]: I1008 18:19:30.314148 4859 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/0902ccb2-78ab-48d0-be02-4c42f03e12fb-audit-dir\") pod \"oauth-openshift-558db77b4-j8gh2\" (UID: \"0902ccb2-78ab-48d0-be02-4c42f03e12fb\") " pod="openshift-authentication/oauth-openshift-558db77b4-j8gh2" Oct 08 18:19:30 crc kubenswrapper[4859]: I1008 18:19:30.314630 4859 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/machine-config-server-7tjh2" Oct 08 18:19:30 crc kubenswrapper[4859]: I1008 18:19:30.315083 4859 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/70f88ce6-b9fe-422e-827e-af2b6840a783-trusted-ca-bundle\") pod \"apiserver-76f77b778f-fv9bp\" (UID: \"70f88ce6-b9fe-422e-827e-af2b6840a783\") " pod="openshift-apiserver/apiserver-76f77b778f-fv9bp" Oct 08 18:19:30 crc kubenswrapper[4859]: I1008 18:19:30.315158 4859 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"available-featuregates\" (UniqueName: \"kubernetes.io/empty-dir/65868bd3-b636-48c5-aac4-fcf734f84e39-available-featuregates\") pod \"openshift-config-operator-7777fb866f-ncj6n\" (UID: \"65868bd3-b636-48c5-aac4-fcf734f84e39\") " pod="openshift-config-operator/openshift-config-operator-7777fb866f-ncj6n" Oct 08 18:19:30 crc kubenswrapper[4859]: I1008 18:19:30.315214 4859 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/460ad6e7-859d-4111-a555-b55c5b7228e0-serving-cert\") pod \"controller-manager-879f6c89f-9gpzj\" (UID: \"460ad6e7-859d-4111-a555-b55c5b7228e0\") " pod="openshift-controller-manager/controller-manager-879f6c89f-9gpzj" Oct 08 18:19:30 crc kubenswrapper[4859]: I1008 18:19:30.315248 4859 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/460ad6e7-859d-4111-a555-b55c5b7228e0-proxy-ca-bundles\") pod \"controller-manager-879f6c89f-9gpzj\" (UID: \"460ad6e7-859d-4111-a555-b55c5b7228e0\") " pod="openshift-controller-manager/controller-manager-879f6c89f-9gpzj" Oct 08 18:19:30 crc kubenswrapper[4859]: I1008 18:19:30.315274 4859 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/a52b9536-1013-482b-9261-e024377687d2-config\") pod \"machine-api-operator-5694c8668f-gmbk9\" (UID: \"a52b9536-1013-482b-9261-e024377687d2\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-gmbk9" Oct 08 18:19:30 crc kubenswrapper[4859]: I1008 18:19:30.315303 4859 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/d407fd47-096d-4f2a-ace1-dab4ddf7192a-audit-policies\") pod \"apiserver-7bbb656c7d-6j5rk\" (UID: \"d407fd47-096d-4f2a-ace1-dab4ddf7192a\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-6j5rk" Oct 08 18:19:30 crc kubenswrapper[4859]: I1008 18:19:30.315343 4859 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/d407fd47-096d-4f2a-ace1-dab4ddf7192a-etcd-serving-ca\") pod \"apiserver-7bbb656c7d-6j5rk\" (UID: \"d407fd47-096d-4f2a-ace1-dab4ddf7192a\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-6j5rk" Oct 08 18:19:30 crc kubenswrapper[4859]: I1008 18:19:30.315385 4859 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/d407fd47-096d-4f2a-ace1-dab4ddf7192a-encryption-config\") pod \"apiserver-7bbb656c7d-6j5rk\" (UID: \"d407fd47-096d-4f2a-ace1-dab4ddf7192a\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-6j5rk" Oct 08 18:19:30 crc kubenswrapper[4859]: I1008 18:19:30.315414 4859 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-sjk7q\" (UniqueName: \"kubernetes.io/projected/65868bd3-b636-48c5-aac4-fcf734f84e39-kube-api-access-sjk7q\") pod \"openshift-config-operator-7777fb866f-ncj6n\" (UID: \"65868bd3-b636-48c5-aac4-fcf734f84e39\") " pod="openshift-config-operator/openshift-config-operator-7777fb866f-ncj6n" Oct 08 18:19:30 crc kubenswrapper[4859]: I1008 18:19:30.315449 4859 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/ca64d874-5d86-4c49-8953-631a75d59788-serving-cert\") pod \"route-controller-manager-6576b87f9c-xlfdd\" (UID: \"ca64d874-5d86-4c49-8953-631a75d59788\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-xlfdd" Oct 08 18:19:30 crc kubenswrapper[4859]: I1008 18:19:30.315486 4859 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etcd-service-ca\" (UniqueName: \"kubernetes.io/configmap/e4b12149-79e9-4b2e-8991-519ad5daa14b-etcd-service-ca\") pod \"etcd-operator-b45778765-tnn6l\" (UID: \"e4b12149-79e9-4b2e-8991-519ad5daa14b\") " pod="openshift-etcd-operator/etcd-operator-b45778765-tnn6l" Oct 08 18:19:30 crc kubenswrapper[4859]: I1008 18:19:30.315515 4859 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-npz7k\" (UniqueName: \"kubernetes.io/projected/0902ccb2-78ab-48d0-be02-4c42f03e12fb-kube-api-access-npz7k\") pod \"oauth-openshift-558db77b4-j8gh2\" (UID: \"0902ccb2-78ab-48d0-be02-4c42f03e12fb\") " pod="openshift-authentication/oauth-openshift-558db77b4-j8gh2" Oct 08 18:19:30 crc kubenswrapper[4859]: I1008 18:19:30.315546 4859 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"image-import-ca\" (UniqueName: \"kubernetes.io/configmap/70f88ce6-b9fe-422e-827e-af2b6840a783-image-import-ca\") pod \"apiserver-76f77b778f-fv9bp\" (UID: \"70f88ce6-b9fe-422e-827e-af2b6840a783\") " pod="openshift-apiserver/apiserver-76f77b778f-fv9bp" Oct 08 18:19:30 crc kubenswrapper[4859]: I1008 18:19:30.315576 4859 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/ca64d874-5d86-4c49-8953-631a75d59788-client-ca\") pod \"route-controller-manager-6576b87f9c-xlfdd\" (UID: \"ca64d874-5d86-4c49-8953-631a75d59788\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-xlfdd" Oct 08 18:19:30 crc kubenswrapper[4859]: I1008 18:19:30.315595 4859 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/b8d9f66c-e7a4-467f-bcf5-701cc0b68d52-service-ca-bundle\") pod \"authentication-operator-69f744f599-jd5kh\" (UID: \"b8d9f66c-e7a4-467f-bcf5-701cc0b68d52\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-jd5kh" Oct 08 18:19:30 crc kubenswrapper[4859]: I1008 18:19:30.315620 4859 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/9a829106-59b1-4389-95bf-2ee4d56e317c-console-config\") pod \"console-f9d7485db-m2xmp\" (UID: \"9a829106-59b1-4389-95bf-2ee4d56e317c\") " pod="openshift-console/console-f9d7485db-m2xmp" Oct 08 18:19:30 crc kubenswrapper[4859]: I1008 18:19:30.315720 4859 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/0902ccb2-78ab-48d0-be02-4c42f03e12fb-v4-0-config-system-serving-cert\") pod \"oauth-openshift-558db77b4-j8gh2\" (UID: \"0902ccb2-78ab-48d0-be02-4c42f03e12fb\") " pod="openshift-authentication/oauth-openshift-558db77b4-j8gh2" Oct 08 18:19:30 crc kubenswrapper[4859]: I1008 18:19:30.315766 4859 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/0902ccb2-78ab-48d0-be02-4c42f03e12fb-v4-0-config-user-template-provider-selection\") pod \"oauth-openshift-558db77b4-j8gh2\" (UID: \"0902ccb2-78ab-48d0-be02-4c42f03e12fb\") " pod="openshift-authentication/oauth-openshift-558db77b4-j8gh2" Oct 08 18:19:30 crc kubenswrapper[4859]: I1008 18:19:30.315803 4859 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/0902ccb2-78ab-48d0-be02-4c42f03e12fb-audit-policies\") pod \"oauth-openshift-558db77b4-j8gh2\" (UID: \"0902ccb2-78ab-48d0-be02-4c42f03e12fb\") " pod="openshift-authentication/oauth-openshift-558db77b4-j8gh2" Oct 08 18:19:30 crc kubenswrapper[4859]: I1008 18:19:30.316405 4859 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver-operator"/"openshift-apiserver-operator-serving-cert" Oct 08 18:19:30 crc kubenswrapper[4859]: I1008 18:19:30.316461 4859 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-console/console-f9d7485db-m2xmp"] Oct 08 18:19:30 crc kubenswrapper[4859]: I1008 18:19:30.317641 4859 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-q9nqj"] Oct 08 18:19:30 crc kubenswrapper[4859]: I1008 18:19:30.319303 4859 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/0902ccb2-78ab-48d0-be02-4c42f03e12fb-v4-0-config-system-router-certs\") pod \"oauth-openshift-558db77b4-j8gh2\" (UID: \"0902ccb2-78ab-48d0-be02-4c42f03e12fb\") " pod="openshift-authentication/oauth-openshift-558db77b4-j8gh2" Oct 08 18:19:30 crc kubenswrapper[4859]: I1008 18:19:30.320947 4859 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-q9nqj" Oct 08 18:19:30 crc kubenswrapper[4859]: I1008 18:19:30.322054 4859 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"images\" (UniqueName: \"kubernetes.io/configmap/a52b9536-1013-482b-9261-e024377687d2-images\") pod \"machine-api-operator-5694c8668f-gmbk9\" (UID: \"a52b9536-1013-482b-9261-e024377687d2\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-gmbk9" Oct 08 18:19:30 crc kubenswrapper[4859]: I1008 18:19:30.326950 4859 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/b8d9f66c-e7a4-467f-bcf5-701cc0b68d52-serving-cert\") pod \"authentication-operator-69f744f599-jd5kh\" (UID: \"b8d9f66c-e7a4-467f-bcf5-701cc0b68d52\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-jd5kh" Oct 08 18:19:30 crc kubenswrapper[4859]: I1008 18:19:30.327017 4859 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/9a829106-59b1-4389-95bf-2ee4d56e317c-oauth-serving-cert\") pod \"console-f9d7485db-m2xmp\" (UID: \"9a829106-59b1-4389-95bf-2ee4d56e317c\") " pod="openshift-console/console-f9d7485db-m2xmp" Oct 08 18:19:30 crc kubenswrapper[4859]: I1008 18:19:30.327041 4859 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/0902ccb2-78ab-48d0-be02-4c42f03e12fb-v4-0-config-system-ocp-branding-template\") pod \"oauth-openshift-558db77b4-j8gh2\" (UID: \"0902ccb2-78ab-48d0-be02-4c42f03e12fb\") " pod="openshift-authentication/oauth-openshift-558db77b4-j8gh2" Oct 08 18:19:30 crc kubenswrapper[4859]: I1008 18:19:30.327075 4859 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-vvbpr\" (UniqueName: \"kubernetes.io/projected/c79c3857-6ecb-4e3c-b59f-6a58ba0f94e9-kube-api-access-vvbpr\") pod \"cluster-image-registry-operator-dc59b4c8b-kq7mx\" (UID: \"c79c3857-6ecb-4e3c-b59f-6a58ba0f94e9\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-kq7mx" Oct 08 18:19:30 crc kubenswrapper[4859]: I1008 18:19:30.327091 4859 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/0902ccb2-78ab-48d0-be02-4c42f03e12fb-v4-0-config-user-template-login\") pod \"oauth-openshift-558db77b4-j8gh2\" (UID: \"0902ccb2-78ab-48d0-be02-4c42f03e12fb\") " pod="openshift-authentication/oauth-openshift-558db77b4-j8gh2" Oct 08 18:19:30 crc kubenswrapper[4859]: I1008 18:19:30.327112 4859 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"machine-approver-tls\" (UniqueName: \"kubernetes.io/secret/39b30c11-d0f6-4302-9ef8-e7fcf1b2170e-machine-approver-tls\") pod \"machine-approver-56656f9798-95s6m\" (UID: \"39b30c11-d0f6-4302-9ef8-e7fcf1b2170e\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-95s6m" Oct 08 18:19:30 crc kubenswrapper[4859]: I1008 18:19:30.327132 4859 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7d9d11d4-9f7b-4f33-a9d2-23db86bb2343-config\") pod \"openshift-apiserver-operator-796bbdcf4f-czj5g\" (UID: \"7d9d11d4-9f7b-4f33-a9d2-23db86bb2343\") " pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-czj5g" Oct 08 18:19:30 crc kubenswrapper[4859]: I1008 18:19:30.327151 4859 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/e4b12149-79e9-4b2e-8991-519ad5daa14b-serving-cert\") pod \"etcd-operator-b45778765-tnn6l\" (UID: \"e4b12149-79e9-4b2e-8991-519ad5daa14b\") " pod="openshift-etcd-operator/etcd-operator-b45778765-tnn6l" Oct 08 18:19:30 crc kubenswrapper[4859]: I1008 18:19:30.327168 4859 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/0902ccb2-78ab-48d0-be02-4c42f03e12fb-v4-0-config-system-trusted-ca-bundle\") pod \"oauth-openshift-558db77b4-j8gh2\" (UID: \"0902ccb2-78ab-48d0-be02-4c42f03e12fb\") " pod="openshift-authentication/oauth-openshift-558db77b4-j8gh2" Oct 08 18:19:30 crc kubenswrapper[4859]: I1008 18:19:30.327188 4859 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"machine-api-operator-tls\" (UniqueName: \"kubernetes.io/secret/a52b9536-1013-482b-9261-e024377687d2-machine-api-operator-tls\") pod \"machine-api-operator-5694c8668f-gmbk9\" (UID: \"a52b9536-1013-482b-9261-e024377687d2\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-gmbk9" Oct 08 18:19:30 crc kubenswrapper[4859]: I1008 18:19:30.327209 4859 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/39b30c11-d0f6-4302-9ef8-e7fcf1b2170e-auth-proxy-config\") pod \"machine-approver-56656f9798-95s6m\" (UID: \"39b30c11-d0f6-4302-9ef8-e7fcf1b2170e\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-95s6m" Oct 08 18:19:30 crc kubenswrapper[4859]: I1008 18:19:30.327256 4859 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-4ppdz\" (UniqueName: \"kubernetes.io/projected/70f88ce6-b9fe-422e-827e-af2b6840a783-kube-api-access-4ppdz\") pod \"apiserver-76f77b778f-fv9bp\" (UID: \"70f88ce6-b9fe-422e-827e-af2b6840a783\") " pod="openshift-apiserver/apiserver-76f77b778f-fv9bp" Oct 08 18:19:30 crc kubenswrapper[4859]: I1008 18:19:30.327279 4859 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/0902ccb2-78ab-48d0-be02-4c42f03e12fb-v4-0-config-user-idp-0-file-data\") pod \"oauth-openshift-558db77b4-j8gh2\" (UID: \"0902ccb2-78ab-48d0-be02-4c42f03e12fb\") " pod="openshift-authentication/oauth-openshift-558db77b4-j8gh2" Oct 08 18:19:30 crc kubenswrapper[4859]: I1008 18:19:30.327315 4859 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"samples-operator-tls\" (UniqueName: \"kubernetes.io/secret/9b239e22-0030-4c68-b0f1-66fbaedd5f9b-samples-operator-tls\") pod \"cluster-samples-operator-665b6dd947-c5p45\" (UID: \"9b239e22-0030-4c68-b0f1-66fbaedd5f9b\") " pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-c5p45" Oct 08 18:19:30 crc kubenswrapper[4859]: I1008 18:19:30.327338 4859 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/70f88ce6-b9fe-422e-827e-af2b6840a783-serving-cert\") pod \"apiserver-76f77b778f-fv9bp\" (UID: \"70f88ce6-b9fe-422e-827e-af2b6840a783\") " pod="openshift-apiserver/apiserver-76f77b778f-fv9bp" Oct 08 18:19:30 crc kubenswrapper[4859]: I1008 18:19:30.327360 4859 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-kfmlz\" (UniqueName: \"kubernetes.io/projected/d80051ae-5521-4f01-b34d-34d06a661177-kube-api-access-kfmlz\") pod \"openshift-controller-manager-operator-756b6f6bc6-82pcg\" (UID: \"d80051ae-5521-4f01-b34d-34d06a661177\") " pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-82pcg" Oct 08 18:19:30 crc kubenswrapper[4859]: I1008 18:19:30.327383 4859 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/bb6dc185-94e3-4cb7-8076-1011c1c60fab-config\") pod \"kube-controller-manager-operator-78b949d7b-kc42p\" (UID: \"bb6dc185-94e3-4cb7-8076-1011c1c60fab\") " pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-kc42p" Oct 08 18:19:30 crc kubenswrapper[4859]: I1008 18:19:30.327402 4859 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/65868bd3-b636-48c5-aac4-fcf734f84e39-serving-cert\") pod \"openshift-config-operator-7777fb866f-ncj6n\" (UID: \"65868bd3-b636-48c5-aac4-fcf734f84e39\") " pod="openshift-config-operator/openshift-config-operator-7777fb866f-ncj6n" Oct 08 18:19:30 crc kubenswrapper[4859]: I1008 18:19:30.327422 4859 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/ca64d874-5d86-4c49-8953-631a75d59788-config\") pod \"route-controller-manager-6576b87f9c-xlfdd\" (UID: \"ca64d874-5d86-4c49-8953-631a75d59788\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-xlfdd" Oct 08 18:19:30 crc kubenswrapper[4859]: I1008 18:19:30.327446 4859 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-hjhn5\" (UniqueName: \"kubernetes.io/projected/7d9d11d4-9f7b-4f33-a9d2-23db86bb2343-kube-api-access-hjhn5\") pod \"openshift-apiserver-operator-796bbdcf4f-czj5g\" (UID: \"7d9d11d4-9f7b-4f33-a9d2-23db86bb2343\") " pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-czj5g" Oct 08 18:19:30 crc kubenswrapper[4859]: I1008 18:19:30.327463 4859 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/e4b12149-79e9-4b2e-8991-519ad5daa14b-config\") pod \"etcd-operator-b45778765-tnn6l\" (UID: \"e4b12149-79e9-4b2e-8991-519ad5daa14b\") " pod="openshift-etcd-operator/etcd-operator-b45778765-tnn6l" Oct 08 18:19:30 crc kubenswrapper[4859]: I1008 18:19:30.327496 4859 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-9hnbb\" (UniqueName: \"kubernetes.io/projected/ca64d874-5d86-4c49-8953-631a75d59788-kube-api-access-9hnbb\") pod \"route-controller-manager-6576b87f9c-xlfdd\" (UID: \"ca64d874-5d86-4c49-8953-631a75d59788\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-xlfdd" Oct 08 18:19:30 crc kubenswrapper[4859]: I1008 18:19:30.327539 4859 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/b8d9f66c-e7a4-467f-bcf5-701cc0b68d52-config\") pod \"authentication-operator-69f744f599-jd5kh\" (UID: \"b8d9f66c-e7a4-467f-bcf5-701cc0b68d52\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-jd5kh" Oct 08 18:19:30 crc kubenswrapper[4859]: I1008 18:19:30.327559 4859 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/0902ccb2-78ab-48d0-be02-4c42f03e12fb-v4-0-config-user-template-error\") pod \"oauth-openshift-558db77b4-j8gh2\" (UID: \"0902ccb2-78ab-48d0-be02-4c42f03e12fb\") " pod="openshift-authentication/oauth-openshift-558db77b4-j8gh2" Oct 08 18:19:30 crc kubenswrapper[4859]: I1008 18:19:30.327574 4859 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-dwj4n\" (UniqueName: \"kubernetes.io/projected/a52b9536-1013-482b-9261-e024377687d2-kube-api-access-dwj4n\") pod \"machine-api-operator-5694c8668f-gmbk9\" (UID: \"a52b9536-1013-482b-9261-e024377687d2\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-gmbk9" Oct 08 18:19:30 crc kubenswrapper[4859]: I1008 18:19:30.327595 4859 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-gbpww\" (UniqueName: \"kubernetes.io/projected/460ad6e7-859d-4111-a555-b55c5b7228e0-kube-api-access-gbpww\") pod \"controller-manager-879f6c89f-9gpzj\" (UID: \"460ad6e7-859d-4111-a555-b55c5b7228e0\") " pod="openshift-controller-manager/controller-manager-879f6c89f-9gpzj" Oct 08 18:19:30 crc kubenswrapper[4859]: I1008 18:19:30.327636 4859 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/0902ccb2-78ab-48d0-be02-4c42f03e12fb-v4-0-config-system-session\") pod \"oauth-openshift-558db77b4-j8gh2\" (UID: \"0902ccb2-78ab-48d0-be02-4c42f03e12fb\") " pod="openshift-authentication/oauth-openshift-558db77b4-j8gh2" Oct 08 18:19:30 crc kubenswrapper[4859]: I1008 18:19:30.327773 4859 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-c5p45"] Oct 08 18:19:30 crc kubenswrapper[4859]: I1008 18:19:30.333767 4859 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-etcd-operator/etcd-operator-b45778765-tnn6l"] Oct 08 18:19:30 crc kubenswrapper[4859]: I1008 18:19:30.336146 4859 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver-operator"/"openshift-apiserver-operator-config" Oct 08 18:19:30 crc kubenswrapper[4859]: I1008 18:19:30.340269 4859 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-ingress-operator/ingress-operator-5b745b69d9-pf7j6"] Oct 08 18:19:30 crc kubenswrapper[4859]: I1008 18:19:30.340321 4859 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-storage-version-migrator/migrator-59844c95c7-pmngm"] Oct 08 18:19:30 crc kubenswrapper[4859]: I1008 18:19:30.341996 4859 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-czj5g"] Oct 08 18:19:30 crc kubenswrapper[4859]: I1008 18:19:30.344801 4859 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-config-operator/openshift-config-operator-7777fb866f-ncj6n"] Oct 08 18:19:30 crc kubenswrapper[4859]: I1008 18:19:30.348670 4859 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-82pcg"] Oct 08 18:19:30 crc kubenswrapper[4859]: I1008 18:19:30.349218 4859 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-image-registry/image-registry-697d97f7c8-kg44m"] Oct 08 18:19:30 crc kubenswrapper[4859]: I1008 18:19:30.350726 4859 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-console/downloads-7954f5f757-8q5jz"] Oct 08 18:19:30 crc kubenswrapper[4859]: I1008 18:19:30.351373 4859 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-apiserver/apiserver-76f77b778f-fv9bp"] Oct 08 18:19:30 crc kubenswrapper[4859]: I1008 18:19:30.353673 4859 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver-operator"/"kube-root-ca.crt" Oct 08 18:19:30 crc kubenswrapper[4859]: I1008 18:19:30.355220 4859 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["hostpath-provisioner/csi-hostpathplugin-q4lkb"] Oct 08 18:19:30 crc kubenswrapper[4859]: I1008 18:19:30.358132 4859 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="hostpath-provisioner/csi-hostpathplugin-q4lkb" Oct 08 18:19:30 crc kubenswrapper[4859]: I1008 18:19:30.367329 4859 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29332455-w9wt8"] Oct 08 18:19:30 crc kubenswrapper[4859]: I1008 18:19:30.368925 4859 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-dns-operator/dns-operator-744455d44c-q222b"] Oct 08 18:19:30 crc kubenswrapper[4859]: I1008 18:19:30.370123 4859 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager-operator"/"openshift-controller-manager-operator-config" Oct 08 18:19:30 crc kubenswrapper[4859]: I1008 18:19:30.370158 4859 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-ingress-canary/ingress-canary-qtvkr"] Oct 08 18:19:30 crc kubenswrapper[4859]: I1008 18:19:30.371036 4859 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ingress-canary/ingress-canary-qtvkr" Oct 08 18:19:30 crc kubenswrapper[4859]: I1008 18:19:30.371518 4859 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-vc7bh"] Oct 08 18:19:30 crc kubenswrapper[4859]: I1008 18:19:30.372713 4859 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-machine-config-operator/machine-config-controller-84d6567774-m7m5n"] Oct 08 18:19:30 crc kubenswrapper[4859]: I1008 18:19:30.374813 4859 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-service-ca/service-ca-9c57cc56f-cs77r"] Oct 08 18:19:30 crc kubenswrapper[4859]: I1008 18:19:30.376733 4859 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-kc42p"] Oct 08 18:19:30 crc kubenswrapper[4859]: I1008 18:19:30.378795 4859 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/catalog-operator-68c6474976-xpnsv"] Oct 08 18:19:30 crc kubenswrapper[4859]: I1008 18:19:30.380236 4859 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-5n49w"] Oct 08 18:19:30 crc kubenswrapper[4859]: I1008 18:19:30.381268 4859 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-multus/multus-admission-controller-857f4d67dd-lrqkp"] Oct 08 18:19:30 crc kubenswrapper[4859]: I1008 18:19:30.382274 4859 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-5f4q5"] Oct 08 18:19:30 crc kubenswrapper[4859]: I1008 18:19:30.383599 4859 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-machine-config-operator/machine-config-operator-74547568cd-xsc6l"] Oct 08 18:19:30 crc kubenswrapper[4859]: I1008 18:19:30.384959 4859 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/marketplace-operator-79b997595-bz6v8"] Oct 08 18:19:30 crc kubenswrapper[4859]: I1008 18:19:30.387268 4859 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-service-ca-operator/service-ca-operator-777779d784-847r8"] Oct 08 18:19:30 crc kubenswrapper[4859]: I1008 18:19:30.390416 4859 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager-operator"/"openshift-controller-manager-operator-serving-cert" Oct 08 18:19:30 crc kubenswrapper[4859]: I1008 18:19:30.391249 4859 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-6dlgn"] Oct 08 18:19:30 crc kubenswrapper[4859]: I1008 18:19:30.393886 4859 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-dns/dns-default-x7pp9"] Oct 08 18:19:30 crc kubenswrapper[4859]: I1008 18:19:30.395344 4859 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-4s5w9"] Oct 08 18:19:30 crc kubenswrapper[4859]: I1008 18:19:30.396621 4859 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["hostpath-provisioner/csi-hostpathplugin-q4lkb"] Oct 08 18:19:30 crc kubenswrapper[4859]: I1008 18:19:30.398159 4859 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-hf2fp"] Oct 08 18:19:30 crc kubenswrapper[4859]: I1008 18:19:30.399109 4859 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-q9nqj"] Oct 08 18:19:30 crc kubenswrapper[4859]: I1008 18:19:30.400519 4859 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-ingress-canary/ingress-canary-qtvkr"] Oct 08 18:19:30 crc kubenswrapper[4859]: I1008 18:19:30.410383 4859 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager-operator"/"kube-root-ca.crt" Oct 08 18:19:30 crc kubenswrapper[4859]: I1008 18:19:30.428231 4859 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/bb6dc185-94e3-4cb7-8076-1011c1c60fab-serving-cert\") pod \"kube-controller-manager-operator-78b949d7b-kc42p\" (UID: \"bb6dc185-94e3-4cb7-8076-1011c1c60fab\") " pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-kc42p" Oct 08 18:19:30 crc kubenswrapper[4859]: I1008 18:19:30.428259 4859 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"available-featuregates\" (UniqueName: \"kubernetes.io/empty-dir/65868bd3-b636-48c5-aac4-fcf734f84e39-available-featuregates\") pod \"openshift-config-operator-7777fb866f-ncj6n\" (UID: \"65868bd3-b636-48c5-aac4-fcf734f84e39\") " pod="openshift-config-operator/openshift-config-operator-7777fb866f-ncj6n" Oct 08 18:19:30 crc kubenswrapper[4859]: I1008 18:19:30.428283 4859 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/460ad6e7-859d-4111-a555-b55c5b7228e0-proxy-ca-bundles\") pod \"controller-manager-879f6c89f-9gpzj\" (UID: \"460ad6e7-859d-4111-a555-b55c5b7228e0\") " pod="openshift-controller-manager/controller-manager-879f6c89f-9gpzj" Oct 08 18:19:30 crc kubenswrapper[4859]: I1008 18:19:30.428306 4859 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/8d4b5b44-816f-44f6-98ab-c5d66aefaa08-marketplace-operator-metrics\") pod \"marketplace-operator-79b997595-bz6v8\" (UID: \"8d4b5b44-816f-44f6-98ab-c5d66aefaa08\") " pod="openshift-marketplace/marketplace-operator-79b997595-bz6v8" Oct 08 18:19:30 crc kubenswrapper[4859]: I1008 18:19:30.428328 4859 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/b2b5fcca-e076-4cbe-bc1e-249fcfd25e53-apiservice-cert\") pod \"packageserver-d55dfcdfc-vc7bh\" (UID: \"b2b5fcca-e076-4cbe-bc1e-249fcfd25e53\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-vc7bh" Oct 08 18:19:30 crc kubenswrapper[4859]: I1008 18:19:30.428374 4859 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/d407fd47-096d-4f2a-ace1-dab4ddf7192a-audit-policies\") pod \"apiserver-7bbb656c7d-6j5rk\" (UID: \"d407fd47-096d-4f2a-ace1-dab4ddf7192a\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-6j5rk" Oct 08 18:19:30 crc kubenswrapper[4859]: I1008 18:19:30.428412 4859 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/d407fd47-096d-4f2a-ace1-dab4ddf7192a-etcd-serving-ca\") pod \"apiserver-7bbb656c7d-6j5rk\" (UID: \"d407fd47-096d-4f2a-ace1-dab4ddf7192a\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-6j5rk" Oct 08 18:19:30 crc kubenswrapper[4859]: I1008 18:19:30.428430 4859 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-sjk7q\" (UniqueName: \"kubernetes.io/projected/65868bd3-b636-48c5-aac4-fcf734f84e39-kube-api-access-sjk7q\") pod \"openshift-config-operator-7777fb866f-ncj6n\" (UID: \"65868bd3-b636-48c5-aac4-fcf734f84e39\") " pod="openshift-config-operator/openshift-config-operator-7777fb866f-ncj6n" Oct 08 18:19:30 crc kubenswrapper[4859]: I1008 18:19:30.428444 4859 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/ca64d874-5d86-4c49-8953-631a75d59788-serving-cert\") pod \"route-controller-manager-6576b87f9c-xlfdd\" (UID: \"ca64d874-5d86-4c49-8953-631a75d59788\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-xlfdd" Oct 08 18:19:30 crc kubenswrapper[4859]: I1008 18:19:30.428757 4859 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"available-featuregates\" (UniqueName: \"kubernetes.io/empty-dir/65868bd3-b636-48c5-aac4-fcf734f84e39-available-featuregates\") pod \"openshift-config-operator-7777fb866f-ncj6n\" (UID: \"65868bd3-b636-48c5-aac4-fcf734f84e39\") " pod="openshift-config-operator/openshift-config-operator-7777fb866f-ncj6n" Oct 08 18:19:30 crc kubenswrapper[4859]: I1008 18:19:30.428868 4859 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etcd-service-ca\" (UniqueName: \"kubernetes.io/configmap/e4b12149-79e9-4b2e-8991-519ad5daa14b-etcd-service-ca\") pod \"etcd-operator-b45778765-tnn6l\" (UID: \"e4b12149-79e9-4b2e-8991-519ad5daa14b\") " pod="openshift-etcd-operator/etcd-operator-b45778765-tnn6l" Oct 08 18:19:30 crc kubenswrapper[4859]: I1008 18:19:30.428965 4859 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-npz7k\" (UniqueName: \"kubernetes.io/projected/0902ccb2-78ab-48d0-be02-4c42f03e12fb-kube-api-access-npz7k\") pod \"oauth-openshift-558db77b4-j8gh2\" (UID: \"0902ccb2-78ab-48d0-be02-4c42f03e12fb\") " pod="openshift-authentication/oauth-openshift-558db77b4-j8gh2" Oct 08 18:19:30 crc kubenswrapper[4859]: I1008 18:19:30.429006 4859 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/ca64d874-5d86-4c49-8953-631a75d59788-client-ca\") pod \"route-controller-manager-6576b87f9c-xlfdd\" (UID: \"ca64d874-5d86-4c49-8953-631a75d59788\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-xlfdd" Oct 08 18:19:30 crc kubenswrapper[4859]: I1008 18:19:30.429028 4859 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/b8d9f66c-e7a4-467f-bcf5-701cc0b68d52-service-ca-bundle\") pod \"authentication-operator-69f744f599-jd5kh\" (UID: \"b8d9f66c-e7a4-467f-bcf5-701cc0b68d52\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-jd5kh" Oct 08 18:19:30 crc kubenswrapper[4859]: I1008 18:19:30.429081 4859 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/9a829106-59b1-4389-95bf-2ee4d56e317c-console-config\") pod \"console-f9d7485db-m2xmp\" (UID: \"9a829106-59b1-4389-95bf-2ee4d56e317c\") " pod="openshift-console/console-f9d7485db-m2xmp" Oct 08 18:19:30 crc kubenswrapper[4859]: I1008 18:19:30.429099 4859 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/0902ccb2-78ab-48d0-be02-4c42f03e12fb-v4-0-config-system-serving-cert\") pod \"oauth-openshift-558db77b4-j8gh2\" (UID: \"0902ccb2-78ab-48d0-be02-4c42f03e12fb\") " pod="openshift-authentication/oauth-openshift-558db77b4-j8gh2" Oct 08 18:19:30 crc kubenswrapper[4859]: I1008 18:19:30.429124 4859 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/01517214-a376-421e-9938-e548d76695d5-config\") pod \"kube-apiserver-operator-766d6c64bb-4s5w9\" (UID: \"01517214-a376-421e-9938-e548d76695d5\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-4s5w9" Oct 08 18:19:30 crc kubenswrapper[4859]: I1008 18:19:30.429146 4859 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"images\" (UniqueName: \"kubernetes.io/configmap/a52b9536-1013-482b-9261-e024377687d2-images\") pod \"machine-api-operator-5694c8668f-gmbk9\" (UID: \"a52b9536-1013-482b-9261-e024377687d2\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-gmbk9" Oct 08 18:19:30 crc kubenswrapper[4859]: I1008 18:19:30.429163 4859 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/01517214-a376-421e-9938-e548d76695d5-serving-cert\") pod \"kube-apiserver-operator-766d6c64bb-4s5w9\" (UID: \"01517214-a376-421e-9938-e548d76695d5\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-4s5w9" Oct 08 18:19:30 crc kubenswrapper[4859]: I1008 18:19:30.429184 4859 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"machine-approver-tls\" (UniqueName: \"kubernetes.io/secret/39b30c11-d0f6-4302-9ef8-e7fcf1b2170e-machine-approver-tls\") pod \"machine-approver-56656f9798-95s6m\" (UID: \"39b30c11-d0f6-4302-9ef8-e7fcf1b2170e\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-95s6m" Oct 08 18:19:30 crc kubenswrapper[4859]: I1008 18:19:30.429209 4859 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/0902ccb2-78ab-48d0-be02-4c42f03e12fb-v4-0-config-system-trusted-ca-bundle\") pod \"oauth-openshift-558db77b4-j8gh2\" (UID: \"0902ccb2-78ab-48d0-be02-4c42f03e12fb\") " pod="openshift-authentication/oauth-openshift-558db77b4-j8gh2" Oct 08 18:19:30 crc kubenswrapper[4859]: I1008 18:19:30.429228 4859 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"machine-api-operator-tls\" (UniqueName: \"kubernetes.io/secret/a52b9536-1013-482b-9261-e024377687d2-machine-api-operator-tls\") pod \"machine-api-operator-5694c8668f-gmbk9\" (UID: \"a52b9536-1013-482b-9261-e024377687d2\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-gmbk9" Oct 08 18:19:30 crc kubenswrapper[4859]: I1008 18:19:30.429252 4859 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/0c9973e7-5c40-463b-aca1-c3c93d7a9d0b-auth-proxy-config\") pod \"machine-config-operator-74547568cd-xsc6l\" (UID: \"0c9973e7-5c40-463b-aca1-c3c93d7a9d0b\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-xsc6l" Oct 08 18:19:30 crc kubenswrapper[4859]: I1008 18:19:30.429273 4859 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/8d4b5b44-816f-44f6-98ab-c5d66aefaa08-marketplace-trusted-ca\") pod \"marketplace-operator-79b997595-bz6v8\" (UID: \"8d4b5b44-816f-44f6-98ab-c5d66aefaa08\") " pod="openshift-marketplace/marketplace-operator-79b997595-bz6v8" Oct 08 18:19:30 crc kubenswrapper[4859]: I1008 18:19:30.429301 4859 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/39b30c11-d0f6-4302-9ef8-e7fcf1b2170e-auth-proxy-config\") pod \"machine-approver-56656f9798-95s6m\" (UID: \"39b30c11-d0f6-4302-9ef8-e7fcf1b2170e\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-95s6m" Oct 08 18:19:30 crc kubenswrapper[4859]: I1008 18:19:30.429321 4859 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/0902ccb2-78ab-48d0-be02-4c42f03e12fb-v4-0-config-user-idp-0-file-data\") pod \"oauth-openshift-558db77b4-j8gh2\" (UID: \"0902ccb2-78ab-48d0-be02-4c42f03e12fb\") " pod="openshift-authentication/oauth-openshift-558db77b4-j8gh2" Oct 08 18:19:30 crc kubenswrapper[4859]: I1008 18:19:30.429380 4859 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"samples-operator-tls\" (UniqueName: \"kubernetes.io/secret/9b239e22-0030-4c68-b0f1-66fbaedd5f9b-samples-operator-tls\") pod \"cluster-samples-operator-665b6dd947-c5p45\" (UID: \"9b239e22-0030-4c68-b0f1-66fbaedd5f9b\") " pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-c5p45" Oct 08 18:19:30 crc kubenswrapper[4859]: I1008 18:19:30.429407 4859 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/70f88ce6-b9fe-422e-827e-af2b6840a783-serving-cert\") pod \"apiserver-76f77b778f-fv9bp\" (UID: \"70f88ce6-b9fe-422e-827e-af2b6840a783\") " pod="openshift-apiserver/apiserver-76f77b778f-fv9bp" Oct 08 18:19:30 crc kubenswrapper[4859]: I1008 18:19:30.429429 4859 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/6f093915-bce6-40eb-917f-ed0058431a79-profile-collector-cert\") pod \"olm-operator-6b444d44fb-5n49w\" (UID: \"6f093915-bce6-40eb-917f-ed0058431a79\") " pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-5n49w" Oct 08 18:19:30 crc kubenswrapper[4859]: I1008 18:19:30.429448 4859 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/e4b12149-79e9-4b2e-8991-519ad5daa14b-config\") pod \"etcd-operator-b45778765-tnn6l\" (UID: \"e4b12149-79e9-4b2e-8991-519ad5daa14b\") " pod="openshift-etcd-operator/etcd-operator-b45778765-tnn6l" Oct 08 18:19:30 crc kubenswrapper[4859]: I1008 18:19:30.429468 4859 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/01517214-a376-421e-9938-e548d76695d5-kube-api-access\") pod \"kube-apiserver-operator-766d6c64bb-4s5w9\" (UID: \"01517214-a376-421e-9938-e548d76695d5\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-4s5w9" Oct 08 18:19:30 crc kubenswrapper[4859]: I1008 18:19:30.429493 4859 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-hjhn5\" (UniqueName: \"kubernetes.io/projected/7d9d11d4-9f7b-4f33-a9d2-23db86bb2343-kube-api-access-hjhn5\") pod \"openshift-apiserver-operator-796bbdcf4f-czj5g\" (UID: \"7d9d11d4-9f7b-4f33-a9d2-23db86bb2343\") " pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-czj5g" Oct 08 18:19:30 crc kubenswrapper[4859]: I1008 18:19:30.429512 4859 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/b8d9f66c-e7a4-467f-bcf5-701cc0b68d52-config\") pod \"authentication-operator-69f744f599-jd5kh\" (UID: \"b8d9f66c-e7a4-467f-bcf5-701cc0b68d52\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-jd5kh" Oct 08 18:19:30 crc kubenswrapper[4859]: I1008 18:19:30.429536 4859 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-gbpww\" (UniqueName: \"kubernetes.io/projected/460ad6e7-859d-4111-a555-b55c5b7228e0-kube-api-access-gbpww\") pod \"controller-manager-879f6c89f-9gpzj\" (UID: \"460ad6e7-859d-4111-a555-b55c5b7228e0\") " pod="openshift-controller-manager/controller-manager-879f6c89f-9gpzj" Oct 08 18:19:30 crc kubenswrapper[4859]: I1008 18:19:30.429558 4859 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-jb26g\" (UniqueName: \"kubernetes.io/projected/6f093915-bce6-40eb-917f-ed0058431a79-kube-api-access-jb26g\") pod \"olm-operator-6b444d44fb-5n49w\" (UID: \"6f093915-bce6-40eb-917f-ed0058431a79\") " pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-5n49w" Oct 08 18:19:30 crc kubenswrapper[4859]: I1008 18:19:30.429591 4859 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-v4vwn\" (UniqueName: \"kubernetes.io/projected/d403dae9-eb8c-4c8d-9daa-bf2a0c815652-kube-api-access-v4vwn\") pod \"collect-profiles-29332455-w9wt8\" (UID: \"d403dae9-eb8c-4c8d-9daa-bf2a0c815652\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29332455-w9wt8" Oct 08 18:19:30 crc kubenswrapper[4859]: I1008 18:19:30.429610 4859 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"node-pullsecrets\" (UniqueName: \"kubernetes.io/host-path/70f88ce6-b9fe-422e-827e-af2b6840a783-node-pullsecrets\") pod \"apiserver-76f77b778f-fv9bp\" (UID: \"70f88ce6-b9fe-422e-827e-af2b6840a783\") " pod="openshift-apiserver/apiserver-76f77b778f-fv9bp" Oct 08 18:19:30 crc kubenswrapper[4859]: I1008 18:19:30.429626 4859 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/d403dae9-eb8c-4c8d-9daa-bf2a0c815652-config-volume\") pod \"collect-profiles-29332455-w9wt8\" (UID: \"d403dae9-eb8c-4c8d-9daa-bf2a0c815652\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29332455-w9wt8" Oct 08 18:19:30 crc kubenswrapper[4859]: I1008 18:19:30.429645 4859 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/70f88ce6-b9fe-422e-827e-af2b6840a783-encryption-config\") pod \"apiserver-76f77b778f-fv9bp\" (UID: \"70f88ce6-b9fe-422e-827e-af2b6840a783\") " pod="openshift-apiserver/apiserver-76f77b778f-fv9bp" Oct 08 18:19:30 crc kubenswrapper[4859]: I1008 18:19:30.429665 4859 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/2736d658-3609-4c30-a010-7b00c6300e2e-trusted-ca\") pod \"console-operator-58897d9998-8sflk\" (UID: \"2736d658-3609-4c30-a010-7b00c6300e2e\") " pod="openshift-console-operator/console-operator-58897d9998-8sflk" Oct 08 18:19:30 crc kubenswrapper[4859]: I1008 18:19:30.429708 4859 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/7d9d11d4-9f7b-4f33-a9d2-23db86bb2343-serving-cert\") pod \"openshift-apiserver-operator-796bbdcf4f-czj5g\" (UID: \"7d9d11d4-9f7b-4f33-a9d2-23db86bb2343\") " pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-czj5g" Oct 08 18:19:30 crc kubenswrapper[4859]: I1008 18:19:30.429718 4859 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/460ad6e7-859d-4111-a555-b55c5b7228e0-proxy-ca-bundles\") pod \"controller-manager-879f6c89f-9gpzj\" (UID: \"460ad6e7-859d-4111-a555-b55c5b7228e0\") " pod="openshift-controller-manager/controller-manager-879f6c89f-9gpzj" Oct 08 18:19:30 crc kubenswrapper[4859]: I1008 18:19:30.429733 4859 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-8vzp2\" (UniqueName: \"kubernetes.io/projected/d407fd47-096d-4f2a-ace1-dab4ddf7192a-kube-api-access-8vzp2\") pod \"apiserver-7bbb656c7d-6j5rk\" (UID: \"d407fd47-096d-4f2a-ace1-dab4ddf7192a\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-6j5rk" Oct 08 18:19:30 crc kubenswrapper[4859]: I1008 18:19:30.429756 4859 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"signing-key\" (UniqueName: \"kubernetes.io/secret/da9f04a1-bb6a-40de-846a-5e2fe04855d1-signing-key\") pod \"service-ca-9c57cc56f-cs77r\" (UID: \"da9f04a1-bb6a-40de-846a-5e2fe04855d1\") " pod="openshift-service-ca/service-ca-9c57cc56f-cs77r" Oct 08 18:19:30 crc kubenswrapper[4859]: I1008 18:19:30.429772 4859 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/b8d9f66c-e7a4-467f-bcf5-701cc0b68d52-service-ca-bundle\") pod \"authentication-operator-69f744f599-jd5kh\" (UID: \"b8d9f66c-e7a4-467f-bcf5-701cc0b68d52\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-jd5kh" Oct 08 18:19:30 crc kubenswrapper[4859]: I1008 18:19:30.429782 4859 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"images\" (UniqueName: \"kubernetes.io/configmap/0c9973e7-5c40-463b-aca1-c3c93d7a9d0b-images\") pod \"machine-config-operator-74547568cd-xsc6l\" (UID: \"0c9973e7-5c40-463b-aca1-c3c93d7a9d0b\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-xsc6l" Oct 08 18:19:30 crc kubenswrapper[4859]: I1008 18:19:30.430427 4859 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"images\" (UniqueName: \"kubernetes.io/configmap/a52b9536-1013-482b-9261-e024377687d2-images\") pod \"machine-api-operator-5694c8668f-gmbk9\" (UID: \"a52b9536-1013-482b-9261-e024377687d2\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-gmbk9" Oct 08 18:19:30 crc kubenswrapper[4859]: I1008 18:19:30.430515 4859 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"node-pullsecrets\" (UniqueName: \"kubernetes.io/host-path/70f88ce6-b9fe-422e-827e-af2b6840a783-node-pullsecrets\") pod \"apiserver-76f77b778f-fv9bp\" (UID: \"70f88ce6-b9fe-422e-827e-af2b6840a783\") " pod="openshift-apiserver/apiserver-76f77b778f-fv9bp" Oct 08 18:19:30 crc kubenswrapper[4859]: I1008 18:19:30.430547 4859 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/9a829106-59b1-4389-95bf-2ee4d56e317c-console-config\") pod \"console-f9d7485db-m2xmp\" (UID: \"9a829106-59b1-4389-95bf-2ee4d56e317c\") " pod="openshift-console/console-f9d7485db-m2xmp" Oct 08 18:19:30 crc kubenswrapper[4859]: I1008 18:19:30.430973 4859 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/0902ccb2-78ab-48d0-be02-4c42f03e12fb-v4-0-config-system-trusted-ca-bundle\") pod \"oauth-openshift-558db77b4-j8gh2\" (UID: \"0902ccb2-78ab-48d0-be02-4c42f03e12fb\") " pod="openshift-authentication/oauth-openshift-558db77b4-j8gh2" Oct 08 18:19:30 crc kubenswrapper[4859]: I1008 18:19:30.431356 4859 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/e4b12149-79e9-4b2e-8991-519ad5daa14b-config\") pod \"etcd-operator-b45778765-tnn6l\" (UID: \"e4b12149-79e9-4b2e-8991-519ad5daa14b\") " pod="openshift-etcd-operator/etcd-operator-b45778765-tnn6l" Oct 08 18:19:30 crc kubenswrapper[4859]: I1008 18:19:30.432055 4859 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/8e2b8a53-4bc3-499b-9746-628c323d2a0c-metrics-certs\") pod \"router-default-5444994796-r5gs2\" (UID: \"8e2b8a53-4bc3-499b-9746-628c323d2a0c\") " pod="openshift-ingress/router-default-5444994796-r5gs2" Oct 08 18:19:30 crc kubenswrapper[4859]: I1008 18:19:30.432085 4859 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-vxgz5\" (UniqueName: \"kubernetes.io/projected/8e2b8a53-4bc3-499b-9746-628c323d2a0c-kube-api-access-vxgz5\") pod \"router-default-5444994796-r5gs2\" (UID: \"8e2b8a53-4bc3-499b-9746-628c323d2a0c\") " pod="openshift-ingress/router-default-5444994796-r5gs2" Oct 08 18:19:30 crc kubenswrapper[4859]: I1008 18:19:30.432207 4859 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/c79c3857-6ecb-4e3c-b59f-6a58ba0f94e9-trusted-ca\") pod \"cluster-image-registry-operator-dc59b4c8b-kq7mx\" (UID: \"c79c3857-6ecb-4e3c-b59f-6a58ba0f94e9\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-kq7mx" Oct 08 18:19:30 crc kubenswrapper[4859]: I1008 18:19:30.432230 4859 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-glqwn\" (UniqueName: \"kubernetes.io/projected/da9f04a1-bb6a-40de-846a-5e2fe04855d1-kube-api-access-glqwn\") pod \"service-ca-9c57cc56f-cs77r\" (UID: \"da9f04a1-bb6a-40de-846a-5e2fe04855d1\") " pod="openshift-service-ca/service-ca-9c57cc56f-cs77r" Oct 08 18:19:30 crc kubenswrapper[4859]: I1008 18:19:30.432250 4859 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-hv4wh\" (UniqueName: \"kubernetes.io/projected/8d4b5b44-816f-44f6-98ab-c5d66aefaa08-kube-api-access-hv4wh\") pod \"marketplace-operator-79b997595-bz6v8\" (UID: \"8d4b5b44-816f-44f6-98ab-c5d66aefaa08\") " pod="openshift-marketplace/marketplace-operator-79b997595-bz6v8" Oct 08 18:19:30 crc kubenswrapper[4859]: I1008 18:19:30.432299 4859 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/460ad6e7-859d-4111-a555-b55c5b7228e0-client-ca\") pod \"controller-manager-879f6c89f-9gpzj\" (UID: \"460ad6e7-859d-4111-a555-b55c5b7228e0\") " pod="openshift-controller-manager/controller-manager-879f6c89f-9gpzj" Oct 08 18:19:30 crc kubenswrapper[4859]: I1008 18:19:30.432319 4859 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/0902ccb2-78ab-48d0-be02-4c42f03e12fb-v4-0-config-system-cliconfig\") pod \"oauth-openshift-558db77b4-j8gh2\" (UID: \"0902ccb2-78ab-48d0-be02-4c42f03e12fb\") " pod="openshift-authentication/oauth-openshift-558db77b4-j8gh2" Oct 08 18:19:30 crc kubenswrapper[4859]: I1008 18:19:30.433093 4859 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-tlm28\" (UniqueName: \"kubernetes.io/projected/e4b12149-79e9-4b2e-8991-519ad5daa14b-kube-api-access-tlm28\") pod \"etcd-operator-b45778765-tnn6l\" (UID: \"e4b12149-79e9-4b2e-8991-519ad5daa14b\") " pod="openshift-etcd-operator/etcd-operator-b45778765-tnn6l" Oct 08 18:19:30 crc kubenswrapper[4859]: I1008 18:19:30.433147 4859 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/c79c3857-6ecb-4e3c-b59f-6a58ba0f94e9-bound-sa-token\") pod \"cluster-image-registry-operator-dc59b4c8b-kq7mx\" (UID: \"c79c3857-6ecb-4e3c-b59f-6a58ba0f94e9\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-kq7mx" Oct 08 18:19:30 crc kubenswrapper[4859]: I1008 18:19:30.433219 4859 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/d407fd47-096d-4f2a-ace1-dab4ddf7192a-etcd-client\") pod \"apiserver-7bbb656c7d-6j5rk\" (UID: \"d407fd47-096d-4f2a-ace1-dab4ddf7192a\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-6j5rk" Oct 08 18:19:30 crc kubenswrapper[4859]: I1008 18:19:30.433238 4859 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/d407fd47-096d-4f2a-ace1-dab4ddf7192a-trusted-ca-bundle\") pod \"apiserver-7bbb656c7d-6j5rk\" (UID: \"d407fd47-096d-4f2a-ace1-dab4ddf7192a\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-6j5rk" Oct 08 18:19:30 crc kubenswrapper[4859]: I1008 18:19:30.433257 4859 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/9a829106-59b1-4389-95bf-2ee4d56e317c-trusted-ca-bundle\") pod \"console-f9d7485db-m2xmp\" (UID: \"9a829106-59b1-4389-95bf-2ee4d56e317c\") " pod="openshift-console/console-f9d7485db-m2xmp" Oct 08 18:19:30 crc kubenswrapper[4859]: I1008 18:19:30.433295 4859 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/460ad6e7-859d-4111-a555-b55c5b7228e0-client-ca\") pod \"controller-manager-879f6c89f-9gpzj\" (UID: \"460ad6e7-859d-4111-a555-b55c5b7228e0\") " pod="openshift-controller-manager/controller-manager-879f6c89f-9gpzj" Oct 08 18:19:30 crc kubenswrapper[4859]: I1008 18:19:30.433397 4859 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/70f88ce6-b9fe-422e-827e-af2b6840a783-audit-dir\") pod \"apiserver-76f77b778f-fv9bp\" (UID: \"70f88ce6-b9fe-422e-827e-af2b6840a783\") " pod="openshift-apiserver/apiserver-76f77b778f-fv9bp" Oct 08 18:19:30 crc kubenswrapper[4859]: I1008 18:19:30.433394 4859 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/0902ccb2-78ab-48d0-be02-4c42f03e12fb-v4-0-config-system-cliconfig\") pod \"oauth-openshift-558db77b4-j8gh2\" (UID: \"0902ccb2-78ab-48d0-be02-4c42f03e12fb\") " pod="openshift-authentication/oauth-openshift-558db77b4-j8gh2" Oct 08 18:19:30 crc kubenswrapper[4859]: I1008 18:19:30.433417 4859 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/b8d9f66c-e7a4-467f-bcf5-701cc0b68d52-config\") pod \"authentication-operator-69f744f599-jd5kh\" (UID: \"b8d9f66c-e7a4-467f-bcf5-701cc0b68d52\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-jd5kh" Oct 08 18:19:30 crc kubenswrapper[4859]: I1008 18:19:30.433419 4859 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"image-registry-operator-tls\" (UniqueName: \"kubernetes.io/secret/c79c3857-6ecb-4e3c-b59f-6a58ba0f94e9-image-registry-operator-tls\") pod \"cluster-image-registry-operator-dc59b4c8b-kq7mx\" (UID: \"c79c3857-6ecb-4e3c-b59f-6a58ba0f94e9\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-kq7mx" Oct 08 18:19:30 crc kubenswrapper[4859]: I1008 18:19:30.433464 4859 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/70f88ce6-b9fe-422e-827e-af2b6840a783-audit-dir\") pod \"apiserver-76f77b778f-fv9bp\" (UID: \"70f88ce6-b9fe-422e-827e-af2b6840a783\") " pod="openshift-apiserver/apiserver-76f77b778f-fv9bp" Oct 08 18:19:30 crc kubenswrapper[4859]: I1008 18:19:30.433479 4859 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/0c9973e7-5c40-463b-aca1-c3c93d7a9d0b-proxy-tls\") pod \"machine-config-operator-74547568cd-xsc6l\" (UID: \"0c9973e7-5c40-463b-aca1-c3c93d7a9d0b\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-xsc6l" Oct 08 18:19:30 crc kubenswrapper[4859]: I1008 18:19:30.433524 4859 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/39b30c11-d0f6-4302-9ef8-e7fcf1b2170e-config\") pod \"machine-approver-56656f9798-95s6m\" (UID: \"39b30c11-d0f6-4302-9ef8-e7fcf1b2170e\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-95s6m" Oct 08 18:19:30 crc kubenswrapper[4859]: I1008 18:19:30.433548 4859 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-w2wf9\" (UniqueName: \"kubernetes.io/projected/9dcb8f76-0fff-436c-9191-fe3f787b7ac0-kube-api-access-w2wf9\") pod \"downloads-7954f5f757-8q5jz\" (UID: \"9dcb8f76-0fff-436c-9191-fe3f787b7ac0\") " pod="openshift-console/downloads-7954f5f757-8q5jz" Oct 08 18:19:30 crc kubenswrapper[4859]: I1008 18:19:30.433572 4859 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/70f88ce6-b9fe-422e-827e-af2b6840a783-etcd-client\") pod \"apiserver-76f77b778f-fv9bp\" (UID: \"70f88ce6-b9fe-422e-827e-af2b6840a783\") " pod="openshift-apiserver/apiserver-76f77b778f-fv9bp" Oct 08 18:19:30 crc kubenswrapper[4859]: I1008 18:19:30.433615 4859 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/c79c3857-6ecb-4e3c-b59f-6a58ba0f94e9-trusted-ca\") pod \"cluster-image-registry-operator-dc59b4c8b-kq7mx\" (UID: \"c79c3857-6ecb-4e3c-b59f-6a58ba0f94e9\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-kq7mx" Oct 08 18:19:30 crc kubenswrapper[4859]: I1008 18:19:30.433735 4859 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/2736d658-3609-4c30-a010-7b00c6300e2e-trusted-ca\") pod \"console-operator-58897d9998-8sflk\" (UID: \"2736d658-3609-4c30-a010-7b00c6300e2e\") " pod="openshift-console-operator/console-operator-58897d9998-8sflk" Oct 08 18:19:30 crc kubenswrapper[4859]: I1008 18:19:30.433735 4859 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/56c0b58f-23d6-4d8e-8b75-9095128ac121-config\") pod \"service-ca-operator-777779d784-847r8\" (UID: \"56c0b58f-23d6-4d8e-8b75-9095128ac121\") " pod="openshift-service-ca-operator/service-ca-operator-777779d784-847r8" Oct 08 18:19:30 crc kubenswrapper[4859]: I1008 18:19:30.433841 4859 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/b8d9f66c-e7a4-467f-bcf5-701cc0b68d52-trusted-ca-bundle\") pod \"authentication-operator-69f744f599-jd5kh\" (UID: \"b8d9f66c-e7a4-467f-bcf5-701cc0b68d52\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-jd5kh" Oct 08 18:19:30 crc kubenswrapper[4859]: I1008 18:19:30.433882 4859 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-9vc6p\" (UniqueName: \"kubernetes.io/projected/2736d658-3609-4c30-a010-7b00c6300e2e-kube-api-access-9vc6p\") pod \"console-operator-58897d9998-8sflk\" (UID: \"2736d658-3609-4c30-a010-7b00c6300e2e\") " pod="openshift-console-operator/console-operator-58897d9998-8sflk" Oct 08 18:19:30 crc kubenswrapper[4859]: I1008 18:19:30.433924 4859 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/9a829106-59b1-4389-95bf-2ee4d56e317c-service-ca\") pod \"console-f9d7485db-m2xmp\" (UID: \"9a829106-59b1-4389-95bf-2ee4d56e317c\") " pod="openshift-console/console-f9d7485db-m2xmp" Oct 08 18:19:30 crc kubenswrapper[4859]: I1008 18:19:30.434138 4859 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/39b30c11-d0f6-4302-9ef8-e7fcf1b2170e-config\") pod \"machine-approver-56656f9798-95s6m\" (UID: \"39b30c11-d0f6-4302-9ef8-e7fcf1b2170e\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-95s6m" Oct 08 18:19:30 crc kubenswrapper[4859]: I1008 18:19:30.434319 4859 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/9a829106-59b1-4389-95bf-2ee4d56e317c-trusted-ca-bundle\") pod \"console-f9d7485db-m2xmp\" (UID: \"9a829106-59b1-4389-95bf-2ee4d56e317c\") " pod="openshift-console/console-f9d7485db-m2xmp" Oct 08 18:19:30 crc kubenswrapper[4859]: I1008 18:19:30.434370 4859 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/39b30c11-d0f6-4302-9ef8-e7fcf1b2170e-auth-proxy-config\") pod \"machine-approver-56656f9798-95s6m\" (UID: \"39b30c11-d0f6-4302-9ef8-e7fcf1b2170e\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-95s6m" Oct 08 18:19:30 crc kubenswrapper[4859]: I1008 18:19:30.434498 4859 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/ca64d874-5d86-4c49-8953-631a75d59788-client-ca\") pod \"route-controller-manager-6576b87f9c-xlfdd\" (UID: \"ca64d874-5d86-4c49-8953-631a75d59788\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-xlfdd" Oct 08 18:19:30 crc kubenswrapper[4859]: I1008 18:19:30.434538 4859 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/d80051ae-5521-4f01-b34d-34d06a661177-config\") pod \"openshift-controller-manager-operator-756b6f6bc6-82pcg\" (UID: \"d80051ae-5521-4f01-b34d-34d06a661177\") " pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-82pcg" Oct 08 18:19:30 crc kubenswrapper[4859]: I1008 18:19:30.434557 4859 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"signing-cabundle\" (UniqueName: \"kubernetes.io/configmap/da9f04a1-bb6a-40de-846a-5e2fe04855d1-signing-cabundle\") pod \"service-ca-9c57cc56f-cs77r\" (UID: \"da9f04a1-bb6a-40de-846a-5e2fe04855d1\") " pod="openshift-service-ca/service-ca-9c57cc56f-cs77r" Oct 08 18:19:30 crc kubenswrapper[4859]: I1008 18:19:30.434590 4859 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"certs\" (UniqueName: \"kubernetes.io/secret/0d04221d-922c-4ee9-9adf-817e295161d9-certs\") pod \"machine-config-server-7tjh2\" (UID: \"0d04221d-922c-4ee9-9adf-817e295161d9\") " pod="openshift-machine-config-operator/machine-config-server-7tjh2" Oct 08 18:19:30 crc kubenswrapper[4859]: I1008 18:19:30.434610 4859 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etcd-ca\" (UniqueName: \"kubernetes.io/configmap/e4b12149-79e9-4b2e-8991-519ad5daa14b-etcd-ca\") pod \"etcd-operator-b45778765-tnn6l\" (UID: \"e4b12149-79e9-4b2e-8991-519ad5daa14b\") " pod="openshift-etcd-operator/etcd-operator-b45778765-tnn6l" Oct 08 18:19:30 crc kubenswrapper[4859]: I1008 18:19:30.434628 4859 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/0902ccb2-78ab-48d0-be02-4c42f03e12fb-audit-dir\") pod \"oauth-openshift-558db77b4-j8gh2\" (UID: \"0902ccb2-78ab-48d0-be02-4c42f03e12fb\") " pod="openshift-authentication/oauth-openshift-558db77b4-j8gh2" Oct 08 18:19:30 crc kubenswrapper[4859]: I1008 18:19:30.434675 4859 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/70f88ce6-b9fe-422e-827e-af2b6840a783-trusted-ca-bundle\") pod \"apiserver-76f77b778f-fv9bp\" (UID: \"70f88ce6-b9fe-422e-827e-af2b6840a783\") " pod="openshift-apiserver/apiserver-76f77b778f-fv9bp" Oct 08 18:19:30 crc kubenswrapper[4859]: I1008 18:19:30.434706 4859 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/460ad6e7-859d-4111-a555-b55c5b7228e0-serving-cert\") pod \"controller-manager-879f6c89f-9gpzj\" (UID: \"460ad6e7-859d-4111-a555-b55c5b7228e0\") " pod="openshift-controller-manager/controller-manager-879f6c89f-9gpzj" Oct 08 18:19:30 crc kubenswrapper[4859]: I1008 18:19:30.434726 4859 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/a52b9536-1013-482b-9261-e024377687d2-config\") pod \"machine-api-operator-5694c8668f-gmbk9\" (UID: \"a52b9536-1013-482b-9261-e024377687d2\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-gmbk9" Oct 08 18:19:30 crc kubenswrapper[4859]: I1008 18:19:30.434767 4859 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"default-certificate\" (UniqueName: \"kubernetes.io/secret/8e2b8a53-4bc3-499b-9746-628c323d2a0c-default-certificate\") pod \"router-default-5444994796-r5gs2\" (UID: \"8e2b8a53-4bc3-499b-9746-628c323d2a0c\") " pod="openshift-ingress/router-default-5444994796-r5gs2" Oct 08 18:19:30 crc kubenswrapper[4859]: I1008 18:19:30.434824 4859 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/d407fd47-096d-4f2a-ace1-dab4ddf7192a-encryption-config\") pod \"apiserver-7bbb656c7d-6j5rk\" (UID: \"d407fd47-096d-4f2a-ace1-dab4ddf7192a\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-6j5rk" Oct 08 18:19:30 crc kubenswrapper[4859]: I1008 18:19:30.434854 4859 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-jbvx4\" (UniqueName: \"kubernetes.io/projected/b2b5fcca-e076-4cbe-bc1e-249fcfd25e53-kube-api-access-jbvx4\") pod \"packageserver-d55dfcdfc-vc7bh\" (UID: \"b2b5fcca-e076-4cbe-bc1e-249fcfd25e53\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-vc7bh" Oct 08 18:19:30 crc kubenswrapper[4859]: I1008 18:19:30.434871 4859 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"stats-auth\" (UniqueName: \"kubernetes.io/secret/8e2b8a53-4bc3-499b-9746-628c323d2a0c-stats-auth\") pod \"router-default-5444994796-r5gs2\" (UID: \"8e2b8a53-4bc3-499b-9746-628c323d2a0c\") " pod="openshift-ingress/router-default-5444994796-r5gs2" Oct 08 18:19:30 crc kubenswrapper[4859]: I1008 18:19:30.434891 4859 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"image-import-ca\" (UniqueName: \"kubernetes.io/configmap/70f88ce6-b9fe-422e-827e-af2b6840a783-image-import-ca\") pod \"apiserver-76f77b778f-fv9bp\" (UID: \"70f88ce6-b9fe-422e-827e-af2b6840a783\") " pod="openshift-apiserver/apiserver-76f77b778f-fv9bp" Oct 08 18:19:30 crc kubenswrapper[4859]: I1008 18:19:30.434911 4859 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/0902ccb2-78ab-48d0-be02-4c42f03e12fb-v4-0-config-user-template-provider-selection\") pod \"oauth-openshift-558db77b4-j8gh2\" (UID: \"0902ccb2-78ab-48d0-be02-4c42f03e12fb\") " pod="openshift-authentication/oauth-openshift-558db77b4-j8gh2" Oct 08 18:19:30 crc kubenswrapper[4859]: I1008 18:19:30.434926 4859 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/0902ccb2-78ab-48d0-be02-4c42f03e12fb-audit-dir\") pod \"oauth-openshift-558db77b4-j8gh2\" (UID: \"0902ccb2-78ab-48d0-be02-4c42f03e12fb\") " pod="openshift-authentication/oauth-openshift-558db77b4-j8gh2" Oct 08 18:19:30 crc kubenswrapper[4859]: I1008 18:19:30.435282 4859 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/0902ccb2-78ab-48d0-be02-4c42f03e12fb-v4-0-config-user-idp-0-file-data\") pod \"oauth-openshift-558db77b4-j8gh2\" (UID: \"0902ccb2-78ab-48d0-be02-4c42f03e12fb\") " pod="openshift-authentication/oauth-openshift-558db77b4-j8gh2" Oct 08 18:19:30 crc kubenswrapper[4859]: I1008 18:19:30.435466 4859 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/d80051ae-5521-4f01-b34d-34d06a661177-config\") pod \"openshift-controller-manager-operator-756b6f6bc6-82pcg\" (UID: \"d80051ae-5521-4f01-b34d-34d06a661177\") " pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-82pcg" Oct 08 18:19:30 crc kubenswrapper[4859]: I1008 18:19:30.435774 4859 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/b8d9f66c-e7a4-467f-bcf5-701cc0b68d52-trusted-ca-bundle\") pod \"authentication-operator-69f744f599-jd5kh\" (UID: \"b8d9f66c-e7a4-467f-bcf5-701cc0b68d52\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-jd5kh" Oct 08 18:19:30 crc kubenswrapper[4859]: I1008 18:19:30.435868 4859 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"samples-operator-tls\" (UniqueName: \"kubernetes.io/secret/9b239e22-0030-4c68-b0f1-66fbaedd5f9b-samples-operator-tls\") pod \"cluster-samples-operator-665b6dd947-c5p45\" (UID: \"9b239e22-0030-4c68-b0f1-66fbaedd5f9b\") " pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-c5p45" Oct 08 18:19:30 crc kubenswrapper[4859]: I1008 18:19:30.435890 4859 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/9a829106-59b1-4389-95bf-2ee4d56e317c-service-ca\") pod \"console-f9d7485db-m2xmp\" (UID: \"9a829106-59b1-4389-95bf-2ee4d56e317c\") " pod="openshift-console/console-f9d7485db-m2xmp" Oct 08 18:19:30 crc kubenswrapper[4859]: I1008 18:19:30.435947 4859 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"machine-approver-tls\" (UniqueName: \"kubernetes.io/secret/39b30c11-d0f6-4302-9ef8-e7fcf1b2170e-machine-approver-tls\") pod \"machine-approver-56656f9798-95s6m\" (UID: \"39b30c11-d0f6-4302-9ef8-e7fcf1b2170e\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-95s6m" Oct 08 18:19:30 crc kubenswrapper[4859]: I1008 18:19:30.435961 4859 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/0902ccb2-78ab-48d0-be02-4c42f03e12fb-audit-policies\") pod \"oauth-openshift-558db77b4-j8gh2\" (UID: \"0902ccb2-78ab-48d0-be02-4c42f03e12fb\") " pod="openshift-authentication/oauth-openshift-558db77b4-j8gh2" Oct 08 18:19:30 crc kubenswrapper[4859]: I1008 18:19:30.436144 4859 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/0902ccb2-78ab-48d0-be02-4c42f03e12fb-v4-0-config-system-router-certs\") pod \"oauth-openshift-558db77b4-j8gh2\" (UID: \"0902ccb2-78ab-48d0-be02-4c42f03e12fb\") " pod="openshift-authentication/oauth-openshift-558db77b4-j8gh2" Oct 08 18:19:30 crc kubenswrapper[4859]: I1008 18:19:30.436183 4859 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/9a829106-59b1-4389-95bf-2ee4d56e317c-oauth-serving-cert\") pod \"console-f9d7485db-m2xmp\" (UID: \"9a829106-59b1-4389-95bf-2ee4d56e317c\") " pod="openshift-console/console-f9d7485db-m2xmp" Oct 08 18:19:30 crc kubenswrapper[4859]: I1008 18:19:30.436186 4859 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etcd-service-ca\" (UniqueName: \"kubernetes.io/configmap/e4b12149-79e9-4b2e-8991-519ad5daa14b-etcd-service-ca\") pod \"etcd-operator-b45778765-tnn6l\" (UID: \"e4b12149-79e9-4b2e-8991-519ad5daa14b\") " pod="openshift-etcd-operator/etcd-operator-b45778765-tnn6l" Oct 08 18:19:30 crc kubenswrapper[4859]: I1008 18:19:30.436546 4859 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/0902ccb2-78ab-48d0-be02-4c42f03e12fb-v4-0-config-system-ocp-branding-template\") pod \"oauth-openshift-558db77b4-j8gh2\" (UID: \"0902ccb2-78ab-48d0-be02-4c42f03e12fb\") " pod="openshift-authentication/oauth-openshift-558db77b4-j8gh2" Oct 08 18:19:30 crc kubenswrapper[4859]: I1008 18:19:30.436724 4859 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/70f88ce6-b9fe-422e-827e-af2b6840a783-trusted-ca-bundle\") pod \"apiserver-76f77b778f-fv9bp\" (UID: \"70f88ce6-b9fe-422e-827e-af2b6840a783\") " pod="openshift-apiserver/apiserver-76f77b778f-fv9bp" Oct 08 18:19:30 crc kubenswrapper[4859]: I1008 18:19:30.436728 4859 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/70f88ce6-b9fe-422e-827e-af2b6840a783-etcd-client\") pod \"apiserver-76f77b778f-fv9bp\" (UID: \"70f88ce6-b9fe-422e-827e-af2b6840a783\") " pod="openshift-apiserver/apiserver-76f77b778f-fv9bp" Oct 08 18:19:30 crc kubenswrapper[4859]: I1008 18:19:30.436779 4859 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/b8d9f66c-e7a4-467f-bcf5-701cc0b68d52-serving-cert\") pod \"authentication-operator-69f744f599-jd5kh\" (UID: \"b8d9f66c-e7a4-467f-bcf5-701cc0b68d52\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-jd5kh" Oct 08 18:19:30 crc kubenswrapper[4859]: I1008 18:19:30.436842 4859 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"image-import-ca\" (UniqueName: \"kubernetes.io/configmap/70f88ce6-b9fe-422e-827e-af2b6840a783-image-import-ca\") pod \"apiserver-76f77b778f-fv9bp\" (UID: \"70f88ce6-b9fe-422e-827e-af2b6840a783\") " pod="openshift-apiserver/apiserver-76f77b778f-fv9bp" Oct 08 18:19:30 crc kubenswrapper[4859]: I1008 18:19:30.436844 4859 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/0902ccb2-78ab-48d0-be02-4c42f03e12fb-v4-0-config-user-template-login\") pod \"oauth-openshift-558db77b4-j8gh2\" (UID: \"0902ccb2-78ab-48d0-be02-4c42f03e12fb\") " pod="openshift-authentication/oauth-openshift-558db77b4-j8gh2" Oct 08 18:19:30 crc kubenswrapper[4859]: I1008 18:19:30.436901 4859 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-vvbpr\" (UniqueName: \"kubernetes.io/projected/c79c3857-6ecb-4e3c-b59f-6a58ba0f94e9-kube-api-access-vvbpr\") pod \"cluster-image-registry-operator-dc59b4c8b-kq7mx\" (UID: \"c79c3857-6ecb-4e3c-b59f-6a58ba0f94e9\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-kq7mx" Oct 08 18:19:30 crc kubenswrapper[4859]: I1008 18:19:30.436948 4859 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7d9d11d4-9f7b-4f33-a9d2-23db86bb2343-config\") pod \"openshift-apiserver-operator-796bbdcf4f-czj5g\" (UID: \"7d9d11d4-9f7b-4f33-a9d2-23db86bb2343\") " pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-czj5g" Oct 08 18:19:30 crc kubenswrapper[4859]: I1008 18:19:30.436998 4859 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/e4b12149-79e9-4b2e-8991-519ad5daa14b-serving-cert\") pod \"etcd-operator-b45778765-tnn6l\" (UID: \"e4b12149-79e9-4b2e-8991-519ad5daa14b\") " pod="openshift-etcd-operator/etcd-operator-b45778765-tnn6l" Oct 08 18:19:30 crc kubenswrapper[4859]: I1008 18:19:30.437023 4859 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/9a829106-59b1-4389-95bf-2ee4d56e317c-oauth-serving-cert\") pod \"console-f9d7485db-m2xmp\" (UID: \"9a829106-59b1-4389-95bf-2ee4d56e317c\") " pod="openshift-console/console-f9d7485db-m2xmp" Oct 08 18:19:30 crc kubenswrapper[4859]: I1008 18:19:30.437049 4859 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-4hdx8\" (UniqueName: \"kubernetes.io/projected/5401cd5b-bc2e-43d4-8d13-b7e9134aa72f-kube-api-access-4hdx8\") pod \"migrator-59844c95c7-pmngm\" (UID: \"5401cd5b-bc2e-43d4-8d13-b7e9134aa72f\") " pod="openshift-kube-storage-version-migrator/migrator-59844c95c7-pmngm" Oct 08 18:19:30 crc kubenswrapper[4859]: I1008 18:19:30.437104 4859 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-4ppdz\" (UniqueName: \"kubernetes.io/projected/70f88ce6-b9fe-422e-827e-af2b6840a783-kube-api-access-4ppdz\") pod \"apiserver-76f77b778f-fv9bp\" (UID: \"70f88ce6-b9fe-422e-827e-af2b6840a783\") " pod="openshift-apiserver/apiserver-76f77b778f-fv9bp" Oct 08 18:19:30 crc kubenswrapper[4859]: I1008 18:19:30.437157 4859 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-5kp2q\" (UniqueName: \"kubernetes.io/projected/0d04221d-922c-4ee9-9adf-817e295161d9-kube-api-access-5kp2q\") pod \"machine-config-server-7tjh2\" (UID: \"0d04221d-922c-4ee9-9adf-817e295161d9\") " pod="openshift-machine-config-operator/machine-config-server-7tjh2" Oct 08 18:19:30 crc kubenswrapper[4859]: I1008 18:19:30.437214 4859 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-kfmlz\" (UniqueName: \"kubernetes.io/projected/d80051ae-5521-4f01-b34d-34d06a661177-kube-api-access-kfmlz\") pod \"openshift-controller-manager-operator-756b6f6bc6-82pcg\" (UID: \"d80051ae-5521-4f01-b34d-34d06a661177\") " pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-82pcg" Oct 08 18:19:30 crc kubenswrapper[4859]: I1008 18:19:30.437261 4859 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/bb6dc185-94e3-4cb7-8076-1011c1c60fab-config\") pod \"kube-controller-manager-operator-78b949d7b-kc42p\" (UID: \"bb6dc185-94e3-4cb7-8076-1011c1c60fab\") " pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-kc42p" Oct 08 18:19:30 crc kubenswrapper[4859]: I1008 18:19:30.437333 4859 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/65868bd3-b636-48c5-aac4-fcf734f84e39-serving-cert\") pod \"openshift-config-operator-7777fb866f-ncj6n\" (UID: \"65868bd3-b636-48c5-aac4-fcf734f84e39\") " pod="openshift-config-operator/openshift-config-operator-7777fb866f-ncj6n" Oct 08 18:19:30 crc kubenswrapper[4859]: I1008 18:19:30.437372 4859 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/ca64d874-5d86-4c49-8953-631a75d59788-config\") pod \"route-controller-manager-6576b87f9c-xlfdd\" (UID: \"ca64d874-5d86-4c49-8953-631a75d59788\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-xlfdd" Oct 08 18:19:30 crc kubenswrapper[4859]: I1008 18:19:30.437395 4859 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etcd-ca\" (UniqueName: \"kubernetes.io/configmap/e4b12149-79e9-4b2e-8991-519ad5daa14b-etcd-ca\") pod \"etcd-operator-b45778765-tnn6l\" (UID: \"e4b12149-79e9-4b2e-8991-519ad5daa14b\") " pod="openshift-etcd-operator/etcd-operator-b45778765-tnn6l" Oct 08 18:19:30 crc kubenswrapper[4859]: I1008 18:19:30.437408 4859 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-68n5n\" (UniqueName: \"kubernetes.io/projected/0c9973e7-5c40-463b-aca1-c3c93d7a9d0b-kube-api-access-68n5n\") pod \"machine-config-operator-74547568cd-xsc6l\" (UID: \"0c9973e7-5c40-463b-aca1-c3c93d7a9d0b\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-xsc6l" Oct 08 18:19:30 crc kubenswrapper[4859]: I1008 18:19:30.437446 4859 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-9hnbb\" (UniqueName: \"kubernetes.io/projected/ca64d874-5d86-4c49-8953-631a75d59788-kube-api-access-9hnbb\") pod \"route-controller-manager-6576b87f9c-xlfdd\" (UID: \"ca64d874-5d86-4c49-8953-631a75d59788\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-xlfdd" Oct 08 18:19:30 crc kubenswrapper[4859]: I1008 18:19:30.437479 4859 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/0902ccb2-78ab-48d0-be02-4c42f03e12fb-v4-0-config-user-template-error\") pod \"oauth-openshift-558db77b4-j8gh2\" (UID: \"0902ccb2-78ab-48d0-be02-4c42f03e12fb\") " pod="openshift-authentication/oauth-openshift-558db77b4-j8gh2" Oct 08 18:19:30 crc kubenswrapper[4859]: I1008 18:19:30.437511 4859 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-dwj4n\" (UniqueName: \"kubernetes.io/projected/a52b9536-1013-482b-9261-e024377687d2-kube-api-access-dwj4n\") pod \"machine-api-operator-5694c8668f-gmbk9\" (UID: \"a52b9536-1013-482b-9261-e024377687d2\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-gmbk9" Oct 08 18:19:30 crc kubenswrapper[4859]: I1008 18:19:30.437540 4859 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/0902ccb2-78ab-48d0-be02-4c42f03e12fb-v4-0-config-system-session\") pod \"oauth-openshift-558db77b4-j8gh2\" (UID: \"0902ccb2-78ab-48d0-be02-4c42f03e12fb\") " pod="openshift-authentication/oauth-openshift-558db77b4-j8gh2" Oct 08 18:19:30 crc kubenswrapper[4859]: I1008 18:19:30.437571 4859 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/d407fd47-096d-4f2a-ace1-dab4ddf7192a-audit-dir\") pod \"apiserver-7bbb656c7d-6j5rk\" (UID: \"d407fd47-096d-4f2a-ace1-dab4ddf7192a\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-6j5rk" Oct 08 18:19:30 crc kubenswrapper[4859]: I1008 18:19:30.437602 4859 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/0902ccb2-78ab-48d0-be02-4c42f03e12fb-v4-0-config-system-service-ca\") pod \"oauth-openshift-558db77b4-j8gh2\" (UID: \"0902ccb2-78ab-48d0-be02-4c42f03e12fb\") " pod="openshift-authentication/oauth-openshift-558db77b4-j8gh2" Oct 08 18:19:30 crc kubenswrapper[4859]: I1008 18:19:30.437638 4859 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-g65gr\" (UniqueName: \"kubernetes.io/projected/9b239e22-0030-4c68-b0f1-66fbaedd5f9b-kube-api-access-g65gr\") pod \"cluster-samples-operator-665b6dd947-c5p45\" (UID: \"9b239e22-0030-4c68-b0f1-66fbaedd5f9b\") " pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-c5p45" Oct 08 18:19:30 crc kubenswrapper[4859]: I1008 18:19:30.437673 4859 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-8cxvk\" (UniqueName: \"kubernetes.io/projected/e431ddbe-2d50-4a0a-9679-5565e0df977f-kube-api-access-8cxvk\") pod \"machine-config-controller-84d6567774-m7m5n\" (UID: \"e431ddbe-2d50-4a0a-9679-5565e0df977f\") " pod="openshift-machine-config-operator/machine-config-controller-84d6567774-m7m5n" Oct 08 18:19:30 crc kubenswrapper[4859]: I1008 18:19:30.437734 4859 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/2736d658-3609-4c30-a010-7b00c6300e2e-serving-cert\") pod \"console-operator-58897d9998-8sflk\" (UID: \"2736d658-3609-4c30-a010-7b00c6300e2e\") " pod="openshift-console-operator/console-operator-58897d9998-8sflk" Oct 08 18:19:30 crc kubenswrapper[4859]: I1008 18:19:30.437767 4859 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit\" (UniqueName: \"kubernetes.io/configmap/70f88ce6-b9fe-422e-827e-af2b6840a783-audit\") pod \"apiserver-76f77b778f-fv9bp\" (UID: \"70f88ce6-b9fe-422e-827e-af2b6840a783\") " pod="openshift-apiserver/apiserver-76f77b778f-fv9bp" Oct 08 18:19:30 crc kubenswrapper[4859]: I1008 18:19:30.437803 4859 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-hf7dh\" (UniqueName: \"kubernetes.io/projected/b8d9f66c-e7a4-467f-bcf5-701cc0b68d52-kube-api-access-hf7dh\") pod \"authentication-operator-69f744f599-jd5kh\" (UID: \"b8d9f66c-e7a4-467f-bcf5-701cc0b68d52\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-jd5kh" Oct 08 18:19:30 crc kubenswrapper[4859]: I1008 18:19:30.437809 4859 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7d9d11d4-9f7b-4f33-a9d2-23db86bb2343-config\") pod \"openshift-apiserver-operator-796bbdcf4f-czj5g\" (UID: \"7d9d11d4-9f7b-4f33-a9d2-23db86bb2343\") " pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-czj5g" Oct 08 18:19:30 crc kubenswrapper[4859]: I1008 18:19:30.437835 4859 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/460ad6e7-859d-4111-a555-b55c5b7228e0-config\") pod \"controller-manager-879f6c89f-9gpzj\" (UID: \"460ad6e7-859d-4111-a555-b55c5b7228e0\") " pod="openshift-controller-manager/controller-manager-879f6c89f-9gpzj" Oct 08 18:19:30 crc kubenswrapper[4859]: I1008 18:19:30.437875 4859 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/9a829106-59b1-4389-95bf-2ee4d56e317c-console-serving-cert\") pod \"console-f9d7485db-m2xmp\" (UID: \"9a829106-59b1-4389-95bf-2ee4d56e317c\") " pod="openshift-console/console-f9d7485db-m2xmp" Oct 08 18:19:30 crc kubenswrapper[4859]: I1008 18:19:30.437907 4859 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-df4rr\" (UniqueName: \"kubernetes.io/projected/9a829106-59b1-4389-95bf-2ee4d56e317c-kube-api-access-df4rr\") pod \"console-f9d7485db-m2xmp\" (UID: \"9a829106-59b1-4389-95bf-2ee4d56e317c\") " pod="openshift-console/console-f9d7485db-m2xmp" Oct 08 18:19:30 crc kubenswrapper[4859]: I1008 18:19:30.437965 4859 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/8e2b8a53-4bc3-499b-9746-628c323d2a0c-service-ca-bundle\") pod \"router-default-5444994796-r5gs2\" (UID: \"8e2b8a53-4bc3-499b-9746-628c323d2a0c\") " pod="openshift-ingress/router-default-5444994796-r5gs2" Oct 08 18:19:30 crc kubenswrapper[4859]: I1008 18:19:30.438002 4859 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/d80051ae-5521-4f01-b34d-34d06a661177-serving-cert\") pod \"openshift-controller-manager-operator-756b6f6bc6-82pcg\" (UID: \"d80051ae-5521-4f01-b34d-34d06a661177\") " pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-82pcg" Oct 08 18:19:30 crc kubenswrapper[4859]: I1008 18:19:30.438032 4859 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/6f093915-bce6-40eb-917f-ed0058431a79-srv-cert\") pod \"olm-operator-6b444d44fb-5n49w\" (UID: \"6f093915-bce6-40eb-917f-ed0058431a79\") " pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-5n49w" Oct 08 18:19:30 crc kubenswrapper[4859]: I1008 18:19:30.438099 4859 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-w5l7z\" (UniqueName: \"kubernetes.io/projected/39b30c11-d0f6-4302-9ef8-e7fcf1b2170e-kube-api-access-w5l7z\") pod \"machine-approver-56656f9798-95s6m\" (UID: \"39b30c11-d0f6-4302-9ef8-e7fcf1b2170e\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-95s6m" Oct 08 18:19:30 crc kubenswrapper[4859]: I1008 18:19:30.438131 4859 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/70f88ce6-b9fe-422e-827e-af2b6840a783-config\") pod \"apiserver-76f77b778f-fv9bp\" (UID: \"70f88ce6-b9fe-422e-827e-af2b6840a783\") " pod="openshift-apiserver/apiserver-76f77b778f-fv9bp" Oct 08 18:19:30 crc kubenswrapper[4859]: I1008 18:19:30.438165 4859 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/70f88ce6-b9fe-422e-827e-af2b6840a783-etcd-serving-ca\") pod \"apiserver-76f77b778f-fv9bp\" (UID: \"70f88ce6-b9fe-422e-827e-af2b6840a783\") " pod="openshift-apiserver/apiserver-76f77b778f-fv9bp" Oct 08 18:19:30 crc kubenswrapper[4859]: I1008 18:19:30.438197 4859 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/2736d658-3609-4c30-a010-7b00c6300e2e-config\") pod \"console-operator-58897d9998-8sflk\" (UID: \"2736d658-3609-4c30-a010-7b00c6300e2e\") " pod="openshift-console-operator/console-operator-58897d9998-8sflk" Oct 08 18:19:30 crc kubenswrapper[4859]: I1008 18:19:30.438228 4859 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"node-bootstrap-token\" (UniqueName: \"kubernetes.io/secret/0d04221d-922c-4ee9-9adf-817e295161d9-node-bootstrap-token\") pod \"machine-config-server-7tjh2\" (UID: \"0d04221d-922c-4ee9-9adf-817e295161d9\") " pod="openshift-machine-config-operator/machine-config-server-7tjh2" Oct 08 18:19:30 crc kubenswrapper[4859]: I1008 18:19:30.438263 4859 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/e4b12149-79e9-4b2e-8991-519ad5daa14b-etcd-client\") pod \"etcd-operator-b45778765-tnn6l\" (UID: \"e4b12149-79e9-4b2e-8991-519ad5daa14b\") " pod="openshift-etcd-operator/etcd-operator-b45778765-tnn6l" Oct 08 18:19:30 crc kubenswrapper[4859]: I1008 18:19:30.438298 4859 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/bb6dc185-94e3-4cb7-8076-1011c1c60fab-kube-api-access\") pod \"kube-controller-manager-operator-78b949d7b-kc42p\" (UID: \"bb6dc185-94e3-4cb7-8076-1011c1c60fab\") " pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-kc42p" Oct 08 18:19:30 crc kubenswrapper[4859]: I1008 18:19:30.438334 4859 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/e431ddbe-2d50-4a0a-9679-5565e0df977f-proxy-tls\") pod \"machine-config-controller-84d6567774-m7m5n\" (UID: \"e431ddbe-2d50-4a0a-9679-5565e0df977f\") " pod="openshift-machine-config-operator/machine-config-controller-84d6567774-m7m5n" Oct 08 18:19:30 crc kubenswrapper[4859]: I1008 18:19:30.438369 4859 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"mcc-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/e431ddbe-2d50-4a0a-9679-5565e0df977f-mcc-auth-proxy-config\") pod \"machine-config-controller-84d6567774-m7m5n\" (UID: \"e431ddbe-2d50-4a0a-9679-5565e0df977f\") " pod="openshift-machine-config-operator/machine-config-controller-84d6567774-m7m5n" Oct 08 18:19:30 crc kubenswrapper[4859]: I1008 18:19:30.438390 4859 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/ca64d874-5d86-4c49-8953-631a75d59788-serving-cert\") pod \"route-controller-manager-6576b87f9c-xlfdd\" (UID: \"ca64d874-5d86-4c49-8953-631a75d59788\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-xlfdd" Oct 08 18:19:30 crc kubenswrapper[4859]: I1008 18:19:30.438402 4859 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/d403dae9-eb8c-4c8d-9daa-bf2a0c815652-secret-volume\") pod \"collect-profiles-29332455-w9wt8\" (UID: \"d403dae9-eb8c-4c8d-9daa-bf2a0c815652\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29332455-w9wt8" Oct 08 18:19:30 crc kubenswrapper[4859]: I1008 18:19:30.438437 4859 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/f114c686-8fde-4560-b8e1-9cd8fea448f8-webhook-certs\") pod \"multus-admission-controller-857f4d67dd-lrqkp\" (UID: \"f114c686-8fde-4560-b8e1-9cd8fea448f8\") " pod="openshift-multus/multus-admission-controller-857f4d67dd-lrqkp" Oct 08 18:19:30 crc kubenswrapper[4859]: I1008 18:19:30.438475 4859 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/d407fd47-096d-4f2a-ace1-dab4ddf7192a-serving-cert\") pod \"apiserver-7bbb656c7d-6j5rk\" (UID: \"d407fd47-096d-4f2a-ace1-dab4ddf7192a\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-6j5rk" Oct 08 18:19:30 crc kubenswrapper[4859]: I1008 18:19:30.438507 4859 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/56c0b58f-23d6-4d8e-8b75-9095128ac121-serving-cert\") pod \"service-ca-operator-777779d784-847r8\" (UID: \"56c0b58f-23d6-4d8e-8b75-9095128ac121\") " pod="openshift-service-ca-operator/service-ca-operator-777779d784-847r8" Oct 08 18:19:30 crc kubenswrapper[4859]: I1008 18:19:30.438778 4859 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-zsgtf\" (UniqueName: \"kubernetes.io/projected/f114c686-8fde-4560-b8e1-9cd8fea448f8-kube-api-access-zsgtf\") pod \"multus-admission-controller-857f4d67dd-lrqkp\" (UID: \"f114c686-8fde-4560-b8e1-9cd8fea448f8\") " pod="openshift-multus/multus-admission-controller-857f4d67dd-lrqkp" Oct 08 18:19:30 crc kubenswrapper[4859]: I1008 18:19:30.438810 4859 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/b2b5fcca-e076-4cbe-bc1e-249fcfd25e53-webhook-cert\") pod \"packageserver-d55dfcdfc-vc7bh\" (UID: \"b2b5fcca-e076-4cbe-bc1e-249fcfd25e53\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-vc7bh" Oct 08 18:19:30 crc kubenswrapper[4859]: I1008 18:19:30.438841 4859 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"tmpfs\" (UniqueName: \"kubernetes.io/empty-dir/b2b5fcca-e076-4cbe-bc1e-249fcfd25e53-tmpfs\") pod \"packageserver-d55dfcdfc-vc7bh\" (UID: \"b2b5fcca-e076-4cbe-bc1e-249fcfd25e53\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-vc7bh" Oct 08 18:19:30 crc kubenswrapper[4859]: I1008 18:19:30.438864 4859 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/ca64d874-5d86-4c49-8953-631a75d59788-config\") pod \"route-controller-manager-6576b87f9c-xlfdd\" (UID: \"ca64d874-5d86-4c49-8953-631a75d59788\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-xlfdd" Oct 08 18:19:30 crc kubenswrapper[4859]: I1008 18:19:30.438891 4859 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/9a829106-59b1-4389-95bf-2ee4d56e317c-console-oauth-config\") pod \"console-f9d7485db-m2xmp\" (UID: \"9a829106-59b1-4389-95bf-2ee4d56e317c\") " pod="openshift-console/console-f9d7485db-m2xmp" Oct 08 18:19:30 crc kubenswrapper[4859]: I1008 18:19:30.438925 4859 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-9tml9\" (UniqueName: \"kubernetes.io/projected/56c0b58f-23d6-4d8e-8b75-9095128ac121-kube-api-access-9tml9\") pod \"service-ca-operator-777779d784-847r8\" (UID: \"56c0b58f-23d6-4d8e-8b75-9095128ac121\") " pod="openshift-service-ca-operator/service-ca-operator-777779d784-847r8" Oct 08 18:19:30 crc kubenswrapper[4859]: I1008 18:19:30.439023 4859 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/70f88ce6-b9fe-422e-827e-af2b6840a783-encryption-config\") pod \"apiserver-76f77b778f-fv9bp\" (UID: \"70f88ce6-b9fe-422e-827e-af2b6840a783\") " pod="openshift-apiserver/apiserver-76f77b778f-fv9bp" Oct 08 18:19:30 crc kubenswrapper[4859]: I1008 18:19:30.439491 4859 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"image-registry-operator-tls\" (UniqueName: \"kubernetes.io/secret/c79c3857-6ecb-4e3c-b59f-6a58ba0f94e9-image-registry-operator-tls\") pod \"cluster-image-registry-operator-dc59b4c8b-kq7mx\" (UID: \"c79c3857-6ecb-4e3c-b59f-6a58ba0f94e9\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-kq7mx" Oct 08 18:19:30 crc kubenswrapper[4859]: I1008 18:19:30.439759 4859 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/d407fd47-096d-4f2a-ace1-dab4ddf7192a-audit-dir\") pod \"apiserver-7bbb656c7d-6j5rk\" (UID: \"d407fd47-096d-4f2a-ace1-dab4ddf7192a\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-6j5rk" Oct 08 18:19:30 crc kubenswrapper[4859]: I1008 18:19:30.440083 4859 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit\" (UniqueName: \"kubernetes.io/configmap/70f88ce6-b9fe-422e-827e-af2b6840a783-audit\") pod \"apiserver-76f77b778f-fv9bp\" (UID: \"70f88ce6-b9fe-422e-827e-af2b6840a783\") " pod="openshift-apiserver/apiserver-76f77b778f-fv9bp" Oct 08 18:19:30 crc kubenswrapper[4859]: I1008 18:19:30.440951 4859 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/70f88ce6-b9fe-422e-827e-af2b6840a783-config\") pod \"apiserver-76f77b778f-fv9bp\" (UID: \"70f88ce6-b9fe-422e-827e-af2b6840a783\") " pod="openshift-apiserver/apiserver-76f77b778f-fv9bp" Oct 08 18:19:30 crc kubenswrapper[4859]: I1008 18:19:30.441432 4859 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/a52b9536-1013-482b-9261-e024377687d2-config\") pod \"machine-api-operator-5694c8668f-gmbk9\" (UID: \"a52b9536-1013-482b-9261-e024377687d2\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-gmbk9" Oct 08 18:19:30 crc kubenswrapper[4859]: I1008 18:19:30.441554 4859 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/70f88ce6-b9fe-422e-827e-af2b6840a783-etcd-serving-ca\") pod \"apiserver-76f77b778f-fv9bp\" (UID: \"70f88ce6-b9fe-422e-827e-af2b6840a783\") " pod="openshift-apiserver/apiserver-76f77b778f-fv9bp" Oct 08 18:19:30 crc kubenswrapper[4859]: I1008 18:19:30.442028 4859 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/65868bd3-b636-48c5-aac4-fcf734f84e39-serving-cert\") pod \"openshift-config-operator-7777fb866f-ncj6n\" (UID: \"65868bd3-b636-48c5-aac4-fcf734f84e39\") " pod="openshift-config-operator/openshift-config-operator-7777fb866f-ncj6n" Oct 08 18:19:30 crc kubenswrapper[4859]: I1008 18:19:30.442250 4859 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/2736d658-3609-4c30-a010-7b00c6300e2e-config\") pod \"console-operator-58897d9998-8sflk\" (UID: \"2736d658-3609-4c30-a010-7b00c6300e2e\") " pod="openshift-console-operator/console-operator-58897d9998-8sflk" Oct 08 18:19:30 crc kubenswrapper[4859]: I1008 18:19:30.442440 4859 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/0902ccb2-78ab-48d0-be02-4c42f03e12fb-v4-0-config-user-template-error\") pod \"oauth-openshift-558db77b4-j8gh2\" (UID: \"0902ccb2-78ab-48d0-be02-4c42f03e12fb\") " pod="openshift-authentication/oauth-openshift-558db77b4-j8gh2" Oct 08 18:19:30 crc kubenswrapper[4859]: I1008 18:19:30.442885 4859 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/b8d9f66c-e7a4-467f-bcf5-701cc0b68d52-serving-cert\") pod \"authentication-operator-69f744f599-jd5kh\" (UID: \"b8d9f66c-e7a4-467f-bcf5-701cc0b68d52\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-jd5kh" Oct 08 18:19:30 crc kubenswrapper[4859]: I1008 18:19:30.443036 4859 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/460ad6e7-859d-4111-a555-b55c5b7228e0-config\") pod \"controller-manager-879f6c89f-9gpzj\" (UID: \"460ad6e7-859d-4111-a555-b55c5b7228e0\") " pod="openshift-controller-manager/controller-manager-879f6c89f-9gpzj" Oct 08 18:19:30 crc kubenswrapper[4859]: I1008 18:19:30.443210 4859 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/7d9d11d4-9f7b-4f33-a9d2-23db86bb2343-serving-cert\") pod \"openshift-apiserver-operator-796bbdcf4f-czj5g\" (UID: \"7d9d11d4-9f7b-4f33-a9d2-23db86bb2343\") " pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-czj5g" Oct 08 18:19:30 crc kubenswrapper[4859]: I1008 18:19:30.443323 4859 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/2736d658-3609-4c30-a010-7b00c6300e2e-serving-cert\") pod \"console-operator-58897d9998-8sflk\" (UID: \"2736d658-3609-4c30-a010-7b00c6300e2e\") " pod="openshift-console-operator/console-operator-58897d9998-8sflk" Oct 08 18:19:30 crc kubenswrapper[4859]: I1008 18:19:30.443886 4859 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/9a829106-59b1-4389-95bf-2ee4d56e317c-console-oauth-config\") pod \"console-f9d7485db-m2xmp\" (UID: \"9a829106-59b1-4389-95bf-2ee4d56e317c\") " pod="openshift-console/console-f9d7485db-m2xmp" Oct 08 18:19:30 crc kubenswrapper[4859]: I1008 18:19:30.444390 4859 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"machine-api-operator-tls\" (UniqueName: \"kubernetes.io/secret/a52b9536-1013-482b-9261-e024377687d2-machine-api-operator-tls\") pod \"machine-api-operator-5694c8668f-gmbk9\" (UID: \"a52b9536-1013-482b-9261-e024377687d2\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-gmbk9" Oct 08 18:19:30 crc kubenswrapper[4859]: I1008 18:19:30.444761 4859 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/e4b12149-79e9-4b2e-8991-519ad5daa14b-etcd-client\") pod \"etcd-operator-b45778765-tnn6l\" (UID: \"e4b12149-79e9-4b2e-8991-519ad5daa14b\") " pod="openshift-etcd-operator/etcd-operator-b45778765-tnn6l" Oct 08 18:19:30 crc kubenswrapper[4859]: I1008 18:19:30.445113 4859 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/0902ccb2-78ab-48d0-be02-4c42f03e12fb-v4-0-config-system-serving-cert\") pod \"oauth-openshift-558db77b4-j8gh2\" (UID: \"0902ccb2-78ab-48d0-be02-4c42f03e12fb\") " pod="openshift-authentication/oauth-openshift-558db77b4-j8gh2" Oct 08 18:19:30 crc kubenswrapper[4859]: I1008 18:19:30.445130 4859 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/460ad6e7-859d-4111-a555-b55c5b7228e0-serving-cert\") pod \"controller-manager-879f6c89f-9gpzj\" (UID: \"460ad6e7-859d-4111-a555-b55c5b7228e0\") " pod="openshift-controller-manager/controller-manager-879f6c89f-9gpzj" Oct 08 18:19:30 crc kubenswrapper[4859]: I1008 18:19:30.445135 4859 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/0902ccb2-78ab-48d0-be02-4c42f03e12fb-v4-0-config-system-router-certs\") pod \"oauth-openshift-558db77b4-j8gh2\" (UID: \"0902ccb2-78ab-48d0-be02-4c42f03e12fb\") " pod="openshift-authentication/oauth-openshift-558db77b4-j8gh2" Oct 08 18:19:30 crc kubenswrapper[4859]: I1008 18:19:30.445115 4859 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/9a829106-59b1-4389-95bf-2ee4d56e317c-console-serving-cert\") pod \"console-f9d7485db-m2xmp\" (UID: \"9a829106-59b1-4389-95bf-2ee4d56e317c\") " pod="openshift-console/console-f9d7485db-m2xmp" Oct 08 18:19:30 crc kubenswrapper[4859]: I1008 18:19:30.445180 4859 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/70f88ce6-b9fe-422e-827e-af2b6840a783-serving-cert\") pod \"apiserver-76f77b778f-fv9bp\" (UID: \"70f88ce6-b9fe-422e-827e-af2b6840a783\") " pod="openshift-apiserver/apiserver-76f77b778f-fv9bp" Oct 08 18:19:30 crc kubenswrapper[4859]: I1008 18:19:30.445560 4859 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/0902ccb2-78ab-48d0-be02-4c42f03e12fb-v4-0-config-user-template-login\") pod \"oauth-openshift-558db77b4-j8gh2\" (UID: \"0902ccb2-78ab-48d0-be02-4c42f03e12fb\") " pod="openshift-authentication/oauth-openshift-558db77b4-j8gh2" Oct 08 18:19:30 crc kubenswrapper[4859]: I1008 18:19:30.445918 4859 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/e4b12149-79e9-4b2e-8991-519ad5daa14b-serving-cert\") pod \"etcd-operator-b45778765-tnn6l\" (UID: \"e4b12149-79e9-4b2e-8991-519ad5daa14b\") " pod="openshift-etcd-operator/etcd-operator-b45778765-tnn6l" Oct 08 18:19:30 crc kubenswrapper[4859]: I1008 18:19:30.445733 4859 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/0902ccb2-78ab-48d0-be02-4c42f03e12fb-v4-0-config-system-ocp-branding-template\") pod \"oauth-openshift-558db77b4-j8gh2\" (UID: \"0902ccb2-78ab-48d0-be02-4c42f03e12fb\") " pod="openshift-authentication/oauth-openshift-558db77b4-j8gh2" Oct 08 18:19:30 crc kubenswrapper[4859]: I1008 18:19:30.446569 4859 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/d80051ae-5521-4f01-b34d-34d06a661177-serving-cert\") pod \"openshift-controller-manager-operator-756b6f6bc6-82pcg\" (UID: \"d80051ae-5521-4f01-b34d-34d06a661177\") " pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-82pcg" Oct 08 18:19:30 crc kubenswrapper[4859]: I1008 18:19:30.446766 4859 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/0902ccb2-78ab-48d0-be02-4c42f03e12fb-v4-0-config-system-service-ca\") pod \"oauth-openshift-558db77b4-j8gh2\" (UID: \"0902ccb2-78ab-48d0-be02-4c42f03e12fb\") " pod="openshift-authentication/oauth-openshift-558db77b4-j8gh2" Oct 08 18:19:30 crc kubenswrapper[4859]: I1008 18:19:30.451321 4859 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-controller-manager-operator"/"kube-controller-manager-operator-serving-cert" Oct 08 18:19:30 crc kubenswrapper[4859]: I1008 18:19:30.462723 4859 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/bb6dc185-94e3-4cb7-8076-1011c1c60fab-serving-cert\") pod \"kube-controller-manager-operator-78b949d7b-kc42p\" (UID: \"bb6dc185-94e3-4cb7-8076-1011c1c60fab\") " pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-kc42p" Oct 08 18:19:30 crc kubenswrapper[4859]: I1008 18:19:30.476065 4859 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-controller-manager-operator"/"kube-controller-manager-operator-dockercfg-gkqpw" Oct 08 18:19:30 crc kubenswrapper[4859]: I1008 18:19:30.490603 4859 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-controller-manager-operator"/"kube-controller-manager-operator-config" Oct 08 18:19:30 crc kubenswrapper[4859]: I1008 18:19:30.498481 4859 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/bb6dc185-94e3-4cb7-8076-1011c1c60fab-config\") pod \"kube-controller-manager-operator-78b949d7b-kc42p\" (UID: \"bb6dc185-94e3-4cb7-8076-1011c1c60fab\") " pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-kc42p" Oct 08 18:19:30 crc kubenswrapper[4859]: I1008 18:19:30.510489 4859 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-controller-manager-operator"/"kube-root-ca.crt" Oct 08 18:19:30 crc kubenswrapper[4859]: I1008 18:19:30.531625 4859 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"image-registry-tls" Oct 08 18:19:30 crc kubenswrapper[4859]: I1008 18:19:30.539703 4859 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"mcc-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/e431ddbe-2d50-4a0a-9679-5565e0df977f-mcc-auth-proxy-config\") pod \"machine-config-controller-84d6567774-m7m5n\" (UID: \"e431ddbe-2d50-4a0a-9679-5565e0df977f\") " pod="openshift-machine-config-operator/machine-config-controller-84d6567774-m7m5n" Oct 08 18:19:30 crc kubenswrapper[4859]: I1008 18:19:30.539751 4859 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/d403dae9-eb8c-4c8d-9daa-bf2a0c815652-secret-volume\") pod \"collect-profiles-29332455-w9wt8\" (UID: \"d403dae9-eb8c-4c8d-9daa-bf2a0c815652\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29332455-w9wt8" Oct 08 18:19:30 crc kubenswrapper[4859]: I1008 18:19:30.539789 4859 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/e431ddbe-2d50-4a0a-9679-5565e0df977f-proxy-tls\") pod \"machine-config-controller-84d6567774-m7m5n\" (UID: \"e431ddbe-2d50-4a0a-9679-5565e0df977f\") " pod="openshift-machine-config-operator/machine-config-controller-84d6567774-m7m5n" Oct 08 18:19:30 crc kubenswrapper[4859]: I1008 18:19:30.539811 4859 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/f114c686-8fde-4560-b8e1-9cd8fea448f8-webhook-certs\") pod \"multus-admission-controller-857f4d67dd-lrqkp\" (UID: \"f114c686-8fde-4560-b8e1-9cd8fea448f8\") " pod="openshift-multus/multus-admission-controller-857f4d67dd-lrqkp" Oct 08 18:19:30 crc kubenswrapper[4859]: I1008 18:19:30.539842 4859 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/56c0b58f-23d6-4d8e-8b75-9095128ac121-serving-cert\") pod \"service-ca-operator-777779d784-847r8\" (UID: \"56c0b58f-23d6-4d8e-8b75-9095128ac121\") " pod="openshift-service-ca-operator/service-ca-operator-777779d784-847r8" Oct 08 18:19:30 crc kubenswrapper[4859]: I1008 18:19:30.539860 4859 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/b2b5fcca-e076-4cbe-bc1e-249fcfd25e53-webhook-cert\") pod \"packageserver-d55dfcdfc-vc7bh\" (UID: \"b2b5fcca-e076-4cbe-bc1e-249fcfd25e53\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-vc7bh" Oct 08 18:19:30 crc kubenswrapper[4859]: I1008 18:19:30.539884 4859 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-zsgtf\" (UniqueName: \"kubernetes.io/projected/f114c686-8fde-4560-b8e1-9cd8fea448f8-kube-api-access-zsgtf\") pod \"multus-admission-controller-857f4d67dd-lrqkp\" (UID: \"f114c686-8fde-4560-b8e1-9cd8fea448f8\") " pod="openshift-multus/multus-admission-controller-857f4d67dd-lrqkp" Oct 08 18:19:30 crc kubenswrapper[4859]: I1008 18:19:30.539912 4859 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-9tml9\" (UniqueName: \"kubernetes.io/projected/56c0b58f-23d6-4d8e-8b75-9095128ac121-kube-api-access-9tml9\") pod \"service-ca-operator-777779d784-847r8\" (UID: \"56c0b58f-23d6-4d8e-8b75-9095128ac121\") " pod="openshift-service-ca-operator/service-ca-operator-777779d784-847r8" Oct 08 18:19:30 crc kubenswrapper[4859]: I1008 18:19:30.539937 4859 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"tmpfs\" (UniqueName: \"kubernetes.io/empty-dir/b2b5fcca-e076-4cbe-bc1e-249fcfd25e53-tmpfs\") pod \"packageserver-d55dfcdfc-vc7bh\" (UID: \"b2b5fcca-e076-4cbe-bc1e-249fcfd25e53\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-vc7bh" Oct 08 18:19:30 crc kubenswrapper[4859]: I1008 18:19:30.539994 4859 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/8d4b5b44-816f-44f6-98ab-c5d66aefaa08-marketplace-operator-metrics\") pod \"marketplace-operator-79b997595-bz6v8\" (UID: \"8d4b5b44-816f-44f6-98ab-c5d66aefaa08\") " pod="openshift-marketplace/marketplace-operator-79b997595-bz6v8" Oct 08 18:19:30 crc kubenswrapper[4859]: I1008 18:19:30.540030 4859 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/b2b5fcca-e076-4cbe-bc1e-249fcfd25e53-apiservice-cert\") pod \"packageserver-d55dfcdfc-vc7bh\" (UID: \"b2b5fcca-e076-4cbe-bc1e-249fcfd25e53\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-vc7bh" Oct 08 18:19:30 crc kubenswrapper[4859]: I1008 18:19:30.540078 4859 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/01517214-a376-421e-9938-e548d76695d5-config\") pod \"kube-apiserver-operator-766d6c64bb-4s5w9\" (UID: \"01517214-a376-421e-9938-e548d76695d5\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-4s5w9" Oct 08 18:19:30 crc kubenswrapper[4859]: I1008 18:19:30.540420 4859 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/01517214-a376-421e-9938-e548d76695d5-serving-cert\") pod \"kube-apiserver-operator-766d6c64bb-4s5w9\" (UID: \"01517214-a376-421e-9938-e548d76695d5\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-4s5w9" Oct 08 18:19:30 crc kubenswrapper[4859]: I1008 18:19:30.541070 4859 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"tmpfs\" (UniqueName: \"kubernetes.io/empty-dir/b2b5fcca-e076-4cbe-bc1e-249fcfd25e53-tmpfs\") pod \"packageserver-d55dfcdfc-vc7bh\" (UID: \"b2b5fcca-e076-4cbe-bc1e-249fcfd25e53\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-vc7bh" Oct 08 18:19:30 crc kubenswrapper[4859]: I1008 18:19:30.541395 4859 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"mcc-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/e431ddbe-2d50-4a0a-9679-5565e0df977f-mcc-auth-proxy-config\") pod \"machine-config-controller-84d6567774-m7m5n\" (UID: \"e431ddbe-2d50-4a0a-9679-5565e0df977f\") " pod="openshift-machine-config-operator/machine-config-controller-84d6567774-m7m5n" Oct 08 18:19:30 crc kubenswrapper[4859]: I1008 18:19:30.541476 4859 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/0c9973e7-5c40-463b-aca1-c3c93d7a9d0b-auth-proxy-config\") pod \"machine-config-operator-74547568cd-xsc6l\" (UID: \"0c9973e7-5c40-463b-aca1-c3c93d7a9d0b\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-xsc6l" Oct 08 18:19:30 crc kubenswrapper[4859]: I1008 18:19:30.541518 4859 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/8d4b5b44-816f-44f6-98ab-c5d66aefaa08-marketplace-trusted-ca\") pod \"marketplace-operator-79b997595-bz6v8\" (UID: \"8d4b5b44-816f-44f6-98ab-c5d66aefaa08\") " pod="openshift-marketplace/marketplace-operator-79b997595-bz6v8" Oct 08 18:19:30 crc kubenswrapper[4859]: I1008 18:19:30.541566 4859 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/6f093915-bce6-40eb-917f-ed0058431a79-profile-collector-cert\") pod \"olm-operator-6b444d44fb-5n49w\" (UID: \"6f093915-bce6-40eb-917f-ed0058431a79\") " pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-5n49w" Oct 08 18:19:30 crc kubenswrapper[4859]: I1008 18:19:30.541598 4859 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/01517214-a376-421e-9938-e548d76695d5-kube-api-access\") pod \"kube-apiserver-operator-766d6c64bb-4s5w9\" (UID: \"01517214-a376-421e-9938-e548d76695d5\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-4s5w9" Oct 08 18:19:30 crc kubenswrapper[4859]: I1008 18:19:30.541634 4859 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-jb26g\" (UniqueName: \"kubernetes.io/projected/6f093915-bce6-40eb-917f-ed0058431a79-kube-api-access-jb26g\") pod \"olm-operator-6b444d44fb-5n49w\" (UID: \"6f093915-bce6-40eb-917f-ed0058431a79\") " pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-5n49w" Oct 08 18:19:30 crc kubenswrapper[4859]: I1008 18:19:30.541751 4859 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-v4vwn\" (UniqueName: \"kubernetes.io/projected/d403dae9-eb8c-4c8d-9daa-bf2a0c815652-kube-api-access-v4vwn\") pod \"collect-profiles-29332455-w9wt8\" (UID: \"d403dae9-eb8c-4c8d-9daa-bf2a0c815652\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29332455-w9wt8" Oct 08 18:19:30 crc kubenswrapper[4859]: I1008 18:19:30.541783 4859 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/d403dae9-eb8c-4c8d-9daa-bf2a0c815652-config-volume\") pod \"collect-profiles-29332455-w9wt8\" (UID: \"d403dae9-eb8c-4c8d-9daa-bf2a0c815652\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29332455-w9wt8" Oct 08 18:19:30 crc kubenswrapper[4859]: I1008 18:19:30.542400 4859 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/0c9973e7-5c40-463b-aca1-c3c93d7a9d0b-auth-proxy-config\") pod \"machine-config-operator-74547568cd-xsc6l\" (UID: \"0c9973e7-5c40-463b-aca1-c3c93d7a9d0b\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-xsc6l" Oct 08 18:19:30 crc kubenswrapper[4859]: I1008 18:19:30.542480 4859 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"signing-key\" (UniqueName: \"kubernetes.io/secret/da9f04a1-bb6a-40de-846a-5e2fe04855d1-signing-key\") pod \"service-ca-9c57cc56f-cs77r\" (UID: \"da9f04a1-bb6a-40de-846a-5e2fe04855d1\") " pod="openshift-service-ca/service-ca-9c57cc56f-cs77r" Oct 08 18:19:30 crc kubenswrapper[4859]: I1008 18:19:30.542535 4859 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"images\" (UniqueName: \"kubernetes.io/configmap/0c9973e7-5c40-463b-aca1-c3c93d7a9d0b-images\") pod \"machine-config-operator-74547568cd-xsc6l\" (UID: \"0c9973e7-5c40-463b-aca1-c3c93d7a9d0b\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-xsc6l" Oct 08 18:19:30 crc kubenswrapper[4859]: I1008 18:19:30.542569 4859 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/8e2b8a53-4bc3-499b-9746-628c323d2a0c-metrics-certs\") pod \"router-default-5444994796-r5gs2\" (UID: \"8e2b8a53-4bc3-499b-9746-628c323d2a0c\") " pod="openshift-ingress/router-default-5444994796-r5gs2" Oct 08 18:19:30 crc kubenswrapper[4859]: I1008 18:19:30.542641 4859 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-vxgz5\" (UniqueName: \"kubernetes.io/projected/8e2b8a53-4bc3-499b-9746-628c323d2a0c-kube-api-access-vxgz5\") pod \"router-default-5444994796-r5gs2\" (UID: \"8e2b8a53-4bc3-499b-9746-628c323d2a0c\") " pod="openshift-ingress/router-default-5444994796-r5gs2" Oct 08 18:19:30 crc kubenswrapper[4859]: I1008 18:19:30.542668 4859 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-hv4wh\" (UniqueName: \"kubernetes.io/projected/8d4b5b44-816f-44f6-98ab-c5d66aefaa08-kube-api-access-hv4wh\") pod \"marketplace-operator-79b997595-bz6v8\" (UID: \"8d4b5b44-816f-44f6-98ab-c5d66aefaa08\") " pod="openshift-marketplace/marketplace-operator-79b997595-bz6v8" Oct 08 18:19:30 crc kubenswrapper[4859]: I1008 18:19:30.544005 4859 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-glqwn\" (UniqueName: \"kubernetes.io/projected/da9f04a1-bb6a-40de-846a-5e2fe04855d1-kube-api-access-glqwn\") pod \"service-ca-9c57cc56f-cs77r\" (UID: \"da9f04a1-bb6a-40de-846a-5e2fe04855d1\") " pod="openshift-service-ca/service-ca-9c57cc56f-cs77r" Oct 08 18:19:30 crc kubenswrapper[4859]: I1008 18:19:30.544201 4859 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/56c0b58f-23d6-4d8e-8b75-9095128ac121-config\") pod \"service-ca-operator-777779d784-847r8\" (UID: \"56c0b58f-23d6-4d8e-8b75-9095128ac121\") " pod="openshift-service-ca-operator/service-ca-operator-777779d784-847r8" Oct 08 18:19:30 crc kubenswrapper[4859]: I1008 18:19:30.544233 4859 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/0c9973e7-5c40-463b-aca1-c3c93d7a9d0b-proxy-tls\") pod \"machine-config-operator-74547568cd-xsc6l\" (UID: \"0c9973e7-5c40-463b-aca1-c3c93d7a9d0b\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-xsc6l" Oct 08 18:19:30 crc kubenswrapper[4859]: I1008 18:19:30.544369 4859 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"signing-cabundle\" (UniqueName: \"kubernetes.io/configmap/da9f04a1-bb6a-40de-846a-5e2fe04855d1-signing-cabundle\") pod \"service-ca-9c57cc56f-cs77r\" (UID: \"da9f04a1-bb6a-40de-846a-5e2fe04855d1\") " pod="openshift-service-ca/service-ca-9c57cc56f-cs77r" Oct 08 18:19:30 crc kubenswrapper[4859]: I1008 18:19:30.544397 4859 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"certs\" (UniqueName: \"kubernetes.io/secret/0d04221d-922c-4ee9-9adf-817e295161d9-certs\") pod \"machine-config-server-7tjh2\" (UID: \"0d04221d-922c-4ee9-9adf-817e295161d9\") " pod="openshift-machine-config-operator/machine-config-server-7tjh2" Oct 08 18:19:30 crc kubenswrapper[4859]: I1008 18:19:30.544518 4859 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"default-certificate\" (UniqueName: \"kubernetes.io/secret/8e2b8a53-4bc3-499b-9746-628c323d2a0c-default-certificate\") pod \"router-default-5444994796-r5gs2\" (UID: \"8e2b8a53-4bc3-499b-9746-628c323d2a0c\") " pod="openshift-ingress/router-default-5444994796-r5gs2" Oct 08 18:19:30 crc kubenswrapper[4859]: I1008 18:19:30.544756 4859 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-jbvx4\" (UniqueName: \"kubernetes.io/projected/b2b5fcca-e076-4cbe-bc1e-249fcfd25e53-kube-api-access-jbvx4\") pod \"packageserver-d55dfcdfc-vc7bh\" (UID: \"b2b5fcca-e076-4cbe-bc1e-249fcfd25e53\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-vc7bh" Oct 08 18:19:30 crc kubenswrapper[4859]: I1008 18:19:30.544780 4859 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"stats-auth\" (UniqueName: \"kubernetes.io/secret/8e2b8a53-4bc3-499b-9746-628c323d2a0c-stats-auth\") pod \"router-default-5444994796-r5gs2\" (UID: \"8e2b8a53-4bc3-499b-9746-628c323d2a0c\") " pod="openshift-ingress/router-default-5444994796-r5gs2" Oct 08 18:19:30 crc kubenswrapper[4859]: I1008 18:19:30.544845 4859 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-4hdx8\" (UniqueName: \"kubernetes.io/projected/5401cd5b-bc2e-43d4-8d13-b7e9134aa72f-kube-api-access-4hdx8\") pod \"migrator-59844c95c7-pmngm\" (UID: \"5401cd5b-bc2e-43d4-8d13-b7e9134aa72f\") " pod="openshift-kube-storage-version-migrator/migrator-59844c95c7-pmngm" Oct 08 18:19:30 crc kubenswrapper[4859]: I1008 18:19:30.544879 4859 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-5kp2q\" (UniqueName: \"kubernetes.io/projected/0d04221d-922c-4ee9-9adf-817e295161d9-kube-api-access-5kp2q\") pod \"machine-config-server-7tjh2\" (UID: \"0d04221d-922c-4ee9-9adf-817e295161d9\") " pod="openshift-machine-config-operator/machine-config-server-7tjh2" Oct 08 18:19:30 crc kubenswrapper[4859]: I1008 18:19:30.544919 4859 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-68n5n\" (UniqueName: \"kubernetes.io/projected/0c9973e7-5c40-463b-aca1-c3c93d7a9d0b-kube-api-access-68n5n\") pod \"machine-config-operator-74547568cd-xsc6l\" (UID: \"0c9973e7-5c40-463b-aca1-c3c93d7a9d0b\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-xsc6l" Oct 08 18:19:30 crc kubenswrapper[4859]: I1008 18:19:30.544986 4859 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-8cxvk\" (UniqueName: \"kubernetes.io/projected/e431ddbe-2d50-4a0a-9679-5565e0df977f-kube-api-access-8cxvk\") pod \"machine-config-controller-84d6567774-m7m5n\" (UID: \"e431ddbe-2d50-4a0a-9679-5565e0df977f\") " pod="openshift-machine-config-operator/machine-config-controller-84d6567774-m7m5n" Oct 08 18:19:30 crc kubenswrapper[4859]: I1008 18:19:30.545061 4859 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/8e2b8a53-4bc3-499b-9746-628c323d2a0c-service-ca-bundle\") pod \"router-default-5444994796-r5gs2\" (UID: \"8e2b8a53-4bc3-499b-9746-628c323d2a0c\") " pod="openshift-ingress/router-default-5444994796-r5gs2" Oct 08 18:19:30 crc kubenswrapper[4859]: I1008 18:19:30.545098 4859 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/6f093915-bce6-40eb-917f-ed0058431a79-srv-cert\") pod \"olm-operator-6b444d44fb-5n49w\" (UID: \"6f093915-bce6-40eb-917f-ed0058431a79\") " pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-5n49w" Oct 08 18:19:30 crc kubenswrapper[4859]: I1008 18:19:30.545161 4859 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"node-bootstrap-token\" (UniqueName: \"kubernetes.io/secret/0d04221d-922c-4ee9-9adf-817e295161d9-node-bootstrap-token\") pod \"machine-config-server-7tjh2\" (UID: \"0d04221d-922c-4ee9-9adf-817e295161d9\") " pod="openshift-machine-config-operator/machine-config-server-7tjh2" Oct 08 18:19:30 crc kubenswrapper[4859]: I1008 18:19:30.550983 4859 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"installation-pull-secrets" Oct 08 18:19:30 crc kubenswrapper[4859]: I1008 18:19:30.574312 4859 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"registry-dockercfg-kzzsd" Oct 08 18:19:30 crc kubenswrapper[4859]: I1008 18:19:30.591304 4859 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-dns-operator"/"openshift-service-ca.crt" Oct 08 18:19:30 crc kubenswrapper[4859]: I1008 18:19:30.613864 4859 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-dns-operator"/"kube-root-ca.crt" Oct 08 18:19:30 crc kubenswrapper[4859]: I1008 18:19:30.630542 4859 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-dns-operator"/"dns-operator-dockercfg-9mqw5" Oct 08 18:19:30 crc kubenswrapper[4859]: I1008 18:19:30.650599 4859 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-dns-operator"/"metrics-tls" Oct 08 18:19:30 crc kubenswrapper[4859]: I1008 18:19:30.671808 4859 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress-operator"/"metrics-tls" Oct 08 18:19:30 crc kubenswrapper[4859]: I1008 18:19:30.690012 4859 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress-operator"/"ingress-operator-dockercfg-7lnqk" Oct 08 18:19:30 crc kubenswrapper[4859]: I1008 18:19:30.716331 4859 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress-operator"/"trusted-ca" Oct 08 18:19:30 crc kubenswrapper[4859]: I1008 18:19:30.730003 4859 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress-operator"/"kube-root-ca.crt" Oct 08 18:19:30 crc kubenswrapper[4859]: I1008 18:19:30.750002 4859 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress-operator"/"openshift-service-ca.crt" Oct 08 18:19:30 crc kubenswrapper[4859]: I1008 18:19:30.770929 4859 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"machine-config-operator-dockercfg-98p87" Oct 08 18:19:30 crc kubenswrapper[4859]: I1008 18:19:30.790447 4859 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-config-operator"/"machine-config-operator-images" Oct 08 18:19:30 crc kubenswrapper[4859]: I1008 18:19:30.794169 4859 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"images\" (UniqueName: \"kubernetes.io/configmap/0c9973e7-5c40-463b-aca1-c3c93d7a9d0b-images\") pod \"machine-config-operator-74547568cd-xsc6l\" (UID: \"0c9973e7-5c40-463b-aca1-c3c93d7a9d0b\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-xsc6l" Oct 08 18:19:30 crc kubenswrapper[4859]: I1008 18:19:30.810592 4859 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"mco-proxy-tls" Oct 08 18:19:30 crc kubenswrapper[4859]: I1008 18:19:30.819488 4859 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/0c9973e7-5c40-463b-aca1-c3c93d7a9d0b-proxy-tls\") pod \"machine-config-operator-74547568cd-xsc6l\" (UID: \"0c9973e7-5c40-463b-aca1-c3c93d7a9d0b\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-xsc6l" Oct 08 18:19:30 crc kubenswrapper[4859]: I1008 18:19:30.832117 4859 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-scheduler-operator"/"kube-root-ca.crt" Oct 08 18:19:30 crc kubenswrapper[4859]: I1008 18:19:30.851125 4859 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-scheduler-operator"/"openshift-kube-scheduler-operator-dockercfg-qt55r" Oct 08 18:19:30 crc kubenswrapper[4859]: I1008 18:19:30.871724 4859 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-scheduler-operator"/"kube-scheduler-operator-serving-cert" Oct 08 18:19:30 crc kubenswrapper[4859]: I1008 18:19:30.891126 4859 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-scheduler-operator"/"openshift-kube-scheduler-operator-config" Oct 08 18:19:30 crc kubenswrapper[4859]: I1008 18:19:30.911096 4859 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-apiserver-operator"/"kube-root-ca.crt" Oct 08 18:19:30 crc kubenswrapper[4859]: I1008 18:19:30.932235 4859 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-apiserver-operator"/"kube-apiserver-operator-dockercfg-x57mr" Oct 08 18:19:30 crc kubenswrapper[4859]: I1008 18:19:30.952294 4859 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-apiserver-operator"/"kube-apiserver-operator-serving-cert" Oct 08 18:19:30 crc kubenswrapper[4859]: I1008 18:19:30.965400 4859 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/01517214-a376-421e-9938-e548d76695d5-serving-cert\") pod \"kube-apiserver-operator-766d6c64bb-4s5w9\" (UID: \"01517214-a376-421e-9938-e548d76695d5\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-4s5w9" Oct 08 18:19:30 crc kubenswrapper[4859]: I1008 18:19:30.970753 4859 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-apiserver-operator"/"kube-apiserver-operator-config" Oct 08 18:19:30 crc kubenswrapper[4859]: I1008 18:19:30.971862 4859 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/01517214-a376-421e-9938-e548d76695d5-config\") pod \"kube-apiserver-operator-766d6c64bb-4s5w9\" (UID: \"01517214-a376-421e-9938-e548d76695d5\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-4s5w9" Oct 08 18:19:30 crc kubenswrapper[4859]: I1008 18:19:30.991289 4859 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"mcc-proxy-tls" Oct 08 18:19:31 crc kubenswrapper[4859]: I1008 18:19:31.004953 4859 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/e431ddbe-2d50-4a0a-9679-5565e0df977f-proxy-tls\") pod \"machine-config-controller-84d6567774-m7m5n\" (UID: \"e431ddbe-2d50-4a0a-9679-5565e0df977f\") " pod="openshift-machine-config-operator/machine-config-controller-84d6567774-m7m5n" Oct 08 18:19:31 crc kubenswrapper[4859]: I1008 18:19:31.010241 4859 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress"/"openshift-service-ca.crt" Oct 08 18:19:31 crc kubenswrapper[4859]: I1008 18:19:31.032951 4859 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"machine-config-controller-dockercfg-c2lfx" Oct 08 18:19:31 crc kubenswrapper[4859]: I1008 18:19:31.050670 4859 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress"/"router-dockercfg-zdk86" Oct 08 18:19:31 crc kubenswrapper[4859]: I1008 18:19:31.071207 4859 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress"/"router-certs-default" Oct 08 18:19:31 crc kubenswrapper[4859]: I1008 18:19:31.078542 4859 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"default-certificate\" (UniqueName: \"kubernetes.io/secret/8e2b8a53-4bc3-499b-9746-628c323d2a0c-default-certificate\") pod \"router-default-5444994796-r5gs2\" (UID: \"8e2b8a53-4bc3-499b-9746-628c323d2a0c\") " pod="openshift-ingress/router-default-5444994796-r5gs2" Oct 08 18:19:31 crc kubenswrapper[4859]: I1008 18:19:31.090542 4859 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress"/"router-stats-default" Oct 08 18:19:31 crc kubenswrapper[4859]: I1008 18:19:31.098411 4859 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"stats-auth\" (UniqueName: \"kubernetes.io/secret/8e2b8a53-4bc3-499b-9746-628c323d2a0c-stats-auth\") pod \"router-default-5444994796-r5gs2\" (UID: \"8e2b8a53-4bc3-499b-9746-628c323d2a0c\") " pod="openshift-ingress/router-default-5444994796-r5gs2" Oct 08 18:19:31 crc kubenswrapper[4859]: I1008 18:19:31.111318 4859 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress"/"router-metrics-certs-default" Oct 08 18:19:31 crc kubenswrapper[4859]: I1008 18:19:31.116378 4859 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/8e2b8a53-4bc3-499b-9746-628c323d2a0c-metrics-certs\") pod \"router-default-5444994796-r5gs2\" (UID: \"8e2b8a53-4bc3-499b-9746-628c323d2a0c\") " pod="openshift-ingress/router-default-5444994796-r5gs2" Oct 08 18:19:31 crc kubenswrapper[4859]: I1008 18:19:31.132031 4859 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress"/"service-ca-bundle" Oct 08 18:19:31 crc kubenswrapper[4859]: I1008 18:19:31.137393 4859 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/8e2b8a53-4bc3-499b-9746-628c323d2a0c-service-ca-bundle\") pod \"router-default-5444994796-r5gs2\" (UID: \"8e2b8a53-4bc3-499b-9746-628c323d2a0c\") " pod="openshift-ingress/router-default-5444994796-r5gs2" Oct 08 18:19:31 crc kubenswrapper[4859]: I1008 18:19:31.151450 4859 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress"/"kube-root-ca.crt" Oct 08 18:19:31 crc kubenswrapper[4859]: I1008 18:19:31.171813 4859 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-storage-version-migrator"/"kube-root-ca.crt" Oct 08 18:19:31 crc kubenswrapper[4859]: I1008 18:19:31.191995 4859 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-storage-version-migrator"/"kube-storage-version-migrator-sa-dockercfg-5xfcg" Oct 08 18:19:31 crc kubenswrapper[4859]: I1008 18:19:31.212370 4859 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-storage-version-migrator"/"openshift-service-ca.crt" Oct 08 18:19:31 crc kubenswrapper[4859]: I1008 18:19:31.231997 4859 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca"/"openshift-service-ca.crt" Oct 08 18:19:31 crc kubenswrapper[4859]: I1008 18:19:31.252321 4859 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-service-ca"/"signing-key" Oct 08 18:19:31 crc kubenswrapper[4859]: I1008 18:19:31.256547 4859 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"signing-key\" (UniqueName: \"kubernetes.io/secret/da9f04a1-bb6a-40de-846a-5e2fe04855d1-signing-key\") pod \"service-ca-9c57cc56f-cs77r\" (UID: \"da9f04a1-bb6a-40de-846a-5e2fe04855d1\") " pod="openshift-service-ca/service-ca-9c57cc56f-cs77r" Oct 08 18:19:31 crc kubenswrapper[4859]: I1008 18:19:31.270879 4859 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-service-ca"/"service-ca-dockercfg-pn86c" Oct 08 18:19:31 crc kubenswrapper[4859]: I1008 18:19:31.288951 4859 request.go:700] Waited for 1.005906867s due to client-side throttling, not priority and fairness, request: GET:https://api-int.crc.testing:6443/api/v1/namespaces/openshift-service-ca/configmaps?fieldSelector=metadata.name%3Dsigning-cabundle&limit=500&resourceVersion=0 Oct 08 18:19:31 crc kubenswrapper[4859]: I1008 18:19:31.291074 4859 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca"/"signing-cabundle" Oct 08 18:19:31 crc kubenswrapper[4859]: I1008 18:19:31.295460 4859 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"signing-cabundle\" (UniqueName: \"kubernetes.io/configmap/da9f04a1-bb6a-40de-846a-5e2fe04855d1-signing-cabundle\") pod \"service-ca-9c57cc56f-cs77r\" (UID: \"da9f04a1-bb6a-40de-846a-5e2fe04855d1\") " pod="openshift-service-ca/service-ca-9c57cc56f-cs77r" Oct 08 18:19:31 crc kubenswrapper[4859]: I1008 18:19:31.311521 4859 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca"/"kube-root-ca.crt" Oct 08 18:19:31 crc kubenswrapper[4859]: I1008 18:19:31.331630 4859 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"openshift-service-ca.crt" Oct 08 18:19:31 crc kubenswrapper[4859]: I1008 18:19:31.352067 4859 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"packageserver-service-cert" Oct 08 18:19:31 crc kubenswrapper[4859]: I1008 18:19:31.365472 4859 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/b2b5fcca-e076-4cbe-bc1e-249fcfd25e53-apiservice-cert\") pod \"packageserver-d55dfcdfc-vc7bh\" (UID: \"b2b5fcca-e076-4cbe-bc1e-249fcfd25e53\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-vc7bh" Oct 08 18:19:31 crc kubenswrapper[4859]: I1008 18:19:31.366949 4859 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/b2b5fcca-e076-4cbe-bc1e-249fcfd25e53-webhook-cert\") pod \"packageserver-d55dfcdfc-vc7bh\" (UID: \"b2b5fcca-e076-4cbe-bc1e-249fcfd25e53\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-vc7bh" Oct 08 18:19:31 crc kubenswrapper[4859]: I1008 18:19:31.371714 4859 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"olm-operator-serviceaccount-dockercfg-rq7zk" Oct 08 18:19:31 crc kubenswrapper[4859]: I1008 18:19:31.391509 4859 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"kube-root-ca.crt" Oct 08 18:19:31 crc kubenswrapper[4859]: I1008 18:19:31.410483 4859 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"collect-profiles-config" Oct 08 18:19:31 crc kubenswrapper[4859]: I1008 18:19:31.414197 4859 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/d403dae9-eb8c-4c8d-9daa-bf2a0c815652-config-volume\") pod \"collect-profiles-29332455-w9wt8\" (UID: \"d403dae9-eb8c-4c8d-9daa-bf2a0c815652\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29332455-w9wt8" Oct 08 18:19:31 crc kubenswrapper[4859]: E1008 18:19:31.429295 4859 configmap.go:193] Couldn't get configMap openshift-oauth-apiserver/audit-1: failed to sync configmap cache: timed out waiting for the condition Oct 08 18:19:31 crc kubenswrapper[4859]: E1008 18:19:31.429297 4859 configmap.go:193] Couldn't get configMap openshift-oauth-apiserver/etcd-serving-ca: failed to sync configmap cache: timed out waiting for the condition Oct 08 18:19:31 crc kubenswrapper[4859]: E1008 18:19:31.429585 4859 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/d407fd47-096d-4f2a-ace1-dab4ddf7192a-audit-policies podName:d407fd47-096d-4f2a-ace1-dab4ddf7192a nodeName:}" failed. No retries permitted until 2025-10-08 18:19:31.929553489 +0000 UTC m=+142.176392908 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "audit-policies" (UniqueName: "kubernetes.io/configmap/d407fd47-096d-4f2a-ace1-dab4ddf7192a-audit-policies") pod "apiserver-7bbb656c7d-6j5rk" (UID: "d407fd47-096d-4f2a-ace1-dab4ddf7192a") : failed to sync configmap cache: timed out waiting for the condition Oct 08 18:19:31 crc kubenswrapper[4859]: E1008 18:19:31.429734 4859 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/d407fd47-096d-4f2a-ace1-dab4ddf7192a-etcd-serving-ca podName:d407fd47-096d-4f2a-ace1-dab4ddf7192a nodeName:}" failed. No retries permitted until 2025-10-08 18:19:31.929704534 +0000 UTC m=+142.176544103 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "etcd-serving-ca" (UniqueName: "kubernetes.io/configmap/d407fd47-096d-4f2a-ace1-dab4ddf7192a-etcd-serving-ca") pod "apiserver-7bbb656c7d-6j5rk" (UID: "d407fd47-096d-4f2a-ace1-dab4ddf7192a") : failed to sync configmap cache: timed out waiting for the condition Oct 08 18:19:31 crc kubenswrapper[4859]: I1008 18:19:31.431485 4859 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"collect-profiles-dockercfg-kzf4t" Oct 08 18:19:31 crc kubenswrapper[4859]: E1008 18:19:31.433773 4859 configmap.go:193] Couldn't get configMap openshift-oauth-apiserver/trusted-ca-bundle: failed to sync configmap cache: timed out waiting for the condition Oct 08 18:19:31 crc kubenswrapper[4859]: E1008 18:19:31.433783 4859 secret.go:188] Couldn't get secret openshift-oauth-apiserver/etcd-client: failed to sync secret cache: timed out waiting for the condition Oct 08 18:19:31 crc kubenswrapper[4859]: E1008 18:19:31.433822 4859 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/d407fd47-096d-4f2a-ace1-dab4ddf7192a-trusted-ca-bundle podName:d407fd47-096d-4f2a-ace1-dab4ddf7192a nodeName:}" failed. No retries permitted until 2025-10-08 18:19:31.933809806 +0000 UTC m=+142.180649185 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "trusted-ca-bundle" (UniqueName: "kubernetes.io/configmap/d407fd47-096d-4f2a-ace1-dab4ddf7192a-trusted-ca-bundle") pod "apiserver-7bbb656c7d-6j5rk" (UID: "d407fd47-096d-4f2a-ace1-dab4ddf7192a") : failed to sync configmap cache: timed out waiting for the condition Oct 08 18:19:31 crc kubenswrapper[4859]: E1008 18:19:31.433915 4859 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/d407fd47-096d-4f2a-ace1-dab4ddf7192a-etcd-client podName:d407fd47-096d-4f2a-ace1-dab4ddf7192a nodeName:}" failed. No retries permitted until 2025-10-08 18:19:31.933881978 +0000 UTC m=+142.180721457 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "etcd-client" (UniqueName: "kubernetes.io/secret/d407fd47-096d-4f2a-ace1-dab4ddf7192a-etcd-client") pod "apiserver-7bbb656c7d-6j5rk" (UID: "d407fd47-096d-4f2a-ace1-dab4ddf7192a") : failed to sync secret cache: timed out waiting for the condition Oct 08 18:19:31 crc kubenswrapper[4859]: E1008 18:19:31.436808 4859 secret.go:188] Couldn't get secret openshift-authentication/v4-0-config-user-template-provider-selection: failed to sync secret cache: timed out waiting for the condition Oct 08 18:19:31 crc kubenswrapper[4859]: E1008 18:19:31.436852 4859 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/0902ccb2-78ab-48d0-be02-4c42f03e12fb-v4-0-config-user-template-provider-selection podName:0902ccb2-78ab-48d0-be02-4c42f03e12fb nodeName:}" failed. No retries permitted until 2025-10-08 18:19:31.936842917 +0000 UTC m=+142.183682296 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "v4-0-config-user-template-provider-selection" (UniqueName: "kubernetes.io/secret/0902ccb2-78ab-48d0-be02-4c42f03e12fb-v4-0-config-user-template-provider-selection") pod "oauth-openshift-558db77b4-j8gh2" (UID: "0902ccb2-78ab-48d0-be02-4c42f03e12fb") : failed to sync secret cache: timed out waiting for the condition Oct 08 18:19:31 crc kubenswrapper[4859]: E1008 18:19:31.436888 4859 secret.go:188] Couldn't get secret openshift-oauth-apiserver/encryption-config-1: failed to sync secret cache: timed out waiting for the condition Oct 08 18:19:31 crc kubenswrapper[4859]: E1008 18:19:31.436964 4859 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/d407fd47-096d-4f2a-ace1-dab4ddf7192a-encryption-config podName:d407fd47-096d-4f2a-ace1-dab4ddf7192a nodeName:}" failed. No retries permitted until 2025-10-08 18:19:31.93694499 +0000 UTC m=+142.183784409 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "encryption-config" (UniqueName: "kubernetes.io/secret/d407fd47-096d-4f2a-ace1-dab4ddf7192a-encryption-config") pod "apiserver-7bbb656c7d-6j5rk" (UID: "d407fd47-096d-4f2a-ace1-dab4ddf7192a") : failed to sync secret cache: timed out waiting for the condition Oct 08 18:19:31 crc kubenswrapper[4859]: E1008 18:19:31.437115 4859 configmap.go:193] Couldn't get configMap openshift-authentication/audit: failed to sync configmap cache: timed out waiting for the condition Oct 08 18:19:31 crc kubenswrapper[4859]: E1008 18:19:31.437554 4859 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/0902ccb2-78ab-48d0-be02-4c42f03e12fb-audit-policies podName:0902ccb2-78ab-48d0-be02-4c42f03e12fb nodeName:}" failed. No retries permitted until 2025-10-08 18:19:31.937226968 +0000 UTC m=+142.184066557 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "audit-policies" (UniqueName: "kubernetes.io/configmap/0902ccb2-78ab-48d0-be02-4c42f03e12fb-audit-policies") pod "oauth-openshift-558db77b4-j8gh2" (UID: "0902ccb2-78ab-48d0-be02-4c42f03e12fb") : failed to sync configmap cache: timed out waiting for the condition Oct 08 18:19:31 crc kubenswrapper[4859]: E1008 18:19:31.441835 4859 secret.go:188] Couldn't get secret openshift-oauth-apiserver/serving-cert: failed to sync secret cache: timed out waiting for the condition Oct 08 18:19:31 crc kubenswrapper[4859]: E1008 18:19:31.441904 4859 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/d407fd47-096d-4f2a-ace1-dab4ddf7192a-serving-cert podName:d407fd47-096d-4f2a-ace1-dab4ddf7192a nodeName:}" failed. No retries permitted until 2025-10-08 18:19:31.941888897 +0000 UTC m=+142.188728466 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "serving-cert" (UniqueName: "kubernetes.io/secret/d407fd47-096d-4f2a-ace1-dab4ddf7192a-serving-cert") pod "apiserver-7bbb656c7d-6j5rk" (UID: "d407fd47-096d-4f2a-ace1-dab4ddf7192a") : failed to sync secret cache: timed out waiting for the condition Oct 08 18:19:31 crc kubenswrapper[4859]: E1008 18:19:31.442035 4859 secret.go:188] Couldn't get secret openshift-authentication/v4-0-config-system-session: failed to sync secret cache: timed out waiting for the condition Oct 08 18:19:31 crc kubenswrapper[4859]: E1008 18:19:31.442165 4859 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/0902ccb2-78ab-48d0-be02-4c42f03e12fb-v4-0-config-system-session podName:0902ccb2-78ab-48d0-be02-4c42f03e12fb nodeName:}" failed. No retries permitted until 2025-10-08 18:19:31.942150805 +0000 UTC m=+142.188990384 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "v4-0-config-system-session" (UniqueName: "kubernetes.io/secret/0902ccb2-78ab-48d0-be02-4c42f03e12fb-v4-0-config-system-session") pod "oauth-openshift-558db77b4-j8gh2" (UID: "0902ccb2-78ab-48d0-be02-4c42f03e12fb") : failed to sync secret cache: timed out waiting for the condition Oct 08 18:19:31 crc kubenswrapper[4859]: I1008 18:19:31.450301 4859 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"pprof-cert" Oct 08 18:19:31 crc kubenswrapper[4859]: I1008 18:19:31.455829 4859 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/d403dae9-eb8c-4c8d-9daa-bf2a0c815652-secret-volume\") pod \"collect-profiles-29332455-w9wt8\" (UID: \"d403dae9-eb8c-4c8d-9daa-bf2a0c815652\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29332455-w9wt8" Oct 08 18:19:31 crc kubenswrapper[4859]: I1008 18:19:31.458065 4859 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/6f093915-bce6-40eb-917f-ed0058431a79-profile-collector-cert\") pod \"olm-operator-6b444d44fb-5n49w\" (UID: \"6f093915-bce6-40eb-917f-ed0058431a79\") " pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-5n49w" Oct 08 18:19:31 crc kubenswrapper[4859]: I1008 18:19:31.471855 4859 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"multus-admission-controller-secret" Oct 08 18:19:31 crc kubenswrapper[4859]: I1008 18:19:31.483938 4859 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/f114c686-8fde-4560-b8e1-9cd8fea448f8-webhook-certs\") pod \"multus-admission-controller-857f4d67dd-lrqkp\" (UID: \"f114c686-8fde-4560-b8e1-9cd8fea448f8\") " pod="openshift-multus/multus-admission-controller-857f4d67dd-lrqkp" Oct 08 18:19:31 crc kubenswrapper[4859]: I1008 18:19:31.490765 4859 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"olm-operator-serving-cert" Oct 08 18:19:31 crc kubenswrapper[4859]: I1008 18:19:31.498864 4859 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/6f093915-bce6-40eb-917f-ed0058431a79-srv-cert\") pod \"olm-operator-6b444d44fb-5n49w\" (UID: \"6f093915-bce6-40eb-917f-ed0058431a79\") " pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-5n49w" Oct 08 18:19:31 crc kubenswrapper[4859]: I1008 18:19:31.510553 4859 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"multus-ac-dockercfg-9lkdf" Oct 08 18:19:31 crc kubenswrapper[4859]: I1008 18:19:31.538291 4859 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-marketplace"/"marketplace-trusted-ca" Oct 08 18:19:31 crc kubenswrapper[4859]: E1008 18:19:31.541243 4859 secret.go:188] Couldn't get secret openshift-marketplace/marketplace-operator-metrics: failed to sync secret cache: timed out waiting for the condition Oct 08 18:19:31 crc kubenswrapper[4859]: E1008 18:19:31.541434 4859 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/8d4b5b44-816f-44f6-98ab-c5d66aefaa08-marketplace-operator-metrics podName:8d4b5b44-816f-44f6-98ab-c5d66aefaa08 nodeName:}" failed. No retries permitted until 2025-10-08 18:19:32.041406337 +0000 UTC m=+142.288245726 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "marketplace-operator-metrics" (UniqueName: "kubernetes.io/secret/8d4b5b44-816f-44f6-98ab-c5d66aefaa08-marketplace-operator-metrics") pod "marketplace-operator-79b997595-bz6v8" (UID: "8d4b5b44-816f-44f6-98ab-c5d66aefaa08") : failed to sync secret cache: timed out waiting for the condition Oct 08 18:19:31 crc kubenswrapper[4859]: E1008 18:19:31.541269 4859 secret.go:188] Couldn't get secret openshift-service-ca-operator/serving-cert: failed to sync secret cache: timed out waiting for the condition Oct 08 18:19:31 crc kubenswrapper[4859]: E1008 18:19:31.541663 4859 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/56c0b58f-23d6-4d8e-8b75-9095128ac121-serving-cert podName:56c0b58f-23d6-4d8e-8b75-9095128ac121 nodeName:}" failed. No retries permitted until 2025-10-08 18:19:32.041647124 +0000 UTC m=+142.288486513 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "serving-cert" (UniqueName: "kubernetes.io/secret/56c0b58f-23d6-4d8e-8b75-9095128ac121-serving-cert") pod "service-ca-operator-777779d784-847r8" (UID: "56c0b58f-23d6-4d8e-8b75-9095128ac121") : failed to sync secret cache: timed out waiting for the condition Oct 08 18:19:31 crc kubenswrapper[4859]: I1008 18:19:31.543317 4859 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/8d4b5b44-816f-44f6-98ab-c5d66aefaa08-marketplace-trusted-ca\") pod \"marketplace-operator-79b997595-bz6v8\" (UID: \"8d4b5b44-816f-44f6-98ab-c5d66aefaa08\") " pod="openshift-marketplace/marketplace-operator-79b997595-bz6v8" Oct 08 18:19:31 crc kubenswrapper[4859]: E1008 18:19:31.544761 4859 configmap.go:193] Couldn't get configMap openshift-service-ca-operator/service-ca-operator-config: failed to sync configmap cache: timed out waiting for the condition Oct 08 18:19:31 crc kubenswrapper[4859]: E1008 18:19:31.544900 4859 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/56c0b58f-23d6-4d8e-8b75-9095128ac121-config podName:56c0b58f-23d6-4d8e-8b75-9095128ac121 nodeName:}" failed. No retries permitted until 2025-10-08 18:19:32.044886301 +0000 UTC m=+142.291725680 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "config" (UniqueName: "kubernetes.io/configmap/56c0b58f-23d6-4d8e-8b75-9095128ac121-config") pod "service-ca-operator-777779d784-847r8" (UID: "56c0b58f-23d6-4d8e-8b75-9095128ac121") : failed to sync configmap cache: timed out waiting for the condition Oct 08 18:19:31 crc kubenswrapper[4859]: E1008 18:19:31.544892 4859 secret.go:188] Couldn't get secret openshift-machine-config-operator/machine-config-server-tls: failed to sync secret cache: timed out waiting for the condition Oct 08 18:19:31 crc kubenswrapper[4859]: E1008 18:19:31.545068 4859 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/0d04221d-922c-4ee9-9adf-817e295161d9-certs podName:0d04221d-922c-4ee9-9adf-817e295161d9 nodeName:}" failed. No retries permitted until 2025-10-08 18:19:32.045059186 +0000 UTC m=+142.291898565 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "certs" (UniqueName: "kubernetes.io/secret/0d04221d-922c-4ee9-9adf-817e295161d9-certs") pod "machine-config-server-7tjh2" (UID: "0d04221d-922c-4ee9-9adf-817e295161d9") : failed to sync secret cache: timed out waiting for the condition Oct 08 18:19:31 crc kubenswrapper[4859]: E1008 18:19:31.545434 4859 secret.go:188] Couldn't get secret openshift-machine-config-operator/node-bootstrapper-token: failed to sync secret cache: timed out waiting for the condition Oct 08 18:19:31 crc kubenswrapper[4859]: E1008 18:19:31.545832 4859 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/0d04221d-922c-4ee9-9adf-817e295161d9-node-bootstrap-token podName:0d04221d-922c-4ee9-9adf-817e295161d9 nodeName:}" failed. No retries permitted until 2025-10-08 18:19:32.045784038 +0000 UTC m=+142.292623497 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "node-bootstrap-token" (UniqueName: "kubernetes.io/secret/0d04221d-922c-4ee9-9adf-817e295161d9-node-bootstrap-token") pod "machine-config-server-7tjh2" (UID: "0d04221d-922c-4ee9-9adf-817e295161d9") : failed to sync secret cache: timed out waiting for the condition Oct 08 18:19:31 crc kubenswrapper[4859]: I1008 18:19:31.551336 4859 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"marketplace-operator-dockercfg-5nsgg" Oct 08 18:19:31 crc kubenswrapper[4859]: I1008 18:19:31.571036 4859 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"marketplace-operator-metrics" Oct 08 18:19:31 crc kubenswrapper[4859]: I1008 18:19:31.590839 4859 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-marketplace"/"kube-root-ca.crt" Oct 08 18:19:31 crc kubenswrapper[4859]: I1008 18:19:31.610994 4859 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"catalog-operator-serving-cert" Oct 08 18:19:31 crc kubenswrapper[4859]: I1008 18:19:31.631061 4859 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-marketplace"/"openshift-service-ca.crt" Oct 08 18:19:31 crc kubenswrapper[4859]: I1008 18:19:31.651616 4859 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"package-server-manager-serving-cert" Oct 08 18:19:31 crc kubenswrapper[4859]: I1008 18:19:31.671219 4859 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-storage-version-migrator-operator"/"kube-storage-version-migrator-operator-dockercfg-2bh8d" Oct 08 18:19:31 crc kubenswrapper[4859]: I1008 18:19:31.691678 4859 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-storage-version-migrator-operator"/"serving-cert" Oct 08 18:19:31 crc kubenswrapper[4859]: I1008 18:19:31.711665 4859 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-storage-version-migrator-operator"/"kube-root-ca.crt" Oct 08 18:19:31 crc kubenswrapper[4859]: I1008 18:19:31.730557 4859 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-storage-version-migrator-operator"/"config" Oct 08 18:19:31 crc kubenswrapper[4859]: I1008 18:19:31.751612 4859 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-service-ca-operator"/"serving-cert" Oct 08 18:19:31 crc kubenswrapper[4859]: I1008 18:19:31.771322 4859 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-service-ca-operator"/"service-ca-operator-dockercfg-rg9jl" Oct 08 18:19:31 crc kubenswrapper[4859]: I1008 18:19:31.791268 4859 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca-operator"/"kube-root-ca.crt" Oct 08 18:19:31 crc kubenswrapper[4859]: I1008 18:19:31.811755 4859 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-storage-version-migrator-operator"/"openshift-service-ca.crt" Oct 08 18:19:31 crc kubenswrapper[4859]: I1008 18:19:31.831228 4859 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca-operator"/"openshift-service-ca.crt" Oct 08 18:19:31 crc kubenswrapper[4859]: I1008 18:19:31.853664 4859 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca-operator"/"service-ca-operator-config" Oct 08 18:19:31 crc kubenswrapper[4859]: I1008 18:19:31.871619 4859 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-dns"/"dns-default" Oct 08 18:19:31 crc kubenswrapper[4859]: I1008 18:19:31.892076 4859 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-dns"/"dns-dockercfg-jwfmh" Oct 08 18:19:31 crc kubenswrapper[4859]: I1008 18:19:31.911927 4859 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-dns"/"dns-default-metrics-tls" Oct 08 18:19:31 crc kubenswrapper[4859]: I1008 18:19:31.931407 4859 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"machine-config-server-tls" Oct 08 18:19:31 crc kubenswrapper[4859]: I1008 18:19:31.951954 4859 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"machine-config-server-dockercfg-qx5rd" Oct 08 18:19:31 crc kubenswrapper[4859]: I1008 18:19:31.968977 4859 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/d407fd47-096d-4f2a-ace1-dab4ddf7192a-serving-cert\") pod \"apiserver-7bbb656c7d-6j5rk\" (UID: \"d407fd47-096d-4f2a-ace1-dab4ddf7192a\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-6j5rk" Oct 08 18:19:31 crc kubenswrapper[4859]: I1008 18:19:31.969121 4859 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/d407fd47-096d-4f2a-ace1-dab4ddf7192a-audit-policies\") pod \"apiserver-7bbb656c7d-6j5rk\" (UID: \"d407fd47-096d-4f2a-ace1-dab4ddf7192a\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-6j5rk" Oct 08 18:19:31 crc kubenswrapper[4859]: I1008 18:19:31.969162 4859 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/d407fd47-096d-4f2a-ace1-dab4ddf7192a-etcd-serving-ca\") pod \"apiserver-7bbb656c7d-6j5rk\" (UID: \"d407fd47-096d-4f2a-ace1-dab4ddf7192a\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-6j5rk" Oct 08 18:19:31 crc kubenswrapper[4859]: I1008 18:19:31.969439 4859 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/d407fd47-096d-4f2a-ace1-dab4ddf7192a-etcd-client\") pod \"apiserver-7bbb656c7d-6j5rk\" (UID: \"d407fd47-096d-4f2a-ace1-dab4ddf7192a\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-6j5rk" Oct 08 18:19:31 crc kubenswrapper[4859]: I1008 18:19:31.969499 4859 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/d407fd47-096d-4f2a-ace1-dab4ddf7192a-trusted-ca-bundle\") pod \"apiserver-7bbb656c7d-6j5rk\" (UID: \"d407fd47-096d-4f2a-ace1-dab4ddf7192a\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-6j5rk" Oct 08 18:19:31 crc kubenswrapper[4859]: I1008 18:19:31.969660 4859 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/d407fd47-096d-4f2a-ace1-dab4ddf7192a-encryption-config\") pod \"apiserver-7bbb656c7d-6j5rk\" (UID: \"d407fd47-096d-4f2a-ace1-dab4ddf7192a\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-6j5rk" Oct 08 18:19:31 crc kubenswrapper[4859]: I1008 18:19:31.969768 4859 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/0902ccb2-78ab-48d0-be02-4c42f03e12fb-v4-0-config-user-template-provider-selection\") pod \"oauth-openshift-558db77b4-j8gh2\" (UID: \"0902ccb2-78ab-48d0-be02-4c42f03e12fb\") " pod="openshift-authentication/oauth-openshift-558db77b4-j8gh2" Oct 08 18:19:31 crc kubenswrapper[4859]: I1008 18:19:31.969824 4859 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/0902ccb2-78ab-48d0-be02-4c42f03e12fb-audit-policies\") pod \"oauth-openshift-558db77b4-j8gh2\" (UID: \"0902ccb2-78ab-48d0-be02-4c42f03e12fb\") " pod="openshift-authentication/oauth-openshift-558db77b4-j8gh2" Oct 08 18:19:31 crc kubenswrapper[4859]: I1008 18:19:31.970029 4859 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/0902ccb2-78ab-48d0-be02-4c42f03e12fb-v4-0-config-system-session\") pod \"oauth-openshift-558db77b4-j8gh2\" (UID: \"0902ccb2-78ab-48d0-be02-4c42f03e12fb\") " pod="openshift-authentication/oauth-openshift-558db77b4-j8gh2" Oct 08 18:19:31 crc kubenswrapper[4859]: I1008 18:19:31.972328 4859 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"node-bootstrapper-token" Oct 08 18:19:31 crc kubenswrapper[4859]: I1008 18:19:31.996382 4859 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-api"/"control-plane-machine-set-operator-dockercfg-k9rxt" Oct 08 18:19:32 crc kubenswrapper[4859]: I1008 18:19:32.010599 4859 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-api"/"control-plane-machine-set-operator-tls" Oct 08 18:19:32 crc kubenswrapper[4859]: I1008 18:19:32.052204 4859 reflector.go:368] Caches populated for *v1.ConfigMap from object-"hostpath-provisioner"/"openshift-service-ca.crt" Oct 08 18:19:32 crc kubenswrapper[4859]: I1008 18:19:32.071191 4859 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"node-bootstrap-token\" (UniqueName: \"kubernetes.io/secret/0d04221d-922c-4ee9-9adf-817e295161d9-node-bootstrap-token\") pod \"machine-config-server-7tjh2\" (UID: \"0d04221d-922c-4ee9-9adf-817e295161d9\") " pod="openshift-machine-config-operator/machine-config-server-7tjh2" Oct 08 18:19:32 crc kubenswrapper[4859]: I1008 18:19:32.071262 4859 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/56c0b58f-23d6-4d8e-8b75-9095128ac121-serving-cert\") pod \"service-ca-operator-777779d784-847r8\" (UID: \"56c0b58f-23d6-4d8e-8b75-9095128ac121\") " pod="openshift-service-ca-operator/service-ca-operator-777779d784-847r8" Oct 08 18:19:32 crc kubenswrapper[4859]: I1008 18:19:32.071304 4859 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/8d4b5b44-816f-44f6-98ab-c5d66aefaa08-marketplace-operator-metrics\") pod \"marketplace-operator-79b997595-bz6v8\" (UID: \"8d4b5b44-816f-44f6-98ab-c5d66aefaa08\") " pod="openshift-marketplace/marketplace-operator-79b997595-bz6v8" Oct 08 18:19:32 crc kubenswrapper[4859]: I1008 18:19:32.071544 4859 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/56c0b58f-23d6-4d8e-8b75-9095128ac121-config\") pod \"service-ca-operator-777779d784-847r8\" (UID: \"56c0b58f-23d6-4d8e-8b75-9095128ac121\") " pod="openshift-service-ca-operator/service-ca-operator-777779d784-847r8" Oct 08 18:19:32 crc kubenswrapper[4859]: I1008 18:19:32.071607 4859 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"certs\" (UniqueName: \"kubernetes.io/secret/0d04221d-922c-4ee9-9adf-817e295161d9-certs\") pod \"machine-config-server-7tjh2\" (UID: \"0d04221d-922c-4ee9-9adf-817e295161d9\") " pod="openshift-machine-config-operator/machine-config-server-7tjh2" Oct 08 18:19:32 crc kubenswrapper[4859]: I1008 18:19:32.072456 4859 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/56c0b58f-23d6-4d8e-8b75-9095128ac121-config\") pod \"service-ca-operator-777779d784-847r8\" (UID: \"56c0b58f-23d6-4d8e-8b75-9095128ac121\") " pod="openshift-service-ca-operator/service-ca-operator-777779d784-847r8" Oct 08 18:19:32 crc kubenswrapper[4859]: I1008 18:19:32.073327 4859 reflector.go:368] Caches populated for *v1.Secret from object-"hostpath-provisioner"/"csi-hostpath-provisioner-sa-dockercfg-qd74k" Oct 08 18:19:32 crc kubenswrapper[4859]: I1008 18:19:32.075283 4859 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"certs\" (UniqueName: \"kubernetes.io/secret/0d04221d-922c-4ee9-9adf-817e295161d9-certs\") pod \"machine-config-server-7tjh2\" (UID: \"0d04221d-922c-4ee9-9adf-817e295161d9\") " pod="openshift-machine-config-operator/machine-config-server-7tjh2" Oct 08 18:19:32 crc kubenswrapper[4859]: I1008 18:19:32.076950 4859 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/8d4b5b44-816f-44f6-98ab-c5d66aefaa08-marketplace-operator-metrics\") pod \"marketplace-operator-79b997595-bz6v8\" (UID: \"8d4b5b44-816f-44f6-98ab-c5d66aefaa08\") " pod="openshift-marketplace/marketplace-operator-79b997595-bz6v8" Oct 08 18:19:32 crc kubenswrapper[4859]: I1008 18:19:32.077226 4859 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/56c0b58f-23d6-4d8e-8b75-9095128ac121-serving-cert\") pod \"service-ca-operator-777779d784-847r8\" (UID: \"56c0b58f-23d6-4d8e-8b75-9095128ac121\") " pod="openshift-service-ca-operator/service-ca-operator-777779d784-847r8" Oct 08 18:19:32 crc kubenswrapper[4859]: I1008 18:19:32.082610 4859 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"node-bootstrap-token\" (UniqueName: \"kubernetes.io/secret/0d04221d-922c-4ee9-9adf-817e295161d9-node-bootstrap-token\") pod \"machine-config-server-7tjh2\" (UID: \"0d04221d-922c-4ee9-9adf-817e295161d9\") " pod="openshift-machine-config-operator/machine-config-server-7tjh2" Oct 08 18:19:32 crc kubenswrapper[4859]: I1008 18:19:32.092758 4859 reflector.go:368] Caches populated for *v1.ConfigMap from object-"hostpath-provisioner"/"kube-root-ca.crt" Oct 08 18:19:32 crc kubenswrapper[4859]: I1008 18:19:32.113432 4859 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress-canary"/"canary-serving-cert" Oct 08 18:19:32 crc kubenswrapper[4859]: I1008 18:19:32.130945 4859 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress-canary"/"default-dockercfg-2llfx" Oct 08 18:19:32 crc kubenswrapper[4859]: I1008 18:19:32.151239 4859 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress-canary"/"kube-root-ca.crt" Oct 08 18:19:32 crc kubenswrapper[4859]: I1008 18:19:32.171833 4859 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress-canary"/"openshift-service-ca.crt" Oct 08 18:19:32 crc kubenswrapper[4859]: I1008 18:19:32.211050 4859 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-sjk7q\" (UniqueName: \"kubernetes.io/projected/65868bd3-b636-48c5-aac4-fcf734f84e39-kube-api-access-sjk7q\") pod \"openshift-config-operator-7777fb866f-ncj6n\" (UID: \"65868bd3-b636-48c5-aac4-fcf734f84e39\") " pod="openshift-config-operator/openshift-config-operator-7777fb866f-ncj6n" Oct 08 18:19:32 crc kubenswrapper[4859]: I1008 18:19:32.242620 4859 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-npz7k\" (UniqueName: \"kubernetes.io/projected/0902ccb2-78ab-48d0-be02-4c42f03e12fb-kube-api-access-npz7k\") pod \"oauth-openshift-558db77b4-j8gh2\" (UID: \"0902ccb2-78ab-48d0-be02-4c42f03e12fb\") " pod="openshift-authentication/oauth-openshift-558db77b4-j8gh2" Oct 08 18:19:32 crc kubenswrapper[4859]: I1008 18:19:32.254576 4859 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-gbpww\" (UniqueName: \"kubernetes.io/projected/460ad6e7-859d-4111-a555-b55c5b7228e0-kube-api-access-gbpww\") pod \"controller-manager-879f6c89f-9gpzj\" (UID: \"460ad6e7-859d-4111-a555-b55c5b7228e0\") " pod="openshift-controller-manager/controller-manager-879f6c89f-9gpzj" Oct 08 18:19:32 crc kubenswrapper[4859]: I1008 18:19:32.289360 4859 request.go:700] Waited for 1.855751355s due to client-side throttling, not priority and fairness, request: POST:https://api-int.crc.testing:6443/api/v1/namespaces/openshift-etcd-operator/serviceaccounts/etcd-operator/token Oct 08 18:19:32 crc kubenswrapper[4859]: I1008 18:19:32.308752 4859 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-tlm28\" (UniqueName: \"kubernetes.io/projected/e4b12149-79e9-4b2e-8991-519ad5daa14b-kube-api-access-tlm28\") pod \"etcd-operator-b45778765-tnn6l\" (UID: \"e4b12149-79e9-4b2e-8991-519ad5daa14b\") " pod="openshift-etcd-operator/etcd-operator-b45778765-tnn6l" Oct 08 18:19:32 crc kubenswrapper[4859]: I1008 18:19:32.325252 4859 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/c79c3857-6ecb-4e3c-b59f-6a58ba0f94e9-bound-sa-token\") pod \"cluster-image-registry-operator-dc59b4c8b-kq7mx\" (UID: \"c79c3857-6ecb-4e3c-b59f-6a58ba0f94e9\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-kq7mx" Oct 08 18:19:32 crc kubenswrapper[4859]: I1008 18:19:32.339271 4859 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-config-operator/openshift-config-operator-7777fb866f-ncj6n" Oct 08 18:19:32 crc kubenswrapper[4859]: I1008 18:19:32.344475 4859 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-w2wf9\" (UniqueName: \"kubernetes.io/projected/9dcb8f76-0fff-436c-9191-fe3f787b7ac0-kube-api-access-w2wf9\") pod \"downloads-7954f5f757-8q5jz\" (UID: \"9dcb8f76-0fff-436c-9191-fe3f787b7ac0\") " pod="openshift-console/downloads-7954f5f757-8q5jz" Oct 08 18:19:32 crc kubenswrapper[4859]: I1008 18:19:32.389761 4859 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-879f6c89f-9gpzj" Oct 08 18:19:32 crc kubenswrapper[4859]: I1008 18:19:32.409710 4859 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-9vc6p\" (UniqueName: \"kubernetes.io/projected/2736d658-3609-4c30-a010-7b00c6300e2e-kube-api-access-9vc6p\") pod \"console-operator-58897d9998-8sflk\" (UID: \"2736d658-3609-4c30-a010-7b00c6300e2e\") " pod="openshift-console-operator/console-operator-58897d9998-8sflk" Oct 08 18:19:32 crc kubenswrapper[4859]: I1008 18:19:32.410969 4859 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-hjhn5\" (UniqueName: \"kubernetes.io/projected/7d9d11d4-9f7b-4f33-a9d2-23db86bb2343-kube-api-access-hjhn5\") pod \"openshift-apiserver-operator-796bbdcf4f-czj5g\" (UID: \"7d9d11d4-9f7b-4f33-a9d2-23db86bb2343\") " pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-czj5g" Oct 08 18:19:32 crc kubenswrapper[4859]: I1008 18:19:32.412499 4859 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-vvbpr\" (UniqueName: \"kubernetes.io/projected/c79c3857-6ecb-4e3c-b59f-6a58ba0f94e9-kube-api-access-vvbpr\") pod \"cluster-image-registry-operator-dc59b4c8b-kq7mx\" (UID: \"c79c3857-6ecb-4e3c-b59f-6a58ba0f94e9\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-kq7mx" Oct 08 18:19:32 crc kubenswrapper[4859]: I1008 18:19:32.430392 4859 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-kfmlz\" (UniqueName: \"kubernetes.io/projected/d80051ae-5521-4f01-b34d-34d06a661177-kube-api-access-kfmlz\") pod \"openshift-controller-manager-operator-756b6f6bc6-82pcg\" (UID: \"d80051ae-5521-4f01-b34d-34d06a661177\") " pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-82pcg" Oct 08 18:19:32 crc kubenswrapper[4859]: I1008 18:19:32.449638 4859 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-console-operator/console-operator-58897d9998-8sflk" Oct 08 18:19:32 crc kubenswrapper[4859]: I1008 18:19:32.453345 4859 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-4ppdz\" (UniqueName: \"kubernetes.io/projected/70f88ce6-b9fe-422e-827e-af2b6840a783-kube-api-access-4ppdz\") pod \"apiserver-76f77b778f-fv9bp\" (UID: \"70f88ce6-b9fe-422e-827e-af2b6840a783\") " pod="openshift-apiserver/apiserver-76f77b778f-fv9bp" Oct 08 18:19:32 crc kubenswrapper[4859]: I1008 18:19:32.453611 4859 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-apiserver/apiserver-76f77b778f-fv9bp" Oct 08 18:19:32 crc kubenswrapper[4859]: I1008 18:19:32.466483 4859 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-kq7mx" Oct 08 18:19:32 crc kubenswrapper[4859]: I1008 18:19:32.473788 4859 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-console/downloads-7954f5f757-8q5jz" Oct 08 18:19:32 crc kubenswrapper[4859]: I1008 18:19:32.481000 4859 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-etcd-operator/etcd-operator-b45778765-tnn6l" Oct 08 18:19:32 crc kubenswrapper[4859]: I1008 18:19:32.488913 4859 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-czj5g" Oct 08 18:19:32 crc kubenswrapper[4859]: I1008 18:19:32.490580 4859 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-9hnbb\" (UniqueName: \"kubernetes.io/projected/ca64d874-5d86-4c49-8953-631a75d59788-kube-api-access-9hnbb\") pod \"route-controller-manager-6576b87f9c-xlfdd\" (UID: \"ca64d874-5d86-4c49-8953-631a75d59788\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-xlfdd" Oct 08 18:19:32 crc kubenswrapper[4859]: I1008 18:19:32.499276 4859 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-82pcg" Oct 08 18:19:32 crc kubenswrapper[4859]: I1008 18:19:32.505316 4859 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-g65gr\" (UniqueName: \"kubernetes.io/projected/9b239e22-0030-4c68-b0f1-66fbaedd5f9b-kube-api-access-g65gr\") pod \"cluster-samples-operator-665b6dd947-c5p45\" (UID: \"9b239e22-0030-4c68-b0f1-66fbaedd5f9b\") " pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-c5p45" Oct 08 18:19:32 crc kubenswrapper[4859]: I1008 18:19:32.517405 4859 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-dwj4n\" (UniqueName: \"kubernetes.io/projected/a52b9536-1013-482b-9261-e024377687d2-kube-api-access-dwj4n\") pod \"machine-api-operator-5694c8668f-gmbk9\" (UID: \"a52b9536-1013-482b-9261-e024377687d2\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-gmbk9" Oct 08 18:19:32 crc kubenswrapper[4859]: I1008 18:19:32.533851 4859 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/bb6dc185-94e3-4cb7-8076-1011c1c60fab-kube-api-access\") pod \"kube-controller-manager-operator-78b949d7b-kc42p\" (UID: \"bb6dc185-94e3-4cb7-8076-1011c1c60fab\") " pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-kc42p" Oct 08 18:19:32 crc kubenswrapper[4859]: I1008 18:19:32.551157 4859 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-w5l7z\" (UniqueName: \"kubernetes.io/projected/39b30c11-d0f6-4302-9ef8-e7fcf1b2170e-kube-api-access-w5l7z\") pod \"machine-approver-56656f9798-95s6m\" (UID: \"39b30c11-d0f6-4302-9ef8-e7fcf1b2170e\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-95s6m" Oct 08 18:19:32 crc kubenswrapper[4859]: I1008 18:19:32.579193 4859 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-df4rr\" (UniqueName: \"kubernetes.io/projected/9a829106-59b1-4389-95bf-2ee4d56e317c-kube-api-access-df4rr\") pod \"console-f9d7485db-m2xmp\" (UID: \"9a829106-59b1-4389-95bf-2ee4d56e317c\") " pod="openshift-console/console-f9d7485db-m2xmp" Oct 08 18:19:32 crc kubenswrapper[4859]: I1008 18:19:32.583572 4859 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-config-operator/openshift-config-operator-7777fb866f-ncj6n"] Oct 08 18:19:32 crc kubenswrapper[4859]: I1008 18:19:32.600457 4859 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-hf7dh\" (UniqueName: \"kubernetes.io/projected/b8d9f66c-e7a4-467f-bcf5-701cc0b68d52-kube-api-access-hf7dh\") pod \"authentication-operator-69f744f599-jd5kh\" (UID: \"b8d9f66c-e7a4-467f-bcf5-701cc0b68d52\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-jd5kh" Oct 08 18:19:32 crc kubenswrapper[4859]: I1008 18:19:32.609483 4859 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-ovn-kubernetes/ovnkube-node-4g8gf" Oct 08 18:19:32 crc kubenswrapper[4859]: I1008 18:19:32.609538 4859 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-xlfdd" Oct 08 18:19:32 crc kubenswrapper[4859]: I1008 18:19:32.613035 4859 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-zsgtf\" (UniqueName: \"kubernetes.io/projected/f114c686-8fde-4560-b8e1-9cd8fea448f8-kube-api-access-zsgtf\") pod \"multus-admission-controller-857f4d67dd-lrqkp\" (UID: \"f114c686-8fde-4560-b8e1-9cd8fea448f8\") " pod="openshift-multus/multus-admission-controller-857f4d67dd-lrqkp" Oct 08 18:19:32 crc kubenswrapper[4859]: I1008 18:19:32.633077 4859 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-9tml9\" (UniqueName: \"kubernetes.io/projected/56c0b58f-23d6-4d8e-8b75-9095128ac121-kube-api-access-9tml9\") pod \"service-ca-operator-777779d784-847r8\" (UID: \"56c0b58f-23d6-4d8e-8b75-9095128ac121\") " pod="openshift-service-ca-operator/service-ca-operator-777779d784-847r8" Oct 08 18:19:32 crc kubenswrapper[4859]: I1008 18:19:32.647192 4859 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/multus-admission-controller-857f4d67dd-lrqkp" Oct 08 18:19:32 crc kubenswrapper[4859]: I1008 18:19:32.649889 4859 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/01517214-a376-421e-9938-e548d76695d5-kube-api-access\") pod \"kube-apiserver-operator-766d6c64bb-4s5w9\" (UID: \"01517214-a376-421e-9938-e548d76695d5\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-4s5w9" Oct 08 18:19:32 crc kubenswrapper[4859]: I1008 18:19:32.668355 4859 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-jb26g\" (UniqueName: \"kubernetes.io/projected/6f093915-bce6-40eb-917f-ed0058431a79-kube-api-access-jb26g\") pod \"olm-operator-6b444d44fb-5n49w\" (UID: \"6f093915-bce6-40eb-917f-ed0058431a79\") " pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-5n49w" Oct 08 18:19:32 crc kubenswrapper[4859]: I1008 18:19:32.668708 4859 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-5n49w" Oct 08 18:19:32 crc kubenswrapper[4859]: I1008 18:19:32.679807 4859 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-authentication-operator/authentication-operator-69f744f599-jd5kh" Oct 08 18:19:32 crc kubenswrapper[4859]: I1008 18:19:32.684984 4859 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-v4vwn\" (UniqueName: \"kubernetes.io/projected/d403dae9-eb8c-4c8d-9daa-bf2a0c815652-kube-api-access-v4vwn\") pod \"collect-profiles-29332455-w9wt8\" (UID: \"d403dae9-eb8c-4c8d-9daa-bf2a0c815652\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29332455-w9wt8" Oct 08 18:19:32 crc kubenswrapper[4859]: I1008 18:19:32.689791 4859 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-cluster-machine-approver/machine-approver-56656f9798-95s6m" Oct 08 18:19:32 crc kubenswrapper[4859]: I1008 18:19:32.699805 4859 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-console/console-f9d7485db-m2xmp" Oct 08 18:19:32 crc kubenswrapper[4859]: I1008 18:19:32.701366 4859 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-apiserver/apiserver-76f77b778f-fv9bp"] Oct 08 18:19:32 crc kubenswrapper[4859]: I1008 18:19:32.707156 4859 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-service-ca-operator/service-ca-operator-777779d784-847r8" Oct 08 18:19:32 crc kubenswrapper[4859]: I1008 18:19:32.717299 4859 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-vxgz5\" (UniqueName: \"kubernetes.io/projected/8e2b8a53-4bc3-499b-9746-628c323d2a0c-kube-api-access-vxgz5\") pod \"router-default-5444994796-r5gs2\" (UID: \"8e2b8a53-4bc3-499b-9746-628c323d2a0c\") " pod="openshift-ingress/router-default-5444994796-r5gs2" Oct 08 18:19:32 crc kubenswrapper[4859]: I1008 18:19:32.717598 4859 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-c5p45" Oct 08 18:19:32 crc kubenswrapper[4859]: I1008 18:19:32.727908 4859 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-hv4wh\" (UniqueName: \"kubernetes.io/projected/8d4b5b44-816f-44f6-98ab-c5d66aefaa08-kube-api-access-hv4wh\") pod \"marketplace-operator-79b997595-bz6v8\" (UID: \"8d4b5b44-816f-44f6-98ab-c5d66aefaa08\") " pod="openshift-marketplace/marketplace-operator-79b997595-bz6v8" Oct 08 18:19:32 crc kubenswrapper[4859]: I1008 18:19:32.730866 4859 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-api/machine-api-operator-5694c8668f-gmbk9" Oct 08 18:19:32 crc kubenswrapper[4859]: I1008 18:19:32.752817 4859 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-glqwn\" (UniqueName: \"kubernetes.io/projected/da9f04a1-bb6a-40de-846a-5e2fe04855d1-kube-api-access-glqwn\") pod \"service-ca-9c57cc56f-cs77r\" (UID: \"da9f04a1-bb6a-40de-846a-5e2fe04855d1\") " pod="openshift-service-ca/service-ca-9c57cc56f-cs77r" Oct 08 18:19:32 crc kubenswrapper[4859]: I1008 18:19:32.790513 4859 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-4hdx8\" (UniqueName: \"kubernetes.io/projected/5401cd5b-bc2e-43d4-8d13-b7e9134aa72f-kube-api-access-4hdx8\") pod \"migrator-59844c95c7-pmngm\" (UID: \"5401cd5b-bc2e-43d4-8d13-b7e9134aa72f\") " pod="openshift-kube-storage-version-migrator/migrator-59844c95c7-pmngm" Oct 08 18:19:32 crc kubenswrapper[4859]: I1008 18:19:32.794216 4859 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-jbvx4\" (UniqueName: \"kubernetes.io/projected/b2b5fcca-e076-4cbe-bc1e-249fcfd25e53-kube-api-access-jbvx4\") pod \"packageserver-d55dfcdfc-vc7bh\" (UID: \"b2b5fcca-e076-4cbe-bc1e-249fcfd25e53\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-vc7bh" Oct 08 18:19:32 crc kubenswrapper[4859]: I1008 18:19:32.795217 4859 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-console-operator/console-operator-58897d9998-8sflk"] Oct 08 18:19:32 crc kubenswrapper[4859]: I1008 18:19:32.805451 4859 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-kc42p" Oct 08 18:19:32 crc kubenswrapper[4859]: I1008 18:19:32.806620 4859 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-5kp2q\" (UniqueName: \"kubernetes.io/projected/0d04221d-922c-4ee9-9adf-817e295161d9-kube-api-access-5kp2q\") pod \"machine-config-server-7tjh2\" (UID: \"0d04221d-922c-4ee9-9adf-817e295161d9\") " pod="openshift-machine-config-operator/machine-config-server-7tjh2" Oct 08 18:19:32 crc kubenswrapper[4859]: I1008 18:19:32.829159 4859 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-8cxvk\" (UniqueName: \"kubernetes.io/projected/e431ddbe-2d50-4a0a-9679-5565e0df977f-kube-api-access-8cxvk\") pod \"machine-config-controller-84d6567774-m7m5n\" (UID: \"e431ddbe-2d50-4a0a-9679-5565e0df977f\") " pod="openshift-machine-config-operator/machine-config-controller-84d6567774-m7m5n" Oct 08 18:19:32 crc kubenswrapper[4859]: I1008 18:19:32.854470 4859 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-system-session" Oct 08 18:19:32 crc kubenswrapper[4859]: I1008 18:19:32.860165 4859 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-68n5n\" (UniqueName: \"kubernetes.io/projected/0c9973e7-5c40-463b-aca1-c3c93d7a9d0b-kube-api-access-68n5n\") pod \"machine-config-operator-74547568cd-xsc6l\" (UID: \"0c9973e7-5c40-463b-aca1-c3c93d7a9d0b\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-xsc6l" Oct 08 18:19:32 crc kubenswrapper[4859]: I1008 18:19:32.865154 4859 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/0902ccb2-78ab-48d0-be02-4c42f03e12fb-v4-0-config-system-session\") pod \"oauth-openshift-558db77b4-j8gh2\" (UID: \"0902ccb2-78ab-48d0-be02-4c42f03e12fb\") " pod="openshift-authentication/oauth-openshift-558db77b4-j8gh2" Oct 08 18:19:32 crc kubenswrapper[4859]: I1008 18:19:32.875290 4859 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-oauth-apiserver"/"encryption-config-1" Oct 08 18:19:32 crc kubenswrapper[4859]: I1008 18:19:32.883133 4859 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/d407fd47-096d-4f2a-ace1-dab4ddf7192a-encryption-config\") pod \"apiserver-7bbb656c7d-6j5rk\" (UID: \"d407fd47-096d-4f2a-ace1-dab4ddf7192a\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-6j5rk" Oct 08 18:19:32 crc kubenswrapper[4859]: I1008 18:19:32.891219 4859 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-oauth-apiserver"/"trusted-ca-bundle" Oct 08 18:19:32 crc kubenswrapper[4859]: I1008 18:19:32.892999 4859 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-4s5w9" Oct 08 18:19:32 crc kubenswrapper[4859]: I1008 18:19:32.894379 4859 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/d407fd47-096d-4f2a-ace1-dab4ddf7192a-trusted-ca-bundle\") pod \"apiserver-7bbb656c7d-6j5rk\" (UID: \"d407fd47-096d-4f2a-ace1-dab4ddf7192a\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-6j5rk" Oct 08 18:19:32 crc kubenswrapper[4859]: I1008 18:19:32.900102 4859 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/machine-config-controller-84d6567774-m7m5n" Oct 08 18:19:32 crc kubenswrapper[4859]: I1008 18:19:32.901490 4859 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-controller-manager/controller-manager-879f6c89f-9gpzj"] Oct 08 18:19:32 crc kubenswrapper[4859]: I1008 18:19:32.911613 4859 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-oauth-apiserver"/"openshift-service-ca.crt" Oct 08 18:19:32 crc kubenswrapper[4859]: I1008 18:19:32.914311 4859 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ingress/router-default-5444994796-r5gs2" Oct 08 18:19:32 crc kubenswrapper[4859]: I1008 18:19:32.916630 4859 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-storage-version-migrator/migrator-59844c95c7-pmngm" Oct 08 18:19:32 crc kubenswrapper[4859]: I1008 18:19:32.925115 4859 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-service-ca/service-ca-9c57cc56f-cs77r" Oct 08 18:19:32 crc kubenswrapper[4859]: I1008 18:19:32.931910 4859 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-vc7bh" Oct 08 18:19:32 crc kubenswrapper[4859]: I1008 18:19:32.931952 4859 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-oauth-apiserver"/"serving-cert" Oct 08 18:19:32 crc kubenswrapper[4859]: I1008 18:19:32.952972 4859 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29332455-w9wt8" Oct 08 18:19:32 crc kubenswrapper[4859]: I1008 18:19:32.955163 4859 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-oauth-apiserver"/"kube-root-ca.crt" Oct 08 18:19:32 crc kubenswrapper[4859]: I1008 18:19:32.960656 4859 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-8vzp2\" (UniqueName: \"kubernetes.io/projected/d407fd47-096d-4f2a-ace1-dab4ddf7192a-kube-api-access-8vzp2\") pod \"apiserver-7bbb656c7d-6j5rk\" (UID: \"d407fd47-096d-4f2a-ace1-dab4ddf7192a\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-6j5rk" Oct 08 18:19:32 crc kubenswrapper[4859]: I1008 18:19:32.963547 4859 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/d407fd47-096d-4f2a-ace1-dab4ddf7192a-serving-cert\") pod \"apiserver-7bbb656c7d-6j5rk\" (UID: \"d407fd47-096d-4f2a-ace1-dab4ddf7192a\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-6j5rk" Oct 08 18:19:32 crc kubenswrapper[4859]: E1008 18:19:32.974081 4859 configmap.go:193] Couldn't get configMap openshift-authentication/audit: failed to sync configmap cache: timed out waiting for the condition Oct 08 18:19:32 crc kubenswrapper[4859]: E1008 18:19:32.974122 4859 secret.go:188] Couldn't get secret openshift-authentication/v4-0-config-user-template-provider-selection: failed to sync secret cache: timed out waiting for the condition Oct 08 18:19:32 crc kubenswrapper[4859]: E1008 18:19:32.974175 4859 configmap.go:193] Couldn't get configMap openshift-oauth-apiserver/etcd-serving-ca: failed to sync configmap cache: timed out waiting for the condition Oct 08 18:19:32 crc kubenswrapper[4859]: E1008 18:19:32.974266 4859 configmap.go:193] Couldn't get configMap openshift-oauth-apiserver/audit-1: failed to sync configmap cache: timed out waiting for the condition Oct 08 18:19:32 crc kubenswrapper[4859]: E1008 18:19:32.974308 4859 secret.go:188] Couldn't get secret openshift-oauth-apiserver/etcd-client: failed to sync secret cache: timed out waiting for the condition Oct 08 18:19:32 crc kubenswrapper[4859]: E1008 18:19:32.974196 4859 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/0902ccb2-78ab-48d0-be02-4c42f03e12fb-audit-policies podName:0902ccb2-78ab-48d0-be02-4c42f03e12fb nodeName:}" failed. No retries permitted until 2025-10-08 18:19:33.97416974 +0000 UTC m=+144.221009119 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "audit-policies" (UniqueName: "kubernetes.io/configmap/0902ccb2-78ab-48d0-be02-4c42f03e12fb-audit-policies") pod "oauth-openshift-558db77b4-j8gh2" (UID: "0902ccb2-78ab-48d0-be02-4c42f03e12fb") : failed to sync configmap cache: timed out waiting for the condition Oct 08 18:19:32 crc kubenswrapper[4859]: E1008 18:19:32.974369 4859 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/0902ccb2-78ab-48d0-be02-4c42f03e12fb-v4-0-config-user-template-provider-selection podName:0902ccb2-78ab-48d0-be02-4c42f03e12fb nodeName:}" failed. No retries permitted until 2025-10-08 18:19:33.974345425 +0000 UTC m=+144.221184984 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "v4-0-config-user-template-provider-selection" (UniqueName: "kubernetes.io/secret/0902ccb2-78ab-48d0-be02-4c42f03e12fb-v4-0-config-user-template-provider-selection") pod "oauth-openshift-558db77b4-j8gh2" (UID: "0902ccb2-78ab-48d0-be02-4c42f03e12fb") : failed to sync secret cache: timed out waiting for the condition Oct 08 18:19:32 crc kubenswrapper[4859]: E1008 18:19:32.974385 4859 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/d407fd47-096d-4f2a-ace1-dab4ddf7192a-etcd-serving-ca podName:d407fd47-096d-4f2a-ace1-dab4ddf7192a nodeName:}" failed. No retries permitted until 2025-10-08 18:19:33.974375546 +0000 UTC m=+144.221215145 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "etcd-serving-ca" (UniqueName: "kubernetes.io/configmap/d407fd47-096d-4f2a-ace1-dab4ddf7192a-etcd-serving-ca") pod "apiserver-7bbb656c7d-6j5rk" (UID: "d407fd47-096d-4f2a-ace1-dab4ddf7192a") : failed to sync configmap cache: timed out waiting for the condition Oct 08 18:19:32 crc kubenswrapper[4859]: E1008 18:19:32.974404 4859 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/d407fd47-096d-4f2a-ace1-dab4ddf7192a-audit-policies podName:d407fd47-096d-4f2a-ace1-dab4ddf7192a nodeName:}" failed. No retries permitted until 2025-10-08 18:19:33.974399087 +0000 UTC m=+144.221238686 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "audit-policies" (UniqueName: "kubernetes.io/configmap/d407fd47-096d-4f2a-ace1-dab4ddf7192a-audit-policies") pod "apiserver-7bbb656c7d-6j5rk" (UID: "d407fd47-096d-4f2a-ace1-dab4ddf7192a") : failed to sync configmap cache: timed out waiting for the condition Oct 08 18:19:32 crc kubenswrapper[4859]: E1008 18:19:32.974423 4859 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/d407fd47-096d-4f2a-ace1-dab4ddf7192a-etcd-client podName:d407fd47-096d-4f2a-ace1-dab4ddf7192a nodeName:}" failed. No retries permitted until 2025-10-08 18:19:33.974415517 +0000 UTC m=+144.221255106 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "etcd-client" (UniqueName: "kubernetes.io/secret/d407fd47-096d-4f2a-ace1-dab4ddf7192a-etcd-client") pod "apiserver-7bbb656c7d-6j5rk" (UID: "d407fd47-096d-4f2a-ace1-dab4ddf7192a") : failed to sync secret cache: timed out waiting for the condition Oct 08 18:19:32 crc kubenswrapper[4859]: I1008 18:19:32.976867 4859 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-oauth-apiserver"/"etcd-serving-ca" Oct 08 18:19:32 crc kubenswrapper[4859]: I1008 18:19:32.985068 4859 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/marketplace-operator-79b997595-bz6v8" Oct 08 18:19:32 crc kubenswrapper[4859]: I1008 18:19:32.992047 4859 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"audit" Oct 08 18:19:33 crc kubenswrapper[4859]: I1008 18:19:33.002766 4859 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-console/downloads-7954f5f757-8q5jz"] Oct 08 18:19:33 crc kubenswrapper[4859]: I1008 18:19:33.018490 4859 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-oauth-apiserver"/"oauth-apiserver-sa-dockercfg-6r2bq" Oct 08 18:19:33 crc kubenswrapper[4859]: I1008 18:19:33.030998 4859 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-oauth-apiserver"/"audit-1" Oct 08 18:19:33 crc kubenswrapper[4859]: I1008 18:19:33.035418 4859 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/machine-config-server-7tjh2" Oct 08 18:19:33 crc kubenswrapper[4859]: I1008 18:19:33.047598 4859 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-kq7mx"] Oct 08 18:19:33 crc kubenswrapper[4859]: I1008 18:19:33.051592 4859 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-oauth-apiserver"/"etcd-client" Oct 08 18:19:33 crc kubenswrapper[4859]: W1008 18:19:33.071242 4859 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod460ad6e7_859d_4111_a555_b55c5b7228e0.slice/crio-1e7c718f9f31e46de3067304f0212127211b5576b88d7881e74ede8e5e3004a7 WatchSource:0}: Error finding container 1e7c718f9f31e46de3067304f0212127211b5576b88d7881e74ede8e5e3004a7: Status 404 returned error can't find the container with id 1e7c718f9f31e46de3067304f0212127211b5576b88d7881e74ede8e5e3004a7 Oct 08 18:19:33 crc kubenswrapper[4859]: I1008 18:19:33.075152 4859 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-user-template-provider-selection" Oct 08 18:19:33 crc kubenswrapper[4859]: I1008 18:19:33.103802 4859 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/2a959183-1370-4723-bb5e-8625689d6121-metrics-tls\") pod \"dns-default-x7pp9\" (UID: \"2a959183-1370-4723-bb5e-8625689d6121\") " pod="openshift-dns/dns-default-x7pp9" Oct 08 18:19:33 crc kubenswrapper[4859]: I1008 18:19:33.103883 4859 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/4afad99b-9bf7-43c9-8dcf-7a3217136be5-trusted-ca\") pod \"image-registry-697d97f7c8-kg44m\" (UID: \"4afad99b-9bf7-43c9-8dcf-7a3217136be5\") " pod="openshift-image-registry/image-registry-697d97f7c8-kg44m" Oct 08 18:19:33 crc kubenswrapper[4859]: I1008 18:19:33.103995 4859 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/4afad99b-9bf7-43c9-8dcf-7a3217136be5-registry-tls\") pod \"image-registry-697d97f7c8-kg44m\" (UID: \"4afad99b-9bf7-43c9-8dcf-7a3217136be5\") " pod="openshift-image-registry/image-registry-697d97f7c8-kg44m" Oct 08 18:19:33 crc kubenswrapper[4859]: I1008 18:19:33.104024 4859 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/46a9f481-0772-4006-99e2-723919011659-config\") pod \"kube-storage-version-migrator-operator-b67b599dd-6dlgn\" (UID: \"46a9f481-0772-4006-99e2-723919011659\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-6dlgn" Oct 08 18:19:33 crc kubenswrapper[4859]: I1008 18:19:33.104047 4859 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-59lvs\" (UniqueName: \"kubernetes.io/projected/63636924-a28b-447e-8568-92a193a69ade-kube-api-access-59lvs\") pod \"dns-operator-744455d44c-q222b\" (UID: \"63636924-a28b-447e-8568-92a193a69ade\") " pod="openshift-dns-operator/dns-operator-744455d44c-q222b" Oct 08 18:19:33 crc kubenswrapper[4859]: I1008 18:19:33.104085 4859 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/ed062c43-61ec-4951-a27c-77f165523d3f-bound-sa-token\") pod \"ingress-operator-5b745b69d9-pf7j6\" (UID: \"ed062c43-61ec-4951-a27c-77f165523d3f\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-pf7j6" Oct 08 18:19:33 crc kubenswrapper[4859]: I1008 18:19:33.104105 4859 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-ls2m9\" (UniqueName: \"kubernetes.io/projected/46a9f481-0772-4006-99e2-723919011659-kube-api-access-ls2m9\") pod \"kube-storage-version-migrator-operator-b67b599dd-6dlgn\" (UID: \"46a9f481-0772-4006-99e2-723919011659\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-6dlgn" Oct 08 18:19:33 crc kubenswrapper[4859]: I1008 18:19:33.104133 4859 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/4afad99b-9bf7-43c9-8dcf-7a3217136be5-ca-trust-extracted\") pod \"image-registry-697d97f7c8-kg44m\" (UID: \"4afad99b-9bf7-43c9-8dcf-7a3217136be5\") " pod="openshift-image-registry/image-registry-697d97f7c8-kg44m" Oct 08 18:19:33 crc kubenswrapper[4859]: I1008 18:19:33.104157 4859 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-rvtkc\" (UniqueName: \"kubernetes.io/projected/ed062c43-61ec-4951-a27c-77f165523d3f-kube-api-access-rvtkc\") pod \"ingress-operator-5b745b69d9-pf7j6\" (UID: \"ed062c43-61ec-4951-a27c-77f165523d3f\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-pf7j6" Oct 08 18:19:33 crc kubenswrapper[4859]: I1008 18:19:33.104191 4859 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-hkbdd\" (UniqueName: \"kubernetes.io/projected/c073e5c1-652c-406a-ad01-9ca626e8c56e-kube-api-access-hkbdd\") pod \"package-server-manager-789f6589d5-hf2fp\" (UID: \"c073e5c1-652c-406a-ad01-9ca626e8c56e\") " pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-hf2fp" Oct 08 18:19:33 crc kubenswrapper[4859]: I1008 18:19:33.104236 4859 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-cs2l8\" (UniqueName: \"kubernetes.io/projected/4afad99b-9bf7-43c9-8dcf-7a3217136be5-kube-api-access-cs2l8\") pod \"image-registry-697d97f7c8-kg44m\" (UID: \"4afad99b-9bf7-43c9-8dcf-7a3217136be5\") " pod="openshift-image-registry/image-registry-697d97f7c8-kg44m" Oct 08 18:19:33 crc kubenswrapper[4859]: I1008 18:19:33.104259 4859 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/ed062c43-61ec-4951-a27c-77f165523d3f-metrics-tls\") pod \"ingress-operator-5b745b69d9-pf7j6\" (UID: \"ed062c43-61ec-4951-a27c-77f165523d3f\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-pf7j6" Oct 08 18:19:33 crc kubenswrapper[4859]: I1008 18:19:33.104298 4859 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/7e23494c-2895-43b3-a35c-47a7b2d33ebc-serving-cert\") pod \"openshift-kube-scheduler-operator-5fdd9b5758-5f4q5\" (UID: \"7e23494c-2895-43b3-a35c-47a7b2d33ebc\") " pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-5f4q5" Oct 08 18:19:33 crc kubenswrapper[4859]: I1008 18:19:33.104333 4859 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/4afad99b-9bf7-43c9-8dcf-7a3217136be5-registry-certificates\") pod \"image-registry-697d97f7c8-kg44m\" (UID: \"4afad99b-9bf7-43c9-8dcf-7a3217136be5\") " pod="openshift-image-registry/image-registry-697d97f7c8-kg44m" Oct 08 18:19:33 crc kubenswrapper[4859]: I1008 18:19:33.104377 4859 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/63636924-a28b-447e-8568-92a193a69ade-metrics-tls\") pod \"dns-operator-744455d44c-q222b\" (UID: \"63636924-a28b-447e-8568-92a193a69ade\") " pod="openshift-dns-operator/dns-operator-744455d44c-q222b" Oct 08 18:19:33 crc kubenswrapper[4859]: I1008 18:19:33.104416 4859 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/4afad99b-9bf7-43c9-8dcf-7a3217136be5-bound-sa-token\") pod \"image-registry-697d97f7c8-kg44m\" (UID: \"4afad99b-9bf7-43c9-8dcf-7a3217136be5\") " pod="openshift-image-registry/image-registry-697d97f7c8-kg44m" Oct 08 18:19:33 crc kubenswrapper[4859]: I1008 18:19:33.104464 4859 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-dzck7\" (UniqueName: \"kubernetes.io/projected/be85b712-b3e4-4535-b14b-672605d02387-kube-api-access-dzck7\") pod \"catalog-operator-68c6474976-xpnsv\" (UID: \"be85b712-b3e4-4535-b14b-672605d02387\") " pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-xpnsv" Oct 08 18:19:33 crc kubenswrapper[4859]: I1008 18:19:33.104498 4859 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"package-server-manager-serving-cert\" (UniqueName: \"kubernetes.io/secret/c073e5c1-652c-406a-ad01-9ca626e8c56e-package-server-manager-serving-cert\") pod \"package-server-manager-789f6589d5-hf2fp\" (UID: \"c073e5c1-652c-406a-ad01-9ca626e8c56e\") " pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-hf2fp" Oct 08 18:19:33 crc kubenswrapper[4859]: I1008 18:19:33.104531 4859 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7e23494c-2895-43b3-a35c-47a7b2d33ebc-config\") pod \"openshift-kube-scheduler-operator-5fdd9b5758-5f4q5\" (UID: \"7e23494c-2895-43b3-a35c-47a7b2d33ebc\") " pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-5f4q5" Oct 08 18:19:33 crc kubenswrapper[4859]: I1008 18:19:33.104566 4859 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/4afad99b-9bf7-43c9-8dcf-7a3217136be5-installation-pull-secrets\") pod \"image-registry-697d97f7c8-kg44m\" (UID: \"4afad99b-9bf7-43c9-8dcf-7a3217136be5\") " pod="openshift-image-registry/image-registry-697d97f7c8-kg44m" Oct 08 18:19:33 crc kubenswrapper[4859]: I1008 18:19:33.104591 4859 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/46a9f481-0772-4006-99e2-723919011659-serving-cert\") pod \"kube-storage-version-migrator-operator-b67b599dd-6dlgn\" (UID: \"46a9f481-0772-4006-99e2-723919011659\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-6dlgn" Oct 08 18:19:33 crc kubenswrapper[4859]: I1008 18:19:33.104625 4859 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/ed062c43-61ec-4951-a27c-77f165523d3f-trusted-ca\") pod \"ingress-operator-5b745b69d9-pf7j6\" (UID: \"ed062c43-61ec-4951-a27c-77f165523d3f\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-pf7j6" Oct 08 18:19:33 crc kubenswrapper[4859]: I1008 18:19:33.104646 4859 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-zcgth\" (UniqueName: \"kubernetes.io/projected/2a959183-1370-4723-bb5e-8625689d6121-kube-api-access-zcgth\") pod \"dns-default-x7pp9\" (UID: \"2a959183-1370-4723-bb5e-8625689d6121\") " pod="openshift-dns/dns-default-x7pp9" Oct 08 18:19:33 crc kubenswrapper[4859]: I1008 18:19:33.104667 4859 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/be85b712-b3e4-4535-b14b-672605d02387-srv-cert\") pod \"catalog-operator-68c6474976-xpnsv\" (UID: \"be85b712-b3e4-4535-b14b-672605d02387\") " pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-xpnsv" Oct 08 18:19:33 crc kubenswrapper[4859]: I1008 18:19:33.104709 4859 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/7e23494c-2895-43b3-a35c-47a7b2d33ebc-kube-api-access\") pod \"openshift-kube-scheduler-operator-5fdd9b5758-5f4q5\" (UID: \"7e23494c-2895-43b3-a35c-47a7b2d33ebc\") " pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-5f4q5" Oct 08 18:19:33 crc kubenswrapper[4859]: I1008 18:19:33.104730 4859 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/2a959183-1370-4723-bb5e-8625689d6121-config-volume\") pod \"dns-default-x7pp9\" (UID: \"2a959183-1370-4723-bb5e-8625689d6121\") " pod="openshift-dns/dns-default-x7pp9" Oct 08 18:19:33 crc kubenswrapper[4859]: I1008 18:19:33.104750 4859 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/be85b712-b3e4-4535-b14b-672605d02387-profile-collector-cert\") pod \"catalog-operator-68c6474976-xpnsv\" (UID: \"be85b712-b3e4-4535-b14b-672605d02387\") " pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-xpnsv" Oct 08 18:19:33 crc kubenswrapper[4859]: I1008 18:19:33.104802 4859 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-kg44m\" (UID: \"4afad99b-9bf7-43c9-8dcf-7a3217136be5\") " pod="openshift-image-registry/image-registry-697d97f7c8-kg44m" Oct 08 18:19:33 crc kubenswrapper[4859]: E1008 18:19:33.105168 4859 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-08 18:19:33.605153739 +0000 UTC m=+143.851993118 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-kg44m" (UID: "4afad99b-9bf7-43c9-8dcf-7a3217136be5") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 08 18:19:33 crc kubenswrapper[4859]: I1008 18:19:33.133832 4859 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/machine-config-operator-74547568cd-xsc6l" Oct 08 18:19:33 crc kubenswrapper[4859]: I1008 18:19:33.142488 4859 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-6576b87f9c-xlfdd"] Oct 08 18:19:33 crc kubenswrapper[4859]: I1008 18:19:33.163986 4859 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-82pcg"] Oct 08 18:19:33 crc kubenswrapper[4859]: I1008 18:19:33.206269 4859 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 08 18:19:33 crc kubenswrapper[4859]: I1008 18:19:33.206441 4859 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"csi-data-dir\" (UniqueName: \"kubernetes.io/host-path/5e418cbb-0072-4767-933d-537c4a274e60-csi-data-dir\") pod \"csi-hostpathplugin-q4lkb\" (UID: \"5e418cbb-0072-4767-933d-537c4a274e60\") " pod="hostpath-provisioner/csi-hostpathplugin-q4lkb" Oct 08 18:19:33 crc kubenswrapper[4859]: E1008 18:19:33.206559 4859 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-08 18:19:33.706535854 +0000 UTC m=+143.953375233 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 08 18:19:33 crc kubenswrapper[4859]: I1008 18:19:33.206742 4859 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-lpdjc\" (UniqueName: \"kubernetes.io/projected/492ef569-6d45-4dd0-89e5-506172689768-kube-api-access-lpdjc\") pod \"control-plane-machine-set-operator-78cbb6b69f-q9nqj\" (UID: \"492ef569-6d45-4dd0-89e5-506172689768\") " pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-q9nqj" Oct 08 18:19:33 crc kubenswrapper[4859]: I1008 18:19:33.206796 4859 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/4afad99b-9bf7-43c9-8dcf-7a3217136be5-registry-tls\") pod \"image-registry-697d97f7c8-kg44m\" (UID: \"4afad99b-9bf7-43c9-8dcf-7a3217136be5\") " pod="openshift-image-registry/image-registry-697d97f7c8-kg44m" Oct 08 18:19:33 crc kubenswrapper[4859]: I1008 18:19:33.206879 4859 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/46a9f481-0772-4006-99e2-723919011659-config\") pod \"kube-storage-version-migrator-operator-b67b599dd-6dlgn\" (UID: \"46a9f481-0772-4006-99e2-723919011659\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-6dlgn" Oct 08 18:19:33 crc kubenswrapper[4859]: I1008 18:19:33.206906 4859 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-59lvs\" (UniqueName: \"kubernetes.io/projected/63636924-a28b-447e-8568-92a193a69ade-kube-api-access-59lvs\") pod \"dns-operator-744455d44c-q222b\" (UID: \"63636924-a28b-447e-8568-92a193a69ade\") " pod="openshift-dns-operator/dns-operator-744455d44c-q222b" Oct 08 18:19:33 crc kubenswrapper[4859]: I1008 18:19:33.206927 4859 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-f5mqx\" (UniqueName: \"kubernetes.io/projected/5e418cbb-0072-4767-933d-537c4a274e60-kube-api-access-f5mqx\") pod \"csi-hostpathplugin-q4lkb\" (UID: \"5e418cbb-0072-4767-933d-537c4a274e60\") " pod="hostpath-provisioner/csi-hostpathplugin-q4lkb" Oct 08 18:19:33 crc kubenswrapper[4859]: I1008 18:19:33.206976 4859 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/ed062c43-61ec-4951-a27c-77f165523d3f-bound-sa-token\") pod \"ingress-operator-5b745b69d9-pf7j6\" (UID: \"ed062c43-61ec-4951-a27c-77f165523d3f\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-pf7j6" Oct 08 18:19:33 crc kubenswrapper[4859]: I1008 18:19:33.207017 4859 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-ls2m9\" (UniqueName: \"kubernetes.io/projected/46a9f481-0772-4006-99e2-723919011659-kube-api-access-ls2m9\") pod \"kube-storage-version-migrator-operator-b67b599dd-6dlgn\" (UID: \"46a9f481-0772-4006-99e2-723919011659\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-6dlgn" Oct 08 18:19:33 crc kubenswrapper[4859]: I1008 18:19:33.207053 4859 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"plugins-dir\" (UniqueName: \"kubernetes.io/host-path/5e418cbb-0072-4767-933d-537c4a274e60-plugins-dir\") pod \"csi-hostpathplugin-q4lkb\" (UID: \"5e418cbb-0072-4767-933d-537c4a274e60\") " pod="hostpath-provisioner/csi-hostpathplugin-q4lkb" Oct 08 18:19:33 crc kubenswrapper[4859]: I1008 18:19:33.207075 4859 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"control-plane-machine-set-operator-tls\" (UniqueName: \"kubernetes.io/secret/492ef569-6d45-4dd0-89e5-506172689768-control-plane-machine-set-operator-tls\") pod \"control-plane-machine-set-operator-78cbb6b69f-q9nqj\" (UID: \"492ef569-6d45-4dd0-89e5-506172689768\") " pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-q9nqj" Oct 08 18:19:33 crc kubenswrapper[4859]: I1008 18:19:33.207135 4859 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/8b8b079c-490b-4fb1-93a6-473855ca319f-cert\") pod \"ingress-canary-qtvkr\" (UID: \"8b8b079c-490b-4fb1-93a6-473855ca319f\") " pod="openshift-ingress-canary/ingress-canary-qtvkr" Oct 08 18:19:33 crc kubenswrapper[4859]: I1008 18:19:33.207171 4859 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/4afad99b-9bf7-43c9-8dcf-7a3217136be5-ca-trust-extracted\") pod \"image-registry-697d97f7c8-kg44m\" (UID: \"4afad99b-9bf7-43c9-8dcf-7a3217136be5\") " pod="openshift-image-registry/image-registry-697d97f7c8-kg44m" Oct 08 18:19:33 crc kubenswrapper[4859]: I1008 18:19:33.207194 4859 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rvtkc\" (UniqueName: \"kubernetes.io/projected/ed062c43-61ec-4951-a27c-77f165523d3f-kube-api-access-rvtkc\") pod \"ingress-operator-5b745b69d9-pf7j6\" (UID: \"ed062c43-61ec-4951-a27c-77f165523d3f\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-pf7j6" Oct 08 18:19:33 crc kubenswrapper[4859]: I1008 18:19:33.207240 4859 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-hkbdd\" (UniqueName: \"kubernetes.io/projected/c073e5c1-652c-406a-ad01-9ca626e8c56e-kube-api-access-hkbdd\") pod \"package-server-manager-789f6589d5-hf2fp\" (UID: \"c073e5c1-652c-406a-ad01-9ca626e8c56e\") " pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-hf2fp" Oct 08 18:19:33 crc kubenswrapper[4859]: I1008 18:19:33.207277 4859 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"socket-dir\" (UniqueName: \"kubernetes.io/host-path/5e418cbb-0072-4767-933d-537c4a274e60-socket-dir\") pod \"csi-hostpathplugin-q4lkb\" (UID: \"5e418cbb-0072-4767-933d-537c4a274e60\") " pod="hostpath-provisioner/csi-hostpathplugin-q4lkb" Oct 08 18:19:33 crc kubenswrapper[4859]: I1008 18:19:33.207300 4859 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cs2l8\" (UniqueName: \"kubernetes.io/projected/4afad99b-9bf7-43c9-8dcf-7a3217136be5-kube-api-access-cs2l8\") pod \"image-registry-697d97f7c8-kg44m\" (UID: \"4afad99b-9bf7-43c9-8dcf-7a3217136be5\") " pod="openshift-image-registry/image-registry-697d97f7c8-kg44m" Oct 08 18:19:33 crc kubenswrapper[4859]: I1008 18:19:33.207362 4859 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/ed062c43-61ec-4951-a27c-77f165523d3f-metrics-tls\") pod \"ingress-operator-5b745b69d9-pf7j6\" (UID: \"ed062c43-61ec-4951-a27c-77f165523d3f\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-pf7j6" Oct 08 18:19:33 crc kubenswrapper[4859]: I1008 18:19:33.207500 4859 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"registration-dir\" (UniqueName: \"kubernetes.io/host-path/5e418cbb-0072-4767-933d-537c4a274e60-registration-dir\") pod \"csi-hostpathplugin-q4lkb\" (UID: \"5e418cbb-0072-4767-933d-537c4a274e60\") " pod="hostpath-provisioner/csi-hostpathplugin-q4lkb" Oct 08 18:19:33 crc kubenswrapper[4859]: I1008 18:19:33.207563 4859 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/7e23494c-2895-43b3-a35c-47a7b2d33ebc-serving-cert\") pod \"openshift-kube-scheduler-operator-5fdd9b5758-5f4q5\" (UID: \"7e23494c-2895-43b3-a35c-47a7b2d33ebc\") " pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-5f4q5" Oct 08 18:19:33 crc kubenswrapper[4859]: I1008 18:19:33.207585 4859 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/4afad99b-9bf7-43c9-8dcf-7a3217136be5-registry-certificates\") pod \"image-registry-697d97f7c8-kg44m\" (UID: \"4afad99b-9bf7-43c9-8dcf-7a3217136be5\") " pod="openshift-image-registry/image-registry-697d97f7c8-kg44m" Oct 08 18:19:33 crc kubenswrapper[4859]: I1008 18:19:33.207725 4859 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/4afad99b-9bf7-43c9-8dcf-7a3217136be5-bound-sa-token\") pod \"image-registry-697d97f7c8-kg44m\" (UID: \"4afad99b-9bf7-43c9-8dcf-7a3217136be5\") " pod="openshift-image-registry/image-registry-697d97f7c8-kg44m" Oct 08 18:19:33 crc kubenswrapper[4859]: I1008 18:19:33.207747 4859 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/63636924-a28b-447e-8568-92a193a69ade-metrics-tls\") pod \"dns-operator-744455d44c-q222b\" (UID: \"63636924-a28b-447e-8568-92a193a69ade\") " pod="openshift-dns-operator/dns-operator-744455d44c-q222b" Oct 08 18:19:33 crc kubenswrapper[4859]: I1008 18:19:33.207808 4859 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-dzck7\" (UniqueName: \"kubernetes.io/projected/be85b712-b3e4-4535-b14b-672605d02387-kube-api-access-dzck7\") pod \"catalog-operator-68c6474976-xpnsv\" (UID: \"be85b712-b3e4-4535-b14b-672605d02387\") " pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-xpnsv" Oct 08 18:19:33 crc kubenswrapper[4859]: I1008 18:19:33.207832 4859 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"package-server-manager-serving-cert\" (UniqueName: \"kubernetes.io/secret/c073e5c1-652c-406a-ad01-9ca626e8c56e-package-server-manager-serving-cert\") pod \"package-server-manager-789f6589d5-hf2fp\" (UID: \"c073e5c1-652c-406a-ad01-9ca626e8c56e\") " pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-hf2fp" Oct 08 18:19:33 crc kubenswrapper[4859]: I1008 18:19:33.207922 4859 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7e23494c-2895-43b3-a35c-47a7b2d33ebc-config\") pod \"openshift-kube-scheduler-operator-5fdd9b5758-5f4q5\" (UID: \"7e23494c-2895-43b3-a35c-47a7b2d33ebc\") " pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-5f4q5" Oct 08 18:19:33 crc kubenswrapper[4859]: I1008 18:19:33.207963 4859 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/4afad99b-9bf7-43c9-8dcf-7a3217136be5-installation-pull-secrets\") pod \"image-registry-697d97f7c8-kg44m\" (UID: \"4afad99b-9bf7-43c9-8dcf-7a3217136be5\") " pod="openshift-image-registry/image-registry-697d97f7c8-kg44m" Oct 08 18:19:33 crc kubenswrapper[4859]: I1008 18:19:33.207980 4859 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/46a9f481-0772-4006-99e2-723919011659-serving-cert\") pod \"kube-storage-version-migrator-operator-b67b599dd-6dlgn\" (UID: \"46a9f481-0772-4006-99e2-723919011659\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-6dlgn" Oct 08 18:19:33 crc kubenswrapper[4859]: I1008 18:19:33.208042 4859 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/ed062c43-61ec-4951-a27c-77f165523d3f-trusted-ca\") pod \"ingress-operator-5b745b69d9-pf7j6\" (UID: \"ed062c43-61ec-4951-a27c-77f165523d3f\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-pf7j6" Oct 08 18:19:33 crc kubenswrapper[4859]: I1008 18:19:33.208058 4859 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-zcgth\" (UniqueName: \"kubernetes.io/projected/2a959183-1370-4723-bb5e-8625689d6121-kube-api-access-zcgth\") pod \"dns-default-x7pp9\" (UID: \"2a959183-1370-4723-bb5e-8625689d6121\") " pod="openshift-dns/dns-default-x7pp9" Oct 08 18:19:33 crc kubenswrapper[4859]: I1008 18:19:33.208085 4859 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/be85b712-b3e4-4535-b14b-672605d02387-srv-cert\") pod \"catalog-operator-68c6474976-xpnsv\" (UID: \"be85b712-b3e4-4535-b14b-672605d02387\") " pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-xpnsv" Oct 08 18:19:33 crc kubenswrapper[4859]: I1008 18:19:33.208120 4859 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/7e23494c-2895-43b3-a35c-47a7b2d33ebc-kube-api-access\") pod \"openshift-kube-scheduler-operator-5fdd9b5758-5f4q5\" (UID: \"7e23494c-2895-43b3-a35c-47a7b2d33ebc\") " pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-5f4q5" Oct 08 18:19:33 crc kubenswrapper[4859]: I1008 18:19:33.208136 4859 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/2a959183-1370-4723-bb5e-8625689d6121-config-volume\") pod \"dns-default-x7pp9\" (UID: \"2a959183-1370-4723-bb5e-8625689d6121\") " pod="openshift-dns/dns-default-x7pp9" Oct 08 18:19:33 crc kubenswrapper[4859]: I1008 18:19:33.208171 4859 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/be85b712-b3e4-4535-b14b-672605d02387-profile-collector-cert\") pod \"catalog-operator-68c6474976-xpnsv\" (UID: \"be85b712-b3e4-4535-b14b-672605d02387\") " pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-xpnsv" Oct 08 18:19:33 crc kubenswrapper[4859]: I1008 18:19:33.208214 4859 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-kg44m\" (UID: \"4afad99b-9bf7-43c9-8dcf-7a3217136be5\") " pod="openshift-image-registry/image-registry-697d97f7c8-kg44m" Oct 08 18:19:33 crc kubenswrapper[4859]: I1008 18:19:33.208230 4859 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/2a959183-1370-4723-bb5e-8625689d6121-metrics-tls\") pod \"dns-default-x7pp9\" (UID: \"2a959183-1370-4723-bb5e-8625689d6121\") " pod="openshift-dns/dns-default-x7pp9" Oct 08 18:19:33 crc kubenswrapper[4859]: I1008 18:19:33.208273 4859 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"mountpoint-dir\" (UniqueName: \"kubernetes.io/host-path/5e418cbb-0072-4767-933d-537c4a274e60-mountpoint-dir\") pod \"csi-hostpathplugin-q4lkb\" (UID: \"5e418cbb-0072-4767-933d-537c4a274e60\") " pod="hostpath-provisioner/csi-hostpathplugin-q4lkb" Oct 08 18:19:33 crc kubenswrapper[4859]: I1008 18:19:33.208361 4859 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/4afad99b-9bf7-43c9-8dcf-7a3217136be5-trusted-ca\") pod \"image-registry-697d97f7c8-kg44m\" (UID: \"4afad99b-9bf7-43c9-8dcf-7a3217136be5\") " pod="openshift-image-registry/image-registry-697d97f7c8-kg44m" Oct 08 18:19:33 crc kubenswrapper[4859]: I1008 18:19:33.208451 4859 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-w7vgd\" (UniqueName: \"kubernetes.io/projected/8b8b079c-490b-4fb1-93a6-473855ca319f-kube-api-access-w7vgd\") pod \"ingress-canary-qtvkr\" (UID: \"8b8b079c-490b-4fb1-93a6-473855ca319f\") " pod="openshift-ingress-canary/ingress-canary-qtvkr" Oct 08 18:19:33 crc kubenswrapper[4859]: I1008 18:19:33.211270 4859 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/4afad99b-9bf7-43c9-8dcf-7a3217136be5-trusted-ca\") pod \"image-registry-697d97f7c8-kg44m\" (UID: \"4afad99b-9bf7-43c9-8dcf-7a3217136be5\") " pod="openshift-image-registry/image-registry-697d97f7c8-kg44m" Oct 08 18:19:33 crc kubenswrapper[4859]: I1008 18:19:33.211344 4859 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7e23494c-2895-43b3-a35c-47a7b2d33ebc-config\") pod \"openshift-kube-scheduler-operator-5fdd9b5758-5f4q5\" (UID: \"7e23494c-2895-43b3-a35c-47a7b2d33ebc\") " pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-5f4q5" Oct 08 18:19:33 crc kubenswrapper[4859]: E1008 18:19:33.212340 4859 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-08 18:19:33.712325077 +0000 UTC m=+143.959164456 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-kg44m" (UID: "4afad99b-9bf7-43c9-8dcf-7a3217136be5") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 08 18:19:33 crc kubenswrapper[4859]: I1008 18:19:33.212471 4859 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/2a959183-1370-4723-bb5e-8625689d6121-config-volume\") pod \"dns-default-x7pp9\" (UID: \"2a959183-1370-4723-bb5e-8625689d6121\") " pod="openshift-dns/dns-default-x7pp9" Oct 08 18:19:33 crc kubenswrapper[4859]: I1008 18:19:33.215166 4859 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/4afad99b-9bf7-43c9-8dcf-7a3217136be5-ca-trust-extracted\") pod \"image-registry-697d97f7c8-kg44m\" (UID: \"4afad99b-9bf7-43c9-8dcf-7a3217136be5\") " pod="openshift-image-registry/image-registry-697d97f7c8-kg44m" Oct 08 18:19:33 crc kubenswrapper[4859]: I1008 18:19:33.216974 4859 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/ed062c43-61ec-4951-a27c-77f165523d3f-trusted-ca\") pod \"ingress-operator-5b745b69d9-pf7j6\" (UID: \"ed062c43-61ec-4951-a27c-77f165523d3f\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-pf7j6" Oct 08 18:19:33 crc kubenswrapper[4859]: I1008 18:19:33.226025 4859 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/63636924-a28b-447e-8568-92a193a69ade-metrics-tls\") pod \"dns-operator-744455d44c-q222b\" (UID: \"63636924-a28b-447e-8568-92a193a69ade\") " pod="openshift-dns-operator/dns-operator-744455d44c-q222b" Oct 08 18:19:33 crc kubenswrapper[4859]: I1008 18:19:33.231048 4859 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/46a9f481-0772-4006-99e2-723919011659-config\") pod \"kube-storage-version-migrator-operator-b67b599dd-6dlgn\" (UID: \"46a9f481-0772-4006-99e2-723919011659\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-6dlgn" Oct 08 18:19:33 crc kubenswrapper[4859]: I1008 18:19:33.231135 4859 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/4afad99b-9bf7-43c9-8dcf-7a3217136be5-registry-tls\") pod \"image-registry-697d97f7c8-kg44m\" (UID: \"4afad99b-9bf7-43c9-8dcf-7a3217136be5\") " pod="openshift-image-registry/image-registry-697d97f7c8-kg44m" Oct 08 18:19:33 crc kubenswrapper[4859]: I1008 18:19:33.232250 4859 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/be85b712-b3e4-4535-b14b-672605d02387-profile-collector-cert\") pod \"catalog-operator-68c6474976-xpnsv\" (UID: \"be85b712-b3e4-4535-b14b-672605d02387\") " pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-xpnsv" Oct 08 18:19:33 crc kubenswrapper[4859]: I1008 18:19:33.234737 4859 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/46a9f481-0772-4006-99e2-723919011659-serving-cert\") pod \"kube-storage-version-migrator-operator-b67b599dd-6dlgn\" (UID: \"46a9f481-0772-4006-99e2-723919011659\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-6dlgn" Oct 08 18:19:33 crc kubenswrapper[4859]: I1008 18:19:33.235782 4859 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-c5p45"] Oct 08 18:19:33 crc kubenswrapper[4859]: I1008 18:19:33.236197 4859 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/7e23494c-2895-43b3-a35c-47a7b2d33ebc-serving-cert\") pod \"openshift-kube-scheduler-operator-5fdd9b5758-5f4q5\" (UID: \"7e23494c-2895-43b3-a35c-47a7b2d33ebc\") " pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-5f4q5" Oct 08 18:19:33 crc kubenswrapper[4859]: I1008 18:19:33.236592 4859 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"package-server-manager-serving-cert\" (UniqueName: \"kubernetes.io/secret/c073e5c1-652c-406a-ad01-9ca626e8c56e-package-server-manager-serving-cert\") pod \"package-server-manager-789f6589d5-hf2fp\" (UID: \"c073e5c1-652c-406a-ad01-9ca626e8c56e\") " pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-hf2fp" Oct 08 18:19:33 crc kubenswrapper[4859]: I1008 18:19:33.239350 4859 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-multus/multus-admission-controller-857f4d67dd-lrqkp"] Oct 08 18:19:33 crc kubenswrapper[4859]: I1008 18:19:33.242722 4859 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/4afad99b-9bf7-43c9-8dcf-7a3217136be5-registry-certificates\") pod \"image-registry-697d97f7c8-kg44m\" (UID: \"4afad99b-9bf7-43c9-8dcf-7a3217136be5\") " pod="openshift-image-registry/image-registry-697d97f7c8-kg44m" Oct 08 18:19:33 crc kubenswrapper[4859]: I1008 18:19:33.248468 4859 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/be85b712-b3e4-4535-b14b-672605d02387-srv-cert\") pod \"catalog-operator-68c6474976-xpnsv\" (UID: \"be85b712-b3e4-4535-b14b-672605d02387\") " pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-xpnsv" Oct 08 18:19:33 crc kubenswrapper[4859]: I1008 18:19:33.249616 4859 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/4afad99b-9bf7-43c9-8dcf-7a3217136be5-installation-pull-secrets\") pod \"image-registry-697d97f7c8-kg44m\" (UID: \"4afad99b-9bf7-43c9-8dcf-7a3217136be5\") " pod="openshift-image-registry/image-registry-697d97f7c8-kg44m" Oct 08 18:19:33 crc kubenswrapper[4859]: I1008 18:19:33.260781 4859 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-authentication-operator/authentication-operator-69f744f599-jd5kh"] Oct 08 18:19:33 crc kubenswrapper[4859]: I1008 18:19:33.262158 4859 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/2a959183-1370-4723-bb5e-8625689d6121-metrics-tls\") pod \"dns-default-x7pp9\" (UID: \"2a959183-1370-4723-bb5e-8625689d6121\") " pod="openshift-dns/dns-default-x7pp9" Oct 08 18:19:33 crc kubenswrapper[4859]: I1008 18:19:33.267400 4859 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-59lvs\" (UniqueName: \"kubernetes.io/projected/63636924-a28b-447e-8568-92a193a69ade-kube-api-access-59lvs\") pod \"dns-operator-744455d44c-q222b\" (UID: \"63636924-a28b-447e-8568-92a193a69ade\") " pod="openshift-dns-operator/dns-operator-744455d44c-q222b" Oct 08 18:19:33 crc kubenswrapper[4859]: I1008 18:19:33.291662 4859 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/ed062c43-61ec-4951-a27c-77f165523d3f-metrics-tls\") pod \"ingress-operator-5b745b69d9-pf7j6\" (UID: \"ed062c43-61ec-4951-a27c-77f165523d3f\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-pf7j6" Oct 08 18:19:33 crc kubenswrapper[4859]: I1008 18:19:33.296817 4859 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-czj5g"] Oct 08 18:19:33 crc kubenswrapper[4859]: I1008 18:19:33.298326 4859 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-cs2l8\" (UniqueName: \"kubernetes.io/projected/4afad99b-9bf7-43c9-8dcf-7a3217136be5-kube-api-access-cs2l8\") pod \"image-registry-697d97f7c8-kg44m\" (UID: \"4afad99b-9bf7-43c9-8dcf-7a3217136be5\") " pod="openshift-image-registry/image-registry-697d97f7c8-kg44m" Oct 08 18:19:33 crc kubenswrapper[4859]: I1008 18:19:33.302800 4859 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-etcd-operator/etcd-operator-b45778765-tnn6l"] Oct 08 18:19:33 crc kubenswrapper[4859]: I1008 18:19:33.302875 4859 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-console/console-f9d7485db-m2xmp"] Oct 08 18:19:33 crc kubenswrapper[4859]: I1008 18:19:33.303867 4859 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-ls2m9\" (UniqueName: \"kubernetes.io/projected/46a9f481-0772-4006-99e2-723919011659-kube-api-access-ls2m9\") pod \"kube-storage-version-migrator-operator-b67b599dd-6dlgn\" (UID: \"46a9f481-0772-4006-99e2-723919011659\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-6dlgn" Oct 08 18:19:33 crc kubenswrapper[4859]: I1008 18:19:33.309634 4859 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 08 18:19:33 crc kubenswrapper[4859]: E1008 18:19:33.309902 4859 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-08 18:19:33.809885418 +0000 UTC m=+144.056724797 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 08 18:19:33 crc kubenswrapper[4859]: I1008 18:19:33.309980 4859 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-lpdjc\" (UniqueName: \"kubernetes.io/projected/492ef569-6d45-4dd0-89e5-506172689768-kube-api-access-lpdjc\") pod \"control-plane-machine-set-operator-78cbb6b69f-q9nqj\" (UID: \"492ef569-6d45-4dd0-89e5-506172689768\") " pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-q9nqj" Oct 08 18:19:33 crc kubenswrapper[4859]: I1008 18:19:33.310013 4859 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-f5mqx\" (UniqueName: \"kubernetes.io/projected/5e418cbb-0072-4767-933d-537c4a274e60-kube-api-access-f5mqx\") pod \"csi-hostpathplugin-q4lkb\" (UID: \"5e418cbb-0072-4767-933d-537c4a274e60\") " pod="hostpath-provisioner/csi-hostpathplugin-q4lkb" Oct 08 18:19:33 crc kubenswrapper[4859]: I1008 18:19:33.310050 4859 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"plugins-dir\" (UniqueName: \"kubernetes.io/host-path/5e418cbb-0072-4767-933d-537c4a274e60-plugins-dir\") pod \"csi-hostpathplugin-q4lkb\" (UID: \"5e418cbb-0072-4767-933d-537c4a274e60\") " pod="hostpath-provisioner/csi-hostpathplugin-q4lkb" Oct 08 18:19:33 crc kubenswrapper[4859]: I1008 18:19:33.310068 4859 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"control-plane-machine-set-operator-tls\" (UniqueName: \"kubernetes.io/secret/492ef569-6d45-4dd0-89e5-506172689768-control-plane-machine-set-operator-tls\") pod \"control-plane-machine-set-operator-78cbb6b69f-q9nqj\" (UID: \"492ef569-6d45-4dd0-89e5-506172689768\") " pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-q9nqj" Oct 08 18:19:33 crc kubenswrapper[4859]: I1008 18:19:33.310093 4859 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/8b8b079c-490b-4fb1-93a6-473855ca319f-cert\") pod \"ingress-canary-qtvkr\" (UID: \"8b8b079c-490b-4fb1-93a6-473855ca319f\") " pod="openshift-ingress-canary/ingress-canary-qtvkr" Oct 08 18:19:33 crc kubenswrapper[4859]: I1008 18:19:33.310129 4859 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"socket-dir\" (UniqueName: \"kubernetes.io/host-path/5e418cbb-0072-4767-933d-537c4a274e60-socket-dir\") pod \"csi-hostpathplugin-q4lkb\" (UID: \"5e418cbb-0072-4767-933d-537c4a274e60\") " pod="hostpath-provisioner/csi-hostpathplugin-q4lkb" Oct 08 18:19:33 crc kubenswrapper[4859]: I1008 18:19:33.310158 4859 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"registration-dir\" (UniqueName: \"kubernetes.io/host-path/5e418cbb-0072-4767-933d-537c4a274e60-registration-dir\") pod \"csi-hostpathplugin-q4lkb\" (UID: \"5e418cbb-0072-4767-933d-537c4a274e60\") " pod="hostpath-provisioner/csi-hostpathplugin-q4lkb" Oct 08 18:19:33 crc kubenswrapper[4859]: I1008 18:19:33.310229 4859 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-kg44m\" (UID: \"4afad99b-9bf7-43c9-8dcf-7a3217136be5\") " pod="openshift-image-registry/image-registry-697d97f7c8-kg44m" Oct 08 18:19:33 crc kubenswrapper[4859]: I1008 18:19:33.310247 4859 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"mountpoint-dir\" (UniqueName: \"kubernetes.io/host-path/5e418cbb-0072-4767-933d-537c4a274e60-mountpoint-dir\") pod \"csi-hostpathplugin-q4lkb\" (UID: \"5e418cbb-0072-4767-933d-537c4a274e60\") " pod="hostpath-provisioner/csi-hostpathplugin-q4lkb" Oct 08 18:19:33 crc kubenswrapper[4859]: I1008 18:19:33.310278 4859 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-w7vgd\" (UniqueName: \"kubernetes.io/projected/8b8b079c-490b-4fb1-93a6-473855ca319f-kube-api-access-w7vgd\") pod \"ingress-canary-qtvkr\" (UID: \"8b8b079c-490b-4fb1-93a6-473855ca319f\") " pod="openshift-ingress-canary/ingress-canary-qtvkr" Oct 08 18:19:33 crc kubenswrapper[4859]: I1008 18:19:33.310293 4859 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"csi-data-dir\" (UniqueName: \"kubernetes.io/host-path/5e418cbb-0072-4767-933d-537c4a274e60-csi-data-dir\") pod \"csi-hostpathplugin-q4lkb\" (UID: \"5e418cbb-0072-4767-933d-537c4a274e60\") " pod="hostpath-provisioner/csi-hostpathplugin-q4lkb" Oct 08 18:19:33 crc kubenswrapper[4859]: I1008 18:19:33.310412 4859 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"csi-data-dir\" (UniqueName: \"kubernetes.io/host-path/5e418cbb-0072-4767-933d-537c4a274e60-csi-data-dir\") pod \"csi-hostpathplugin-q4lkb\" (UID: \"5e418cbb-0072-4767-933d-537c4a274e60\") " pod="hostpath-provisioner/csi-hostpathplugin-q4lkb" Oct 08 18:19:33 crc kubenswrapper[4859]: I1008 18:19:33.310812 4859 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"plugins-dir\" (UniqueName: \"kubernetes.io/host-path/5e418cbb-0072-4767-933d-537c4a274e60-plugins-dir\") pod \"csi-hostpathplugin-q4lkb\" (UID: \"5e418cbb-0072-4767-933d-537c4a274e60\") " pod="hostpath-provisioner/csi-hostpathplugin-q4lkb" Oct 08 18:19:33 crc kubenswrapper[4859]: I1008 18:19:33.311106 4859 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"mountpoint-dir\" (UniqueName: \"kubernetes.io/host-path/5e418cbb-0072-4767-933d-537c4a274e60-mountpoint-dir\") pod \"csi-hostpathplugin-q4lkb\" (UID: \"5e418cbb-0072-4767-933d-537c4a274e60\") " pod="hostpath-provisioner/csi-hostpathplugin-q4lkb" Oct 08 18:19:33 crc kubenswrapper[4859]: I1008 18:19:33.311165 4859 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"registration-dir\" (UniqueName: \"kubernetes.io/host-path/5e418cbb-0072-4767-933d-537c4a274e60-registration-dir\") pod \"csi-hostpathplugin-q4lkb\" (UID: \"5e418cbb-0072-4767-933d-537c4a274e60\") " pod="hostpath-provisioner/csi-hostpathplugin-q4lkb" Oct 08 18:19:33 crc kubenswrapper[4859]: E1008 18:19:33.311349 4859 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-08 18:19:33.811341401 +0000 UTC m=+144.058180780 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-kg44m" (UID: "4afad99b-9bf7-43c9-8dcf-7a3217136be5") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 08 18:19:33 crc kubenswrapper[4859]: I1008 18:19:33.311876 4859 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"socket-dir\" (UniqueName: \"kubernetes.io/host-path/5e418cbb-0072-4767-933d-537c4a274e60-socket-dir\") pod \"csi-hostpathplugin-q4lkb\" (UID: \"5e418cbb-0072-4767-933d-537c4a274e60\") " pod="hostpath-provisioner/csi-hostpathplugin-q4lkb" Oct 08 18:19:33 crc kubenswrapper[4859]: I1008 18:19:33.314629 4859 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-5n49w"] Oct 08 18:19:33 crc kubenswrapper[4859]: I1008 18:19:33.314724 4859 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-machine-api/machine-api-operator-5694c8668f-gmbk9"] Oct 08 18:19:33 crc kubenswrapper[4859]: I1008 18:19:33.318307 4859 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/ed062c43-61ec-4951-a27c-77f165523d3f-bound-sa-token\") pod \"ingress-operator-5b745b69d9-pf7j6\" (UID: \"ed062c43-61ec-4951-a27c-77f165523d3f\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-pf7j6" Oct 08 18:19:33 crc kubenswrapper[4859]: I1008 18:19:33.318788 4859 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-6dlgn" Oct 08 18:19:33 crc kubenswrapper[4859]: I1008 18:19:33.320081 4859 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert\" (UniqueName: \"kubernetes.io/secret/8b8b079c-490b-4fb1-93a6-473855ca319f-cert\") pod \"ingress-canary-qtvkr\" (UID: \"8b8b079c-490b-4fb1-93a6-473855ca319f\") " pod="openshift-ingress-canary/ingress-canary-qtvkr" Oct 08 18:19:33 crc kubenswrapper[4859]: I1008 18:19:33.321327 4859 generic.go:334] "Generic (PLEG): container finished" podID="65868bd3-b636-48c5-aac4-fcf734f84e39" containerID="908f88db05f6ce7bd489c7d5f0709c05ec7448cbef6699282b83e6154be0f141" exitCode=0 Oct 08 18:19:33 crc kubenswrapper[4859]: I1008 18:19:33.321397 4859 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-config-operator/openshift-config-operator-7777fb866f-ncj6n" event={"ID":"65868bd3-b636-48c5-aac4-fcf734f84e39","Type":"ContainerDied","Data":"908f88db05f6ce7bd489c7d5f0709c05ec7448cbef6699282b83e6154be0f141"} Oct 08 18:19:33 crc kubenswrapper[4859]: I1008 18:19:33.321428 4859 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-config-operator/openshift-config-operator-7777fb866f-ncj6n" event={"ID":"65868bd3-b636-48c5-aac4-fcf734f84e39","Type":"ContainerStarted","Data":"40083dbafda971eb4610c89f7c577055c4f1016bbaad8aba9dfc3c209df0f914"} Oct 08 18:19:33 crc kubenswrapper[4859]: I1008 18:19:33.323137 4859 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"control-plane-machine-set-operator-tls\" (UniqueName: \"kubernetes.io/secret/492ef569-6d45-4dd0-89e5-506172689768-control-plane-machine-set-operator-tls\") pod \"control-plane-machine-set-operator-78cbb6b69f-q9nqj\" (UID: \"492ef569-6d45-4dd0-89e5-506172689768\") " pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-q9nqj" Oct 08 18:19:33 crc kubenswrapper[4859]: I1008 18:19:33.325501 4859 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-service-ca-operator/service-ca-operator-777779d784-847r8"] Oct 08 18:19:33 crc kubenswrapper[4859]: I1008 18:19:33.326343 4859 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-dzck7\" (UniqueName: \"kubernetes.io/projected/be85b712-b3e4-4535-b14b-672605d02387-kube-api-access-dzck7\") pod \"catalog-operator-68c6474976-xpnsv\" (UID: \"be85b712-b3e4-4535-b14b-672605d02387\") " pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-xpnsv" Oct 08 18:19:33 crc kubenswrapper[4859]: I1008 18:19:33.328326 4859 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-cluster-machine-approver/machine-approver-56656f9798-95s6m" event={"ID":"39b30c11-d0f6-4302-9ef8-e7fcf1b2170e","Type":"ContainerStarted","Data":"5e08c2a1cbbe921831a86b0d0153c5edce0e9c9e13f06fd09f20a3cb2756d83f"} Oct 08 18:19:33 crc kubenswrapper[4859]: I1008 18:19:33.329959 4859 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console-operator/console-operator-58897d9998-8sflk" event={"ID":"2736d658-3609-4c30-a010-7b00c6300e2e","Type":"ContainerStarted","Data":"16eeb95934696b67aef514fb9601e547c6ec37ac0e7b550818c5713a85f9f6d4"} Oct 08 18:19:33 crc kubenswrapper[4859]: I1008 18:19:33.330009 4859 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console-operator/console-operator-58897d9998-8sflk" event={"ID":"2736d658-3609-4c30-a010-7b00c6300e2e","Type":"ContainerStarted","Data":"336bd022a1ee8b75ec8eeac3c78a81f1fdb4fd74e135177484856cb11973ee15"} Oct 08 18:19:33 crc kubenswrapper[4859]: I1008 18:19:33.330971 4859 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-console-operator/console-operator-58897d9998-8sflk" Oct 08 18:19:33 crc kubenswrapper[4859]: I1008 18:19:33.337575 4859 patch_prober.go:28] interesting pod/console-operator-58897d9998-8sflk container/console-operator namespace/openshift-console-operator: Readiness probe status=failure output="Get \"https://10.217.0.12:8443/readyz\": dial tcp 10.217.0.12:8443: connect: connection refused" start-of-body= Oct 08 18:19:33 crc kubenswrapper[4859]: I1008 18:19:33.337641 4859 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-console-operator/console-operator-58897d9998-8sflk" podUID="2736d658-3609-4c30-a010-7b00c6300e2e" containerName="console-operator" probeResult="failure" output="Get \"https://10.217.0.12:8443/readyz\": dial tcp 10.217.0.12:8443: connect: connection refused" Oct 08 18:19:33 crc kubenswrapper[4859]: I1008 18:19:33.338566 4859 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-kq7mx" event={"ID":"c79c3857-6ecb-4e3c-b59f-6a58ba0f94e9","Type":"ContainerStarted","Data":"4931c310b2586ff1ddd4bd39a979f5c0bea0b6915cfb62c4eb356203e16ef1aa"} Oct 08 18:19:33 crc kubenswrapper[4859]: I1008 18:19:33.344317 4859 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/4afad99b-9bf7-43c9-8dcf-7a3217136be5-bound-sa-token\") pod \"image-registry-697d97f7c8-kg44m\" (UID: \"4afad99b-9bf7-43c9-8dcf-7a3217136be5\") " pod="openshift-image-registry/image-registry-697d97f7c8-kg44m" Oct 08 18:19:33 crc kubenswrapper[4859]: I1008 18:19:33.347171 4859 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-apiserver/apiserver-76f77b778f-fv9bp" event={"ID":"70f88ce6-b9fe-422e-827e-af2b6840a783","Type":"ContainerStarted","Data":"58b9888d533f693b3f3c8166a8c788e73750cf92ac0a4aa9e7c4aea57e25bddc"} Oct 08 18:19:33 crc kubenswrapper[4859]: I1008 18:19:33.349552 4859 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-879f6c89f-9gpzj" event={"ID":"460ad6e7-859d-4111-a555-b55c5b7228e0","Type":"ContainerStarted","Data":"1e7c718f9f31e46de3067304f0212127211b5576b88d7881e74ede8e5e3004a7"} Oct 08 18:19:33 crc kubenswrapper[4859]: I1008 18:19:33.363912 4859 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-machine-config-operator/machine-config-controller-84d6567774-m7m5n"] Oct 08 18:19:33 crc kubenswrapper[4859]: I1008 18:19:33.365781 4859 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-zcgth\" (UniqueName: \"kubernetes.io/projected/2a959183-1370-4723-bb5e-8625689d6121-kube-api-access-zcgth\") pod \"dns-default-x7pp9\" (UID: \"2a959183-1370-4723-bb5e-8625689d6121\") " pod="openshift-dns/dns-default-x7pp9" Oct 08 18:19:33 crc kubenswrapper[4859]: W1008 18:19:33.368119 4859 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podf114c686_8fde_4560_b8e1_9cd8fea448f8.slice/crio-55703680ce636d8de7afba38777f65e3fd30d78c717fe2a9384bd46927288d87 WatchSource:0}: Error finding container 55703680ce636d8de7afba38777f65e3fd30d78c717fe2a9384bd46927288d87: Status 404 returned error can't find the container with id 55703680ce636d8de7afba38777f65e3fd30d78c717fe2a9384bd46927288d87 Oct 08 18:19:33 crc kubenswrapper[4859]: W1008 18:19:33.369943 4859 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-poda52b9536_1013_482b_9261_e024377687d2.slice/crio-7b6a6c818be35d1bb96148bb9cf45bba330d4e3450e22fe5d4719d65cf5a946b WatchSource:0}: Error finding container 7b6a6c818be35d1bb96148bb9cf45bba330d4e3450e22fe5d4719d65cf5a946b: Status 404 returned error can't find the container with id 7b6a6c818be35d1bb96148bb9cf45bba330d4e3450e22fe5d4719d65cf5a946b Oct 08 18:19:33 crc kubenswrapper[4859]: W1008 18:19:33.375787 4859 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod6f093915_bce6_40eb_917f_ed0058431a79.slice/crio-8f278f321c76e92f6996c0e678fb2cec355fb405d66a86450910a417a63f46bb WatchSource:0}: Error finding container 8f278f321c76e92f6996c0e678fb2cec355fb405d66a86450910a417a63f46bb: Status 404 returned error can't find the container with id 8f278f321c76e92f6996c0e678fb2cec355fb405d66a86450910a417a63f46bb Oct 08 18:19:33 crc kubenswrapper[4859]: I1008 18:19:33.388286 4859 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/7e23494c-2895-43b3-a35c-47a7b2d33ebc-kube-api-access\") pod \"openshift-kube-scheduler-operator-5fdd9b5758-5f4q5\" (UID: \"7e23494c-2895-43b3-a35c-47a7b2d33ebc\") " pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-5f4q5" Oct 08 18:19:33 crc kubenswrapper[4859]: W1008 18:19:33.391460 4859 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pode431ddbe_2d50_4a0a_9679_5565e0df977f.slice/crio-695e3eea1167688cc367354ffb90c737f654fb0904ebfaca4ae1196abff91d90 WatchSource:0}: Error finding container 695e3eea1167688cc367354ffb90c737f654fb0904ebfaca4ae1196abff91d90: Status 404 returned error can't find the container with id 695e3eea1167688cc367354ffb90c737f654fb0904ebfaca4ae1196abff91d90 Oct 08 18:19:33 crc kubenswrapper[4859]: I1008 18:19:33.407772 4859 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rvtkc\" (UniqueName: \"kubernetes.io/projected/ed062c43-61ec-4951-a27c-77f165523d3f-kube-api-access-rvtkc\") pod \"ingress-operator-5b745b69d9-pf7j6\" (UID: \"ed062c43-61ec-4951-a27c-77f165523d3f\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-pf7j6" Oct 08 18:19:33 crc kubenswrapper[4859]: I1008 18:19:33.412979 4859 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 08 18:19:33 crc kubenswrapper[4859]: E1008 18:19:33.413893 4859 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-08 18:19:33.913874031 +0000 UTC m=+144.160713410 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 08 18:19:33 crc kubenswrapper[4859]: I1008 18:19:33.418875 4859 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-dns-operator/dns-operator-744455d44c-q222b" Oct 08 18:19:33 crc kubenswrapper[4859]: I1008 18:19:33.426110 4859 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ingress-operator/ingress-operator-5b745b69d9-pf7j6" Oct 08 18:19:33 crc kubenswrapper[4859]: I1008 18:19:33.428294 4859 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-hkbdd\" (UniqueName: \"kubernetes.io/projected/c073e5c1-652c-406a-ad01-9ca626e8c56e-kube-api-access-hkbdd\") pod \"package-server-manager-789f6589d5-hf2fp\" (UID: \"c073e5c1-652c-406a-ad01-9ca626e8c56e\") " pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-hf2fp" Oct 08 18:19:33 crc kubenswrapper[4859]: I1008 18:19:33.483928 4859 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-lpdjc\" (UniqueName: \"kubernetes.io/projected/492ef569-6d45-4dd0-89e5-506172689768-kube-api-access-lpdjc\") pod \"control-plane-machine-set-operator-78cbb6b69f-q9nqj\" (UID: \"492ef569-6d45-4dd0-89e5-506172689768\") " pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-q9nqj" Oct 08 18:19:33 crc kubenswrapper[4859]: I1008 18:19:33.484187 4859 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-5f4q5" Oct 08 18:19:33 crc kubenswrapper[4859]: I1008 18:19:33.489448 4859 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-f5mqx\" (UniqueName: \"kubernetes.io/projected/5e418cbb-0072-4767-933d-537c4a274e60-kube-api-access-f5mqx\") pod \"csi-hostpathplugin-q4lkb\" (UID: \"5e418cbb-0072-4767-933d-537c4a274e60\") " pod="hostpath-provisioner/csi-hostpathplugin-q4lkb" Oct 08 18:19:33 crc kubenswrapper[4859]: I1008 18:19:33.490199 4859 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-kc42p"] Oct 08 18:19:33 crc kubenswrapper[4859]: I1008 18:19:33.514355 4859 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-kg44m\" (UID: \"4afad99b-9bf7-43c9-8dcf-7a3217136be5\") " pod="openshift-image-registry/image-registry-697d97f7c8-kg44m" Oct 08 18:19:33 crc kubenswrapper[4859]: I1008 18:19:33.516205 4859 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-w7vgd\" (UniqueName: \"kubernetes.io/projected/8b8b079c-490b-4fb1-93a6-473855ca319f-kube-api-access-w7vgd\") pod \"ingress-canary-qtvkr\" (UID: \"8b8b079c-490b-4fb1-93a6-473855ca319f\") " pod="openshift-ingress-canary/ingress-canary-qtvkr" Oct 08 18:19:33 crc kubenswrapper[4859]: I1008 18:19:33.525469 4859 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-4s5w9"] Oct 08 18:19:33 crc kubenswrapper[4859]: E1008 18:19:33.534143 4859 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-08 18:19:34.034121569 +0000 UTC m=+144.280960948 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-kg44m" (UID: "4afad99b-9bf7-43c9-8dcf-7a3217136be5") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 08 18:19:33 crc kubenswrapper[4859]: I1008 18:19:33.590980 4859 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-xpnsv" Oct 08 18:19:33 crc kubenswrapper[4859]: I1008 18:19:33.601301 4859 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-hf2fp" Oct 08 18:19:33 crc kubenswrapper[4859]: I1008 18:19:33.615182 4859 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 08 18:19:33 crc kubenswrapper[4859]: E1008 18:19:33.615367 4859 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-08 18:19:34.115341383 +0000 UTC m=+144.362180762 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 08 18:19:33 crc kubenswrapper[4859]: I1008 18:19:33.615580 4859 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-kg44m\" (UID: \"4afad99b-9bf7-43c9-8dcf-7a3217136be5\") " pod="openshift-image-registry/image-registry-697d97f7c8-kg44m" Oct 08 18:19:33 crc kubenswrapper[4859]: E1008 18:19:33.615995 4859 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-08 18:19:34.115980342 +0000 UTC m=+144.362819721 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-kg44m" (UID: "4afad99b-9bf7-43c9-8dcf-7a3217136be5") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 08 18:19:33 crc kubenswrapper[4859]: I1008 18:19:33.624898 4859 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-service-ca/service-ca-9c57cc56f-cs77r"] Oct 08 18:19:33 crc kubenswrapper[4859]: I1008 18:19:33.625299 4859 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-dns/dns-default-x7pp9" Oct 08 18:19:33 crc kubenswrapper[4859]: I1008 18:19:33.626373 4859 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-vc7bh"] Oct 08 18:19:33 crc kubenswrapper[4859]: I1008 18:19:33.643905 4859 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-q9nqj" Oct 08 18:19:33 crc kubenswrapper[4859]: I1008 18:19:33.662679 4859 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="hostpath-provisioner/csi-hostpathplugin-q4lkb" Oct 08 18:19:33 crc kubenswrapper[4859]: I1008 18:19:33.670530 4859 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ingress-canary/ingress-canary-qtvkr" Oct 08 18:19:33 crc kubenswrapper[4859]: W1008 18:19:33.693746 4859 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podbb6dc185_94e3_4cb7_8076_1011c1c60fab.slice/crio-4f39207ac47873b7bef2cade61826664c57d5556da080f248f0eb036436d3c6a WatchSource:0}: Error finding container 4f39207ac47873b7bef2cade61826664c57d5556da080f248f0eb036436d3c6a: Status 404 returned error can't find the container with id 4f39207ac47873b7bef2cade61826664c57d5556da080f248f0eb036436d3c6a Oct 08 18:19:33 crc kubenswrapper[4859]: W1008 18:19:33.703550 4859 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod01517214_a376_421e_9938_e548d76695d5.slice/crio-3348e36546eb9fc76222194a019a46203cd9eec8e62992d5cfe47a729366eb48 WatchSource:0}: Error finding container 3348e36546eb9fc76222194a019a46203cd9eec8e62992d5cfe47a729366eb48: Status 404 returned error can't find the container with id 3348e36546eb9fc76222194a019a46203cd9eec8e62992d5cfe47a729366eb48 Oct 08 18:19:33 crc kubenswrapper[4859]: I1008 18:19:33.721802 4859 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 08 18:19:33 crc kubenswrapper[4859]: E1008 18:19:33.722168 4859 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-08 18:19:34.22214291 +0000 UTC m=+144.468982289 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 08 18:19:33 crc kubenswrapper[4859]: I1008 18:19:33.734164 4859 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-machine-config-operator/machine-config-operator-74547568cd-xsc6l"] Oct 08 18:19:33 crc kubenswrapper[4859]: I1008 18:19:33.800245 4859 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-6dlgn"] Oct 08 18:19:33 crc kubenswrapper[4859]: I1008 18:19:33.823969 4859 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-kg44m\" (UID: \"4afad99b-9bf7-43c9-8dcf-7a3217136be5\") " pod="openshift-image-registry/image-registry-697d97f7c8-kg44m" Oct 08 18:19:33 crc kubenswrapper[4859]: E1008 18:19:33.824834 4859 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-08 18:19:34.324814804 +0000 UTC m=+144.571654183 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-kg44m" (UID: "4afad99b-9bf7-43c9-8dcf-7a3217136be5") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 08 18:19:33 crc kubenswrapper[4859]: I1008 18:19:33.858005 4859 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29332455-w9wt8"] Oct 08 18:19:33 crc kubenswrapper[4859]: I1008 18:19:33.871881 4859 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-storage-version-migrator/migrator-59844c95c7-pmngm"] Oct 08 18:19:33 crc kubenswrapper[4859]: I1008 18:19:33.873999 4859 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/marketplace-operator-79b997595-bz6v8"] Oct 08 18:19:33 crc kubenswrapper[4859]: I1008 18:19:33.925016 4859 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 08 18:19:33 crc kubenswrapper[4859]: E1008 18:19:33.925139 4859 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-08 18:19:34.425110406 +0000 UTC m=+144.671949775 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 08 18:19:33 crc kubenswrapper[4859]: I1008 18:19:33.925917 4859 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-kg44m\" (UID: \"4afad99b-9bf7-43c9-8dcf-7a3217136be5\") " pod="openshift-image-registry/image-registry-697d97f7c8-kg44m" Oct 08 18:19:33 crc kubenswrapper[4859]: E1008 18:19:33.926381 4859 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-08 18:19:34.426367044 +0000 UTC m=+144.673206423 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-kg44m" (UID: "4afad99b-9bf7-43c9-8dcf-7a3217136be5") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 08 18:19:34 crc kubenswrapper[4859]: I1008 18:19:34.028222 4859 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 08 18:19:34 crc kubenswrapper[4859]: I1008 18:19:34.028480 4859 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/d407fd47-096d-4f2a-ace1-dab4ddf7192a-etcd-client\") pod \"apiserver-7bbb656c7d-6j5rk\" (UID: \"d407fd47-096d-4f2a-ace1-dab4ddf7192a\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-6j5rk" Oct 08 18:19:34 crc kubenswrapper[4859]: I1008 18:19:34.028524 4859 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/0902ccb2-78ab-48d0-be02-4c42f03e12fb-v4-0-config-user-template-provider-selection\") pod \"oauth-openshift-558db77b4-j8gh2\" (UID: \"0902ccb2-78ab-48d0-be02-4c42f03e12fb\") " pod="openshift-authentication/oauth-openshift-558db77b4-j8gh2" Oct 08 18:19:34 crc kubenswrapper[4859]: I1008 18:19:34.028545 4859 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/0902ccb2-78ab-48d0-be02-4c42f03e12fb-audit-policies\") pod \"oauth-openshift-558db77b4-j8gh2\" (UID: \"0902ccb2-78ab-48d0-be02-4c42f03e12fb\") " pod="openshift-authentication/oauth-openshift-558db77b4-j8gh2" Oct 08 18:19:34 crc kubenswrapper[4859]: I1008 18:19:34.028620 4859 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/d407fd47-096d-4f2a-ace1-dab4ddf7192a-audit-policies\") pod \"apiserver-7bbb656c7d-6j5rk\" (UID: \"d407fd47-096d-4f2a-ace1-dab4ddf7192a\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-6j5rk" Oct 08 18:19:34 crc kubenswrapper[4859]: I1008 18:19:34.028635 4859 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/d407fd47-096d-4f2a-ace1-dab4ddf7192a-etcd-serving-ca\") pod \"apiserver-7bbb656c7d-6j5rk\" (UID: \"d407fd47-096d-4f2a-ace1-dab4ddf7192a\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-6j5rk" Oct 08 18:19:34 crc kubenswrapper[4859]: I1008 18:19:34.030807 4859 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/d407fd47-096d-4f2a-ace1-dab4ddf7192a-etcd-serving-ca\") pod \"apiserver-7bbb656c7d-6j5rk\" (UID: \"d407fd47-096d-4f2a-ace1-dab4ddf7192a\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-6j5rk" Oct 08 18:19:34 crc kubenswrapper[4859]: I1008 18:19:34.031287 4859 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/0902ccb2-78ab-48d0-be02-4c42f03e12fb-audit-policies\") pod \"oauth-openshift-558db77b4-j8gh2\" (UID: \"0902ccb2-78ab-48d0-be02-4c42f03e12fb\") " pod="openshift-authentication/oauth-openshift-558db77b4-j8gh2" Oct 08 18:19:34 crc kubenswrapper[4859]: I1008 18:19:34.031407 4859 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/d407fd47-096d-4f2a-ace1-dab4ddf7192a-audit-policies\") pod \"apiserver-7bbb656c7d-6j5rk\" (UID: \"d407fd47-096d-4f2a-ace1-dab4ddf7192a\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-6j5rk" Oct 08 18:19:34 crc kubenswrapper[4859]: E1008 18:19:34.031936 4859 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-08 18:19:34.531900933 +0000 UTC m=+144.778740322 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 08 18:19:34 crc kubenswrapper[4859]: I1008 18:19:34.050594 4859 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/d407fd47-096d-4f2a-ace1-dab4ddf7192a-etcd-client\") pod \"apiserver-7bbb656c7d-6j5rk\" (UID: \"d407fd47-096d-4f2a-ace1-dab4ddf7192a\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-6j5rk" Oct 08 18:19:34 crc kubenswrapper[4859]: I1008 18:19:34.054601 4859 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/0902ccb2-78ab-48d0-be02-4c42f03e12fb-v4-0-config-user-template-provider-selection\") pod \"oauth-openshift-558db77b4-j8gh2\" (UID: \"0902ccb2-78ab-48d0-be02-4c42f03e12fb\") " pod="openshift-authentication/oauth-openshift-558db77b4-j8gh2" Oct 08 18:19:34 crc kubenswrapper[4859]: I1008 18:19:34.087041 4859 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-authentication/oauth-openshift-558db77b4-j8gh2" Oct 08 18:19:34 crc kubenswrapper[4859]: I1008 18:19:34.105359 4859 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-dns-operator/dns-operator-744455d44c-q222b"] Oct 08 18:19:34 crc kubenswrapper[4859]: I1008 18:19:34.130253 4859 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-kg44m\" (UID: \"4afad99b-9bf7-43c9-8dcf-7a3217136be5\") " pod="openshift-image-registry/image-registry-697d97f7c8-kg44m" Oct 08 18:19:34 crc kubenswrapper[4859]: E1008 18:19:34.130753 4859 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-08 18:19:34.630739212 +0000 UTC m=+144.877578591 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-kg44m" (UID: "4afad99b-9bf7-43c9-8dcf-7a3217136be5") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 08 18:19:34 crc kubenswrapper[4859]: I1008 18:19:34.201097 4859 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-ingress-operator/ingress-operator-5b745b69d9-pf7j6"] Oct 08 18:19:34 crc kubenswrapper[4859]: W1008 18:19:34.229651 4859 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-poded062c43_61ec_4951_a27c_77f165523d3f.slice/crio-0d037f831407eee5f18ac3d7e67af8029c8e60fb9d8a3bb537e7f8d32daba2e9 WatchSource:0}: Error finding container 0d037f831407eee5f18ac3d7e67af8029c8e60fb9d8a3bb537e7f8d32daba2e9: Status 404 returned error can't find the container with id 0d037f831407eee5f18ac3d7e67af8029c8e60fb9d8a3bb537e7f8d32daba2e9 Oct 08 18:19:34 crc kubenswrapper[4859]: I1008 18:19:34.233439 4859 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 08 18:19:34 crc kubenswrapper[4859]: E1008 18:19:34.233657 4859 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-08 18:19:34.733630913 +0000 UTC m=+144.980470292 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 08 18:19:34 crc kubenswrapper[4859]: I1008 18:19:34.234028 4859 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-kg44m\" (UID: \"4afad99b-9bf7-43c9-8dcf-7a3217136be5\") " pod="openshift-image-registry/image-registry-697d97f7c8-kg44m" Oct 08 18:19:34 crc kubenswrapper[4859]: E1008 18:19:34.234543 4859 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-08 18:19:34.734525909 +0000 UTC m=+144.981365288 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-kg44m" (UID: "4afad99b-9bf7-43c9-8dcf-7a3217136be5") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 08 18:19:34 crc kubenswrapper[4859]: I1008 18:19:34.334486 4859 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 08 18:19:34 crc kubenswrapper[4859]: E1008 18:19:34.334788 4859 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-08 18:19:34.83475531 +0000 UTC m=+145.081594709 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 08 18:19:34 crc kubenswrapper[4859]: I1008 18:19:34.335033 4859 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-kg44m\" (UID: \"4afad99b-9bf7-43c9-8dcf-7a3217136be5\") " pod="openshift-image-registry/image-registry-697d97f7c8-kg44m" Oct 08 18:19:34 crc kubenswrapper[4859]: E1008 18:19:34.335550 4859 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-08 18:19:34.835537974 +0000 UTC m=+145.082377353 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-kg44m" (UID: "4afad99b-9bf7-43c9-8dcf-7a3217136be5") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 08 18:19:34 crc kubenswrapper[4859]: I1008 18:19:34.357583 4859 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-6j5rk" Oct 08 18:19:34 crc kubenswrapper[4859]: I1008 18:19:34.370464 4859 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-vc7bh" event={"ID":"b2b5fcca-e076-4cbe-bc1e-249fcfd25e53","Type":"ContainerStarted","Data":"e37d373abb60d955f1b152b9f3c40a8648585de1c7c5f0f487e6ddde5659ab44"} Oct 08 18:19:34 crc kubenswrapper[4859]: I1008 18:19:34.378898 4859 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-config-operator/openshift-config-operator-7777fb866f-ncj6n" event={"ID":"65868bd3-b636-48c5-aac4-fcf734f84e39","Type":"ContainerStarted","Data":"c27fae04378c57f2d58bb3587dd9fcc22b3432b16e9458114798571d9794d1b3"} Oct 08 18:19:34 crc kubenswrapper[4859]: I1008 18:19:34.379828 4859 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-config-operator/openshift-config-operator-7777fb866f-ncj6n" Oct 08 18:19:34 crc kubenswrapper[4859]: I1008 18:19:34.400452 4859 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-admission-controller-857f4d67dd-lrqkp" event={"ID":"f114c686-8fde-4560-b8e1-9cd8fea448f8","Type":"ContainerStarted","Data":"2a2a2d5e1b72fe5ea93da37e2c01a8fe933651745d39c9b01e29d1896b95c2ac"} Oct 08 18:19:34 crc kubenswrapper[4859]: I1008 18:19:34.400501 4859 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-admission-controller-857f4d67dd-lrqkp" event={"ID":"f114c686-8fde-4560-b8e1-9cd8fea448f8","Type":"ContainerStarted","Data":"55703680ce636d8de7afba38777f65e3fd30d78c717fe2a9384bd46927288d87"} Oct 08 18:19:34 crc kubenswrapper[4859]: I1008 18:19:34.422898 4859 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-service-ca/service-ca-9c57cc56f-cs77r" event={"ID":"da9f04a1-bb6a-40de-846a-5e2fe04855d1","Type":"ContainerStarted","Data":"74f1574c36bbd5f68083945b6009a7159eae66c5d6c52ae1ac89e556bd6e9676"} Oct 08 18:19:34 crc kubenswrapper[4859]: I1008 18:19:34.426556 4859 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-kq7mx" event={"ID":"c79c3857-6ecb-4e3c-b59f-6a58ba0f94e9","Type":"ContainerStarted","Data":"0801b540624334eb96eb6602b4d95276974933142f154a4a032da1ece2e45517"} Oct 08 18:19:34 crc kubenswrapper[4859]: I1008 18:19:34.430871 4859 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-authentication-operator/authentication-operator-69f744f599-jd5kh" event={"ID":"b8d9f66c-e7a4-467f-bcf5-701cc0b68d52","Type":"ContainerStarted","Data":"e287c282f81a59fb6082bb38f8acbe2143226bff07293aceafbcceb4d111f0b9"} Oct 08 18:19:34 crc kubenswrapper[4859]: I1008 18:19:34.430936 4859 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-authentication-operator/authentication-operator-69f744f599-jd5kh" event={"ID":"b8d9f66c-e7a4-467f-bcf5-701cc0b68d52","Type":"ContainerStarted","Data":"0f42ba40bca269b2823a35c502c24ca6bd77ab247751b3c5f5098e97319bc856"} Oct 08 18:19:34 crc kubenswrapper[4859]: I1008 18:19:34.435610 4859 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 08 18:19:34 crc kubenswrapper[4859]: I1008 18:19:34.435627 4859 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-dns-operator/dns-operator-744455d44c-q222b" event={"ID":"63636924-a28b-447e-8568-92a193a69ade","Type":"ContainerStarted","Data":"96b56466b1d6b6eb4b0e793c4ed08080e014a9494be4d285fbba3d599fcee2a9"} Oct 08 18:19:34 crc kubenswrapper[4859]: E1008 18:19:34.435888 4859 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-08 18:19:34.935867128 +0000 UTC m=+145.182706507 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 08 18:19:34 crc kubenswrapper[4859]: I1008 18:19:34.435931 4859 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-kg44m\" (UID: \"4afad99b-9bf7-43c9-8dcf-7a3217136be5\") " pod="openshift-image-registry/image-registry-697d97f7c8-kg44m" Oct 08 18:19:34 crc kubenswrapper[4859]: E1008 18:19:34.436282 4859 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-08 18:19:34.93627504 +0000 UTC m=+145.183114419 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-kg44m" (UID: "4afad99b-9bf7-43c9-8dcf-7a3217136be5") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 08 18:19:34 crc kubenswrapper[4859]: I1008 18:19:34.452563 4859 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-5n49w" event={"ID":"6f093915-bce6-40eb-917f-ed0058431a79","Type":"ContainerStarted","Data":"8e968eb95da598888aaec40c1b2b69d70b6612d635fc1e037980d7413e527a77"} Oct 08 18:19:34 crc kubenswrapper[4859]: I1008 18:19:34.452629 4859 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-5n49w" event={"ID":"6f093915-bce6-40eb-917f-ed0058431a79","Type":"ContainerStarted","Data":"8f278f321c76e92f6996c0e678fb2cec355fb405d66a86450910a417a63f46bb"} Oct 08 18:19:34 crc kubenswrapper[4859]: I1008 18:19:34.453365 4859 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-5n49w" Oct 08 18:19:34 crc kubenswrapper[4859]: I1008 18:19:34.453898 4859 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-storage-version-migrator/migrator-59844c95c7-pmngm" event={"ID":"5401cd5b-bc2e-43d4-8d13-b7e9134aa72f","Type":"ContainerStarted","Data":"3f58e13e0c19bb2b9b34902812fa92599226282cbc1498c53e4fde764115ff83"} Oct 08 18:19:34 crc kubenswrapper[4859]: I1008 18:19:34.461054 4859 patch_prober.go:28] interesting pod/olm-operator-6b444d44fb-5n49w container/olm-operator namespace/openshift-operator-lifecycle-manager: Readiness probe status=failure output="Get \"https://10.217.0.31:8443/healthz\": dial tcp 10.217.0.31:8443: connect: connection refused" start-of-body= Oct 08 18:19:34 crc kubenswrapper[4859]: I1008 18:19:34.464194 4859 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-czj5g" event={"ID":"7d9d11d4-9f7b-4f33-a9d2-23db86bb2343","Type":"ContainerStarted","Data":"c7d58cafd1658e6f19729d05ad2d0838494f2d4b162f77cce7896b184f9d233a"} Oct 08 18:19:34 crc kubenswrapper[4859]: I1008 18:19:34.464239 4859 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-czj5g" event={"ID":"7d9d11d4-9f7b-4f33-a9d2-23db86bb2343","Type":"ContainerStarted","Data":"a0102584a3d8da9259d4e842059a0fee4583b07885c6a7dc8cad6acc65e8d28b"} Oct 08 18:19:34 crc kubenswrapper[4859]: I1008 18:19:34.461592 4859 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-5n49w" podUID="6f093915-bce6-40eb-917f-ed0058431a79" containerName="olm-operator" probeResult="failure" output="Get \"https://10.217.0.31:8443/healthz\": dial tcp 10.217.0.31:8443: connect: connection refused" Oct 08 18:19:34 crc kubenswrapper[4859]: I1008 18:19:34.504441 4859 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-controller-manager/controller-manager-879f6c89f-9gpzj" Oct 08 18:19:34 crc kubenswrapper[4859]: I1008 18:19:34.504475 4859 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-xlfdd" Oct 08 18:19:34 crc kubenswrapper[4859]: I1008 18:19:34.504486 4859 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-controller-84d6567774-m7m5n" event={"ID":"e431ddbe-2d50-4a0a-9679-5565e0df977f","Type":"ContainerStarted","Data":"3a2542e3f0cef4166668c9fad2e3b7b40390cd9ff7e1adb44217da9c1327bbbb"} Oct 08 18:19:34 crc kubenswrapper[4859]: I1008 18:19:34.504505 4859 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-controller-84d6567774-m7m5n" event={"ID":"e431ddbe-2d50-4a0a-9679-5565e0df977f","Type":"ContainerStarted","Data":"695e3eea1167688cc367354ffb90c737f654fb0904ebfaca4ae1196abff91d90"} Oct 08 18:19:34 crc kubenswrapper[4859]: I1008 18:19:34.504515 4859 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-4s5w9" event={"ID":"01517214-a376-421e-9938-e548d76695d5","Type":"ContainerStarted","Data":"3348e36546eb9fc76222194a019a46203cd9eec8e62992d5cfe47a729366eb48"} Oct 08 18:19:34 crc kubenswrapper[4859]: I1008 18:19:34.504524 4859 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-879f6c89f-9gpzj" event={"ID":"460ad6e7-859d-4111-a555-b55c5b7228e0","Type":"ContainerStarted","Data":"46ee8cb842f339e6b4ba065cab54a4e76c9bac684c499161345db073e276bd89"} Oct 08 18:19:34 crc kubenswrapper[4859]: I1008 18:19:34.504533 4859 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-82pcg" event={"ID":"d80051ae-5521-4f01-b34d-34d06a661177","Type":"ContainerStarted","Data":"76cb317f4d15c5103c86e0c51756f52859807d22bea5aa5ed62de35c8d7a20cc"} Oct 08 18:19:34 crc kubenswrapper[4859]: I1008 18:19:34.504547 4859 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-82pcg" event={"ID":"d80051ae-5521-4f01-b34d-34d06a661177","Type":"ContainerStarted","Data":"c1b92d5ce1bc408c3da8236e90069121428e6ce8ffb4892b237037364b4035b1"} Oct 08 18:19:34 crc kubenswrapper[4859]: I1008 18:19:34.504557 4859 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-server-7tjh2" event={"ID":"0d04221d-922c-4ee9-9adf-817e295161d9","Type":"ContainerStarted","Data":"c6017a502ba8696026cd7239ab6bbf47899625ed1477207f037d3d64488def3c"} Oct 08 18:19:34 crc kubenswrapper[4859]: I1008 18:19:34.504569 4859 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-server-7tjh2" event={"ID":"0d04221d-922c-4ee9-9adf-817e295161d9","Type":"ContainerStarted","Data":"6c1db8d1f77b07a3da9b23d4da5a79436190f1d5a765b4cbca2f130f105967cd"} Oct 08 18:19:34 crc kubenswrapper[4859]: I1008 18:19:34.504578 4859 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-xlfdd" event={"ID":"ca64d874-5d86-4c49-8953-631a75d59788","Type":"ContainerStarted","Data":"44d292882e24bdc475f139a3b68beb09c235ab83c773cff48466e5d53d3b8a19"} Oct 08 18:19:34 crc kubenswrapper[4859]: I1008 18:19:34.504587 4859 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-xlfdd" event={"ID":"ca64d874-5d86-4c49-8953-631a75d59788","Type":"ContainerStarted","Data":"ba63a47285e8dd8deca0f8b02ae365771ce71a87162bf7b4eeedeeead2dd4b5b"} Oct 08 18:19:34 crc kubenswrapper[4859]: I1008 18:19:34.507792 4859 patch_prober.go:28] interesting pod/route-controller-manager-6576b87f9c-xlfdd container/route-controller-manager namespace/openshift-route-controller-manager: Readiness probe status=failure output="Get \"https://10.217.0.7:8443/healthz\": dial tcp 10.217.0.7:8443: connect: connection refused" start-of-body= Oct 08 18:19:34 crc kubenswrapper[4859]: I1008 18:19:34.508082 4859 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-xlfdd" podUID="ca64d874-5d86-4c49-8953-631a75d59788" containerName="route-controller-manager" probeResult="failure" output="Get \"https://10.217.0.7:8443/healthz\": dial tcp 10.217.0.7:8443: connect: connection refused" Oct 08 18:19:34 crc kubenswrapper[4859]: I1008 18:19:34.512035 4859 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console/console-f9d7485db-m2xmp" event={"ID":"9a829106-59b1-4389-95bf-2ee4d56e317c","Type":"ContainerStarted","Data":"18ddf7cb370db54cf37e40e1f580d3e3a64ea5cacb148bb385217940676b8053"} Oct 08 18:19:34 crc kubenswrapper[4859]: I1008 18:19:34.512081 4859 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console/console-f9d7485db-m2xmp" event={"ID":"9a829106-59b1-4389-95bf-2ee4d56e317c","Type":"ContainerStarted","Data":"d678c5240a9fe6aca79f30a7a28b1340b3c082b8b6cdeee98fbb95d197a14f0a"} Oct 08 18:19:34 crc kubenswrapper[4859]: I1008 18:19:34.513921 4859 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-kc42p" event={"ID":"bb6dc185-94e3-4cb7-8076-1011c1c60fab","Type":"ContainerStarted","Data":"4f39207ac47873b7bef2cade61826664c57d5556da080f248f0eb036436d3c6a"} Oct 08 18:19:34 crc kubenswrapper[4859]: I1008 18:19:34.519343 4859 patch_prober.go:28] interesting pod/controller-manager-879f6c89f-9gpzj container/controller-manager namespace/openshift-controller-manager: Readiness probe status=failure output="Get \"https://10.217.0.8:8443/healthz\": dial tcp 10.217.0.8:8443: connect: connection refused" start-of-body= Oct 08 18:19:34 crc kubenswrapper[4859]: I1008 18:19:34.519394 4859 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-controller-manager/controller-manager-879f6c89f-9gpzj" podUID="460ad6e7-859d-4111-a555-b55c5b7228e0" containerName="controller-manager" probeResult="failure" output="Get \"https://10.217.0.8:8443/healthz\": dial tcp 10.217.0.8:8443: connect: connection refused" Oct 08 18:19:34 crc kubenswrapper[4859]: I1008 18:19:34.536866 4859 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 08 18:19:34 crc kubenswrapper[4859]: E1008 18:19:34.540975 4859 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-08 18:19:35.040949153 +0000 UTC m=+145.287788532 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 08 18:19:34 crc kubenswrapper[4859]: I1008 18:19:34.543823 4859 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-kg44m\" (UID: \"4afad99b-9bf7-43c9-8dcf-7a3217136be5\") " pod="openshift-image-registry/image-registry-697d97f7c8-kg44m" Oct 08 18:19:34 crc kubenswrapper[4859]: E1008 18:19:34.547658 4859 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-08 18:19:35.047640733 +0000 UTC m=+145.294480112 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-kg44m" (UID: "4afad99b-9bf7-43c9-8dcf-7a3217136be5") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 08 18:19:34 crc kubenswrapper[4859]: I1008 18:19:34.573632 4859 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-api/machine-api-operator-5694c8668f-gmbk9" event={"ID":"a52b9536-1013-482b-9261-e024377687d2","Type":"ContainerStarted","Data":"4e76529ed4dbde52d7d14dd35964180537550cfb36e5da33fbe492a7754dd5f0"} Oct 08 18:19:34 crc kubenswrapper[4859]: I1008 18:19:34.573711 4859 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-api/machine-api-operator-5694c8668f-gmbk9" event={"ID":"a52b9536-1013-482b-9261-e024377687d2","Type":"ContainerStarted","Data":"7b6a6c818be35d1bb96148bb9cf45bba330d4e3450e22fe5d4719d65cf5a946b"} Oct 08 18:19:34 crc kubenswrapper[4859]: I1008 18:19:34.583119 4859 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ingress/router-default-5444994796-r5gs2" event={"ID":"8e2b8a53-4bc3-499b-9746-628c323d2a0c","Type":"ContainerStarted","Data":"330846f4a2874ee4f831b39ed3b4cd3c1ca8624f75940a0838be1acf003a1fbb"} Oct 08 18:19:34 crc kubenswrapper[4859]: I1008 18:19:34.583173 4859 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ingress/router-default-5444994796-r5gs2" event={"ID":"8e2b8a53-4bc3-499b-9746-628c323d2a0c","Type":"ContainerStarted","Data":"537e3afc299ce71830284e1cf5963b367dd40ee6a2c6511fcaf5d71d45704e73"} Oct 08 18:19:34 crc kubenswrapper[4859]: I1008 18:19:34.586428 4859 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-c5p45" event={"ID":"9b239e22-0030-4c68-b0f1-66fbaedd5f9b","Type":"ContainerStarted","Data":"06e67146dba214f00d0f6cdc71c37714af2ee4e7dc546bd455d44ad8a4748934"} Oct 08 18:19:34 crc kubenswrapper[4859]: I1008 18:19:34.594044 4859 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-operator-74547568cd-xsc6l" event={"ID":"0c9973e7-5c40-463b-aca1-c3c93d7a9d0b","Type":"ContainerStarted","Data":"089a398dda64a0e3bb35d3866b731461678c59a5c6d262151317daf3ea8c9fcd"} Oct 08 18:19:34 crc kubenswrapper[4859]: I1008 18:19:34.625855 4859 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-6dlgn" event={"ID":"46a9f481-0772-4006-99e2-723919011659","Type":"ContainerStarted","Data":"b54c2d6c17458509ad8fc154ff00a2ea507016a17c3c5b01b112af5f516665ac"} Oct 08 18:19:34 crc kubenswrapper[4859]: I1008 18:19:34.627646 4859 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ingress-operator/ingress-operator-5b745b69d9-pf7j6" event={"ID":"ed062c43-61ec-4951-a27c-77f165523d3f","Type":"ContainerStarted","Data":"0d037f831407eee5f18ac3d7e67af8029c8e60fb9d8a3bb537e7f8d32daba2e9"} Oct 08 18:19:34 crc kubenswrapper[4859]: I1008 18:19:34.629456 4859 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/marketplace-operator-79b997595-bz6v8" event={"ID":"8d4b5b44-816f-44f6-98ab-c5d66aefaa08","Type":"ContainerStarted","Data":"30e8f4a72ca43a49d25154d0078616d084b696a53c76f98a22049f48584ecae0"} Oct 08 18:19:34 crc kubenswrapper[4859]: I1008 18:19:34.641665 4859 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29332455-w9wt8" event={"ID":"d403dae9-eb8c-4c8d-9daa-bf2a0c815652","Type":"ContainerStarted","Data":"cbfeda28315cc27b0db4ba339316e0c0f8c58fd68dec107d7c70d3e13e83c4ab"} Oct 08 18:19:34 crc kubenswrapper[4859]: I1008 18:19:34.644588 4859 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 08 18:19:34 crc kubenswrapper[4859]: E1008 18:19:34.644878 4859 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-08 18:19:35.144846114 +0000 UTC m=+145.391685493 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 08 18:19:34 crc kubenswrapper[4859]: I1008 18:19:34.644936 4859 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-kg44m\" (UID: \"4afad99b-9bf7-43c9-8dcf-7a3217136be5\") " pod="openshift-image-registry/image-registry-697d97f7c8-kg44m" Oct 08 18:19:34 crc kubenswrapper[4859]: E1008 18:19:34.645547 4859 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-08 18:19:35.145537774 +0000 UTC m=+145.392377143 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-kg44m" (UID: "4afad99b-9bf7-43c9-8dcf-7a3217136be5") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 08 18:19:34 crc kubenswrapper[4859]: I1008 18:19:34.659865 4859 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/catalog-operator-68c6474976-xpnsv"] Oct 08 18:19:34 crc kubenswrapper[4859]: I1008 18:19:34.660047 4859 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd-operator/etcd-operator-b45778765-tnn6l" event={"ID":"e4b12149-79e9-4b2e-8991-519ad5daa14b","Type":"ContainerStarted","Data":"35fa26dc6b77f8c0ae77b64879bb6850c736d9cc833f7c8a6b5daa3f2a110953"} Oct 08 18:19:34 crc kubenswrapper[4859]: I1008 18:19:34.660088 4859 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd-operator/etcd-operator-b45778765-tnn6l" event={"ID":"e4b12149-79e9-4b2e-8991-519ad5daa14b","Type":"ContainerStarted","Data":"a55cd826807382c082280734b02f007c1294f1422e0d6b9bece8739634a13234"} Oct 08 18:19:34 crc kubenswrapper[4859]: I1008 18:19:34.662810 4859 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-cluster-machine-approver/machine-approver-56656f9798-95s6m" event={"ID":"39b30c11-d0f6-4302-9ef8-e7fcf1b2170e","Type":"ContainerStarted","Data":"a31de6df140db8c94c12f0bf82edbcbab35599166e059c5d2d13bc0837e7c303"} Oct 08 18:19:34 crc kubenswrapper[4859]: I1008 18:19:34.666208 4859 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-5f4q5"] Oct 08 18:19:34 crc kubenswrapper[4859]: I1008 18:19:34.672835 4859 generic.go:334] "Generic (PLEG): container finished" podID="70f88ce6-b9fe-422e-827e-af2b6840a783" containerID="543c1283ddf357972ee2f461f029103672b0cae1f4f7eb39b32f25ca85d863eb" exitCode=0 Oct 08 18:19:34 crc kubenswrapper[4859]: I1008 18:19:34.672944 4859 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-apiserver/apiserver-76f77b778f-fv9bp" event={"ID":"70f88ce6-b9fe-422e-827e-af2b6840a783","Type":"ContainerDied","Data":"543c1283ddf357972ee2f461f029103672b0cae1f4f7eb39b32f25ca85d863eb"} Oct 08 18:19:34 crc kubenswrapper[4859]: I1008 18:19:34.672982 4859 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-apiserver/apiserver-76f77b778f-fv9bp" event={"ID":"70f88ce6-b9fe-422e-827e-af2b6840a783","Type":"ContainerStarted","Data":"9ca5214a7055380284b3383cb5985d32cc2531a6a7ed5ded06032aff5db181db"} Oct 08 18:19:34 crc kubenswrapper[4859]: I1008 18:19:34.690229 4859 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-service-ca-operator/service-ca-operator-777779d784-847r8" event={"ID":"56c0b58f-23d6-4d8e-8b75-9095128ac121","Type":"ContainerStarted","Data":"d121774248f62c4cd1de2cd079f3fdc8a37c96464303d0eca33719cd2a4c496d"} Oct 08 18:19:34 crc kubenswrapper[4859]: I1008 18:19:34.690298 4859 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-service-ca-operator/service-ca-operator-777779d784-847r8" event={"ID":"56c0b58f-23d6-4d8e-8b75-9095128ac121","Type":"ContainerStarted","Data":"b6b63a48e02ad893a3a0efdcd56eab6fd6c497ebbce787cce6ab41c93fdfb9f1"} Oct 08 18:19:34 crc kubenswrapper[4859]: I1008 18:19:34.703608 4859 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console/downloads-7954f5f757-8q5jz" event={"ID":"9dcb8f76-0fff-436c-9191-fe3f787b7ac0","Type":"ContainerStarted","Data":"a78796a9e9ca192409c67d03b9ff1155fc06628735e7e84ae2dfe508964be7d9"} Oct 08 18:19:34 crc kubenswrapper[4859]: I1008 18:19:34.703664 4859 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console/downloads-7954f5f757-8q5jz" event={"ID":"9dcb8f76-0fff-436c-9191-fe3f787b7ac0","Type":"ContainerStarted","Data":"a053326b1482d92cdfa32301f2477297d99e00920a67cfeda9a11cffc858f276"} Oct 08 18:19:34 crc kubenswrapper[4859]: I1008 18:19:34.717742 4859 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-console-operator/console-operator-58897d9998-8sflk" Oct 08 18:19:34 crc kubenswrapper[4859]: I1008 18:19:34.747504 4859 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 08 18:19:34 crc kubenswrapper[4859]: E1008 18:19:34.748597 4859 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-08 18:19:35.248562849 +0000 UTC m=+145.495402288 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 08 18:19:34 crc kubenswrapper[4859]: W1008 18:19:34.752033 4859 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podbe85b712_b3e4_4535_b14b_672605d02387.slice/crio-0289eb7286a940f52eb6d1b2ee38185230594371fd5146aa57111cb7e54fb058 WatchSource:0}: Error finding container 0289eb7286a940f52eb6d1b2ee38185230594371fd5146aa57111cb7e54fb058: Status 404 returned error can't find the container with id 0289eb7286a940f52eb6d1b2ee38185230594371fd5146aa57111cb7e54fb058 Oct 08 18:19:34 crc kubenswrapper[4859]: I1008 18:19:34.772255 4859 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-console-operator/console-operator-58897d9998-8sflk" podStartSLOduration=124.772234975 podStartE2EDuration="2m4.772234975s" podCreationTimestamp="2025-10-08 18:17:30 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-08 18:19:34.766070321 +0000 UTC m=+145.012909700" watchObservedRunningTime="2025-10-08 18:19:34.772234975 +0000 UTC m=+145.019074354" Oct 08 18:19:34 crc kubenswrapper[4859]: W1008 18:19:34.820184 4859 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod7e23494c_2895_43b3_a35c_47a7b2d33ebc.slice/crio-0e1a75c2970c0c9cabac42007f9102f4114a37b02303e138fab00a12a2fdf707 WatchSource:0}: Error finding container 0e1a75c2970c0c9cabac42007f9102f4114a37b02303e138fab00a12a2fdf707: Status 404 returned error can't find the container with id 0e1a75c2970c0c9cabac42007f9102f4114a37b02303e138fab00a12a2fdf707 Oct 08 18:19:34 crc kubenswrapper[4859]: I1008 18:19:34.848850 4859 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-kg44m\" (UID: \"4afad99b-9bf7-43c9-8dcf-7a3217136be5\") " pod="openshift-image-registry/image-registry-697d97f7c8-kg44m" Oct 08 18:19:34 crc kubenswrapper[4859]: E1008 18:19:34.849422 4859 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-08 18:19:35.349373627 +0000 UTC m=+145.596213006 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-kg44m" (UID: "4afad99b-9bf7-43c9-8dcf-7a3217136be5") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 08 18:19:34 crc kubenswrapper[4859]: I1008 18:19:34.879016 4859 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-hf2fp"] Oct 08 18:19:34 crc kubenswrapper[4859]: I1008 18:19:34.890179 4859 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-q9nqj"] Oct 08 18:19:34 crc kubenswrapper[4859]: I1008 18:19:34.900224 4859 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-dns/dns-default-x7pp9"] Oct 08 18:19:34 crc kubenswrapper[4859]: I1008 18:19:34.909934 4859 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-ingress-canary/ingress-canary-qtvkr"] Oct 08 18:19:34 crc kubenswrapper[4859]: I1008 18:19:34.915672 4859 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-ingress/router-default-5444994796-r5gs2" Oct 08 18:19:34 crc kubenswrapper[4859]: I1008 18:19:34.917542 4859 patch_prober.go:28] interesting pod/router-default-5444994796-r5gs2 container/router namespace/openshift-ingress: Startup probe status=failure output="Get \"http://localhost:1936/healthz/ready\": dial tcp [::1]:1936: connect: connection refused" start-of-body= Oct 08 18:19:34 crc kubenswrapper[4859]: I1008 18:19:34.917598 4859 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-r5gs2" podUID="8e2b8a53-4bc3-499b-9746-628c323d2a0c" containerName="router" probeResult="failure" output="Get \"http://localhost:1936/healthz/ready\": dial tcp [::1]:1936: connect: connection refused" Oct 08 18:19:34 crc kubenswrapper[4859]: I1008 18:19:34.951572 4859 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 08 18:19:34 crc kubenswrapper[4859]: E1008 18:19:34.951789 4859 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-08 18:19:35.451743262 +0000 UTC m=+145.698582641 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 08 18:19:34 crc kubenswrapper[4859]: I1008 18:19:34.952203 4859 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-kg44m\" (UID: \"4afad99b-9bf7-43c9-8dcf-7a3217136be5\") " pod="openshift-image-registry/image-registry-697d97f7c8-kg44m" Oct 08 18:19:34 crc kubenswrapper[4859]: E1008 18:19:34.952627 4859 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-08 18:19:35.452607287 +0000 UTC m=+145.699446666 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-kg44m" (UID: "4afad99b-9bf7-43c9-8dcf-7a3217136be5") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 08 18:19:35 crc kubenswrapper[4859]: I1008 18:19:35.026701 4859 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["hostpath-provisioner/csi-hostpathplugin-q4lkb"] Oct 08 18:19:35 crc kubenswrapper[4859]: I1008 18:19:35.055722 4859 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 08 18:19:35 crc kubenswrapper[4859]: E1008 18:19:35.056039 4859 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-08 18:19:35.556021873 +0000 UTC m=+145.802861252 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 08 18:19:35 crc kubenswrapper[4859]: W1008 18:19:35.135964 4859 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod5e418cbb_0072_4767_933d_537c4a274e60.slice/crio-f4b3048b62c0cfcadcb8fded4f558807ae7eb5a4b16875c6002b356fdf8adb29 WatchSource:0}: Error finding container f4b3048b62c0cfcadcb8fded4f558807ae7eb5a4b16875c6002b356fdf8adb29: Status 404 returned error can't find the container with id f4b3048b62c0cfcadcb8fded4f558807ae7eb5a4b16875c6002b356fdf8adb29 Oct 08 18:19:35 crc kubenswrapper[4859]: I1008 18:19:35.156853 4859 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-kg44m\" (UID: \"4afad99b-9bf7-43c9-8dcf-7a3217136be5\") " pod="openshift-image-registry/image-registry-697d97f7c8-kg44m" Oct 08 18:19:35 crc kubenswrapper[4859]: E1008 18:19:35.157478 4859 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-08 18:19:35.65746407 +0000 UTC m=+145.904303449 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-kg44m" (UID: "4afad99b-9bf7-43c9-8dcf-7a3217136be5") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 08 18:19:35 crc kubenswrapper[4859]: I1008 18:19:35.170480 4859 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-xlfdd" podStartSLOduration=124.170451218 podStartE2EDuration="2m4.170451218s" podCreationTimestamp="2025-10-08 18:17:31 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-08 18:19:35.165087488 +0000 UTC m=+145.411926867" watchObservedRunningTime="2025-10-08 18:19:35.170451218 +0000 UTC m=+145.417290597" Oct 08 18:19:35 crc kubenswrapper[4859]: I1008 18:19:35.209205 4859 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-controller-manager/controller-manager-879f6c89f-9gpzj" podStartSLOduration=125.209183744 podStartE2EDuration="2m5.209183744s" podCreationTimestamp="2025-10-08 18:17:30 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-08 18:19:35.208173574 +0000 UTC m=+145.455012953" watchObservedRunningTime="2025-10-08 18:19:35.209183744 +0000 UTC m=+145.456023113" Oct 08 18:19:35 crc kubenswrapper[4859]: I1008 18:19:35.251996 4859 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-console/console-f9d7485db-m2xmp" podStartSLOduration=125.251974411 podStartE2EDuration="2m5.251974411s" podCreationTimestamp="2025-10-08 18:17:30 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-08 18:19:35.249504587 +0000 UTC m=+145.496343976" watchObservedRunningTime="2025-10-08 18:19:35.251974411 +0000 UTC m=+145.498813790" Oct 08 18:19:35 crc kubenswrapper[4859]: I1008 18:19:35.258650 4859 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 08 18:19:35 crc kubenswrapper[4859]: E1008 18:19:35.259087 4859 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-08 18:19:35.759071132 +0000 UTC m=+146.005910501 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 08 18:19:35 crc kubenswrapper[4859]: I1008 18:19:35.311346 4859 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-machine-config-operator/machine-config-server-7tjh2" podStartSLOduration=5.311326812 podStartE2EDuration="5.311326812s" podCreationTimestamp="2025-10-08 18:19:30 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-08 18:19:35.308445556 +0000 UTC m=+145.555284945" watchObservedRunningTime="2025-10-08 18:19:35.311326812 +0000 UTC m=+145.558166191" Oct 08 18:19:35 crc kubenswrapper[4859]: I1008 18:19:35.328646 4859 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-authentication/oauth-openshift-558db77b4-j8gh2"] Oct 08 18:19:35 crc kubenswrapper[4859]: I1008 18:19:35.360836 4859 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-kg44m\" (UID: \"4afad99b-9bf7-43c9-8dcf-7a3217136be5\") " pod="openshift-image-registry/image-registry-697d97f7c8-kg44m" Oct 08 18:19:35 crc kubenswrapper[4859]: E1008 18:19:35.361328 4859 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-08 18:19:35.861311493 +0000 UTC m=+146.108150872 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-kg44m" (UID: "4afad99b-9bf7-43c9-8dcf-7a3217136be5") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 08 18:19:35 crc kubenswrapper[4859]: I1008 18:19:35.391827 4859 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-oauth-apiserver/apiserver-7bbb656c7d-6j5rk"] Oct 08 18:19:35 crc kubenswrapper[4859]: I1008 18:19:35.437828 4859 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-authentication-operator/authentication-operator-69f744f599-jd5kh" podStartSLOduration=125.437804366 podStartE2EDuration="2m5.437804366s" podCreationTimestamp="2025-10-08 18:17:30 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-08 18:19:35.405211123 +0000 UTC m=+145.652050502" watchObservedRunningTime="2025-10-08 18:19:35.437804366 +0000 UTC m=+145.684643745" Oct 08 18:19:35 crc kubenswrapper[4859]: I1008 18:19:35.438610 4859 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-czj5g" podStartSLOduration=125.43860098 podStartE2EDuration="2m5.43860098s" podCreationTimestamp="2025-10-08 18:17:30 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-08 18:19:35.437909529 +0000 UTC m=+145.684748908" watchObservedRunningTime="2025-10-08 18:19:35.43860098 +0000 UTC m=+145.685440359" Oct 08 18:19:35 crc kubenswrapper[4859]: I1008 18:19:35.462585 4859 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 08 18:19:35 crc kubenswrapper[4859]: E1008 18:19:35.463041 4859 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-08 18:19:35.963023928 +0000 UTC m=+146.209863307 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 08 18:19:35 crc kubenswrapper[4859]: W1008 18:19:35.494438 4859 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podd407fd47_096d_4f2a_ace1_dab4ddf7192a.slice/crio-ccde7798e02b8915fb95d5812af95ec91c29e19028c4db4057fa194aed59b582 WatchSource:0}: Error finding container ccde7798e02b8915fb95d5812af95ec91c29e19028c4db4057fa194aed59b582: Status 404 returned error can't find the container with id ccde7798e02b8915fb95d5812af95ec91c29e19028c4db4057fa194aed59b582 Oct 08 18:19:35 crc kubenswrapper[4859]: I1008 18:19:35.556020 4859 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-service-ca-operator/service-ca-operator-777779d784-847r8" podStartSLOduration=124.556000173 podStartE2EDuration="2m4.556000173s" podCreationTimestamp="2025-10-08 18:17:31 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-08 18:19:35.554595651 +0000 UTC m=+145.801435020" watchObservedRunningTime="2025-10-08 18:19:35.556000173 +0000 UTC m=+145.802839552" Oct 08 18:19:35 crc kubenswrapper[4859]: I1008 18:19:35.563878 4859 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-kg44m\" (UID: \"4afad99b-9bf7-43c9-8dcf-7a3217136be5\") " pod="openshift-image-registry/image-registry-697d97f7c8-kg44m" Oct 08 18:19:35 crc kubenswrapper[4859]: E1008 18:19:35.564324 4859 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-08 18:19:36.064308591 +0000 UTC m=+146.311147970 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-kg44m" (UID: "4afad99b-9bf7-43c9-8dcf-7a3217136be5") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 08 18:19:35 crc kubenswrapper[4859]: I1008 18:19:35.592985 4859 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-kq7mx" podStartSLOduration=125.592966096 podStartE2EDuration="2m5.592966096s" podCreationTimestamp="2025-10-08 18:17:30 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-08 18:19:35.581911286 +0000 UTC m=+145.828750665" watchObservedRunningTime="2025-10-08 18:19:35.592966096 +0000 UTC m=+145.839805475" Oct 08 18:19:35 crc kubenswrapper[4859]: I1008 18:19:35.617412 4859 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-console/downloads-7954f5f757-8q5jz" podStartSLOduration=125.617376364 podStartE2EDuration="2m5.617376364s" podCreationTimestamp="2025-10-08 18:17:30 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-08 18:19:35.604333565 +0000 UTC m=+145.851172964" watchObservedRunningTime="2025-10-08 18:19:35.617376364 +0000 UTC m=+145.864215743" Oct 08 18:19:35 crc kubenswrapper[4859]: I1008 18:19:35.665363 4859 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 08 18:19:35 crc kubenswrapper[4859]: E1008 18:19:35.666123 4859 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-08 18:19:36.166105758 +0000 UTC m=+146.412945127 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 08 18:19:35 crc kubenswrapper[4859]: I1008 18:19:35.677018 4859 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-etcd-operator/etcd-operator-b45778765-tnn6l" podStartSLOduration=125.676995203 podStartE2EDuration="2m5.676995203s" podCreationTimestamp="2025-10-08 18:17:30 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-08 18:19:35.675991933 +0000 UTC m=+145.922831312" watchObservedRunningTime="2025-10-08 18:19:35.676995203 +0000 UTC m=+145.923834592" Oct 08 18:19:35 crc kubenswrapper[4859]: I1008 18:19:35.753094 4859 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-dns/dns-default-x7pp9" event={"ID":"2a959183-1370-4723-bb5e-8625689d6121","Type":"ContainerStarted","Data":"f95da224f0e4e3bfa4deadf8fd7086126bd857b23d6a2746b0ae17af5ee0779b"} Oct 08 18:19:35 crc kubenswrapper[4859]: I1008 18:19:35.774748 4859 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-kg44m\" (UID: \"4afad99b-9bf7-43c9-8dcf-7a3217136be5\") " pod="openshift-image-registry/image-registry-697d97f7c8-kg44m" Oct 08 18:19:35 crc kubenswrapper[4859]: E1008 18:19:35.775224 4859 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-08 18:19:36.275210804 +0000 UTC m=+146.522050173 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-kg44m" (UID: "4afad99b-9bf7-43c9-8dcf-7a3217136be5") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 08 18:19:35 crc kubenswrapper[4859]: I1008 18:19:35.790371 4859 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-api/machine-api-operator-5694c8668f-gmbk9" event={"ID":"a52b9536-1013-482b-9261-e024377687d2","Type":"ContainerStarted","Data":"c5dbab1c9549b600fd7c5f247d6ed774a1b8ccff8a5b259f983e33f8e33b73d0"} Oct 08 18:19:35 crc kubenswrapper[4859]: I1008 18:19:35.805621 4859 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-controller-84d6567774-m7m5n" event={"ID":"e431ddbe-2d50-4a0a-9679-5565e0df977f","Type":"ContainerStarted","Data":"a9211ea7564cb048b2adc94a1817d52b50bcff02cc1aa008c33a6961b5225b78"} Oct 08 18:19:35 crc kubenswrapper[4859]: I1008 18:19:35.831794 4859 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-config-operator/openshift-config-operator-7777fb866f-ncj6n" podStartSLOduration=125.831762932 podStartE2EDuration="2m5.831762932s" podCreationTimestamp="2025-10-08 18:17:30 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-08 18:19:35.817849396 +0000 UTC m=+146.064688775" watchObservedRunningTime="2025-10-08 18:19:35.831762932 +0000 UTC m=+146.078602311" Oct 08 18:19:35 crc kubenswrapper[4859]: I1008 18:19:35.832097 4859 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-authentication/oauth-openshift-558db77b4-j8gh2" event={"ID":"0902ccb2-78ab-48d0-be02-4c42f03e12fb","Type":"ContainerStarted","Data":"d787d40d8d2e2389a64edbee99189e2b4ac2bd21b7ae58311a0299d5ea1445a4"} Oct 08 18:19:35 crc kubenswrapper[4859]: I1008 18:19:35.832791 4859 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-ingress/router-default-5444994796-r5gs2" podStartSLOduration=124.832780572 podStartE2EDuration="2m4.832780572s" podCreationTimestamp="2025-10-08 18:17:31 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-08 18:19:35.773848253 +0000 UTC m=+146.020687632" watchObservedRunningTime="2025-10-08 18:19:35.832780572 +0000 UTC m=+146.079619981" Oct 08 18:19:35 crc kubenswrapper[4859]: I1008 18:19:35.857287 4859 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-cluster-machine-approver/machine-approver-56656f9798-95s6m" event={"ID":"39b30c11-d0f6-4302-9ef8-e7fcf1b2170e","Type":"ContainerStarted","Data":"112f609837e90fe1a8a0d191e9aa39ac5d48bacf27026c994a5d87632840f279"} Oct 08 18:19:35 crc kubenswrapper[4859]: I1008 18:19:35.870148 4859 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-4s5w9" event={"ID":"01517214-a376-421e-9938-e548d76695d5","Type":"ContainerStarted","Data":"1e97b3dbac5ac96ce0d1152627d76b420aad2c297c46aa4943acc620eb9344bc"} Oct 08 18:19:35 crc kubenswrapper[4859]: I1008 18:19:35.872534 4859 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-6dlgn" event={"ID":"46a9f481-0772-4006-99e2-723919011659","Type":"ContainerStarted","Data":"15606cda09e263b8be370a1966cf30d757321051593f217ae6a1613a3d06ed40"} Oct 08 18:19:35 crc kubenswrapper[4859]: I1008 18:19:35.876185 4859 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 08 18:19:35 crc kubenswrapper[4859]: E1008 18:19:35.877359 4859 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-08 18:19:36.377341181 +0000 UTC m=+146.624180560 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 08 18:19:35 crc kubenswrapper[4859]: I1008 18:19:35.881881 4859 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/marketplace-operator-79b997595-bz6v8" event={"ID":"8d4b5b44-816f-44f6-98ab-c5d66aefaa08","Type":"ContainerStarted","Data":"b6f47fb2611d35c2a1a47f8cdcf8133fc94d28820b6d8a1908a935ecfa0d3e2d"} Oct 08 18:19:35 crc kubenswrapper[4859]: I1008 18:19:35.904627 4859 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/marketplace-operator-79b997595-bz6v8" Oct 08 18:19:35 crc kubenswrapper[4859]: I1008 18:19:35.904781 4859 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-storage-version-migrator/migrator-59844c95c7-pmngm" event={"ID":"5401cd5b-bc2e-43d4-8d13-b7e9134aa72f","Type":"ContainerStarted","Data":"48d2d17354112a7b79a47de4780ea40da9ede10405420e23229a3fa65b877735"} Oct 08 18:19:35 crc kubenswrapper[4859]: I1008 18:19:35.889916 4859 patch_prober.go:28] interesting pod/marketplace-operator-79b997595-bz6v8 container/marketplace-operator namespace/openshift-marketplace: Readiness probe status=failure output="Get \"http://10.217.0.37:8080/healthz\": dial tcp 10.217.0.37:8080: connect: connection refused" start-of-body= Oct 08 18:19:35 crc kubenswrapper[4859]: I1008 18:19:35.904980 4859 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-marketplace/marketplace-operator-79b997595-bz6v8" podUID="8d4b5b44-816f-44f6-98ab-c5d66aefaa08" containerName="marketplace-operator" probeResult="failure" output="Get \"http://10.217.0.37:8080/healthz\": dial tcp 10.217.0.37:8080: connect: connection refused" Oct 08 18:19:35 crc kubenswrapper[4859]: I1008 18:19:35.938996 4859 patch_prober.go:28] interesting pod/router-default-5444994796-r5gs2 container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Oct 08 18:19:35 crc kubenswrapper[4859]: [-]has-synced failed: reason withheld Oct 08 18:19:35 crc kubenswrapper[4859]: [+]process-running ok Oct 08 18:19:35 crc kubenswrapper[4859]: healthz check failed Oct 08 18:19:35 crc kubenswrapper[4859]: I1008 18:19:35.939282 4859 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-r5gs2" podUID="8e2b8a53-4bc3-499b-9746-628c323d2a0c" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Oct 08 18:19:35 crc kubenswrapper[4859]: I1008 18:19:35.971672 4859 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-5n49w" podStartSLOduration=124.971650965 podStartE2EDuration="2m4.971650965s" podCreationTimestamp="2025-10-08 18:17:31 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-08 18:19:35.860293843 +0000 UTC m=+146.107133222" watchObservedRunningTime="2025-10-08 18:19:35.971650965 +0000 UTC m=+146.218490344" Oct 08 18:19:35 crc kubenswrapper[4859]: I1008 18:19:35.991147 4859 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-kg44m\" (UID: \"4afad99b-9bf7-43c9-8dcf-7a3217136be5\") " pod="openshift-image-registry/image-registry-697d97f7c8-kg44m" Oct 08 18:19:35 crc kubenswrapper[4859]: E1008 18:19:35.995525 4859 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-08 18:19:36.495509977 +0000 UTC m=+146.742349356 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-kg44m" (UID: "4afad99b-9bf7-43c9-8dcf-7a3217136be5") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 08 18:19:36 crc kubenswrapper[4859]: I1008 18:19:36.006634 4859 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-vc7bh" event={"ID":"b2b5fcca-e076-4cbe-bc1e-249fcfd25e53","Type":"ContainerStarted","Data":"7e988f5b00ca7204b3d4d69db4887b5d1c36eeb107c32f8c69ff739c6aae1b30"} Oct 08 18:19:36 crc kubenswrapper[4859]: I1008 18:19:36.007742 4859 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-vc7bh" Oct 08 18:19:36 crc kubenswrapper[4859]: I1008 18:19:36.066872 4859 patch_prober.go:28] interesting pod/packageserver-d55dfcdfc-vc7bh container/packageserver namespace/openshift-operator-lifecycle-manager: Readiness probe status=failure output="Get \"https://10.217.0.24:5443/healthz\": dial tcp 10.217.0.24:5443: connect: connection refused" start-of-body= Oct 08 18:19:36 crc kubenswrapper[4859]: I1008 18:19:36.066940 4859 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-vc7bh" podUID="b2b5fcca-e076-4cbe-bc1e-249fcfd25e53" containerName="packageserver" probeResult="failure" output="Get \"https://10.217.0.24:5443/healthz\": dial tcp 10.217.0.24:5443: connect: connection refused" Oct 08 18:19:36 crc kubenswrapper[4859]: I1008 18:19:36.069820 4859 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-82pcg" podStartSLOduration=126.069807404 podStartE2EDuration="2m6.069807404s" podCreationTimestamp="2025-10-08 18:17:30 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-08 18:19:36.031036177 +0000 UTC m=+146.277875566" watchObservedRunningTime="2025-10-08 18:19:36.069807404 +0000 UTC m=+146.316646783" Oct 08 18:19:36 crc kubenswrapper[4859]: I1008 18:19:36.079263 4859 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-5f4q5" event={"ID":"7e23494c-2895-43b3-a35c-47a7b2d33ebc","Type":"ContainerStarted","Data":"0e1a75c2970c0c9cabac42007f9102f4114a37b02303e138fab00a12a2fdf707"} Oct 08 18:19:36 crc kubenswrapper[4859]: I1008 18:19:36.095242 4859 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 08 18:19:36 crc kubenswrapper[4859]: E1008 18:19:36.096122 4859 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-08 18:19:36.596094678 +0000 UTC m=+146.842934057 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 08 18:19:36 crc kubenswrapper[4859]: I1008 18:19:36.097304 4859 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="hostpath-provisioner/csi-hostpathplugin-q4lkb" event={"ID":"5e418cbb-0072-4767-933d-537c4a274e60","Type":"ContainerStarted","Data":"f4b3048b62c0cfcadcb8fded4f558807ae7eb5a4b16875c6002b356fdf8adb29"} Oct 08 18:19:36 crc kubenswrapper[4859]: I1008 18:19:36.100993 4859 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ingress-canary/ingress-canary-qtvkr" event={"ID":"8b8b079c-490b-4fb1-93a6-473855ca319f","Type":"ContainerStarted","Data":"ebd1ff0198c284b3c92ae79f1b71bca125fecb05f693265a79103d8448f73691"} Oct 08 18:19:36 crc kubenswrapper[4859]: I1008 18:19:36.102915 4859 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-6j5rk" event={"ID":"d407fd47-096d-4f2a-ace1-dab4ddf7192a","Type":"ContainerStarted","Data":"ccde7798e02b8915fb95d5812af95ec91c29e19028c4db4057fa194aed59b582"} Oct 08 18:19:36 crc kubenswrapper[4859]: I1008 18:19:36.103956 4859 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29332455-w9wt8" event={"ID":"d403dae9-eb8c-4c8d-9daa-bf2a0c815652","Type":"ContainerStarted","Data":"562f5b1c2a23fe3445feaf2252f9159c42be63c462f3daaa37a9a6a551fc6acc"} Oct 08 18:19:36 crc kubenswrapper[4859]: I1008 18:19:36.105543 4859 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ingress-operator/ingress-operator-5b745b69d9-pf7j6" event={"ID":"ed062c43-61ec-4951-a27c-77f165523d3f","Type":"ContainerStarted","Data":"a4942ddbc61ef7906397a1d56baf9f9947c35449622640212fb99d7d3c2cf776"} Oct 08 18:19:36 crc kubenswrapper[4859]: I1008 18:19:36.106645 4859 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-service-ca/service-ca-9c57cc56f-cs77r" event={"ID":"da9f04a1-bb6a-40de-846a-5e2fe04855d1","Type":"ContainerStarted","Data":"51460dfe5d766f41cb5cb7b79c80bfa36db6b1743aa62200b8adc65dd4b732b8"} Oct 08 18:19:36 crc kubenswrapper[4859]: I1008 18:19:36.111190 4859 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-q9nqj" event={"ID":"492ef569-6d45-4dd0-89e5-506172689768","Type":"ContainerStarted","Data":"42f52c55385dbea9c674a2530fe9cdd60fbeccfd57d3580f840b549b28c97a41"} Oct 08 18:19:36 crc kubenswrapper[4859]: I1008 18:19:36.124661 4859 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-cluster-machine-approver/machine-approver-56656f9798-95s6m" podStartSLOduration=126.12463323 podStartE2EDuration="2m6.12463323s" podCreationTimestamp="2025-10-08 18:17:30 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-08 18:19:36.121236499 +0000 UTC m=+146.368075878" watchObservedRunningTime="2025-10-08 18:19:36.12463323 +0000 UTC m=+146.371472599" Oct 08 18:19:36 crc kubenswrapper[4859]: I1008 18:19:36.143611 4859 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-operator-74547568cd-xsc6l" event={"ID":"0c9973e7-5c40-463b-aca1-c3c93d7a9d0b","Type":"ContainerStarted","Data":"78f7c9af012ffd0e8974372be2cc8bdcbd5947d8e8d2b062687b875224117a16"} Oct 08 18:19:36 crc kubenswrapper[4859]: I1008 18:19:36.143666 4859 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-operator-74547568cd-xsc6l" event={"ID":"0c9973e7-5c40-463b-aca1-c3c93d7a9d0b","Type":"ContainerStarted","Data":"36d27b10da2dd37ebd577d92a962aa3e8f5fae22f5c373f8cec414fa0da54d29"} Oct 08 18:19:36 crc kubenswrapper[4859]: I1008 18:19:36.144291 4859 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-6dlgn" podStartSLOduration=125.144273826 podStartE2EDuration="2m5.144273826s" podCreationTimestamp="2025-10-08 18:17:31 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-08 18:19:36.143751421 +0000 UTC m=+146.390590800" watchObservedRunningTime="2025-10-08 18:19:36.144273826 +0000 UTC m=+146.391113205" Oct 08 18:19:36 crc kubenswrapper[4859]: I1008 18:19:36.179015 4859 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-vc7bh" podStartSLOduration=125.178999552 podStartE2EDuration="2m5.178999552s" podCreationTimestamp="2025-10-08 18:17:31 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-08 18:19:36.177996502 +0000 UTC m=+146.424835881" watchObservedRunningTime="2025-10-08 18:19:36.178999552 +0000 UTC m=+146.425838931" Oct 08 18:19:36 crc kubenswrapper[4859]: I1008 18:19:36.190386 4859 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-c5p45" event={"ID":"9b239e22-0030-4c68-b0f1-66fbaedd5f9b","Type":"ContainerStarted","Data":"c7edda8bbc9d066ea2cb92ec4f72ae1add2ee9ad40cb4975a8d354c97be0ab29"} Oct 08 18:19:36 crc kubenswrapper[4859]: I1008 18:19:36.196360 4859 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-kg44m\" (UID: \"4afad99b-9bf7-43c9-8dcf-7a3217136be5\") " pod="openshift-image-registry/image-registry-697d97f7c8-kg44m" Oct 08 18:19:36 crc kubenswrapper[4859]: E1008 18:19:36.196661 4859 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-08 18:19:36.696647369 +0000 UTC m=+146.943486748 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-kg44m" (UID: "4afad99b-9bf7-43c9-8dcf-7a3217136be5") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 08 18:19:36 crc kubenswrapper[4859]: I1008 18:19:36.214552 4859 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-hf2fp" event={"ID":"c073e5c1-652c-406a-ad01-9ca626e8c56e","Type":"ContainerStarted","Data":"da2e2d35c46ca71d52156621a190437a8449fe93ed08bb1e8bd2a0c5e729c27e"} Oct 08 18:19:36 crc kubenswrapper[4859]: I1008 18:19:36.215213 4859 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-machine-config-operator/machine-config-controller-84d6567774-m7m5n" podStartSLOduration=125.215193212 podStartE2EDuration="2m5.215193212s" podCreationTimestamp="2025-10-08 18:17:31 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-08 18:19:36.214999987 +0000 UTC m=+146.461839376" watchObservedRunningTime="2025-10-08 18:19:36.215193212 +0000 UTC m=+146.462032591" Oct 08 18:19:36 crc kubenswrapper[4859]: I1008 18:19:36.250123 4859 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-xpnsv" event={"ID":"be85b712-b3e4-4535-b14b-672605d02387","Type":"ContainerStarted","Data":"0289eb7286a940f52eb6d1b2ee38185230594371fd5146aa57111cb7e54fb058"} Oct 08 18:19:36 crc kubenswrapper[4859]: I1008 18:19:36.250673 4859 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-xpnsv" Oct 08 18:19:36 crc kubenswrapper[4859]: I1008 18:19:36.252116 4859 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-machine-api/machine-api-operator-5694c8668f-gmbk9" podStartSLOduration=125.252103754 podStartE2EDuration="2m5.252103754s" podCreationTimestamp="2025-10-08 18:17:31 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-08 18:19:36.250186586 +0000 UTC m=+146.497025965" watchObservedRunningTime="2025-10-08 18:19:36.252103754 +0000 UTC m=+146.498943133" Oct 08 18:19:36 crc kubenswrapper[4859]: I1008 18:19:36.269933 4859 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-admission-controller-857f4d67dd-lrqkp" event={"ID":"f114c686-8fde-4560-b8e1-9cd8fea448f8","Type":"ContainerStarted","Data":"aacdae4eeeafd8e57e87d20865090f491c49e665abd1c91d07209ab2db97ff77"} Oct 08 18:19:36 crc kubenswrapper[4859]: I1008 18:19:36.297479 4859 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 08 18:19:36 crc kubenswrapper[4859]: E1008 18:19:36.298783 4859 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-08 18:19:36.798763906 +0000 UTC m=+147.045603285 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 08 18:19:36 crc kubenswrapper[4859]: I1008 18:19:36.306666 4859 patch_prober.go:28] interesting pod/catalog-operator-68c6474976-xpnsv container/catalog-operator namespace/openshift-operator-lifecycle-manager: Readiness probe status=failure output="Get \"https://10.217.0.32:8443/healthz\": dial tcp 10.217.0.32:8443: connect: connection refused" start-of-body= Oct 08 18:19:36 crc kubenswrapper[4859]: I1008 18:19:36.307067 4859 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-xpnsv" podUID="be85b712-b3e4-4535-b14b-672605d02387" containerName="catalog-operator" probeResult="failure" output="Get \"https://10.217.0.32:8443/healthz\": dial tcp 10.217.0.32:8443: connect: connection refused" Oct 08 18:19:36 crc kubenswrapper[4859]: I1008 18:19:36.318990 4859 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-kc42p" event={"ID":"bb6dc185-94e3-4cb7-8076-1011c1c60fab","Type":"ContainerStarted","Data":"5a80493c2b740c108d681123a559543b768da07086e879d8efded42dffd50319"} Oct 08 18:19:36 crc kubenswrapper[4859]: I1008 18:19:36.343187 4859 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-4s5w9" podStartSLOduration=125.343167811 podStartE2EDuration="2m5.343167811s" podCreationTimestamp="2025-10-08 18:17:31 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-08 18:19:36.291091417 +0000 UTC m=+146.537930796" watchObservedRunningTime="2025-10-08 18:19:36.343167811 +0000 UTC m=+146.590007190" Oct 08 18:19:36 crc kubenswrapper[4859]: I1008 18:19:36.354123 4859 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-dns-operator/dns-operator-744455d44c-q222b" event={"ID":"63636924-a28b-447e-8568-92a193a69ade","Type":"ContainerStarted","Data":"b85f078c42af7b53d815f1c44a95ec495470f60d874a3008807aca14a8399fbd"} Oct 08 18:19:36 crc kubenswrapper[4859]: I1008 18:19:36.357222 4859 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/marketplace-operator-79b997595-bz6v8" podStartSLOduration=125.35719482 podStartE2EDuration="2m5.35719482s" podCreationTimestamp="2025-10-08 18:17:31 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-08 18:19:36.329976097 +0000 UTC m=+146.576815466" watchObservedRunningTime="2025-10-08 18:19:36.35719482 +0000 UTC m=+146.604034199" Oct 08 18:19:36 crc kubenswrapper[4859]: I1008 18:19:36.360519 4859 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-console/downloads-7954f5f757-8q5jz" Oct 08 18:19:36 crc kubenswrapper[4859]: I1008 18:19:36.368902 4859 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-xpnsv" podStartSLOduration=125.368878718 podStartE2EDuration="2m5.368878718s" podCreationTimestamp="2025-10-08 18:17:31 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-08 18:19:36.353094037 +0000 UTC m=+146.599933416" watchObservedRunningTime="2025-10-08 18:19:36.368878718 +0000 UTC m=+146.615718097" Oct 08 18:19:36 crc kubenswrapper[4859]: I1008 18:19:36.386901 4859 patch_prober.go:28] interesting pod/downloads-7954f5f757-8q5jz container/download-server namespace/openshift-console: Readiness probe status=failure output="Get \"http://10.217.0.20:8080/\": dial tcp 10.217.0.20:8080: connect: connection refused" start-of-body= Oct 08 18:19:36 crc kubenswrapper[4859]: I1008 18:19:36.386960 4859 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-console/downloads-7954f5f757-8q5jz" podUID="9dcb8f76-0fff-436c-9191-fe3f787b7ac0" containerName="download-server" probeResult="failure" output="Get \"http://10.217.0.20:8080/\": dial tcp 10.217.0.20:8080: connect: connection refused" Oct 08 18:19:36 crc kubenswrapper[4859]: I1008 18:19:36.387569 4859 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-5n49w" Oct 08 18:19:36 crc kubenswrapper[4859]: I1008 18:19:36.393230 4859 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-controller-manager/controller-manager-879f6c89f-9gpzj" Oct 08 18:19:36 crc kubenswrapper[4859]: I1008 18:19:36.402338 4859 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-ingress-canary/ingress-canary-qtvkr" podStartSLOduration=6.402317816 podStartE2EDuration="6.402317816s" podCreationTimestamp="2025-10-08 18:19:30 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-08 18:19:36.401924974 +0000 UTC m=+146.648764353" watchObservedRunningTime="2025-10-08 18:19:36.402317816 +0000 UTC m=+146.649157195" Oct 08 18:19:36 crc kubenswrapper[4859]: I1008 18:19:36.402980 4859 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-kg44m\" (UID: \"4afad99b-9bf7-43c9-8dcf-7a3217136be5\") " pod="openshift-image-registry/image-registry-697d97f7c8-kg44m" Oct 08 18:19:36 crc kubenswrapper[4859]: E1008 18:19:36.405555 4859 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-08 18:19:36.905534992 +0000 UTC m=+147.152374371 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-kg44m" (UID: "4afad99b-9bf7-43c9-8dcf-7a3217136be5") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 08 18:19:36 crc kubenswrapper[4859]: I1008 18:19:36.408310 4859 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-xlfdd" Oct 08 18:19:36 crc kubenswrapper[4859]: I1008 18:19:36.491442 4859 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-kc42p" podStartSLOduration=125.491421925 podStartE2EDuration="2m5.491421925s" podCreationTimestamp="2025-10-08 18:17:31 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-08 18:19:36.491363513 +0000 UTC m=+146.738202902" watchObservedRunningTime="2025-10-08 18:19:36.491421925 +0000 UTC m=+146.738261304" Oct 08 18:19:36 crc kubenswrapper[4859]: I1008 18:19:36.491622 4859 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-machine-config-operator/machine-config-operator-74547568cd-xsc6l" podStartSLOduration=125.491617241 podStartE2EDuration="2m5.491617241s" podCreationTimestamp="2025-10-08 18:17:31 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-08 18:19:36.445318499 +0000 UTC m=+146.692157878" watchObservedRunningTime="2025-10-08 18:19:36.491617241 +0000 UTC m=+146.738456620" Oct 08 18:19:36 crc kubenswrapper[4859]: I1008 18:19:36.506307 4859 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 08 18:19:36 crc kubenswrapper[4859]: E1008 18:19:36.508014 4859 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-08 18:19:37.007981639 +0000 UTC m=+147.254821018 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 08 18:19:36 crc kubenswrapper[4859]: I1008 18:19:36.580285 4859 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-operator-lifecycle-manager/collect-profiles-29332455-w9wt8" podStartSLOduration=126.580259976 podStartE2EDuration="2m6.580259976s" podCreationTimestamp="2025-10-08 18:17:30 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-08 18:19:36.547081026 +0000 UTC m=+146.793920405" watchObservedRunningTime="2025-10-08 18:19:36.580259976 +0000 UTC m=+146.827099355" Oct 08 18:19:36 crc kubenswrapper[4859]: I1008 18:19:36.611008 4859 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-kg44m\" (UID: \"4afad99b-9bf7-43c9-8dcf-7a3217136be5\") " pod="openshift-image-registry/image-registry-697d97f7c8-kg44m" Oct 08 18:19:36 crc kubenswrapper[4859]: E1008 18:19:36.611379 4859 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-08 18:19:37.111367274 +0000 UTC m=+147.358206643 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-kg44m" (UID: "4afad99b-9bf7-43c9-8dcf-7a3217136be5") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 08 18:19:36 crc kubenswrapper[4859]: I1008 18:19:36.620888 4859 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-service-ca/service-ca-9c57cc56f-cs77r" podStartSLOduration=125.620869858 podStartE2EDuration="2m5.620869858s" podCreationTimestamp="2025-10-08 18:17:31 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-08 18:19:36.620571499 +0000 UTC m=+146.867410888" watchObservedRunningTime="2025-10-08 18:19:36.620869858 +0000 UTC m=+146.867709237" Oct 08 18:19:36 crc kubenswrapper[4859]: I1008 18:19:36.677087 4859 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-multus/multus-admission-controller-857f4d67dd-lrqkp" podStartSLOduration=125.677064975 podStartE2EDuration="2m5.677064975s" podCreationTimestamp="2025-10-08 18:17:31 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-08 18:19:36.675357924 +0000 UTC m=+146.922197303" watchObservedRunningTime="2025-10-08 18:19:36.677064975 +0000 UTC m=+146.923904354" Oct 08 18:19:36 crc kubenswrapper[4859]: I1008 18:19:36.740781 4859 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 08 18:19:36 crc kubenswrapper[4859]: E1008 18:19:36.741453 4859 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-08 18:19:37.241436756 +0000 UTC m=+147.488276135 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 08 18:19:36 crc kubenswrapper[4859]: I1008 18:19:36.844069 4859 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-kg44m\" (UID: \"4afad99b-9bf7-43c9-8dcf-7a3217136be5\") " pod="openshift-image-registry/image-registry-697d97f7c8-kg44m" Oct 08 18:19:36 crc kubenswrapper[4859]: E1008 18:19:36.844525 4859 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-08 18:19:37.344508151 +0000 UTC m=+147.591347530 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-kg44m" (UID: "4afad99b-9bf7-43c9-8dcf-7a3217136be5") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 08 18:19:36 crc kubenswrapper[4859]: I1008 18:19:36.933000 4859 patch_prober.go:28] interesting pod/router-default-5444994796-r5gs2 container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Oct 08 18:19:36 crc kubenswrapper[4859]: [-]has-synced failed: reason withheld Oct 08 18:19:36 crc kubenswrapper[4859]: [+]process-running ok Oct 08 18:19:36 crc kubenswrapper[4859]: healthz check failed Oct 08 18:19:36 crc kubenswrapper[4859]: I1008 18:19:36.933085 4859 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-r5gs2" podUID="8e2b8a53-4bc3-499b-9746-628c323d2a0c" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Oct 08 18:19:36 crc kubenswrapper[4859]: I1008 18:19:36.949336 4859 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 08 18:19:36 crc kubenswrapper[4859]: E1008 18:19:36.949707 4859 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-08 18:19:37.449671919 +0000 UTC m=+147.696511298 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 08 18:19:37 crc kubenswrapper[4859]: I1008 18:19:37.050584 4859 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-kg44m\" (UID: \"4afad99b-9bf7-43c9-8dcf-7a3217136be5\") " pod="openshift-image-registry/image-registry-697d97f7c8-kg44m" Oct 08 18:19:37 crc kubenswrapper[4859]: E1008 18:19:37.051065 4859 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-08 18:19:37.551049094 +0000 UTC m=+147.797888463 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-kg44m" (UID: "4afad99b-9bf7-43c9-8dcf-7a3217136be5") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 08 18:19:37 crc kubenswrapper[4859]: I1008 18:19:37.151881 4859 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 08 18:19:37 crc kubenswrapper[4859]: E1008 18:19:37.152194 4859 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-08 18:19:37.652166622 +0000 UTC m=+147.899006001 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 08 18:19:37 crc kubenswrapper[4859]: I1008 18:19:37.152411 4859 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-kg44m\" (UID: \"4afad99b-9bf7-43c9-8dcf-7a3217136be5\") " pod="openshift-image-registry/image-registry-697d97f7c8-kg44m" Oct 08 18:19:37 crc kubenswrapper[4859]: E1008 18:19:37.152828 4859 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-08 18:19:37.652815961 +0000 UTC m=+147.899655340 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-kg44m" (UID: "4afad99b-9bf7-43c9-8dcf-7a3217136be5") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 08 18:19:37 crc kubenswrapper[4859]: I1008 18:19:37.212012 4859 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-config-operator/openshift-config-operator-7777fb866f-ncj6n" Oct 08 18:19:37 crc kubenswrapper[4859]: I1008 18:19:37.253363 4859 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 08 18:19:37 crc kubenswrapper[4859]: E1008 18:19:37.253799 4859 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-08 18:19:37.753783284 +0000 UTC m=+148.000622663 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 08 18:19:37 crc kubenswrapper[4859]: I1008 18:19:37.355088 4859 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-kg44m\" (UID: \"4afad99b-9bf7-43c9-8dcf-7a3217136be5\") " pod="openshift-image-registry/image-registry-697d97f7c8-kg44m" Oct 08 18:19:37 crc kubenswrapper[4859]: E1008 18:19:37.355444 4859 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-08 18:19:37.855422437 +0000 UTC m=+148.102261816 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-kg44m" (UID: "4afad99b-9bf7-43c9-8dcf-7a3217136be5") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 08 18:19:37 crc kubenswrapper[4859]: I1008 18:19:37.370441 4859 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ingress-operator/ingress-operator-5b745b69d9-pf7j6" event={"ID":"ed062c43-61ec-4951-a27c-77f165523d3f","Type":"ContainerStarted","Data":"62fd58415be5a7378c9dbe85aa3637f617483c6731a64396babbe3f88d415542"} Oct 08 18:19:37 crc kubenswrapper[4859]: I1008 18:19:37.372523 4859 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-q9nqj" event={"ID":"492ef569-6d45-4dd0-89e5-506172689768","Type":"ContainerStarted","Data":"38fdb422287af705123e4e875182a3c7ea389d90b9ea0cd7eb0de4909b11a28e"} Oct 08 18:19:37 crc kubenswrapper[4859]: I1008 18:19:37.374374 4859 generic.go:334] "Generic (PLEG): container finished" podID="d407fd47-096d-4f2a-ace1-dab4ddf7192a" containerID="0b2476b7b5d1cbb1881d5a53a097bc63e421769019453b9b9966e4a831f66a20" exitCode=0 Oct 08 18:19:37 crc kubenswrapper[4859]: I1008 18:19:37.374834 4859 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-6j5rk" event={"ID":"d407fd47-096d-4f2a-ace1-dab4ddf7192a","Type":"ContainerDied","Data":"0b2476b7b5d1cbb1881d5a53a097bc63e421769019453b9b9966e4a831f66a20"} Oct 08 18:19:37 crc kubenswrapper[4859]: I1008 18:19:37.377326 4859 generic.go:334] "Generic (PLEG): container finished" podID="d403dae9-eb8c-4c8d-9daa-bf2a0c815652" containerID="562f5b1c2a23fe3445feaf2252f9159c42be63c462f3daaa37a9a6a551fc6acc" exitCode=0 Oct 08 18:19:37 crc kubenswrapper[4859]: I1008 18:19:37.377429 4859 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29332455-w9wt8" event={"ID":"d403dae9-eb8c-4c8d-9daa-bf2a0c815652","Type":"ContainerDied","Data":"562f5b1c2a23fe3445feaf2252f9159c42be63c462f3daaa37a9a6a551fc6acc"} Oct 08 18:19:37 crc kubenswrapper[4859]: I1008 18:19:37.380483 4859 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-dns-operator/dns-operator-744455d44c-q222b" event={"ID":"63636924-a28b-447e-8568-92a193a69ade","Type":"ContainerStarted","Data":"c5e8d49da38a99f42ec5084da9a4db2179b350ef84e5a45dff98cc9c3709920a"} Oct 08 18:19:37 crc kubenswrapper[4859]: I1008 18:19:37.399611 4859 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-dns/dns-default-x7pp9" event={"ID":"2a959183-1370-4723-bb5e-8625689d6121","Type":"ContainerStarted","Data":"7a13c3b364044be050f283535f99a861f3fd7096ea790a20078057322ae65e3b"} Oct 08 18:19:37 crc kubenswrapper[4859]: I1008 18:19:37.399674 4859 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-dns/dns-default-x7pp9" Oct 08 18:19:37 crc kubenswrapper[4859]: I1008 18:19:37.402984 4859 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ingress-canary/ingress-canary-qtvkr" event={"ID":"8b8b079c-490b-4fb1-93a6-473855ca319f","Type":"ContainerStarted","Data":"60795d1c585f2aa99452489593a4ad282036dc0372d5cf447e0845bdfae941d3"} Oct 08 18:19:37 crc kubenswrapper[4859]: I1008 18:19:37.409502 4859 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-c5p45" event={"ID":"9b239e22-0030-4c68-b0f1-66fbaedd5f9b","Type":"ContainerStarted","Data":"b9ab2e760d70c7070625e3e5ae8668a4dfab1f5accbeca857625e4416a3befb2"} Oct 08 18:19:37 crc kubenswrapper[4859]: I1008 18:19:37.422350 4859 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-apiserver/apiserver-76f77b778f-fv9bp" event={"ID":"70f88ce6-b9fe-422e-827e-af2b6840a783","Type":"ContainerStarted","Data":"91e75cc938cdf082fcc725cffd8d83ddf90909314a95e11330cfe14def2c9147"} Oct 08 18:19:37 crc kubenswrapper[4859]: I1008 18:19:37.427200 4859 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-hf2fp" event={"ID":"c073e5c1-652c-406a-ad01-9ca626e8c56e","Type":"ContainerStarted","Data":"3b56401e70af68aabd46086aedd80c7cfe24b3261b4e6d2ed1770dcd4a262f33"} Oct 08 18:19:37 crc kubenswrapper[4859]: I1008 18:19:37.427238 4859 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-hf2fp" event={"ID":"c073e5c1-652c-406a-ad01-9ca626e8c56e","Type":"ContainerStarted","Data":"6d64bbfba5cb2e0f03533dad28b2ce0baf37c000d7c127fd2cab70619a117718"} Oct 08 18:19:37 crc kubenswrapper[4859]: I1008 18:19:37.427673 4859 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-hf2fp" Oct 08 18:19:37 crc kubenswrapper[4859]: I1008 18:19:37.437031 4859 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-storage-version-migrator/migrator-59844c95c7-pmngm" event={"ID":"5401cd5b-bc2e-43d4-8d13-b7e9134aa72f","Type":"ContainerStarted","Data":"38e137f0cc2fb29a35952b522458202e72fef7f108f1cc344dff9a959f3e9143"} Oct 08 18:19:37 crc kubenswrapper[4859]: I1008 18:19:37.448500 4859 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-5f4q5" event={"ID":"7e23494c-2895-43b3-a35c-47a7b2d33ebc","Type":"ContainerStarted","Data":"3a708ac7ff5d31f0186974bd4d59abf6482fc65989cbde5f016c174b92702abc"} Oct 08 18:19:37 crc kubenswrapper[4859]: I1008 18:19:37.450378 4859 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-ingress-operator/ingress-operator-5b745b69d9-pf7j6" podStartSLOduration=126.4503656 podStartE2EDuration="2m6.4503656s" podCreationTimestamp="2025-10-08 18:17:31 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-08 18:19:37.447231567 +0000 UTC m=+147.694070936" watchObservedRunningTime="2025-10-08 18:19:37.4503656 +0000 UTC m=+147.697204979" Oct 08 18:19:37 crc kubenswrapper[4859]: I1008 18:19:37.456852 4859 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-apiserver/apiserver-76f77b778f-fv9bp" Oct 08 18:19:37 crc kubenswrapper[4859]: I1008 18:19:37.457044 4859 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-apiserver/apiserver-76f77b778f-fv9bp" Oct 08 18:19:37 crc kubenswrapper[4859]: I1008 18:19:37.457666 4859 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 08 18:19:37 crc kubenswrapper[4859]: E1008 18:19:37.458061 4859 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-08 18:19:37.958045019 +0000 UTC m=+148.204884388 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 08 18:19:37 crc kubenswrapper[4859]: I1008 18:19:37.464043 4859 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-xpnsv" event={"ID":"be85b712-b3e4-4535-b14b-672605d02387","Type":"ContainerStarted","Data":"7eabf561ddd4c50d9d14d4303a724cff96aa530d6611218d164308ffd8b0fcc0"} Oct 08 18:19:37 crc kubenswrapper[4859]: I1008 18:19:37.464371 4859 patch_prober.go:28] interesting pod/catalog-operator-68c6474976-xpnsv container/catalog-operator namespace/openshift-operator-lifecycle-manager: Readiness probe status=failure output="Get \"https://10.217.0.32:8443/healthz\": dial tcp 10.217.0.32:8443: connect: connection refused" start-of-body= Oct 08 18:19:37 crc kubenswrapper[4859]: I1008 18:19:37.464415 4859 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-xpnsv" podUID="be85b712-b3e4-4535-b14b-672605d02387" containerName="catalog-operator" probeResult="failure" output="Get \"https://10.217.0.32:8443/healthz\": dial tcp 10.217.0.32:8443: connect: connection refused" Oct 08 18:19:37 crc kubenswrapper[4859]: I1008 18:19:37.478743 4859 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-authentication/oauth-openshift-558db77b4-j8gh2" event={"ID":"0902ccb2-78ab-48d0-be02-4c42f03e12fb","Type":"ContainerStarted","Data":"7023b5c72b539780ca0fa2a5d77a71496bfa187a6666d8c3d8fd2b6e898c0c70"} Oct 08 18:19:37 crc kubenswrapper[4859]: I1008 18:19:37.478785 4859 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-authentication/oauth-openshift-558db77b4-j8gh2" Oct 08 18:19:37 crc kubenswrapper[4859]: I1008 18:19:37.480935 4859 patch_prober.go:28] interesting pod/marketplace-operator-79b997595-bz6v8 container/marketplace-operator namespace/openshift-marketplace: Readiness probe status=failure output="Get \"http://10.217.0.37:8080/healthz\": dial tcp 10.217.0.37:8080: connect: connection refused" start-of-body= Oct 08 18:19:37 crc kubenswrapper[4859]: I1008 18:19:37.480984 4859 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-marketplace/marketplace-operator-79b997595-bz6v8" podUID="8d4b5b44-816f-44f6-98ab-c5d66aefaa08" containerName="marketplace-operator" probeResult="failure" output="Get \"http://10.217.0.37:8080/healthz\": dial tcp 10.217.0.37:8080: connect: connection refused" Oct 08 18:19:37 crc kubenswrapper[4859]: I1008 18:19:37.481037 4859 patch_prober.go:28] interesting pod/downloads-7954f5f757-8q5jz container/download-server namespace/openshift-console: Readiness probe status=failure output="Get \"http://10.217.0.20:8080/\": dial tcp 10.217.0.20:8080: connect: connection refused" start-of-body= Oct 08 18:19:37 crc kubenswrapper[4859]: I1008 18:19:37.481049 4859 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-console/downloads-7954f5f757-8q5jz" podUID="9dcb8f76-0fff-436c-9191-fe3f787b7ac0" containerName="download-server" probeResult="failure" output="Get \"http://10.217.0.20:8080/\": dial tcp 10.217.0.20:8080: connect: connection refused" Oct 08 18:19:37 crc kubenswrapper[4859]: I1008 18:19:37.481105 4859 patch_prober.go:28] interesting pod/oauth-openshift-558db77b4-j8gh2 container/oauth-openshift namespace/openshift-authentication: Readiness probe status=failure output="Get \"https://10.217.0.6:6443/healthz\": dial tcp 10.217.0.6:6443: connect: connection refused" start-of-body= Oct 08 18:19:37 crc kubenswrapper[4859]: I1008 18:19:37.481117 4859 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-authentication/oauth-openshift-558db77b4-j8gh2" podUID="0902ccb2-78ab-48d0-be02-4c42f03e12fb" containerName="oauth-openshift" probeResult="failure" output="Get \"https://10.217.0.6:6443/healthz\": dial tcp 10.217.0.6:6443: connect: connection refused" Oct 08 18:19:37 crc kubenswrapper[4859]: I1008 18:19:37.562540 4859 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-dns/dns-default-x7pp9" podStartSLOduration=7.562518677 podStartE2EDuration="7.562518677s" podCreationTimestamp="2025-10-08 18:19:30 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-08 18:19:37.562252769 +0000 UTC m=+147.809092148" watchObservedRunningTime="2025-10-08 18:19:37.562518677 +0000 UTC m=+147.809358056" Oct 08 18:19:37 crc kubenswrapper[4859]: I1008 18:19:37.563372 4859 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-kg44m\" (UID: \"4afad99b-9bf7-43c9-8dcf-7a3217136be5\") " pod="openshift-image-registry/image-registry-697d97f7c8-kg44m" Oct 08 18:19:37 crc kubenswrapper[4859]: E1008 18:19:37.567677 4859 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-08 18:19:38.0676551 +0000 UTC m=+148.314494689 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-kg44m" (UID: "4afad99b-9bf7-43c9-8dcf-7a3217136be5") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 08 18:19:37 crc kubenswrapper[4859]: I1008 18:19:37.665254 4859 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 08 18:19:37 crc kubenswrapper[4859]: I1008 18:19:37.670671 4859 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-apiserver/apiserver-76f77b778f-fv9bp" podStartSLOduration=127.670643023 podStartE2EDuration="2m7.670643023s" podCreationTimestamp="2025-10-08 18:17:30 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-08 18:19:37.652300826 +0000 UTC m=+147.899140225" watchObservedRunningTime="2025-10-08 18:19:37.670643023 +0000 UTC m=+147.917482412" Oct 08 18:19:37 crc kubenswrapper[4859]: E1008 18:19:37.671795 4859 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-08 18:19:38.171775057 +0000 UTC m=+148.418614436 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 08 18:19:37 crc kubenswrapper[4859]: I1008 18:19:37.729057 4859 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-c5p45" podStartSLOduration=127.729036716 podStartE2EDuration="2m7.729036716s" podCreationTimestamp="2025-10-08 18:17:30 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-08 18:19:37.728092867 +0000 UTC m=+147.974932256" watchObservedRunningTime="2025-10-08 18:19:37.729036716 +0000 UTC m=+147.975876095" Oct 08 18:19:37 crc kubenswrapper[4859]: I1008 18:19:37.773816 4859 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-kg44m\" (UID: \"4afad99b-9bf7-43c9-8dcf-7a3217136be5\") " pod="openshift-image-registry/image-registry-697d97f7c8-kg44m" Oct 08 18:19:37 crc kubenswrapper[4859]: E1008 18:19:37.774471 4859 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-08 18:19:38.274453961 +0000 UTC m=+148.521293340 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-kg44m" (UID: "4afad99b-9bf7-43c9-8dcf-7a3217136be5") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 08 18:19:37 crc kubenswrapper[4859]: I1008 18:19:37.879389 4859 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 08 18:19:37 crc kubenswrapper[4859]: E1008 18:19:37.879503 4859 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-08 18:19:38.379482325 +0000 UTC m=+148.626321704 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 08 18:19:37 crc kubenswrapper[4859]: I1008 18:19:37.879928 4859 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-kg44m\" (UID: \"4afad99b-9bf7-43c9-8dcf-7a3217136be5\") " pod="openshift-image-registry/image-registry-697d97f7c8-kg44m" Oct 08 18:19:37 crc kubenswrapper[4859]: E1008 18:19:37.880266 4859 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-08 18:19:38.380256538 +0000 UTC m=+148.627095917 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-kg44m" (UID: "4afad99b-9bf7-43c9-8dcf-7a3217136be5") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 08 18:19:37 crc kubenswrapper[4859]: I1008 18:19:37.938917 4859 patch_prober.go:28] interesting pod/router-default-5444994796-r5gs2 container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Oct 08 18:19:37 crc kubenswrapper[4859]: [-]has-synced failed: reason withheld Oct 08 18:19:37 crc kubenswrapper[4859]: [+]process-running ok Oct 08 18:19:37 crc kubenswrapper[4859]: healthz check failed Oct 08 18:19:37 crc kubenswrapper[4859]: I1008 18:19:37.938988 4859 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-r5gs2" podUID="8e2b8a53-4bc3-499b-9746-628c323d2a0c" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Oct 08 18:19:37 crc kubenswrapper[4859]: I1008 18:19:37.940561 4859 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-hf2fp" podStartSLOduration=126.940546567 podStartE2EDuration="2m6.940546567s" podCreationTimestamp="2025-10-08 18:17:31 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-08 18:19:37.929222319 +0000 UTC m=+148.176061698" watchObservedRunningTime="2025-10-08 18:19:37.940546567 +0000 UTC m=+148.187385946" Oct 08 18:19:37 crc kubenswrapper[4859]: I1008 18:19:37.941731 4859 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-q9nqj" podStartSLOduration=126.941723122 podStartE2EDuration="2m6.941723122s" podCreationTimestamp="2025-10-08 18:17:31 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-08 18:19:37.878143825 +0000 UTC m=+148.124983204" watchObservedRunningTime="2025-10-08 18:19:37.941723122 +0000 UTC m=+148.188562501" Oct 08 18:19:37 crc kubenswrapper[4859]: I1008 18:19:37.981287 4859 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 08 18:19:37 crc kubenswrapper[4859]: E1008 18:19:37.981757 4859 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-08 18:19:38.481738836 +0000 UTC m=+148.728578215 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 08 18:19:38 crc kubenswrapper[4859]: I1008 18:19:38.005951 4859 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-dns-operator/dns-operator-744455d44c-q222b" podStartSLOduration=128.005928838 podStartE2EDuration="2m8.005928838s" podCreationTimestamp="2025-10-08 18:17:30 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-08 18:19:38.005121024 +0000 UTC m=+148.251960393" watchObservedRunningTime="2025-10-08 18:19:38.005928838 +0000 UTC m=+148.252768207" Oct 08 18:19:38 crc kubenswrapper[4859]: I1008 18:19:38.007410 4859 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-storage-version-migrator/migrator-59844c95c7-pmngm" podStartSLOduration=127.007402912 podStartE2EDuration="2m7.007402912s" podCreationTimestamp="2025-10-08 18:17:31 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-08 18:19:37.975973854 +0000 UTC m=+148.222813243" watchObservedRunningTime="2025-10-08 18:19:38.007402912 +0000 UTC m=+148.254242291" Oct 08 18:19:38 crc kubenswrapper[4859]: I1008 18:19:38.082557 4859 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-kg44m\" (UID: \"4afad99b-9bf7-43c9-8dcf-7a3217136be5\") " pod="openshift-image-registry/image-registry-697d97f7c8-kg44m" Oct 08 18:19:38 crc kubenswrapper[4859]: E1008 18:19:38.082929 4859 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-08 18:19:38.582914896 +0000 UTC m=+148.829754275 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-kg44m" (UID: "4afad99b-9bf7-43c9-8dcf-7a3217136be5") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 08 18:19:38 crc kubenswrapper[4859]: I1008 18:19:38.124775 4859 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-authentication/oauth-openshift-558db77b4-j8gh2" podStartSLOduration=128.124747574 podStartE2EDuration="2m8.124747574s" podCreationTimestamp="2025-10-08 18:17:30 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-08 18:19:38.121308771 +0000 UTC m=+148.368148150" watchObservedRunningTime="2025-10-08 18:19:38.124747574 +0000 UTC m=+148.371586953" Oct 08 18:19:38 crc kubenswrapper[4859]: I1008 18:19:38.154411 4859 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-5f4q5" podStartSLOduration=127.154387848 podStartE2EDuration="2m7.154387848s" podCreationTimestamp="2025-10-08 18:17:31 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-08 18:19:38.151524123 +0000 UTC m=+148.398363502" watchObservedRunningTime="2025-10-08 18:19:38.154387848 +0000 UTC m=+148.401227227" Oct 08 18:19:38 crc kubenswrapper[4859]: I1008 18:19:38.183285 4859 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 08 18:19:38 crc kubenswrapper[4859]: E1008 18:19:38.183892 4859 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-08 18:19:38.683874938 +0000 UTC m=+148.930714317 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 08 18:19:38 crc kubenswrapper[4859]: I1008 18:19:38.285532 4859 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-kg44m\" (UID: \"4afad99b-9bf7-43c9-8dcf-7a3217136be5\") " pod="openshift-image-registry/image-registry-697d97f7c8-kg44m" Oct 08 18:19:38 crc kubenswrapper[4859]: E1008 18:19:38.285938 4859 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-08 18:19:38.785924193 +0000 UTC m=+149.032763572 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-kg44m" (UID: "4afad99b-9bf7-43c9-8dcf-7a3217136be5") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 08 18:19:38 crc kubenswrapper[4859]: I1008 18:19:38.386881 4859 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 08 18:19:38 crc kubenswrapper[4859]: E1008 18:19:38.386982 4859 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-08 18:19:38.886962488 +0000 UTC m=+149.133801867 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 08 18:19:38 crc kubenswrapper[4859]: I1008 18:19:38.387112 4859 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-kg44m\" (UID: \"4afad99b-9bf7-43c9-8dcf-7a3217136be5\") " pod="openshift-image-registry/image-registry-697d97f7c8-kg44m" Oct 08 18:19:38 crc kubenswrapper[4859]: E1008 18:19:38.387430 4859 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-08 18:19:38.887422472 +0000 UTC m=+149.134261841 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-kg44m" (UID: "4afad99b-9bf7-43c9-8dcf-7a3217136be5") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 08 18:19:38 crc kubenswrapper[4859]: I1008 18:19:38.478650 4859 patch_prober.go:28] interesting pod/packageserver-d55dfcdfc-vc7bh container/packageserver namespace/openshift-operator-lifecycle-manager: Readiness probe status=failure output="Get \"https://10.217.0.24:5443/healthz\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" start-of-body= Oct 08 18:19:38 crc kubenswrapper[4859]: I1008 18:19:38.478762 4859 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-vc7bh" podUID="b2b5fcca-e076-4cbe-bc1e-249fcfd25e53" containerName="packageserver" probeResult="failure" output="Get \"https://10.217.0.24:5443/healthz\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" Oct 08 18:19:38 crc kubenswrapper[4859]: I1008 18:19:38.486437 4859 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="hostpath-provisioner/csi-hostpathplugin-q4lkb" event={"ID":"5e418cbb-0072-4767-933d-537c4a274e60","Type":"ContainerStarted","Data":"2e68a7e24c335af69fb490123af242f6753f93086da9fb151a940c870b054b24"} Oct 08 18:19:38 crc kubenswrapper[4859]: I1008 18:19:38.487710 4859 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 08 18:19:38 crc kubenswrapper[4859]: E1008 18:19:38.487840 4859 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-08 18:19:38.987822318 +0000 UTC m=+149.234661697 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 08 18:19:38 crc kubenswrapper[4859]: I1008 18:19:38.488047 4859 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-kg44m\" (UID: \"4afad99b-9bf7-43c9-8dcf-7a3217136be5\") " pod="openshift-image-registry/image-registry-697d97f7c8-kg44m" Oct 08 18:19:38 crc kubenswrapper[4859]: E1008 18:19:38.488340 4859 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-08 18:19:38.988332353 +0000 UTC m=+149.235171732 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-kg44m" (UID: "4afad99b-9bf7-43c9-8dcf-7a3217136be5") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 08 18:19:38 crc kubenswrapper[4859]: I1008 18:19:38.490895 4859 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-6j5rk" event={"ID":"d407fd47-096d-4f2a-ace1-dab4ddf7192a","Type":"ContainerStarted","Data":"54195f64b7c453256e94fb4077b122bbab9d5296fd62e07cac7b1d4bf1d047bb"} Oct 08 18:19:38 crc kubenswrapper[4859]: I1008 18:19:38.492901 4859 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-dns/dns-default-x7pp9" event={"ID":"2a959183-1370-4723-bb5e-8625689d6121","Type":"ContainerStarted","Data":"3b4afe98329c467fa314e3a1a96afba683c0a8ebd714cf9133aaad1819ce3a5d"} Oct 08 18:19:38 crc kubenswrapper[4859]: I1008 18:19:38.505957 4859 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-xpnsv" Oct 08 18:19:38 crc kubenswrapper[4859]: I1008 18:19:38.588600 4859 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 08 18:19:38 crc kubenswrapper[4859]: E1008 18:19:38.588876 4859 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-08 18:19:39.088829802 +0000 UTC m=+149.335669231 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 08 18:19:38 crc kubenswrapper[4859]: I1008 18:19:38.589062 4859 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 08 18:19:38 crc kubenswrapper[4859]: I1008 18:19:38.589192 4859 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 08 18:19:38 crc kubenswrapper[4859]: I1008 18:19:38.589669 4859 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-kg44m\" (UID: \"4afad99b-9bf7-43c9-8dcf-7a3217136be5\") " pod="openshift-image-registry/image-registry-697d97f7c8-kg44m" Oct 08 18:19:38 crc kubenswrapper[4859]: I1008 18:19:38.589745 4859 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 08 18:19:38 crc kubenswrapper[4859]: I1008 18:19:38.589767 4859 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 08 18:19:38 crc kubenswrapper[4859]: E1008 18:19:38.591447 4859 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-08 18:19:39.09142953 +0000 UTC m=+149.338268909 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-kg44m" (UID: "4afad99b-9bf7-43c9-8dcf-7a3217136be5") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 08 18:19:38 crc kubenswrapper[4859]: I1008 18:19:38.596399 4859 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 08 18:19:38 crc kubenswrapper[4859]: I1008 18:19:38.599407 4859 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 08 18:19:38 crc kubenswrapper[4859]: I1008 18:19:38.599545 4859 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 08 18:19:38 crc kubenswrapper[4859]: I1008 18:19:38.600879 4859 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-6j5rk" podStartSLOduration=127.600867541 podStartE2EDuration="2m7.600867541s" podCreationTimestamp="2025-10-08 18:17:31 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-08 18:19:38.587897934 +0000 UTC m=+148.834737313" watchObservedRunningTime="2025-10-08 18:19:38.600867541 +0000 UTC m=+148.847706910" Oct 08 18:19:38 crc kubenswrapper[4859]: I1008 18:19:38.604380 4859 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 08 18:19:38 crc kubenswrapper[4859]: I1008 18:19:38.612045 4859 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 08 18:19:38 crc kubenswrapper[4859]: I1008 18:19:38.624965 4859 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 08 18:19:38 crc kubenswrapper[4859]: I1008 18:19:38.700472 4859 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 08 18:19:38 crc kubenswrapper[4859]: E1008 18:19:38.700923 4859 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-08 18:19:39.200902636 +0000 UTC m=+149.447742015 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 08 18:19:38 crc kubenswrapper[4859]: I1008 18:19:38.802551 4859 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-kg44m\" (UID: \"4afad99b-9bf7-43c9-8dcf-7a3217136be5\") " pod="openshift-image-registry/image-registry-697d97f7c8-kg44m" Oct 08 18:19:38 crc kubenswrapper[4859]: E1008 18:19:38.802944 4859 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-08 18:19:39.302917501 +0000 UTC m=+149.549756880 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-kg44m" (UID: "4afad99b-9bf7-43c9-8dcf-7a3217136be5") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 08 18:19:38 crc kubenswrapper[4859]: I1008 18:19:38.893999 4859 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 08 18:19:38 crc kubenswrapper[4859]: I1008 18:19:38.903275 4859 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 08 18:19:38 crc kubenswrapper[4859]: E1008 18:19:38.903789 4859 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-08 18:19:39.40377196 +0000 UTC m=+149.650611329 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 08 18:19:38 crc kubenswrapper[4859]: I1008 18:19:38.922383 4859 patch_prober.go:28] interesting pod/router-default-5444994796-r5gs2 container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Oct 08 18:19:38 crc kubenswrapper[4859]: [-]has-synced failed: reason withheld Oct 08 18:19:38 crc kubenswrapper[4859]: [+]process-running ok Oct 08 18:19:38 crc kubenswrapper[4859]: healthz check failed Oct 08 18:19:38 crc kubenswrapper[4859]: I1008 18:19:38.922453 4859 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-r5gs2" podUID="8e2b8a53-4bc3-499b-9746-628c323d2a0c" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Oct 08 18:19:38 crc kubenswrapper[4859]: I1008 18:19:38.934963 4859 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-vc7bh" Oct 08 18:19:39 crc kubenswrapper[4859]: I1008 18:19:39.007357 4859 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-kg44m\" (UID: \"4afad99b-9bf7-43c9-8dcf-7a3217136be5\") " pod="openshift-image-registry/image-registry-697d97f7c8-kg44m" Oct 08 18:19:39 crc kubenswrapper[4859]: E1008 18:19:39.007837 4859 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-08 18:19:39.507822515 +0000 UTC m=+149.754661884 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-kg44m" (UID: "4afad99b-9bf7-43c9-8dcf-7a3217136be5") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 08 18:19:39 crc kubenswrapper[4859]: I1008 18:19:39.112093 4859 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 08 18:19:39 crc kubenswrapper[4859]: E1008 18:19:39.113064 4859 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-08 18:19:39.613043405 +0000 UTC m=+149.859882784 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 08 18:19:39 crc kubenswrapper[4859]: I1008 18:19:39.113927 4859 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-kg44m\" (UID: \"4afad99b-9bf7-43c9-8dcf-7a3217136be5\") " pod="openshift-image-registry/image-registry-697d97f7c8-kg44m" Oct 08 18:19:39 crc kubenswrapper[4859]: E1008 18:19:39.114412 4859 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-08 18:19:39.614400655 +0000 UTC m=+149.861240024 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-kg44m" (UID: "4afad99b-9bf7-43c9-8dcf-7a3217136be5") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 08 18:19:39 crc kubenswrapper[4859]: I1008 18:19:39.116894 4859 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29332455-w9wt8" Oct 08 18:19:39 crc kubenswrapper[4859]: I1008 18:19:39.153173 4859 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-authentication/oauth-openshift-558db77b4-j8gh2" Oct 08 18:19:39 crc kubenswrapper[4859]: I1008 18:19:39.228080 4859 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 08 18:19:39 crc kubenswrapper[4859]: I1008 18:19:39.228132 4859 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/d403dae9-eb8c-4c8d-9daa-bf2a0c815652-config-volume\") pod \"d403dae9-eb8c-4c8d-9daa-bf2a0c815652\" (UID: \"d403dae9-eb8c-4c8d-9daa-bf2a0c815652\") " Oct 08 18:19:39 crc kubenswrapper[4859]: I1008 18:19:39.228156 4859 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-v4vwn\" (UniqueName: \"kubernetes.io/projected/d403dae9-eb8c-4c8d-9daa-bf2a0c815652-kube-api-access-v4vwn\") pod \"d403dae9-eb8c-4c8d-9daa-bf2a0c815652\" (UID: \"d403dae9-eb8c-4c8d-9daa-bf2a0c815652\") " Oct 08 18:19:39 crc kubenswrapper[4859]: I1008 18:19:39.228209 4859 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/d403dae9-eb8c-4c8d-9daa-bf2a0c815652-secret-volume\") pod \"d403dae9-eb8c-4c8d-9daa-bf2a0c815652\" (UID: \"d403dae9-eb8c-4c8d-9daa-bf2a0c815652\") " Oct 08 18:19:39 crc kubenswrapper[4859]: E1008 18:19:39.229381 4859 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-08 18:19:39.729358806 +0000 UTC m=+149.976198185 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 08 18:19:39 crc kubenswrapper[4859]: I1008 18:19:39.229579 4859 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/d403dae9-eb8c-4c8d-9daa-bf2a0c815652-config-volume" (OuterVolumeSpecName: "config-volume") pod "d403dae9-eb8c-4c8d-9daa-bf2a0c815652" (UID: "d403dae9-eb8c-4c8d-9daa-bf2a0c815652"). InnerVolumeSpecName "config-volume". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 08 18:19:39 crc kubenswrapper[4859]: I1008 18:19:39.248415 4859 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d403dae9-eb8c-4c8d-9daa-bf2a0c815652-secret-volume" (OuterVolumeSpecName: "secret-volume") pod "d403dae9-eb8c-4c8d-9daa-bf2a0c815652" (UID: "d403dae9-eb8c-4c8d-9daa-bf2a0c815652"). InnerVolumeSpecName "secret-volume". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 08 18:19:39 crc kubenswrapper[4859]: I1008 18:19:39.249473 4859 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/d403dae9-eb8c-4c8d-9daa-bf2a0c815652-kube-api-access-v4vwn" (OuterVolumeSpecName: "kube-api-access-v4vwn") pod "d403dae9-eb8c-4c8d-9daa-bf2a0c815652" (UID: "d403dae9-eb8c-4c8d-9daa-bf2a0c815652"). InnerVolumeSpecName "kube-api-access-v4vwn". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 08 18:19:39 crc kubenswrapper[4859]: I1008 18:19:39.331585 4859 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-kg44m\" (UID: \"4afad99b-9bf7-43c9-8dcf-7a3217136be5\") " pod="openshift-image-registry/image-registry-697d97f7c8-kg44m" Oct 08 18:19:39 crc kubenswrapper[4859]: I1008 18:19:39.332073 4859 reconciler_common.go:293] "Volume detached for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/d403dae9-eb8c-4c8d-9daa-bf2a0c815652-config-volume\") on node \"crc\" DevicePath \"\"" Oct 08 18:19:39 crc kubenswrapper[4859]: I1008 18:19:39.332087 4859 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-v4vwn\" (UniqueName: \"kubernetes.io/projected/d403dae9-eb8c-4c8d-9daa-bf2a0c815652-kube-api-access-v4vwn\") on node \"crc\" DevicePath \"\"" Oct 08 18:19:39 crc kubenswrapper[4859]: I1008 18:19:39.332098 4859 reconciler_common.go:293] "Volume detached for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/d403dae9-eb8c-4c8d-9daa-bf2a0c815652-secret-volume\") on node \"crc\" DevicePath \"\"" Oct 08 18:19:39 crc kubenswrapper[4859]: E1008 18:19:39.332255 4859 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-08 18:19:39.832236486 +0000 UTC m=+150.079075865 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-kg44m" (UID: "4afad99b-9bf7-43c9-8dcf-7a3217136be5") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 08 18:19:39 crc kubenswrapper[4859]: I1008 18:19:39.361930 4859 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-6j5rk" Oct 08 18:19:39 crc kubenswrapper[4859]: I1008 18:19:39.362330 4859 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-6j5rk" Oct 08 18:19:39 crc kubenswrapper[4859]: I1008 18:19:39.435146 4859 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 08 18:19:39 crc kubenswrapper[4859]: E1008 18:19:39.435340 4859 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-08 18:19:39.935312872 +0000 UTC m=+150.182152251 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 08 18:19:39 crc kubenswrapper[4859]: I1008 18:19:39.435425 4859 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-kg44m\" (UID: \"4afad99b-9bf7-43c9-8dcf-7a3217136be5\") " pod="openshift-image-registry/image-registry-697d97f7c8-kg44m" Oct 08 18:19:39 crc kubenswrapper[4859]: E1008 18:19:39.435990 4859 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-08 18:19:39.935977162 +0000 UTC m=+150.182816541 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-kg44m" (UID: "4afad99b-9bf7-43c9-8dcf-7a3217136be5") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 08 18:19:39 crc kubenswrapper[4859]: I1008 18:19:39.484738 4859 patch_prober.go:28] interesting pod/apiserver-76f77b778f-fv9bp container/openshift-apiserver namespace/openshift-apiserver: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[+]ping ok Oct 08 18:19:39 crc kubenswrapper[4859]: [+]log ok Oct 08 18:19:39 crc kubenswrapper[4859]: [+]etcd ok Oct 08 18:19:39 crc kubenswrapper[4859]: [+]poststarthook/start-apiserver-admission-initializer ok Oct 08 18:19:39 crc kubenswrapper[4859]: [+]poststarthook/generic-apiserver-start-informers ok Oct 08 18:19:39 crc kubenswrapper[4859]: [+]poststarthook/max-in-flight-filter ok Oct 08 18:19:39 crc kubenswrapper[4859]: [+]poststarthook/storage-object-count-tracker-hook ok Oct 08 18:19:39 crc kubenswrapper[4859]: [+]poststarthook/image.openshift.io-apiserver-caches ok Oct 08 18:19:39 crc kubenswrapper[4859]: [-]poststarthook/authorization.openshift.io-bootstrapclusterroles failed: reason withheld Oct 08 18:19:39 crc kubenswrapper[4859]: [-]poststarthook/authorization.openshift.io-ensurenodebootstrap-sa failed: reason withheld Oct 08 18:19:39 crc kubenswrapper[4859]: [+]poststarthook/project.openshift.io-projectcache ok Oct 08 18:19:39 crc kubenswrapper[4859]: [+]poststarthook/project.openshift.io-projectauthorizationcache ok Oct 08 18:19:39 crc kubenswrapper[4859]: [-]poststarthook/openshift.io-startinformers failed: reason withheld Oct 08 18:19:39 crc kubenswrapper[4859]: [+]poststarthook/openshift.io-restmapperupdater ok Oct 08 18:19:39 crc kubenswrapper[4859]: [+]poststarthook/quota.openshift.io-clusterquotamapping ok Oct 08 18:19:39 crc kubenswrapper[4859]: livez check failed Oct 08 18:19:39 crc kubenswrapper[4859]: I1008 18:19:39.484813 4859 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-apiserver/apiserver-76f77b778f-fv9bp" podUID="70f88ce6-b9fe-422e-827e-af2b6840a783" containerName="openshift-apiserver" probeResult="failure" output="HTTP probe failed with statuscode: 500" Oct 08 18:19:39 crc kubenswrapper[4859]: I1008 18:19:39.498307 4859 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29332455-w9wt8" event={"ID":"d403dae9-eb8c-4c8d-9daa-bf2a0c815652","Type":"ContainerDied","Data":"cbfeda28315cc27b0db4ba339316e0c0f8c58fd68dec107d7c70d3e13e83c4ab"} Oct 08 18:19:39 crc kubenswrapper[4859]: I1008 18:19:39.498357 4859 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="cbfeda28315cc27b0db4ba339316e0c0f8c58fd68dec107d7c70d3e13e83c4ab" Oct 08 18:19:39 crc kubenswrapper[4859]: I1008 18:19:39.498421 4859 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29332455-w9wt8" Oct 08 18:19:39 crc kubenswrapper[4859]: I1008 18:19:39.511510 4859 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="hostpath-provisioner/csi-hostpathplugin-q4lkb" event={"ID":"5e418cbb-0072-4767-933d-537c4a274e60","Type":"ContainerStarted","Data":"c69610b32f9cfda3864322bc3dee14ed893e2cc99de5eea8e84f85c09a088c5c"} Oct 08 18:19:39 crc kubenswrapper[4859]: I1008 18:19:39.536082 4859 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 08 18:19:39 crc kubenswrapper[4859]: E1008 18:19:39.536424 4859 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-08 18:19:40.036407887 +0000 UTC m=+150.283247266 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 08 18:19:39 crc kubenswrapper[4859]: I1008 18:19:39.638556 4859 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-kg44m\" (UID: \"4afad99b-9bf7-43c9-8dcf-7a3217136be5\") " pod="openshift-image-registry/image-registry-697d97f7c8-kg44m" Oct 08 18:19:39 crc kubenswrapper[4859]: E1008 18:19:39.639162 4859 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-08 18:19:40.139127313 +0000 UTC m=+150.385966692 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-kg44m" (UID: "4afad99b-9bf7-43c9-8dcf-7a3217136be5") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 08 18:19:39 crc kubenswrapper[4859]: I1008 18:19:39.740715 4859 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 08 18:19:39 crc kubenswrapper[4859]: E1008 18:19:39.740960 4859 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-08 18:19:40.24092516 +0000 UTC m=+150.487764529 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 08 18:19:39 crc kubenswrapper[4859]: I1008 18:19:39.741456 4859 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-kg44m\" (UID: \"4afad99b-9bf7-43c9-8dcf-7a3217136be5\") " pod="openshift-image-registry/image-registry-697d97f7c8-kg44m" Oct 08 18:19:39 crc kubenswrapper[4859]: E1008 18:19:39.741954 4859 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-08 18:19:40.241942801 +0000 UTC m=+150.488782260 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-kg44m" (UID: "4afad99b-9bf7-43c9-8dcf-7a3217136be5") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 08 18:19:39 crc kubenswrapper[4859]: W1008 18:19:39.835183 4859 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod5fe485a1_e14f_4c09_b5b9_f252bc42b7e8.slice/crio-7ca1aca3b4a17c914f1f9b541c98b09aabe30990ae83583da91bc4830f5a95ff WatchSource:0}: Error finding container 7ca1aca3b4a17c914f1f9b541c98b09aabe30990ae83583da91bc4830f5a95ff: Status 404 returned error can't find the container with id 7ca1aca3b4a17c914f1f9b541c98b09aabe30990ae83583da91bc4830f5a95ff Oct 08 18:19:39 crc kubenswrapper[4859]: I1008 18:19:39.842259 4859 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 08 18:19:39 crc kubenswrapper[4859]: E1008 18:19:39.842726 4859 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-08 18:19:40.342708467 +0000 UTC m=+150.589547846 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 08 18:19:39 crc kubenswrapper[4859]: I1008 18:19:39.920135 4859 patch_prober.go:28] interesting pod/router-default-5444994796-r5gs2 container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Oct 08 18:19:39 crc kubenswrapper[4859]: [-]has-synced failed: reason withheld Oct 08 18:19:39 crc kubenswrapper[4859]: [+]process-running ok Oct 08 18:19:39 crc kubenswrapper[4859]: healthz check failed Oct 08 18:19:39 crc kubenswrapper[4859]: I1008 18:19:39.920220 4859 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-r5gs2" podUID="8e2b8a53-4bc3-499b-9746-628c323d2a0c" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Oct 08 18:19:39 crc kubenswrapper[4859]: I1008 18:19:39.940577 4859 plugin_watcher.go:194] "Adding socket path or updating timestamp to desired state cache" path="/var/lib/kubelet/plugins_registry/kubevirt.io.hostpath-provisioner-reg.sock" Oct 08 18:19:39 crc kubenswrapper[4859]: I1008 18:19:39.943451 4859 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-kg44m\" (UID: \"4afad99b-9bf7-43c9-8dcf-7a3217136be5\") " pod="openshift-image-registry/image-registry-697d97f7c8-kg44m" Oct 08 18:19:39 crc kubenswrapper[4859]: E1008 18:19:39.943865 4859 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-08 18:19:40.443847045 +0000 UTC m=+150.690686424 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-kg44m" (UID: "4afad99b-9bf7-43c9-8dcf-7a3217136be5") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 08 18:19:40 crc kubenswrapper[4859]: I1008 18:19:40.011572 4859 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-controller-manager/revision-pruner-9-crc"] Oct 08 18:19:40 crc kubenswrapper[4859]: E1008 18:19:40.011899 4859 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d403dae9-eb8c-4c8d-9daa-bf2a0c815652" containerName="collect-profiles" Oct 08 18:19:40 crc kubenswrapper[4859]: I1008 18:19:40.011915 4859 state_mem.go:107] "Deleted CPUSet assignment" podUID="d403dae9-eb8c-4c8d-9daa-bf2a0c815652" containerName="collect-profiles" Oct 08 18:19:40 crc kubenswrapper[4859]: I1008 18:19:40.012009 4859 memory_manager.go:354] "RemoveStaleState removing state" podUID="d403dae9-eb8c-4c8d-9daa-bf2a0c815652" containerName="collect-profiles" Oct 08 18:19:40 crc kubenswrapper[4859]: I1008 18:19:40.012401 4859 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-controller-manager/revision-pruner-9-crc" Oct 08 18:19:40 crc kubenswrapper[4859]: I1008 18:19:40.019199 4859 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-controller-manager"/"kube-root-ca.crt" Oct 08 18:19:40 crc kubenswrapper[4859]: I1008 18:19:40.019324 4859 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-controller-manager"/"installer-sa-dockercfg-kjl2n" Oct 08 18:19:40 crc kubenswrapper[4859]: I1008 18:19:40.028983 4859 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-controller-manager/revision-pruner-9-crc"] Oct 08 18:19:40 crc kubenswrapper[4859]: I1008 18:19:40.044273 4859 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 08 18:19:40 crc kubenswrapper[4859]: E1008 18:19:40.044462 4859 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-08 18:19:40.544428017 +0000 UTC m=+150.791267396 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 08 18:19:40 crc kubenswrapper[4859]: I1008 18:19:40.044569 4859 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-kg44m\" (UID: \"4afad99b-9bf7-43c9-8dcf-7a3217136be5\") " pod="openshift-image-registry/image-registry-697d97f7c8-kg44m" Oct 08 18:19:40 crc kubenswrapper[4859]: E1008 18:19:40.045046 4859 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-08 18:19:40.545024805 +0000 UTC m=+150.791864304 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-kg44m" (UID: "4afad99b-9bf7-43c9-8dcf-7a3217136be5") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 08 18:19:40 crc kubenswrapper[4859]: I1008 18:19:40.146070 4859 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 08 18:19:40 crc kubenswrapper[4859]: E1008 18:19:40.146375 4859 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-08 18:19:40.646345748 +0000 UTC m=+150.893185127 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 08 18:19:40 crc kubenswrapper[4859]: I1008 18:19:40.146698 4859 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-kg44m\" (UID: \"4afad99b-9bf7-43c9-8dcf-7a3217136be5\") " pod="openshift-image-registry/image-registry-697d97f7c8-kg44m" Oct 08 18:19:40 crc kubenswrapper[4859]: I1008 18:19:40.146743 4859 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/8bdda932-cfce-4df3-be03-e23d05f19679-kube-api-access\") pod \"revision-pruner-9-crc\" (UID: \"8bdda932-cfce-4df3-be03-e23d05f19679\") " pod="openshift-kube-controller-manager/revision-pruner-9-crc" Oct 08 18:19:40 crc kubenswrapper[4859]: I1008 18:19:40.146792 4859 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/8bdda932-cfce-4df3-be03-e23d05f19679-kubelet-dir\") pod \"revision-pruner-9-crc\" (UID: \"8bdda932-cfce-4df3-be03-e23d05f19679\") " pod="openshift-kube-controller-manager/revision-pruner-9-crc" Oct 08 18:19:40 crc kubenswrapper[4859]: E1008 18:19:40.147176 4859 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-08 18:19:40.647163133 +0000 UTC m=+150.894002512 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-kg44m" (UID: "4afad99b-9bf7-43c9-8dcf-7a3217136be5") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 08 18:19:40 crc kubenswrapper[4859]: I1008 18:19:40.177017 4859 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-zb82c"] Oct 08 18:19:40 crc kubenswrapper[4859]: I1008 18:19:40.178023 4859 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-zb82c" Oct 08 18:19:40 crc kubenswrapper[4859]: I1008 18:19:40.180376 4859 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"community-operators-dockercfg-dmngl" Oct 08 18:19:40 crc kubenswrapper[4859]: I1008 18:19:40.207784 4859 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-zb82c"] Oct 08 18:19:40 crc kubenswrapper[4859]: I1008 18:19:40.247366 4859 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 08 18:19:40 crc kubenswrapper[4859]: I1008 18:19:40.247608 4859 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/8bdda932-cfce-4df3-be03-e23d05f19679-kubelet-dir\") pod \"revision-pruner-9-crc\" (UID: \"8bdda932-cfce-4df3-be03-e23d05f19679\") " pod="openshift-kube-controller-manager/revision-pruner-9-crc" Oct 08 18:19:40 crc kubenswrapper[4859]: E1008 18:19:40.247665 4859 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-08 18:19:40.747635141 +0000 UTC m=+150.994474520 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 08 18:19:40 crc kubenswrapper[4859]: I1008 18:19:40.247720 4859 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/8bdda932-cfce-4df3-be03-e23d05f19679-kubelet-dir\") pod \"revision-pruner-9-crc\" (UID: \"8bdda932-cfce-4df3-be03-e23d05f19679\") " pod="openshift-kube-controller-manager/revision-pruner-9-crc" Oct 08 18:19:40 crc kubenswrapper[4859]: I1008 18:19:40.247747 4859 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-kg44m\" (UID: \"4afad99b-9bf7-43c9-8dcf-7a3217136be5\") " pod="openshift-image-registry/image-registry-697d97f7c8-kg44m" Oct 08 18:19:40 crc kubenswrapper[4859]: I1008 18:19:40.247834 4859 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/8bdda932-cfce-4df3-be03-e23d05f19679-kube-api-access\") pod \"revision-pruner-9-crc\" (UID: \"8bdda932-cfce-4df3-be03-e23d05f19679\") " pod="openshift-kube-controller-manager/revision-pruner-9-crc" Oct 08 18:19:40 crc kubenswrapper[4859]: E1008 18:19:40.248224 4859 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-08 18:19:40.748211718 +0000 UTC m=+150.995051097 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-kg44m" (UID: "4afad99b-9bf7-43c9-8dcf-7a3217136be5") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 08 18:19:40 crc kubenswrapper[4859]: I1008 18:19:40.290476 4859 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/8bdda932-cfce-4df3-be03-e23d05f19679-kube-api-access\") pod \"revision-pruner-9-crc\" (UID: \"8bdda932-cfce-4df3-be03-e23d05f19679\") " pod="openshift-kube-controller-manager/revision-pruner-9-crc" Oct 08 18:19:40 crc kubenswrapper[4859]: I1008 18:19:40.342795 4859 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-6j5rk" Oct 08 18:19:40 crc kubenswrapper[4859]: I1008 18:19:40.346347 4859 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-controller-manager/revision-pruner-9-crc" Oct 08 18:19:40 crc kubenswrapper[4859]: I1008 18:19:40.349366 4859 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 08 18:19:40 crc kubenswrapper[4859]: I1008 18:19:40.349801 4859 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-7g9pg\" (UniqueName: \"kubernetes.io/projected/5b68d592-4a46-4fbc-bcec-a8bca2b973aa-kube-api-access-7g9pg\") pod \"community-operators-zb82c\" (UID: \"5b68d592-4a46-4fbc-bcec-a8bca2b973aa\") " pod="openshift-marketplace/community-operators-zb82c" Oct 08 18:19:40 crc kubenswrapper[4859]: I1008 18:19:40.349874 4859 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/5b68d592-4a46-4fbc-bcec-a8bca2b973aa-catalog-content\") pod \"community-operators-zb82c\" (UID: \"5b68d592-4a46-4fbc-bcec-a8bca2b973aa\") " pod="openshift-marketplace/community-operators-zb82c" Oct 08 18:19:40 crc kubenswrapper[4859]: I1008 18:19:40.349922 4859 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/5b68d592-4a46-4fbc-bcec-a8bca2b973aa-utilities\") pod \"community-operators-zb82c\" (UID: \"5b68d592-4a46-4fbc-bcec-a8bca2b973aa\") " pod="openshift-marketplace/community-operators-zb82c" Oct 08 18:19:40 crc kubenswrapper[4859]: E1008 18:19:40.350090 4859 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-08 18:19:40.850050087 +0000 UTC m=+151.096889466 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 08 18:19:40 crc kubenswrapper[4859]: I1008 18:19:40.355272 4859 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-c5tcx"] Oct 08 18:19:40 crc kubenswrapper[4859]: I1008 18:19:40.356305 4859 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-c5tcx" Oct 08 18:19:40 crc kubenswrapper[4859]: I1008 18:19:40.359090 4859 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"certified-operators-dockercfg-4rs5g" Oct 08 18:19:40 crc kubenswrapper[4859]: I1008 18:19:40.378995 4859 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-c5tcx"] Oct 08 18:19:40 crc kubenswrapper[4859]: I1008 18:19:40.452417 4859 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-7g9pg\" (UniqueName: \"kubernetes.io/projected/5b68d592-4a46-4fbc-bcec-a8bca2b973aa-kube-api-access-7g9pg\") pod \"community-operators-zb82c\" (UID: \"5b68d592-4a46-4fbc-bcec-a8bca2b973aa\") " pod="openshift-marketplace/community-operators-zb82c" Oct 08 18:19:40 crc kubenswrapper[4859]: I1008 18:19:40.452730 4859 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/5b68d592-4a46-4fbc-bcec-a8bca2b973aa-catalog-content\") pod \"community-operators-zb82c\" (UID: \"5b68d592-4a46-4fbc-bcec-a8bca2b973aa\") " pod="openshift-marketplace/community-operators-zb82c" Oct 08 18:19:40 crc kubenswrapper[4859]: I1008 18:19:40.452755 4859 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-kg44m\" (UID: \"4afad99b-9bf7-43c9-8dcf-7a3217136be5\") " pod="openshift-image-registry/image-registry-697d97f7c8-kg44m" Oct 08 18:19:40 crc kubenswrapper[4859]: I1008 18:19:40.452774 4859 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/5b68d592-4a46-4fbc-bcec-a8bca2b973aa-utilities\") pod \"community-operators-zb82c\" (UID: \"5b68d592-4a46-4fbc-bcec-a8bca2b973aa\") " pod="openshift-marketplace/community-operators-zb82c" Oct 08 18:19:40 crc kubenswrapper[4859]: I1008 18:19:40.452802 4859 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/438a5b9a-aaee-4e06-9d22-691862b9207e-utilities\") pod \"certified-operators-c5tcx\" (UID: \"438a5b9a-aaee-4e06-9d22-691862b9207e\") " pod="openshift-marketplace/certified-operators-c5tcx" Oct 08 18:19:40 crc kubenswrapper[4859]: I1008 18:19:40.452821 4859 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/438a5b9a-aaee-4e06-9d22-691862b9207e-catalog-content\") pod \"certified-operators-c5tcx\" (UID: \"438a5b9a-aaee-4e06-9d22-691862b9207e\") " pod="openshift-marketplace/certified-operators-c5tcx" Oct 08 18:19:40 crc kubenswrapper[4859]: I1008 18:19:40.452850 4859 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-rp9cl\" (UniqueName: \"kubernetes.io/projected/438a5b9a-aaee-4e06-9d22-691862b9207e-kube-api-access-rp9cl\") pod \"certified-operators-c5tcx\" (UID: \"438a5b9a-aaee-4e06-9d22-691862b9207e\") " pod="openshift-marketplace/certified-operators-c5tcx" Oct 08 18:19:40 crc kubenswrapper[4859]: I1008 18:19:40.453630 4859 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/5b68d592-4a46-4fbc-bcec-a8bca2b973aa-catalog-content\") pod \"community-operators-zb82c\" (UID: \"5b68d592-4a46-4fbc-bcec-a8bca2b973aa\") " pod="openshift-marketplace/community-operators-zb82c" Oct 08 18:19:40 crc kubenswrapper[4859]: E1008 18:19:40.453963 4859 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-08 18:19:40.953948857 +0000 UTC m=+151.200788236 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-kg44m" (UID: "4afad99b-9bf7-43c9-8dcf-7a3217136be5") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 08 18:19:40 crc kubenswrapper[4859]: I1008 18:19:40.454391 4859 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/5b68d592-4a46-4fbc-bcec-a8bca2b973aa-utilities\") pod \"community-operators-zb82c\" (UID: \"5b68d592-4a46-4fbc-bcec-a8bca2b973aa\") " pod="openshift-marketplace/community-operators-zb82c" Oct 08 18:19:40 crc kubenswrapper[4859]: I1008 18:19:40.486074 4859 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-7g9pg\" (UniqueName: \"kubernetes.io/projected/5b68d592-4a46-4fbc-bcec-a8bca2b973aa-kube-api-access-7g9pg\") pod \"community-operators-zb82c\" (UID: \"5b68d592-4a46-4fbc-bcec-a8bca2b973aa\") " pod="openshift-marketplace/community-operators-zb82c" Oct 08 18:19:40 crc kubenswrapper[4859]: I1008 18:19:40.496609 4859 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-zb82c" Oct 08 18:19:40 crc kubenswrapper[4859]: I1008 18:19:40.518732 4859 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" event={"ID":"9d751cbb-f2e2-430d-9754-c882a5e924a5","Type":"ContainerStarted","Data":"d2b9116dc26e9cff7142975d1ca948961416a9cc32a1ded5142c9c6fd5430c05"} Oct 08 18:19:40 crc kubenswrapper[4859]: I1008 18:19:40.518781 4859 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" event={"ID":"9d751cbb-f2e2-430d-9754-c882a5e924a5","Type":"ContainerStarted","Data":"ba7f1d77e67bab2f36dd1a9aa737becfca60041611ff7853921fff787f776c95"} Oct 08 18:19:40 crc kubenswrapper[4859]: I1008 18:19:40.527210 4859 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="hostpath-provisioner/csi-hostpathplugin-q4lkb" event={"ID":"5e418cbb-0072-4767-933d-537c4a274e60","Type":"ContainerStarted","Data":"87765cf429fdb41736be953f47120e8b63348f23589c540fe06607d873deac1f"} Oct 08 18:19:40 crc kubenswrapper[4859]: I1008 18:19:40.527268 4859 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="hostpath-provisioner/csi-hostpathplugin-q4lkb" event={"ID":"5e418cbb-0072-4767-933d-537c4a274e60","Type":"ContainerStarted","Data":"92cffb6098c83437d11e8f2828094166807f90e9e30991414e555f3bc656d36f"} Oct 08 18:19:40 crc kubenswrapper[4859]: I1008 18:19:40.558459 4859 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 08 18:19:40 crc kubenswrapper[4859]: I1008 18:19:40.558795 4859 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/438a5b9a-aaee-4e06-9d22-691862b9207e-utilities\") pod \"certified-operators-c5tcx\" (UID: \"438a5b9a-aaee-4e06-9d22-691862b9207e\") " pod="openshift-marketplace/certified-operators-c5tcx" Oct 08 18:19:40 crc kubenswrapper[4859]: I1008 18:19:40.558821 4859 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/438a5b9a-aaee-4e06-9d22-691862b9207e-catalog-content\") pod \"certified-operators-c5tcx\" (UID: \"438a5b9a-aaee-4e06-9d22-691862b9207e\") " pod="openshift-marketplace/certified-operators-c5tcx" Oct 08 18:19:40 crc kubenswrapper[4859]: I1008 18:19:40.558850 4859 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rp9cl\" (UniqueName: \"kubernetes.io/projected/438a5b9a-aaee-4e06-9d22-691862b9207e-kube-api-access-rp9cl\") pod \"certified-operators-c5tcx\" (UID: \"438a5b9a-aaee-4e06-9d22-691862b9207e\") " pod="openshift-marketplace/certified-operators-c5tcx" Oct 08 18:19:40 crc kubenswrapper[4859]: I1008 18:19:40.559590 4859 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/438a5b9a-aaee-4e06-9d22-691862b9207e-catalog-content\") pod \"certified-operators-c5tcx\" (UID: \"438a5b9a-aaee-4e06-9d22-691862b9207e\") " pod="openshift-marketplace/certified-operators-c5tcx" Oct 08 18:19:40 crc kubenswrapper[4859]: E1008 18:19:40.559707 4859 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-08 18:19:41.059667532 +0000 UTC m=+151.306506911 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 08 18:19:40 crc kubenswrapper[4859]: I1008 18:19:40.559730 4859 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/438a5b9a-aaee-4e06-9d22-691862b9207e-utilities\") pod \"certified-operators-c5tcx\" (UID: \"438a5b9a-aaee-4e06-9d22-691862b9207e\") " pod="openshift-marketplace/certified-operators-c5tcx" Oct 08 18:19:40 crc kubenswrapper[4859]: I1008 18:19:40.568987 4859 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" event={"ID":"3b6479f0-333b-4a96-9adf-2099afdc2447","Type":"ContainerStarted","Data":"c2b84e34be125307e68b4e03a16a5d75ddf15a034c5ce319e68f5c18cbde7116"} Oct 08 18:19:40 crc kubenswrapper[4859]: I1008 18:19:40.569039 4859 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" event={"ID":"3b6479f0-333b-4a96-9adf-2099afdc2447","Type":"ContainerStarted","Data":"97a423096391403313d4dc7cf9963285ba9b7833c992a42c9c17cf6cae5a8264"} Oct 08 18:19:40 crc kubenswrapper[4859]: I1008 18:19:40.569554 4859 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 08 18:19:40 crc kubenswrapper[4859]: I1008 18:19:40.569622 4859 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-pxq2g"] Oct 08 18:19:40 crc kubenswrapper[4859]: I1008 18:19:40.574151 4859 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-pxq2g" Oct 08 18:19:40 crc kubenswrapper[4859]: I1008 18:19:40.580169 4859 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" event={"ID":"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8","Type":"ContainerStarted","Data":"b82b973736f263fde130d4254887e5e1588574f815e34542f8d5d337c256f913"} Oct 08 18:19:40 crc kubenswrapper[4859]: I1008 18:19:40.580220 4859 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" event={"ID":"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8","Type":"ContainerStarted","Data":"7ca1aca3b4a17c914f1f9b541c98b09aabe30990ae83583da91bc4830f5a95ff"} Oct 08 18:19:40 crc kubenswrapper[4859]: I1008 18:19:40.597474 4859 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rp9cl\" (UniqueName: \"kubernetes.io/projected/438a5b9a-aaee-4e06-9d22-691862b9207e-kube-api-access-rp9cl\") pod \"certified-operators-c5tcx\" (UID: \"438a5b9a-aaee-4e06-9d22-691862b9207e\") " pod="openshift-marketplace/certified-operators-c5tcx" Oct 08 18:19:40 crc kubenswrapper[4859]: I1008 18:19:40.601486 4859 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-pxq2g"] Oct 08 18:19:40 crc kubenswrapper[4859]: I1008 18:19:40.603871 4859 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-6j5rk" Oct 08 18:19:40 crc kubenswrapper[4859]: I1008 18:19:40.617554 4859 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="hostpath-provisioner/csi-hostpathplugin-q4lkb" podStartSLOduration=10.617527648 podStartE2EDuration="10.617527648s" podCreationTimestamp="2025-10-08 18:19:30 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-08 18:19:40.615701684 +0000 UTC m=+150.862541083" watchObservedRunningTime="2025-10-08 18:19:40.617527648 +0000 UTC m=+150.864367027" Oct 08 18:19:40 crc kubenswrapper[4859]: I1008 18:19:40.639989 4859 reconciler.go:161] "OperationExecutor.RegisterPlugin started" plugin={"SocketPath":"/var/lib/kubelet/plugins_registry/kubevirt.io.hostpath-provisioner-reg.sock","Timestamp":"2025-10-08T18:19:39.940606209Z","Handler":null,"Name":""} Oct 08 18:19:40 crc kubenswrapper[4859]: I1008 18:19:40.651003 4859 csi_plugin.go:100] kubernetes.io/csi: Trying to validate a new CSI Driver with name: kubevirt.io.hostpath-provisioner endpoint: /var/lib/kubelet/plugins/csi-hostpath/csi.sock versions: 1.0.0 Oct 08 18:19:40 crc kubenswrapper[4859]: I1008 18:19:40.651045 4859 csi_plugin.go:113] kubernetes.io/csi: Register new plugin with name: kubevirt.io.hostpath-provisioner at endpoint: /var/lib/kubelet/plugins/csi-hostpath/csi.sock Oct 08 18:19:40 crc kubenswrapper[4859]: I1008 18:19:40.663597 4859 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-kg44m\" (UID: \"4afad99b-9bf7-43c9-8dcf-7a3217136be5\") " pod="openshift-image-registry/image-registry-697d97f7c8-kg44m" Oct 08 18:19:40 crc kubenswrapper[4859]: I1008 18:19:40.671015 4859 csi_attacher.go:380] kubernetes.io/csi: attacher.MountDevice STAGE_UNSTAGE_VOLUME capability not set. Skipping MountDevice... Oct 08 18:19:40 crc kubenswrapper[4859]: I1008 18:19:40.671053 4859 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-kg44m\" (UID: \"4afad99b-9bf7-43c9-8dcf-7a3217136be5\") device mount path \"/var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner/1f4776af88835e41c12b831b4c9fed40233456d14189815a54dbe7f892fc1983/globalmount\"" pod="openshift-image-registry/image-registry-697d97f7c8-kg44m" Oct 08 18:19:40 crc kubenswrapper[4859]: I1008 18:19:40.678019 4859 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-c5tcx" Oct 08 18:19:40 crc kubenswrapper[4859]: I1008 18:19:40.705861 4859 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-kg44m\" (UID: \"4afad99b-9bf7-43c9-8dcf-7a3217136be5\") " pod="openshift-image-registry/image-registry-697d97f7c8-kg44m" Oct 08 18:19:40 crc kubenswrapper[4859]: I1008 18:19:40.726889 4859 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-controller-manager/revision-pruner-9-crc"] Oct 08 18:19:40 crc kubenswrapper[4859]: I1008 18:19:40.772934 4859 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-q6ktk"] Oct 08 18:19:40 crc kubenswrapper[4859]: I1008 18:19:40.773667 4859 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 08 18:19:40 crc kubenswrapper[4859]: I1008 18:19:40.774321 4859 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/742016ca-eb57-474a-b4d9-dbfc682628be-catalog-content\") pod \"community-operators-pxq2g\" (UID: \"742016ca-eb57-474a-b4d9-dbfc682628be\") " pod="openshift-marketplace/community-operators-pxq2g" Oct 08 18:19:40 crc kubenswrapper[4859]: I1008 18:19:40.774355 4859 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/742016ca-eb57-474a-b4d9-dbfc682628be-utilities\") pod \"community-operators-pxq2g\" (UID: \"742016ca-eb57-474a-b4d9-dbfc682628be\") " pod="openshift-marketplace/community-operators-pxq2g" Oct 08 18:19:40 crc kubenswrapper[4859]: I1008 18:19:40.774393 4859 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-694ds\" (UniqueName: \"kubernetes.io/projected/742016ca-eb57-474a-b4d9-dbfc682628be-kube-api-access-694ds\") pod \"community-operators-pxq2g\" (UID: \"742016ca-eb57-474a-b4d9-dbfc682628be\") " pod="openshift-marketplace/community-operators-pxq2g" Oct 08 18:19:40 crc kubenswrapper[4859]: I1008 18:19:40.776177 4859 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-q6ktk" Oct 08 18:19:40 crc kubenswrapper[4859]: I1008 18:19:40.786165 4859 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-q6ktk"] Oct 08 18:19:40 crc kubenswrapper[4859]: I1008 18:19:40.808034 4859 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (OuterVolumeSpecName: "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b"). InnerVolumeSpecName "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8". PluginName "kubernetes.io/csi", VolumeGidValue "" Oct 08 18:19:40 crc kubenswrapper[4859]: I1008 18:19:40.875628 4859 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/742016ca-eb57-474a-b4d9-dbfc682628be-catalog-content\") pod \"community-operators-pxq2g\" (UID: \"742016ca-eb57-474a-b4d9-dbfc682628be\") " pod="openshift-marketplace/community-operators-pxq2g" Oct 08 18:19:40 crc kubenswrapper[4859]: I1008 18:19:40.875675 4859 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/742016ca-eb57-474a-b4d9-dbfc682628be-utilities\") pod \"community-operators-pxq2g\" (UID: \"742016ca-eb57-474a-b4d9-dbfc682628be\") " pod="openshift-marketplace/community-operators-pxq2g" Oct 08 18:19:40 crc kubenswrapper[4859]: I1008 18:19:40.875710 4859 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-694ds\" (UniqueName: \"kubernetes.io/projected/742016ca-eb57-474a-b4d9-dbfc682628be-kube-api-access-694ds\") pod \"community-operators-pxq2g\" (UID: \"742016ca-eb57-474a-b4d9-dbfc682628be\") " pod="openshift-marketplace/community-operators-pxq2g" Oct 08 18:19:40 crc kubenswrapper[4859]: I1008 18:19:40.875777 4859 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/ae751baf-6f53-4a05-b28d-0888f7682db7-catalog-content\") pod \"certified-operators-q6ktk\" (UID: \"ae751baf-6f53-4a05-b28d-0888f7682db7\") " pod="openshift-marketplace/certified-operators-q6ktk" Oct 08 18:19:40 crc kubenswrapper[4859]: I1008 18:19:40.875811 4859 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/ae751baf-6f53-4a05-b28d-0888f7682db7-utilities\") pod \"certified-operators-q6ktk\" (UID: \"ae751baf-6f53-4a05-b28d-0888f7682db7\") " pod="openshift-marketplace/certified-operators-q6ktk" Oct 08 18:19:40 crc kubenswrapper[4859]: I1008 18:19:40.875835 4859 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-t9kgm\" (UniqueName: \"kubernetes.io/projected/ae751baf-6f53-4a05-b28d-0888f7682db7-kube-api-access-t9kgm\") pod \"certified-operators-q6ktk\" (UID: \"ae751baf-6f53-4a05-b28d-0888f7682db7\") " pod="openshift-marketplace/certified-operators-q6ktk" Oct 08 18:19:40 crc kubenswrapper[4859]: I1008 18:19:40.876348 4859 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/742016ca-eb57-474a-b4d9-dbfc682628be-catalog-content\") pod \"community-operators-pxq2g\" (UID: \"742016ca-eb57-474a-b4d9-dbfc682628be\") " pod="openshift-marketplace/community-operators-pxq2g" Oct 08 18:19:40 crc kubenswrapper[4859]: I1008 18:19:40.876595 4859 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/742016ca-eb57-474a-b4d9-dbfc682628be-utilities\") pod \"community-operators-pxq2g\" (UID: \"742016ca-eb57-474a-b4d9-dbfc682628be\") " pod="openshift-marketplace/community-operators-pxq2g" Oct 08 18:19:40 crc kubenswrapper[4859]: I1008 18:19:40.903916 4859 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-694ds\" (UniqueName: \"kubernetes.io/projected/742016ca-eb57-474a-b4d9-dbfc682628be-kube-api-access-694ds\") pod \"community-operators-pxq2g\" (UID: \"742016ca-eb57-474a-b4d9-dbfc682628be\") " pod="openshift-marketplace/community-operators-pxq2g" Oct 08 18:19:40 crc kubenswrapper[4859]: I1008 18:19:40.912290 4859 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/image-registry-697d97f7c8-kg44m" Oct 08 18:19:40 crc kubenswrapper[4859]: I1008 18:19:40.920447 4859 patch_prober.go:28] interesting pod/router-default-5444994796-r5gs2 container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Oct 08 18:19:40 crc kubenswrapper[4859]: [-]has-synced failed: reason withheld Oct 08 18:19:40 crc kubenswrapper[4859]: [+]process-running ok Oct 08 18:19:40 crc kubenswrapper[4859]: healthz check failed Oct 08 18:19:40 crc kubenswrapper[4859]: I1008 18:19:40.920511 4859 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-r5gs2" podUID="8e2b8a53-4bc3-499b-9746-628c323d2a0c" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Oct 08 18:19:40 crc kubenswrapper[4859]: I1008 18:19:40.967816 4859 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-zb82c"] Oct 08 18:19:40 crc kubenswrapper[4859]: I1008 18:19:40.976582 4859 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/ae751baf-6f53-4a05-b28d-0888f7682db7-catalog-content\") pod \"certified-operators-q6ktk\" (UID: \"ae751baf-6f53-4a05-b28d-0888f7682db7\") " pod="openshift-marketplace/certified-operators-q6ktk" Oct 08 18:19:40 crc kubenswrapper[4859]: I1008 18:19:40.976738 4859 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/ae751baf-6f53-4a05-b28d-0888f7682db7-utilities\") pod \"certified-operators-q6ktk\" (UID: \"ae751baf-6f53-4a05-b28d-0888f7682db7\") " pod="openshift-marketplace/certified-operators-q6ktk" Oct 08 18:19:40 crc kubenswrapper[4859]: I1008 18:19:40.976779 4859 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-t9kgm\" (UniqueName: \"kubernetes.io/projected/ae751baf-6f53-4a05-b28d-0888f7682db7-kube-api-access-t9kgm\") pod \"certified-operators-q6ktk\" (UID: \"ae751baf-6f53-4a05-b28d-0888f7682db7\") " pod="openshift-marketplace/certified-operators-q6ktk" Oct 08 18:19:40 crc kubenswrapper[4859]: I1008 18:19:40.978232 4859 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/ae751baf-6f53-4a05-b28d-0888f7682db7-catalog-content\") pod \"certified-operators-q6ktk\" (UID: \"ae751baf-6f53-4a05-b28d-0888f7682db7\") " pod="openshift-marketplace/certified-operators-q6ktk" Oct 08 18:19:40 crc kubenswrapper[4859]: I1008 18:19:40.978455 4859 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/ae751baf-6f53-4a05-b28d-0888f7682db7-utilities\") pod \"certified-operators-q6ktk\" (UID: \"ae751baf-6f53-4a05-b28d-0888f7682db7\") " pod="openshift-marketplace/certified-operators-q6ktk" Oct 08 18:19:41 crc kubenswrapper[4859]: I1008 18:19:41.002908 4859 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-t9kgm\" (UniqueName: \"kubernetes.io/projected/ae751baf-6f53-4a05-b28d-0888f7682db7-kube-api-access-t9kgm\") pod \"certified-operators-q6ktk\" (UID: \"ae751baf-6f53-4a05-b28d-0888f7682db7\") " pod="openshift-marketplace/certified-operators-q6ktk" Oct 08 18:19:41 crc kubenswrapper[4859]: I1008 18:19:41.117197 4859 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-q6ktk" Oct 08 18:19:41 crc kubenswrapper[4859]: I1008 18:19:41.196777 4859 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-image-registry/image-registry-697d97f7c8-kg44m"] Oct 08 18:19:41 crc kubenswrapper[4859]: I1008 18:19:41.200757 4859 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-pxq2g" Oct 08 18:19:41 crc kubenswrapper[4859]: I1008 18:19:41.241314 4859 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-c5tcx"] Oct 08 18:19:41 crc kubenswrapper[4859]: W1008 18:19:41.254551 4859 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod438a5b9a_aaee_4e06_9d22_691862b9207e.slice/crio-018a03ca21455674a1ca2e2f60ca8e66b58ce6fc32a1698043d2788e8c022286 WatchSource:0}: Error finding container 018a03ca21455674a1ca2e2f60ca8e66b58ce6fc32a1698043d2788e8c022286: Status 404 returned error can't find the container with id 018a03ca21455674a1ca2e2f60ca8e66b58ce6fc32a1698043d2788e8c022286 Oct 08 18:19:41 crc kubenswrapper[4859]: I1008 18:19:41.352412 4859 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-q6ktk"] Oct 08 18:19:41 crc kubenswrapper[4859]: W1008 18:19:41.368029 4859 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podae751baf_6f53_4a05_b28d_0888f7682db7.slice/crio-e4c394bb7faf6da21cb8bcdc695d6518a4f4e91eacd37fe96ec9b270dedafb83 WatchSource:0}: Error finding container e4c394bb7faf6da21cb8bcdc695d6518a4f4e91eacd37fe96ec9b270dedafb83: Status 404 returned error can't find the container with id e4c394bb7faf6da21cb8bcdc695d6518a4f4e91eacd37fe96ec9b270dedafb83 Oct 08 18:19:41 crc kubenswrapper[4859]: I1008 18:19:41.587111 4859 generic.go:334] "Generic (PLEG): container finished" podID="ae751baf-6f53-4a05-b28d-0888f7682db7" containerID="042432da094dcaa29f61e3777c2ac6bcaea4b5a376e5ca646acc8d6ca678a053" exitCode=0 Oct 08 18:19:41 crc kubenswrapper[4859]: I1008 18:19:41.587197 4859 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-q6ktk" event={"ID":"ae751baf-6f53-4a05-b28d-0888f7682db7","Type":"ContainerDied","Data":"042432da094dcaa29f61e3777c2ac6bcaea4b5a376e5ca646acc8d6ca678a053"} Oct 08 18:19:41 crc kubenswrapper[4859]: I1008 18:19:41.587606 4859 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-q6ktk" event={"ID":"ae751baf-6f53-4a05-b28d-0888f7682db7","Type":"ContainerStarted","Data":"e4c394bb7faf6da21cb8bcdc695d6518a4f4e91eacd37fe96ec9b270dedafb83"} Oct 08 18:19:41 crc kubenswrapper[4859]: I1008 18:19:41.589253 4859 generic.go:334] "Generic (PLEG): container finished" podID="5b68d592-4a46-4fbc-bcec-a8bca2b973aa" containerID="e38579761dd4da27a2aec9a1a2a6138684ba3f36629898992d39947e3fd12c35" exitCode=0 Oct 08 18:19:41 crc kubenswrapper[4859]: I1008 18:19:41.589321 4859 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-zb82c" event={"ID":"5b68d592-4a46-4fbc-bcec-a8bca2b973aa","Type":"ContainerDied","Data":"e38579761dd4da27a2aec9a1a2a6138684ba3f36629898992d39947e3fd12c35"} Oct 08 18:19:41 crc kubenswrapper[4859]: I1008 18:19:41.589350 4859 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-zb82c" event={"ID":"5b68d592-4a46-4fbc-bcec-a8bca2b973aa","Type":"ContainerStarted","Data":"da8c59c1f62a80084e100ca36e84217397827dfa0c3fd9cac0e1b5bc10f153d0"} Oct 08 18:19:41 crc kubenswrapper[4859]: I1008 18:19:41.590229 4859 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Oct 08 18:19:41 crc kubenswrapper[4859]: I1008 18:19:41.592058 4859 generic.go:334] "Generic (PLEG): container finished" podID="438a5b9a-aaee-4e06-9d22-691862b9207e" containerID="223b3d675a30b7c8d3aba21fe8f0a9147dca08dbc807080e4d52889f52dd1da4" exitCode=0 Oct 08 18:19:41 crc kubenswrapper[4859]: I1008 18:19:41.592132 4859 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-c5tcx" event={"ID":"438a5b9a-aaee-4e06-9d22-691862b9207e","Type":"ContainerDied","Data":"223b3d675a30b7c8d3aba21fe8f0a9147dca08dbc807080e4d52889f52dd1da4"} Oct 08 18:19:41 crc kubenswrapper[4859]: I1008 18:19:41.592188 4859 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-c5tcx" event={"ID":"438a5b9a-aaee-4e06-9d22-691862b9207e","Type":"ContainerStarted","Data":"018a03ca21455674a1ca2e2f60ca8e66b58ce6fc32a1698043d2788e8c022286"} Oct 08 18:19:41 crc kubenswrapper[4859]: I1008 18:19:41.593588 4859 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/image-registry-697d97f7c8-kg44m" event={"ID":"4afad99b-9bf7-43c9-8dcf-7a3217136be5","Type":"ContainerStarted","Data":"c4e4ac2065ed20944a9de0ebaa0ce1d4d3c970b586c0a4da7abe8fb14445e685"} Oct 08 18:19:41 crc kubenswrapper[4859]: I1008 18:19:41.593618 4859 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/image-registry-697d97f7c8-kg44m" event={"ID":"4afad99b-9bf7-43c9-8dcf-7a3217136be5","Type":"ContainerStarted","Data":"e936f0856e9768e37d0620582beb93df7b58bedfc12fd790f00e445203415533"} Oct 08 18:19:41 crc kubenswrapper[4859]: I1008 18:19:41.594055 4859 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-image-registry/image-registry-697d97f7c8-kg44m" Oct 08 18:19:41 crc kubenswrapper[4859]: I1008 18:19:41.596277 4859 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/revision-pruner-9-crc" event={"ID":"8bdda932-cfce-4df3-be03-e23d05f19679","Type":"ContainerStarted","Data":"e86684103c0102db6aeb03666e4b3187703e7f83a18169a5b0543e8cb30b863b"} Oct 08 18:19:41 crc kubenswrapper[4859]: I1008 18:19:41.596307 4859 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/revision-pruner-9-crc" event={"ID":"8bdda932-cfce-4df3-be03-e23d05f19679","Type":"ContainerStarted","Data":"648d9a0c5732ccf512ff636d748dc4ac27ffb37df3e6ae0072cca948aec44c3d"} Oct 08 18:19:41 crc kubenswrapper[4859]: I1008 18:19:41.656205 4859 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-pxq2g"] Oct 08 18:19:41 crc kubenswrapper[4859]: I1008 18:19:41.686041 4859 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-image-registry/image-registry-697d97f7c8-kg44m" podStartSLOduration=131.686016622 podStartE2EDuration="2m11.686016622s" podCreationTimestamp="2025-10-08 18:17:30 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-08 18:19:41.68425229 +0000 UTC m=+151.931091679" watchObservedRunningTime="2025-10-08 18:19:41.686016622 +0000 UTC m=+151.932855991" Oct 08 18:19:41 crc kubenswrapper[4859]: I1008 18:19:41.725224 4859 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-controller-manager/revision-pruner-9-crc" podStartSLOduration=2.725203232 podStartE2EDuration="2.725203232s" podCreationTimestamp="2025-10-08 18:19:39 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-08 18:19:41.722674416 +0000 UTC m=+151.969513795" watchObservedRunningTime="2025-10-08 18:19:41.725203232 +0000 UTC m=+151.972042611" Oct 08 18:19:41 crc kubenswrapper[4859]: I1008 18:19:41.919100 4859 patch_prober.go:28] interesting pod/router-default-5444994796-r5gs2 container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Oct 08 18:19:41 crc kubenswrapper[4859]: [-]has-synced failed: reason withheld Oct 08 18:19:41 crc kubenswrapper[4859]: [+]process-running ok Oct 08 18:19:41 crc kubenswrapper[4859]: healthz check failed Oct 08 18:19:41 crc kubenswrapper[4859]: I1008 18:19:41.919207 4859 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-r5gs2" podUID="8e2b8a53-4bc3-499b-9746-628c323d2a0c" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Oct 08 18:19:42 crc kubenswrapper[4859]: I1008 18:19:42.361593 4859 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-8z5ch"] Oct 08 18:19:42 crc kubenswrapper[4859]: I1008 18:19:42.364293 4859 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-8z5ch" Oct 08 18:19:42 crc kubenswrapper[4859]: I1008 18:19:42.367951 4859 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"redhat-marketplace-dockercfg-x2ctb" Oct 08 18:19:42 crc kubenswrapper[4859]: I1008 18:19:42.370594 4859 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-8z5ch"] Oct 08 18:19:42 crc kubenswrapper[4859]: I1008 18:19:42.459371 4859 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-apiserver/apiserver-76f77b778f-fv9bp" Oct 08 18:19:42 crc kubenswrapper[4859]: I1008 18:19:42.464121 4859 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-apiserver/apiserver-76f77b778f-fv9bp" Oct 08 18:19:42 crc kubenswrapper[4859]: I1008 18:19:42.474760 4859 patch_prober.go:28] interesting pod/downloads-7954f5f757-8q5jz container/download-server namespace/openshift-console: Readiness probe status=failure output="Get \"http://10.217.0.20:8080/\": dial tcp 10.217.0.20:8080: connect: connection refused" start-of-body= Oct 08 18:19:42 crc kubenswrapper[4859]: I1008 18:19:42.474816 4859 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-console/downloads-7954f5f757-8q5jz" podUID="9dcb8f76-0fff-436c-9191-fe3f787b7ac0" containerName="download-server" probeResult="failure" output="Get \"http://10.217.0.20:8080/\": dial tcp 10.217.0.20:8080: connect: connection refused" Oct 08 18:19:42 crc kubenswrapper[4859]: I1008 18:19:42.475461 4859 patch_prober.go:28] interesting pod/downloads-7954f5f757-8q5jz container/download-server namespace/openshift-console: Liveness probe status=failure output="Get \"http://10.217.0.20:8080/\": dial tcp 10.217.0.20:8080: connect: connection refused" start-of-body= Oct 08 18:19:42 crc kubenswrapper[4859]: I1008 18:19:42.475519 4859 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-console/downloads-7954f5f757-8q5jz" podUID="9dcb8f76-0fff-436c-9191-fe3f787b7ac0" containerName="download-server" probeResult="failure" output="Get \"http://10.217.0.20:8080/\": dial tcp 10.217.0.20:8080: connect: connection refused" Oct 08 18:19:42 crc kubenswrapper[4859]: I1008 18:19:42.486868 4859 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="8f668bae-612b-4b75-9490-919e737c6a3b" path="/var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes" Oct 08 18:19:42 crc kubenswrapper[4859]: I1008 18:19:42.499235 4859 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/557e8291-7e96-4634-b744-bcf6981d3a05-utilities\") pod \"redhat-marketplace-8z5ch\" (UID: \"557e8291-7e96-4634-b744-bcf6981d3a05\") " pod="openshift-marketplace/redhat-marketplace-8z5ch" Oct 08 18:19:42 crc kubenswrapper[4859]: I1008 18:19:42.499288 4859 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-r5mcr\" (UniqueName: \"kubernetes.io/projected/557e8291-7e96-4634-b744-bcf6981d3a05-kube-api-access-r5mcr\") pod \"redhat-marketplace-8z5ch\" (UID: \"557e8291-7e96-4634-b744-bcf6981d3a05\") " pod="openshift-marketplace/redhat-marketplace-8z5ch" Oct 08 18:19:42 crc kubenswrapper[4859]: I1008 18:19:42.499316 4859 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/557e8291-7e96-4634-b744-bcf6981d3a05-catalog-content\") pod \"redhat-marketplace-8z5ch\" (UID: \"557e8291-7e96-4634-b744-bcf6981d3a05\") " pod="openshift-marketplace/redhat-marketplace-8z5ch" Oct 08 18:19:42 crc kubenswrapper[4859]: I1008 18:19:42.600761 4859 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/557e8291-7e96-4634-b744-bcf6981d3a05-catalog-content\") pod \"redhat-marketplace-8z5ch\" (UID: \"557e8291-7e96-4634-b744-bcf6981d3a05\") " pod="openshift-marketplace/redhat-marketplace-8z5ch" Oct 08 18:19:42 crc kubenswrapper[4859]: I1008 18:19:42.600920 4859 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/557e8291-7e96-4634-b744-bcf6981d3a05-utilities\") pod \"redhat-marketplace-8z5ch\" (UID: \"557e8291-7e96-4634-b744-bcf6981d3a05\") " pod="openshift-marketplace/redhat-marketplace-8z5ch" Oct 08 18:19:42 crc kubenswrapper[4859]: I1008 18:19:42.600971 4859 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-r5mcr\" (UniqueName: \"kubernetes.io/projected/557e8291-7e96-4634-b744-bcf6981d3a05-kube-api-access-r5mcr\") pod \"redhat-marketplace-8z5ch\" (UID: \"557e8291-7e96-4634-b744-bcf6981d3a05\") " pod="openshift-marketplace/redhat-marketplace-8z5ch" Oct 08 18:19:42 crc kubenswrapper[4859]: I1008 18:19:42.602945 4859 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/557e8291-7e96-4634-b744-bcf6981d3a05-catalog-content\") pod \"redhat-marketplace-8z5ch\" (UID: \"557e8291-7e96-4634-b744-bcf6981d3a05\") " pod="openshift-marketplace/redhat-marketplace-8z5ch" Oct 08 18:19:42 crc kubenswrapper[4859]: I1008 18:19:42.603928 4859 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/557e8291-7e96-4634-b744-bcf6981d3a05-utilities\") pod \"redhat-marketplace-8z5ch\" (UID: \"557e8291-7e96-4634-b744-bcf6981d3a05\") " pod="openshift-marketplace/redhat-marketplace-8z5ch" Oct 08 18:19:42 crc kubenswrapper[4859]: I1008 18:19:42.620498 4859 generic.go:334] "Generic (PLEG): container finished" podID="8bdda932-cfce-4df3-be03-e23d05f19679" containerID="e86684103c0102db6aeb03666e4b3187703e7f83a18169a5b0543e8cb30b863b" exitCode=0 Oct 08 18:19:42 crc kubenswrapper[4859]: I1008 18:19:42.620644 4859 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/revision-pruner-9-crc" event={"ID":"8bdda932-cfce-4df3-be03-e23d05f19679","Type":"ContainerDied","Data":"e86684103c0102db6aeb03666e4b3187703e7f83a18169a5b0543e8cb30b863b"} Oct 08 18:19:42 crc kubenswrapper[4859]: I1008 18:19:42.625012 4859 generic.go:334] "Generic (PLEG): container finished" podID="742016ca-eb57-474a-b4d9-dbfc682628be" containerID="9deae894e6ab21ee78584acb5fbccdcc8d79c9b4f9fd58f0624cdec092b55ed7" exitCode=0 Oct 08 18:19:42 crc kubenswrapper[4859]: I1008 18:19:42.626160 4859 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-pxq2g" event={"ID":"742016ca-eb57-474a-b4d9-dbfc682628be","Type":"ContainerDied","Data":"9deae894e6ab21ee78584acb5fbccdcc8d79c9b4f9fd58f0624cdec092b55ed7"} Oct 08 18:19:42 crc kubenswrapper[4859]: I1008 18:19:42.626185 4859 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-pxq2g" event={"ID":"742016ca-eb57-474a-b4d9-dbfc682628be","Type":"ContainerStarted","Data":"fcb4c394362aaa21fd76341ed436c077abbad9409f43f8d78851b1dc45652d94"} Oct 08 18:19:42 crc kubenswrapper[4859]: I1008 18:19:42.630638 4859 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-r5mcr\" (UniqueName: \"kubernetes.io/projected/557e8291-7e96-4634-b744-bcf6981d3a05-kube-api-access-r5mcr\") pod \"redhat-marketplace-8z5ch\" (UID: \"557e8291-7e96-4634-b744-bcf6981d3a05\") " pod="openshift-marketplace/redhat-marketplace-8z5ch" Oct 08 18:19:42 crc kubenswrapper[4859]: I1008 18:19:42.693333 4859 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-8z5ch" Oct 08 18:19:42 crc kubenswrapper[4859]: I1008 18:19:42.701348 4859 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-console/console-f9d7485db-m2xmp" Oct 08 18:19:42 crc kubenswrapper[4859]: I1008 18:19:42.701389 4859 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-console/console-f9d7485db-m2xmp" Oct 08 18:19:42 crc kubenswrapper[4859]: I1008 18:19:42.711761 4859 patch_prober.go:28] interesting pod/console-f9d7485db-m2xmp container/console namespace/openshift-console: Startup probe status=failure output="Get \"https://10.217.0.10:8443/health\": dial tcp 10.217.0.10:8443: connect: connection refused" start-of-body= Oct 08 18:19:42 crc kubenswrapper[4859]: I1008 18:19:42.711875 4859 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-console/console-f9d7485db-m2xmp" podUID="9a829106-59b1-4389-95bf-2ee4d56e317c" containerName="console" probeResult="failure" output="Get \"https://10.217.0.10:8443/health\": dial tcp 10.217.0.10:8443: connect: connection refused" Oct 08 18:19:42 crc kubenswrapper[4859]: I1008 18:19:42.773958 4859 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-b7vxm"] Oct 08 18:19:42 crc kubenswrapper[4859]: I1008 18:19:42.776594 4859 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-b7vxm" Oct 08 18:19:42 crc kubenswrapper[4859]: I1008 18:19:42.786094 4859 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-b7vxm"] Oct 08 18:19:42 crc kubenswrapper[4859]: I1008 18:19:42.915562 4859 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ingress/router-default-5444994796-r5gs2" Oct 08 18:19:42 crc kubenswrapper[4859]: I1008 18:19:42.918422 4859 patch_prober.go:28] interesting pod/router-default-5444994796-r5gs2 container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Oct 08 18:19:42 crc kubenswrapper[4859]: [-]has-synced failed: reason withheld Oct 08 18:19:42 crc kubenswrapper[4859]: [+]process-running ok Oct 08 18:19:42 crc kubenswrapper[4859]: healthz check failed Oct 08 18:19:42 crc kubenswrapper[4859]: I1008 18:19:42.918467 4859 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-r5gs2" podUID="8e2b8a53-4bc3-499b-9746-628c323d2a0c" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Oct 08 18:19:42 crc kubenswrapper[4859]: I1008 18:19:42.930247 4859 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/eb5e42da-6186-4208-8ceb-348959b9e1da-utilities\") pod \"redhat-marketplace-b7vxm\" (UID: \"eb5e42da-6186-4208-8ceb-348959b9e1da\") " pod="openshift-marketplace/redhat-marketplace-b7vxm" Oct 08 18:19:42 crc kubenswrapper[4859]: I1008 18:19:42.930298 4859 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/eb5e42da-6186-4208-8ceb-348959b9e1da-catalog-content\") pod \"redhat-marketplace-b7vxm\" (UID: \"eb5e42da-6186-4208-8ceb-348959b9e1da\") " pod="openshift-marketplace/redhat-marketplace-b7vxm" Oct 08 18:19:42 crc kubenswrapper[4859]: I1008 18:19:42.930317 4859 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-7jm2v\" (UniqueName: \"kubernetes.io/projected/eb5e42da-6186-4208-8ceb-348959b9e1da-kube-api-access-7jm2v\") pod \"redhat-marketplace-b7vxm\" (UID: \"eb5e42da-6186-4208-8ceb-348959b9e1da\") " pod="openshift-marketplace/redhat-marketplace-b7vxm" Oct 08 18:19:43 crc kubenswrapper[4859]: I1008 18:19:43.010249 4859 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/marketplace-operator-79b997595-bz6v8" Oct 08 18:19:43 crc kubenswrapper[4859]: I1008 18:19:43.031894 4859 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/eb5e42da-6186-4208-8ceb-348959b9e1da-catalog-content\") pod \"redhat-marketplace-b7vxm\" (UID: \"eb5e42da-6186-4208-8ceb-348959b9e1da\") " pod="openshift-marketplace/redhat-marketplace-b7vxm" Oct 08 18:19:43 crc kubenswrapper[4859]: I1008 18:19:43.031953 4859 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-7jm2v\" (UniqueName: \"kubernetes.io/projected/eb5e42da-6186-4208-8ceb-348959b9e1da-kube-api-access-7jm2v\") pod \"redhat-marketplace-b7vxm\" (UID: \"eb5e42da-6186-4208-8ceb-348959b9e1da\") " pod="openshift-marketplace/redhat-marketplace-b7vxm" Oct 08 18:19:43 crc kubenswrapper[4859]: I1008 18:19:43.032109 4859 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/eb5e42da-6186-4208-8ceb-348959b9e1da-utilities\") pod \"redhat-marketplace-b7vxm\" (UID: \"eb5e42da-6186-4208-8ceb-348959b9e1da\") " pod="openshift-marketplace/redhat-marketplace-b7vxm" Oct 08 18:19:43 crc kubenswrapper[4859]: I1008 18:19:43.032871 4859 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/eb5e42da-6186-4208-8ceb-348959b9e1da-utilities\") pod \"redhat-marketplace-b7vxm\" (UID: \"eb5e42da-6186-4208-8ceb-348959b9e1da\") " pod="openshift-marketplace/redhat-marketplace-b7vxm" Oct 08 18:19:43 crc kubenswrapper[4859]: I1008 18:19:43.033112 4859 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/eb5e42da-6186-4208-8ceb-348959b9e1da-catalog-content\") pod \"redhat-marketplace-b7vxm\" (UID: \"eb5e42da-6186-4208-8ceb-348959b9e1da\") " pod="openshift-marketplace/redhat-marketplace-b7vxm" Oct 08 18:19:43 crc kubenswrapper[4859]: I1008 18:19:43.040827 4859 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-8z5ch"] Oct 08 18:19:43 crc kubenswrapper[4859]: I1008 18:19:43.091540 4859 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-7jm2v\" (UniqueName: \"kubernetes.io/projected/eb5e42da-6186-4208-8ceb-348959b9e1da-kube-api-access-7jm2v\") pod \"redhat-marketplace-b7vxm\" (UID: \"eb5e42da-6186-4208-8ceb-348959b9e1da\") " pod="openshift-marketplace/redhat-marketplace-b7vxm" Oct 08 18:19:43 crc kubenswrapper[4859]: I1008 18:19:43.129024 4859 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-b7vxm" Oct 08 18:19:43 crc kubenswrapper[4859]: I1008 18:19:43.381528 4859 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-xxgmr"] Oct 08 18:19:43 crc kubenswrapper[4859]: I1008 18:19:43.383063 4859 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-xxgmr"] Oct 08 18:19:43 crc kubenswrapper[4859]: I1008 18:19:43.383162 4859 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-xxgmr" Oct 08 18:19:43 crc kubenswrapper[4859]: I1008 18:19:43.389314 4859 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"redhat-operators-dockercfg-ct8rh" Oct 08 18:19:43 crc kubenswrapper[4859]: I1008 18:19:43.499142 4859 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-b7vxm"] Oct 08 18:19:43 crc kubenswrapper[4859]: I1008 18:19:43.544094 4859 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-9c6gz\" (UniqueName: \"kubernetes.io/projected/aeadfa85-a482-40a8-b947-09d2f6613900-kube-api-access-9c6gz\") pod \"redhat-operators-xxgmr\" (UID: \"aeadfa85-a482-40a8-b947-09d2f6613900\") " pod="openshift-marketplace/redhat-operators-xxgmr" Oct 08 18:19:43 crc kubenswrapper[4859]: I1008 18:19:43.544154 4859 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/aeadfa85-a482-40a8-b947-09d2f6613900-utilities\") pod \"redhat-operators-xxgmr\" (UID: \"aeadfa85-a482-40a8-b947-09d2f6613900\") " pod="openshift-marketplace/redhat-operators-xxgmr" Oct 08 18:19:43 crc kubenswrapper[4859]: I1008 18:19:43.544213 4859 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/aeadfa85-a482-40a8-b947-09d2f6613900-catalog-content\") pod \"redhat-operators-xxgmr\" (UID: \"aeadfa85-a482-40a8-b947-09d2f6613900\") " pod="openshift-marketplace/redhat-operators-xxgmr" Oct 08 18:19:43 crc kubenswrapper[4859]: I1008 18:19:43.645451 4859 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-9c6gz\" (UniqueName: \"kubernetes.io/projected/aeadfa85-a482-40a8-b947-09d2f6613900-kube-api-access-9c6gz\") pod \"redhat-operators-xxgmr\" (UID: \"aeadfa85-a482-40a8-b947-09d2f6613900\") " pod="openshift-marketplace/redhat-operators-xxgmr" Oct 08 18:19:43 crc kubenswrapper[4859]: I1008 18:19:43.645532 4859 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/aeadfa85-a482-40a8-b947-09d2f6613900-utilities\") pod \"redhat-operators-xxgmr\" (UID: \"aeadfa85-a482-40a8-b947-09d2f6613900\") " pod="openshift-marketplace/redhat-operators-xxgmr" Oct 08 18:19:43 crc kubenswrapper[4859]: I1008 18:19:43.645582 4859 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/aeadfa85-a482-40a8-b947-09d2f6613900-catalog-content\") pod \"redhat-operators-xxgmr\" (UID: \"aeadfa85-a482-40a8-b947-09d2f6613900\") " pod="openshift-marketplace/redhat-operators-xxgmr" Oct 08 18:19:43 crc kubenswrapper[4859]: I1008 18:19:43.646174 4859 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/aeadfa85-a482-40a8-b947-09d2f6613900-catalog-content\") pod \"redhat-operators-xxgmr\" (UID: \"aeadfa85-a482-40a8-b947-09d2f6613900\") " pod="openshift-marketplace/redhat-operators-xxgmr" Oct 08 18:19:43 crc kubenswrapper[4859]: I1008 18:19:43.646894 4859 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/aeadfa85-a482-40a8-b947-09d2f6613900-utilities\") pod \"redhat-operators-xxgmr\" (UID: \"aeadfa85-a482-40a8-b947-09d2f6613900\") " pod="openshift-marketplace/redhat-operators-xxgmr" Oct 08 18:19:43 crc kubenswrapper[4859]: I1008 18:19:43.647112 4859 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-b7vxm" event={"ID":"eb5e42da-6186-4208-8ceb-348959b9e1da","Type":"ContainerStarted","Data":"e509d18e7a8e545af82fd281904c29063c4c0510deca7c581e1d73bd4dcda742"} Oct 08 18:19:43 crc kubenswrapper[4859]: I1008 18:19:43.651233 4859 generic.go:334] "Generic (PLEG): container finished" podID="557e8291-7e96-4634-b744-bcf6981d3a05" containerID="af8756f796c45216e3a3536e2f9927a86ac35371e46f68a4b91bb5c60b9e22b3" exitCode=0 Oct 08 18:19:43 crc kubenswrapper[4859]: I1008 18:19:43.651634 4859 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-8z5ch" event={"ID":"557e8291-7e96-4634-b744-bcf6981d3a05","Type":"ContainerDied","Data":"af8756f796c45216e3a3536e2f9927a86ac35371e46f68a4b91bb5c60b9e22b3"} Oct 08 18:19:43 crc kubenswrapper[4859]: I1008 18:19:43.651779 4859 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-8z5ch" event={"ID":"557e8291-7e96-4634-b744-bcf6981d3a05","Type":"ContainerStarted","Data":"bf3646c1c9400204a8151253765e05527f1418f9ec46904a6062307a265566b3"} Oct 08 18:19:43 crc kubenswrapper[4859]: I1008 18:19:43.675997 4859 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-9c6gz\" (UniqueName: \"kubernetes.io/projected/aeadfa85-a482-40a8-b947-09d2f6613900-kube-api-access-9c6gz\") pod \"redhat-operators-xxgmr\" (UID: \"aeadfa85-a482-40a8-b947-09d2f6613900\") " pod="openshift-marketplace/redhat-operators-xxgmr" Oct 08 18:19:43 crc kubenswrapper[4859]: I1008 18:19:43.718015 4859 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-xxgmr" Oct 08 18:19:43 crc kubenswrapper[4859]: I1008 18:19:43.756878 4859 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-9ml9h"] Oct 08 18:19:43 crc kubenswrapper[4859]: I1008 18:19:43.761857 4859 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-9ml9h" Oct 08 18:19:43 crc kubenswrapper[4859]: I1008 18:19:43.829601 4859 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-9ml9h"] Oct 08 18:19:43 crc kubenswrapper[4859]: I1008 18:19:43.851770 4859 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/1b84a3b0-464e-4d4a-b87f-2341db3b73cc-utilities\") pod \"redhat-operators-9ml9h\" (UID: \"1b84a3b0-464e-4d4a-b87f-2341db3b73cc\") " pod="openshift-marketplace/redhat-operators-9ml9h" Oct 08 18:19:43 crc kubenswrapper[4859]: I1008 18:19:43.851846 4859 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/1b84a3b0-464e-4d4a-b87f-2341db3b73cc-catalog-content\") pod \"redhat-operators-9ml9h\" (UID: \"1b84a3b0-464e-4d4a-b87f-2341db3b73cc\") " pod="openshift-marketplace/redhat-operators-9ml9h" Oct 08 18:19:43 crc kubenswrapper[4859]: I1008 18:19:43.851904 4859 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-4fbtb\" (UniqueName: \"kubernetes.io/projected/1b84a3b0-464e-4d4a-b87f-2341db3b73cc-kube-api-access-4fbtb\") pod \"redhat-operators-9ml9h\" (UID: \"1b84a3b0-464e-4d4a-b87f-2341db3b73cc\") " pod="openshift-marketplace/redhat-operators-9ml9h" Oct 08 18:19:43 crc kubenswrapper[4859]: I1008 18:19:43.918951 4859 patch_prober.go:28] interesting pod/router-default-5444994796-r5gs2 container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Oct 08 18:19:43 crc kubenswrapper[4859]: [-]has-synced failed: reason withheld Oct 08 18:19:43 crc kubenswrapper[4859]: [+]process-running ok Oct 08 18:19:43 crc kubenswrapper[4859]: healthz check failed Oct 08 18:19:43 crc kubenswrapper[4859]: I1008 18:19:43.920562 4859 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-r5gs2" podUID="8e2b8a53-4bc3-499b-9746-628c323d2a0c" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Oct 08 18:19:43 crc kubenswrapper[4859]: I1008 18:19:43.954151 4859 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/1b84a3b0-464e-4d4a-b87f-2341db3b73cc-utilities\") pod \"redhat-operators-9ml9h\" (UID: \"1b84a3b0-464e-4d4a-b87f-2341db3b73cc\") " pod="openshift-marketplace/redhat-operators-9ml9h" Oct 08 18:19:43 crc kubenswrapper[4859]: I1008 18:19:43.954254 4859 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/1b84a3b0-464e-4d4a-b87f-2341db3b73cc-catalog-content\") pod \"redhat-operators-9ml9h\" (UID: \"1b84a3b0-464e-4d4a-b87f-2341db3b73cc\") " pod="openshift-marketplace/redhat-operators-9ml9h" Oct 08 18:19:43 crc kubenswrapper[4859]: I1008 18:19:43.954299 4859 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-4fbtb\" (UniqueName: \"kubernetes.io/projected/1b84a3b0-464e-4d4a-b87f-2341db3b73cc-kube-api-access-4fbtb\") pod \"redhat-operators-9ml9h\" (UID: \"1b84a3b0-464e-4d4a-b87f-2341db3b73cc\") " pod="openshift-marketplace/redhat-operators-9ml9h" Oct 08 18:19:43 crc kubenswrapper[4859]: I1008 18:19:43.955491 4859 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/1b84a3b0-464e-4d4a-b87f-2341db3b73cc-utilities\") pod \"redhat-operators-9ml9h\" (UID: \"1b84a3b0-464e-4d4a-b87f-2341db3b73cc\") " pod="openshift-marketplace/redhat-operators-9ml9h" Oct 08 18:19:43 crc kubenswrapper[4859]: I1008 18:19:43.956183 4859 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/1b84a3b0-464e-4d4a-b87f-2341db3b73cc-catalog-content\") pod \"redhat-operators-9ml9h\" (UID: \"1b84a3b0-464e-4d4a-b87f-2341db3b73cc\") " pod="openshift-marketplace/redhat-operators-9ml9h" Oct 08 18:19:43 crc kubenswrapper[4859]: I1008 18:19:43.977173 4859 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-4fbtb\" (UniqueName: \"kubernetes.io/projected/1b84a3b0-464e-4d4a-b87f-2341db3b73cc-kube-api-access-4fbtb\") pod \"redhat-operators-9ml9h\" (UID: \"1b84a3b0-464e-4d4a-b87f-2341db3b73cc\") " pod="openshift-marketplace/redhat-operators-9ml9h" Oct 08 18:19:44 crc kubenswrapper[4859]: I1008 18:19:44.006127 4859 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-controller-manager/revision-pruner-9-crc" Oct 08 18:19:44 crc kubenswrapper[4859]: I1008 18:19:44.119740 4859 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-xxgmr"] Oct 08 18:19:44 crc kubenswrapper[4859]: I1008 18:19:44.142598 4859 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-9ml9h" Oct 08 18:19:44 crc kubenswrapper[4859]: I1008 18:19:44.157664 4859 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/8bdda932-cfce-4df3-be03-e23d05f19679-kubelet-dir\") pod \"8bdda932-cfce-4df3-be03-e23d05f19679\" (UID: \"8bdda932-cfce-4df3-be03-e23d05f19679\") " Oct 08 18:19:44 crc kubenswrapper[4859]: I1008 18:19:44.157741 4859 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/8bdda932-cfce-4df3-be03-e23d05f19679-kube-api-access\") pod \"8bdda932-cfce-4df3-be03-e23d05f19679\" (UID: \"8bdda932-cfce-4df3-be03-e23d05f19679\") " Oct 08 18:19:44 crc kubenswrapper[4859]: I1008 18:19:44.157812 4859 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/8bdda932-cfce-4df3-be03-e23d05f19679-kubelet-dir" (OuterVolumeSpecName: "kubelet-dir") pod "8bdda932-cfce-4df3-be03-e23d05f19679" (UID: "8bdda932-cfce-4df3-be03-e23d05f19679"). InnerVolumeSpecName "kubelet-dir". PluginName "kubernetes.io/host-path", VolumeGidValue "" Oct 08 18:19:44 crc kubenswrapper[4859]: I1008 18:19:44.158381 4859 reconciler_common.go:293] "Volume detached for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/8bdda932-cfce-4df3-be03-e23d05f19679-kubelet-dir\") on node \"crc\" DevicePath \"\"" Oct 08 18:19:44 crc kubenswrapper[4859]: I1008 18:19:44.178919 4859 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8bdda932-cfce-4df3-be03-e23d05f19679-kube-api-access" (OuterVolumeSpecName: "kube-api-access") pod "8bdda932-cfce-4df3-be03-e23d05f19679" (UID: "8bdda932-cfce-4df3-be03-e23d05f19679"). InnerVolumeSpecName "kube-api-access". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 08 18:19:44 crc kubenswrapper[4859]: I1008 18:19:44.260592 4859 reconciler_common.go:293] "Volume detached for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/8bdda932-cfce-4df3-be03-e23d05f19679-kube-api-access\") on node \"crc\" DevicePath \"\"" Oct 08 18:19:44 crc kubenswrapper[4859]: I1008 18:19:44.516797 4859 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-apiserver/revision-pruner-8-crc"] Oct 08 18:19:44 crc kubenswrapper[4859]: E1008 18:19:44.517400 4859 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8bdda932-cfce-4df3-be03-e23d05f19679" containerName="pruner" Oct 08 18:19:44 crc kubenswrapper[4859]: I1008 18:19:44.517414 4859 state_mem.go:107] "Deleted CPUSet assignment" podUID="8bdda932-cfce-4df3-be03-e23d05f19679" containerName="pruner" Oct 08 18:19:44 crc kubenswrapper[4859]: I1008 18:19:44.517506 4859 memory_manager.go:354] "RemoveStaleState removing state" podUID="8bdda932-cfce-4df3-be03-e23d05f19679" containerName="pruner" Oct 08 18:19:44 crc kubenswrapper[4859]: I1008 18:19:44.521154 4859 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/revision-pruner-8-crc" Oct 08 18:19:44 crc kubenswrapper[4859]: I1008 18:19:44.527484 4859 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-apiserver"/"kube-root-ca.crt" Oct 08 18:19:44 crc kubenswrapper[4859]: I1008 18:19:44.527535 4859 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-apiserver"/"installer-sa-dockercfg-5pr6n" Oct 08 18:19:44 crc kubenswrapper[4859]: I1008 18:19:44.544025 4859 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-apiserver/revision-pruner-8-crc"] Oct 08 18:19:44 crc kubenswrapper[4859]: I1008 18:19:44.672273 4859 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/c488e0ea-436c-4766-a311-56b54bf0fb10-kubelet-dir\") pod \"revision-pruner-8-crc\" (UID: \"c488e0ea-436c-4766-a311-56b54bf0fb10\") " pod="openshift-kube-apiserver/revision-pruner-8-crc" Oct 08 18:19:44 crc kubenswrapper[4859]: I1008 18:19:44.672349 4859 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/c488e0ea-436c-4766-a311-56b54bf0fb10-kube-api-access\") pod \"revision-pruner-8-crc\" (UID: \"c488e0ea-436c-4766-a311-56b54bf0fb10\") " pod="openshift-kube-apiserver/revision-pruner-8-crc" Oct 08 18:19:44 crc kubenswrapper[4859]: I1008 18:19:44.689165 4859 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/revision-pruner-9-crc" event={"ID":"8bdda932-cfce-4df3-be03-e23d05f19679","Type":"ContainerDied","Data":"648d9a0c5732ccf512ff636d748dc4ac27ffb37df3e6ae0072cca948aec44c3d"} Oct 08 18:19:44 crc kubenswrapper[4859]: I1008 18:19:44.689250 4859 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="648d9a0c5732ccf512ff636d748dc4ac27ffb37df3e6ae0072cca948aec44c3d" Oct 08 18:19:44 crc kubenswrapper[4859]: I1008 18:19:44.689191 4859 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-controller-manager/revision-pruner-9-crc" Oct 08 18:19:44 crc kubenswrapper[4859]: I1008 18:19:44.696119 4859 generic.go:334] "Generic (PLEG): container finished" podID="aeadfa85-a482-40a8-b947-09d2f6613900" containerID="86728cd32e96b87dd733a9b0e8329867e46b270d05e3305e491159fe58b2f03c" exitCode=0 Oct 08 18:19:44 crc kubenswrapper[4859]: I1008 18:19:44.696189 4859 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-xxgmr" event={"ID":"aeadfa85-a482-40a8-b947-09d2f6613900","Type":"ContainerDied","Data":"86728cd32e96b87dd733a9b0e8329867e46b270d05e3305e491159fe58b2f03c"} Oct 08 18:19:44 crc kubenswrapper[4859]: I1008 18:19:44.696222 4859 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-xxgmr" event={"ID":"aeadfa85-a482-40a8-b947-09d2f6613900","Type":"ContainerStarted","Data":"343ad2ab2149bab5b80f2c2b5e34738d0110669b70b2f5af343ad42515c23e78"} Oct 08 18:19:44 crc kubenswrapper[4859]: I1008 18:19:44.709018 4859 generic.go:334] "Generic (PLEG): container finished" podID="eb5e42da-6186-4208-8ceb-348959b9e1da" containerID="5f21941ab9e1b6d9e394e6c1da4946f357475da962ec69127f65dd8e6c344899" exitCode=0 Oct 08 18:19:44 crc kubenswrapper[4859]: I1008 18:19:44.709064 4859 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-b7vxm" event={"ID":"eb5e42da-6186-4208-8ceb-348959b9e1da","Type":"ContainerDied","Data":"5f21941ab9e1b6d9e394e6c1da4946f357475da962ec69127f65dd8e6c344899"} Oct 08 18:19:44 crc kubenswrapper[4859]: I1008 18:19:44.729853 4859 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-9ml9h"] Oct 08 18:19:44 crc kubenswrapper[4859]: I1008 18:19:44.773311 4859 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/c488e0ea-436c-4766-a311-56b54bf0fb10-kubelet-dir\") pod \"revision-pruner-8-crc\" (UID: \"c488e0ea-436c-4766-a311-56b54bf0fb10\") " pod="openshift-kube-apiserver/revision-pruner-8-crc" Oct 08 18:19:44 crc kubenswrapper[4859]: I1008 18:19:44.773378 4859 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/c488e0ea-436c-4766-a311-56b54bf0fb10-kube-api-access\") pod \"revision-pruner-8-crc\" (UID: \"c488e0ea-436c-4766-a311-56b54bf0fb10\") " pod="openshift-kube-apiserver/revision-pruner-8-crc" Oct 08 18:19:44 crc kubenswrapper[4859]: I1008 18:19:44.773454 4859 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/c488e0ea-436c-4766-a311-56b54bf0fb10-kubelet-dir\") pod \"revision-pruner-8-crc\" (UID: \"c488e0ea-436c-4766-a311-56b54bf0fb10\") " pod="openshift-kube-apiserver/revision-pruner-8-crc" Oct 08 18:19:44 crc kubenswrapper[4859]: I1008 18:19:44.791345 4859 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/c488e0ea-436c-4766-a311-56b54bf0fb10-kube-api-access\") pod \"revision-pruner-8-crc\" (UID: \"c488e0ea-436c-4766-a311-56b54bf0fb10\") " pod="openshift-kube-apiserver/revision-pruner-8-crc" Oct 08 18:19:44 crc kubenswrapper[4859]: I1008 18:19:44.851218 4859 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/revision-pruner-8-crc" Oct 08 18:19:44 crc kubenswrapper[4859]: I1008 18:19:44.955035 4859 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-ingress/router-default-5444994796-r5gs2" Oct 08 18:19:44 crc kubenswrapper[4859]: I1008 18:19:44.962862 4859 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-ingress/router-default-5444994796-r5gs2" Oct 08 18:19:45 crc kubenswrapper[4859]: I1008 18:19:45.399829 4859 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-apiserver/revision-pruner-8-crc"] Oct 08 18:19:45 crc kubenswrapper[4859]: I1008 18:19:45.720252 4859 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-9ml9h" event={"ID":"1b84a3b0-464e-4d4a-b87f-2341db3b73cc","Type":"ContainerStarted","Data":"ed9287be68925e52499da1bf0708e71fae2564190b4ed03c2c031a97428b5434"} Oct 08 18:19:45 crc kubenswrapper[4859]: I1008 18:19:45.720304 4859 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-9ml9h" event={"ID":"1b84a3b0-464e-4d4a-b87f-2341db3b73cc","Type":"ContainerStarted","Data":"934a532c858325e3db47a11b5ffd60c54e62c546dd66e7c6c4157bd333c86765"} Oct 08 18:19:45 crc kubenswrapper[4859]: I1008 18:19:45.725217 4859 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/revision-pruner-8-crc" event={"ID":"c488e0ea-436c-4766-a311-56b54bf0fb10","Type":"ContainerStarted","Data":"d930249945d6d8d261b76a47c3f80783dc8067c8e5ccec02cedd453a5a5a8c00"} Oct 08 18:19:46 crc kubenswrapper[4859]: I1008 18:19:46.755916 4859 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/revision-pruner-8-crc" event={"ID":"c488e0ea-436c-4766-a311-56b54bf0fb10","Type":"ContainerStarted","Data":"507e1a3faf7a9e7105b280f4ea86990eeb229148b230c87d2f6a533520fcc872"} Oct 08 18:19:46 crc kubenswrapper[4859]: I1008 18:19:46.771763 4859 generic.go:334] "Generic (PLEG): container finished" podID="1b84a3b0-464e-4d4a-b87f-2341db3b73cc" containerID="ed9287be68925e52499da1bf0708e71fae2564190b4ed03c2c031a97428b5434" exitCode=0 Oct 08 18:19:46 crc kubenswrapper[4859]: I1008 18:19:46.772009 4859 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-9ml9h" event={"ID":"1b84a3b0-464e-4d4a-b87f-2341db3b73cc","Type":"ContainerDied","Data":"ed9287be68925e52499da1bf0708e71fae2564190b4ed03c2c031a97428b5434"} Oct 08 18:19:47 crc kubenswrapper[4859]: I1008 18:19:47.788030 4859 generic.go:334] "Generic (PLEG): container finished" podID="c488e0ea-436c-4766-a311-56b54bf0fb10" containerID="507e1a3faf7a9e7105b280f4ea86990eeb229148b230c87d2f6a533520fcc872" exitCode=0 Oct 08 18:19:47 crc kubenswrapper[4859]: I1008 18:19:47.788126 4859 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/revision-pruner-8-crc" event={"ID":"c488e0ea-436c-4766-a311-56b54bf0fb10","Type":"ContainerDied","Data":"507e1a3faf7a9e7105b280f4ea86990eeb229148b230c87d2f6a533520fcc872"} Oct 08 18:19:47 crc kubenswrapper[4859]: I1008 18:19:47.924331 4859 patch_prober.go:28] interesting pod/machine-config-daemon-82s52 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 08 18:19:47 crc kubenswrapper[4859]: I1008 18:19:47.924383 4859 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-82s52" podUID="b23a6a6c-9d92-4e7b-840e-55cfda873a2d" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 08 18:19:48 crc kubenswrapper[4859]: I1008 18:19:48.628396 4859 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-dns/dns-default-x7pp9" Oct 08 18:19:51 crc kubenswrapper[4859]: I1008 18:19:51.732078 4859 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/revision-pruner-8-crc" Oct 08 18:19:51 crc kubenswrapper[4859]: I1008 18:19:51.783021 4859 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/c488e0ea-436c-4766-a311-56b54bf0fb10-kube-api-access\") pod \"c488e0ea-436c-4766-a311-56b54bf0fb10\" (UID: \"c488e0ea-436c-4766-a311-56b54bf0fb10\") " Oct 08 18:19:51 crc kubenswrapper[4859]: I1008 18:19:51.783087 4859 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/c488e0ea-436c-4766-a311-56b54bf0fb10-kubelet-dir\") pod \"c488e0ea-436c-4766-a311-56b54bf0fb10\" (UID: \"c488e0ea-436c-4766-a311-56b54bf0fb10\") " Oct 08 18:19:51 crc kubenswrapper[4859]: I1008 18:19:51.783159 4859 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/c488e0ea-436c-4766-a311-56b54bf0fb10-kubelet-dir" (OuterVolumeSpecName: "kubelet-dir") pod "c488e0ea-436c-4766-a311-56b54bf0fb10" (UID: "c488e0ea-436c-4766-a311-56b54bf0fb10"). InnerVolumeSpecName "kubelet-dir". PluginName "kubernetes.io/host-path", VolumeGidValue "" Oct 08 18:19:51 crc kubenswrapper[4859]: I1008 18:19:51.783403 4859 reconciler_common.go:293] "Volume detached for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/c488e0ea-436c-4766-a311-56b54bf0fb10-kubelet-dir\") on node \"crc\" DevicePath \"\"" Oct 08 18:19:51 crc kubenswrapper[4859]: I1008 18:19:51.789857 4859 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/c488e0ea-436c-4766-a311-56b54bf0fb10-kube-api-access" (OuterVolumeSpecName: "kube-api-access") pod "c488e0ea-436c-4766-a311-56b54bf0fb10" (UID: "c488e0ea-436c-4766-a311-56b54bf0fb10"). InnerVolumeSpecName "kube-api-access". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 08 18:19:51 crc kubenswrapper[4859]: I1008 18:19:51.821889 4859 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/revision-pruner-8-crc" event={"ID":"c488e0ea-436c-4766-a311-56b54bf0fb10","Type":"ContainerDied","Data":"d930249945d6d8d261b76a47c3f80783dc8067c8e5ccec02cedd453a5a5a8c00"} Oct 08 18:19:51 crc kubenswrapper[4859]: I1008 18:19:51.821918 4859 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/revision-pruner-8-crc" Oct 08 18:19:51 crc kubenswrapper[4859]: I1008 18:19:51.821939 4859 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="d930249945d6d8d261b76a47c3f80783dc8067c8e5ccec02cedd453a5a5a8c00" Oct 08 18:19:51 crc kubenswrapper[4859]: I1008 18:19:51.884859 4859 reconciler_common.go:293] "Volume detached for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/c488e0ea-436c-4766-a311-56b54bf0fb10-kube-api-access\") on node \"crc\" DevicePath \"\"" Oct 08 18:19:52 crc kubenswrapper[4859]: I1008 18:19:52.479749 4859 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-console/downloads-7954f5f757-8q5jz" Oct 08 18:19:52 crc kubenswrapper[4859]: I1008 18:19:52.733795 4859 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-console/console-f9d7485db-m2xmp" Oct 08 18:19:52 crc kubenswrapper[4859]: I1008 18:19:52.738664 4859 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-console/console-f9d7485db-m2xmp" Oct 08 18:19:54 crc kubenswrapper[4859]: I1008 18:19:54.115768 4859 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/2e53cd9b-64f7-4e07-8b96-fb77847c9ac6-metrics-certs\") pod \"network-metrics-daemon-4b4sz\" (UID: \"2e53cd9b-64f7-4e07-8b96-fb77847c9ac6\") " pod="openshift-multus/network-metrics-daemon-4b4sz" Oct 08 18:19:54 crc kubenswrapper[4859]: I1008 18:19:54.134312 4859 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/2e53cd9b-64f7-4e07-8b96-fb77847c9ac6-metrics-certs\") pod \"network-metrics-daemon-4b4sz\" (UID: \"2e53cd9b-64f7-4e07-8b96-fb77847c9ac6\") " pod="openshift-multus/network-metrics-daemon-4b4sz" Oct 08 18:19:54 crc kubenswrapper[4859]: I1008 18:19:54.387054 4859 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-4b4sz" Oct 08 18:20:00 crc kubenswrapper[4859]: I1008 18:20:00.919344 4859 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-image-registry/image-registry-697d97f7c8-kg44m" Oct 08 18:20:08 crc kubenswrapper[4859]: E1008 18:20:08.421220 4859 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" image="registry.redhat.io/redhat/community-operator-index:v4.18" Oct 08 18:20:08 crc kubenswrapper[4859]: E1008 18:20:08.422029 4859 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:extract-content,Image:registry.redhat.io/redhat/community-operator-index:v4.18,Command:[/utilities/copy-content],Args:[--catalog.from=/configs --catalog.to=/extracted-catalog/catalog --cache.from=/tmp/cache --cache.to=/extracted-catalog/cache],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:utilities,ReadOnly:false,MountPath:/utilities,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:catalog-content,ReadOnly:false,MountPath:/extracted-catalog,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-7g9pg,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:Always,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000170000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:FallbackToLogsOnError,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod community-operators-zb82c_openshift-marketplace(5b68d592-4a46-4fbc-bcec-a8bca2b973aa): ErrImagePull: rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" logger="UnhandledError" Oct 08 18:20:08 crc kubenswrapper[4859]: E1008 18:20:08.423430 4859 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ErrImagePull: \"rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled\"" pod="openshift-marketplace/community-operators-zb82c" podUID="5b68d592-4a46-4fbc-bcec-a8bca2b973aa" Oct 08 18:20:11 crc kubenswrapper[4859]: E1008 18:20:11.651519 4859 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"\"" pod="openshift-marketplace/community-operators-zb82c" podUID="5b68d592-4a46-4fbc-bcec-a8bca2b973aa" Oct 08 18:20:11 crc kubenswrapper[4859]: E1008 18:20:11.743196 4859 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" image="registry.redhat.io/redhat/redhat-operator-index:v4.18" Oct 08 18:20:11 crc kubenswrapper[4859]: E1008 18:20:11.743955 4859 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:extract-content,Image:registry.redhat.io/redhat/redhat-operator-index:v4.18,Command:[/utilities/copy-content],Args:[--catalog.from=/configs --catalog.to=/extracted-catalog/catalog --cache.from=/tmp/cache --cache.to=/extracted-catalog/cache],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:utilities,ReadOnly:false,MountPath:/utilities,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:catalog-content,ReadOnly:false,MountPath:/extracted-catalog,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-9c6gz,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:Always,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000170000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:FallbackToLogsOnError,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod redhat-operators-xxgmr_openshift-marketplace(aeadfa85-a482-40a8-b947-09d2f6613900): ErrImagePull: rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" logger="UnhandledError" Oct 08 18:20:11 crc kubenswrapper[4859]: E1008 18:20:11.745639 4859 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ErrImagePull: \"rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled\"" pod="openshift-marketplace/redhat-operators-xxgmr" podUID="aeadfa85-a482-40a8-b947-09d2f6613900" Oct 08 18:20:11 crc kubenswrapper[4859]: E1008 18:20:11.772802 4859 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" image="registry.redhat.io/redhat/redhat-marketplace-index:v4.18" Oct 08 18:20:11 crc kubenswrapper[4859]: E1008 18:20:11.772968 4859 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:extract-content,Image:registry.redhat.io/redhat/redhat-marketplace-index:v4.18,Command:[/utilities/copy-content],Args:[--catalog.from=/configs --catalog.to=/extracted-catalog/catalog --cache.from=/tmp/cache --cache.to=/extracted-catalog/cache],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:utilities,ReadOnly:false,MountPath:/utilities,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:catalog-content,ReadOnly:false,MountPath:/extracted-catalog,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-r5mcr,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:Always,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000170000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:FallbackToLogsOnError,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod redhat-marketplace-8z5ch_openshift-marketplace(557e8291-7e96-4634-b744-bcf6981d3a05): ErrImagePull: rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" logger="UnhandledError" Oct 08 18:20:11 crc kubenswrapper[4859]: E1008 18:20:11.774122 4859 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ErrImagePull: \"rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled\"" pod="openshift-marketplace/redhat-marketplace-8z5ch" podUID="557e8291-7e96-4634-b744-bcf6981d3a05" Oct 08 18:20:11 crc kubenswrapper[4859]: I1008 18:20:11.915190 4859 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-multus/network-metrics-daemon-4b4sz"] Oct 08 18:20:11 crc kubenswrapper[4859]: W1008 18:20:11.942953 4859 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod2e53cd9b_64f7_4e07_8b96_fb77847c9ac6.slice/crio-0b55b4cecdac083ec0f1bbccd6f7524124b488be07ba2f0f6dea8b9270d2f77f WatchSource:0}: Error finding container 0b55b4cecdac083ec0f1bbccd6f7524124b488be07ba2f0f6dea8b9270d2f77f: Status 404 returned error can't find the container with id 0b55b4cecdac083ec0f1bbccd6f7524124b488be07ba2f0f6dea8b9270d2f77f Oct 08 18:20:11 crc kubenswrapper[4859]: I1008 18:20:11.960710 4859 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/network-metrics-daemon-4b4sz" event={"ID":"2e53cd9b-64f7-4e07-8b96-fb77847c9ac6","Type":"ContainerStarted","Data":"0b55b4cecdac083ec0f1bbccd6f7524124b488be07ba2f0f6dea8b9270d2f77f"} Oct 08 18:20:11 crc kubenswrapper[4859]: I1008 18:20:11.962850 4859 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-c5tcx" event={"ID":"438a5b9a-aaee-4e06-9d22-691862b9207e","Type":"ContainerStarted","Data":"3a5c39dbc2e6abe1c86f584dff2ab8edde1ccbdd76be2b5d7d17cef9c1396da8"} Oct 08 18:20:11 crc kubenswrapper[4859]: E1008 18:20:11.967151 4859 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"\"" pod="openshift-marketplace/redhat-operators-xxgmr" podUID="aeadfa85-a482-40a8-b947-09d2f6613900" Oct 08 18:20:11 crc kubenswrapper[4859]: E1008 18:20:11.968321 4859 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"\"" pod="openshift-marketplace/redhat-marketplace-8z5ch" podUID="557e8291-7e96-4634-b744-bcf6981d3a05" Oct 08 18:20:12 crc kubenswrapper[4859]: I1008 18:20:12.975506 4859 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/network-metrics-daemon-4b4sz" event={"ID":"2e53cd9b-64f7-4e07-8b96-fb77847c9ac6","Type":"ContainerStarted","Data":"b27c41fb4e4ca3e43d7a0a583c7f298dc9f69d9c9b1cc972f1377662c430f00f"} Oct 08 18:20:12 crc kubenswrapper[4859]: I1008 18:20:12.975910 4859 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/network-metrics-daemon-4b4sz" event={"ID":"2e53cd9b-64f7-4e07-8b96-fb77847c9ac6","Type":"ContainerStarted","Data":"bdaad4a19a9560b4f8331a4ee3e8b5454a55d0e29d9873a2f921e5e3248ffcd1"} Oct 08 18:20:12 crc kubenswrapper[4859]: I1008 18:20:12.979330 4859 generic.go:334] "Generic (PLEG): container finished" podID="ae751baf-6f53-4a05-b28d-0888f7682db7" containerID="8720d9a552ab4d000922475090494de59806953a5afadcf848cde2c83bb63de3" exitCode=0 Oct 08 18:20:12 crc kubenswrapper[4859]: I1008 18:20:12.979476 4859 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-q6ktk" event={"ID":"ae751baf-6f53-4a05-b28d-0888f7682db7","Type":"ContainerDied","Data":"8720d9a552ab4d000922475090494de59806953a5afadcf848cde2c83bb63de3"} Oct 08 18:20:12 crc kubenswrapper[4859]: I1008 18:20:12.982372 4859 generic.go:334] "Generic (PLEG): container finished" podID="1b84a3b0-464e-4d4a-b87f-2341db3b73cc" containerID="ce4a61e3e561228ff7b8c66698b5e93c28edd16df3b59bfc8ccb5c99c0811594" exitCode=0 Oct 08 18:20:12 crc kubenswrapper[4859]: I1008 18:20:12.982437 4859 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-9ml9h" event={"ID":"1b84a3b0-464e-4d4a-b87f-2341db3b73cc","Type":"ContainerDied","Data":"ce4a61e3e561228ff7b8c66698b5e93c28edd16df3b59bfc8ccb5c99c0811594"} Oct 08 18:20:12 crc kubenswrapper[4859]: I1008 18:20:12.986095 4859 generic.go:334] "Generic (PLEG): container finished" podID="438a5b9a-aaee-4e06-9d22-691862b9207e" containerID="3a5c39dbc2e6abe1c86f584dff2ab8edde1ccbdd76be2b5d7d17cef9c1396da8" exitCode=0 Oct 08 18:20:12 crc kubenswrapper[4859]: I1008 18:20:12.986293 4859 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-c5tcx" event={"ID":"438a5b9a-aaee-4e06-9d22-691862b9207e","Type":"ContainerDied","Data":"3a5c39dbc2e6abe1c86f584dff2ab8edde1ccbdd76be2b5d7d17cef9c1396da8"} Oct 08 18:20:12 crc kubenswrapper[4859]: I1008 18:20:12.990157 4859 generic.go:334] "Generic (PLEG): container finished" podID="eb5e42da-6186-4208-8ceb-348959b9e1da" containerID="1c11ea9059b1720d52e070b6dfc6db266a57ebb6f9f2094bae86f980f45f6bc5" exitCode=0 Oct 08 18:20:12 crc kubenswrapper[4859]: I1008 18:20:12.990200 4859 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-b7vxm" event={"ID":"eb5e42da-6186-4208-8ceb-348959b9e1da","Type":"ContainerDied","Data":"1c11ea9059b1720d52e070b6dfc6db266a57ebb6f9f2094bae86f980f45f6bc5"} Oct 08 18:20:12 crc kubenswrapper[4859]: I1008 18:20:12.995505 4859 generic.go:334] "Generic (PLEG): container finished" podID="742016ca-eb57-474a-b4d9-dbfc682628be" containerID="afba94ebc1a6f3924cbdb193588fcdeeecbdfca1175af7d119e40a1d2224aa4c" exitCode=0 Oct 08 18:20:12 crc kubenswrapper[4859]: I1008 18:20:12.995541 4859 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-pxq2g" event={"ID":"742016ca-eb57-474a-b4d9-dbfc682628be","Type":"ContainerDied","Data":"afba94ebc1a6f3924cbdb193588fcdeeecbdfca1175af7d119e40a1d2224aa4c"} Oct 08 18:20:13 crc kubenswrapper[4859]: I1008 18:20:13.006257 4859 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-multus/network-metrics-daemon-4b4sz" podStartSLOduration=163.006243131 podStartE2EDuration="2m43.006243131s" podCreationTimestamp="2025-10-08 18:17:30 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-08 18:20:13.004778617 +0000 UTC m=+183.251618006" watchObservedRunningTime="2025-10-08 18:20:13.006243131 +0000 UTC m=+183.253082510" Oct 08 18:20:13 crc kubenswrapper[4859]: I1008 18:20:13.615950 4859 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-hf2fp" Oct 08 18:20:14 crc kubenswrapper[4859]: I1008 18:20:14.004136 4859 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-b7vxm" event={"ID":"eb5e42da-6186-4208-8ceb-348959b9e1da","Type":"ContainerStarted","Data":"a8c3827c2b47958d14b87bde8a6b3061aedfb311fd7c96da23198d5b80c3869b"} Oct 08 18:20:14 crc kubenswrapper[4859]: I1008 18:20:14.007987 4859 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-pxq2g" event={"ID":"742016ca-eb57-474a-b4d9-dbfc682628be","Type":"ContainerStarted","Data":"1d90316bbffaa725abf95beeb54bcd8eab5e2bc82d8bc3b9c5aa016801b9775e"} Oct 08 18:20:14 crc kubenswrapper[4859]: I1008 18:20:14.011587 4859 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-q6ktk" event={"ID":"ae751baf-6f53-4a05-b28d-0888f7682db7","Type":"ContainerStarted","Data":"837340b192e24abc2fa902de0e90df4b1d18563db537b98272dd37dd177459cf"} Oct 08 18:20:14 crc kubenswrapper[4859]: I1008 18:20:14.014671 4859 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-9ml9h" event={"ID":"1b84a3b0-464e-4d4a-b87f-2341db3b73cc","Type":"ContainerStarted","Data":"8a47ec34fbda7524552f5a47b071ad6395bb48ac13fe587b4fb3e55f085ec428"} Oct 08 18:20:14 crc kubenswrapper[4859]: I1008 18:20:14.017155 4859 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-c5tcx" event={"ID":"438a5b9a-aaee-4e06-9d22-691862b9207e","Type":"ContainerStarted","Data":"8b6f3d58eacd7cf2bce7bfad476971fff5b2db73d55e67cbe99f2fe89dea2ac6"} Oct 08 18:20:14 crc kubenswrapper[4859]: I1008 18:20:14.032835 4859 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-b7vxm" podStartSLOduration=3.051015436 podStartE2EDuration="32.032806754s" podCreationTimestamp="2025-10-08 18:19:42 +0000 UTC" firstStartedPulling="2025-10-08 18:19:44.71035501 +0000 UTC m=+154.957194389" lastFinishedPulling="2025-10-08 18:20:13.692146328 +0000 UTC m=+183.938985707" observedRunningTime="2025-10-08 18:20:14.031645699 +0000 UTC m=+184.278485088" watchObservedRunningTime="2025-10-08 18:20:14.032806754 +0000 UTC m=+184.279646143" Oct 08 18:20:14 crc kubenswrapper[4859]: I1008 18:20:14.054452 4859 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-q6ktk" podStartSLOduration=2.194513786 podStartE2EDuration="34.054435019s" podCreationTimestamp="2025-10-08 18:19:40 +0000 UTC" firstStartedPulling="2025-10-08 18:19:41.589966866 +0000 UTC m=+151.836806245" lastFinishedPulling="2025-10-08 18:20:13.449888039 +0000 UTC m=+183.696727478" observedRunningTime="2025-10-08 18:20:14.050242034 +0000 UTC m=+184.297081413" watchObservedRunningTime="2025-10-08 18:20:14.054435019 +0000 UTC m=+184.301274398" Oct 08 18:20:14 crc kubenswrapper[4859]: I1008 18:20:14.077628 4859 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-c5tcx" podStartSLOduration=2.263483344 podStartE2EDuration="34.07760195s" podCreationTimestamp="2025-10-08 18:19:40 +0000 UTC" firstStartedPulling="2025-10-08 18:19:41.59411638 +0000 UTC m=+151.840955749" lastFinishedPulling="2025-10-08 18:20:13.408234936 +0000 UTC m=+183.655074355" observedRunningTime="2025-10-08 18:20:14.074365404 +0000 UTC m=+184.321204783" watchObservedRunningTime="2025-10-08 18:20:14.07760195 +0000 UTC m=+184.324441329" Oct 08 18:20:14 crc kubenswrapper[4859]: I1008 18:20:14.096197 4859 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-9ml9h" podStartSLOduration=9.020191205 podStartE2EDuration="31.096179335s" podCreationTimestamp="2025-10-08 18:19:43 +0000 UTC" firstStartedPulling="2025-10-08 18:19:51.68892519 +0000 UTC m=+161.935764569" lastFinishedPulling="2025-10-08 18:20:13.76491332 +0000 UTC m=+184.011752699" observedRunningTime="2025-10-08 18:20:14.096090212 +0000 UTC m=+184.342929591" watchObservedRunningTime="2025-10-08 18:20:14.096179335 +0000 UTC m=+184.343018714" Oct 08 18:20:14 crc kubenswrapper[4859]: I1008 18:20:14.119411 4859 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-pxq2g" podStartSLOduration=2.97038635 podStartE2EDuration="34.119390878s" podCreationTimestamp="2025-10-08 18:19:40 +0000 UTC" firstStartedPulling="2025-10-08 18:19:42.627962511 +0000 UTC m=+152.874801890" lastFinishedPulling="2025-10-08 18:20:13.776967039 +0000 UTC m=+184.023806418" observedRunningTime="2025-10-08 18:20:14.115567743 +0000 UTC m=+184.362407152" watchObservedRunningTime="2025-10-08 18:20:14.119390878 +0000 UTC m=+184.366230257" Oct 08 18:20:14 crc kubenswrapper[4859]: I1008 18:20:14.142853 4859 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-9ml9h" Oct 08 18:20:14 crc kubenswrapper[4859]: I1008 18:20:14.143087 4859 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-9ml9h" Oct 08 18:20:15 crc kubenswrapper[4859]: I1008 18:20:15.313697 4859 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/redhat-operators-9ml9h" podUID="1b84a3b0-464e-4d4a-b87f-2341db3b73cc" containerName="registry-server" probeResult="failure" output=< Oct 08 18:20:15 crc kubenswrapper[4859]: timeout: failed to connect service ":50051" within 1s Oct 08 18:20:15 crc kubenswrapper[4859]: > Oct 08 18:20:17 crc kubenswrapper[4859]: I1008 18:20:17.924664 4859 patch_prober.go:28] interesting pod/machine-config-daemon-82s52 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 08 18:20:17 crc kubenswrapper[4859]: I1008 18:20:17.925270 4859 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-82s52" podUID="b23a6a6c-9d92-4e7b-840e-55cfda873a2d" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 08 18:20:18 crc kubenswrapper[4859]: I1008 18:20:18.903287 4859 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 08 18:20:20 crc kubenswrapper[4859]: I1008 18:20:20.679278 4859 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-c5tcx" Oct 08 18:20:20 crc kubenswrapper[4859]: I1008 18:20:20.679647 4859 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-c5tcx" Oct 08 18:20:20 crc kubenswrapper[4859]: I1008 18:20:20.736861 4859 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-c5tcx" Oct 08 18:20:21 crc kubenswrapper[4859]: I1008 18:20:21.112398 4859 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-c5tcx" Oct 08 18:20:21 crc kubenswrapper[4859]: I1008 18:20:21.117523 4859 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-q6ktk" Oct 08 18:20:21 crc kubenswrapper[4859]: I1008 18:20:21.117573 4859 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-q6ktk" Oct 08 18:20:21 crc kubenswrapper[4859]: I1008 18:20:21.178335 4859 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-q6ktk" Oct 08 18:20:21 crc kubenswrapper[4859]: I1008 18:20:21.201472 4859 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-pxq2g" Oct 08 18:20:21 crc kubenswrapper[4859]: I1008 18:20:21.201608 4859 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-pxq2g" Oct 08 18:20:21 crc kubenswrapper[4859]: I1008 18:20:21.238824 4859 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-pxq2g" Oct 08 18:20:22 crc kubenswrapper[4859]: I1008 18:20:22.105902 4859 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-q6ktk" Oct 08 18:20:22 crc kubenswrapper[4859]: I1008 18:20:22.119190 4859 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-pxq2g" Oct 08 18:20:23 crc kubenswrapper[4859]: I1008 18:20:23.133322 4859 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-b7vxm" Oct 08 18:20:23 crc kubenswrapper[4859]: I1008 18:20:23.133476 4859 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-b7vxm" Oct 08 18:20:23 crc kubenswrapper[4859]: I1008 18:20:23.189613 4859 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-b7vxm" Oct 08 18:20:23 crc kubenswrapper[4859]: I1008 18:20:23.572148 4859 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-pxq2g"] Oct 08 18:20:24 crc kubenswrapper[4859]: I1008 18:20:24.137971 4859 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-b7vxm" Oct 08 18:20:24 crc kubenswrapper[4859]: I1008 18:20:24.178124 4859 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-q6ktk"] Oct 08 18:20:24 crc kubenswrapper[4859]: I1008 18:20:24.178379 4859 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/certified-operators-q6ktk" podUID="ae751baf-6f53-4a05-b28d-0888f7682db7" containerName="registry-server" containerID="cri-o://837340b192e24abc2fa902de0e90df4b1d18563db537b98272dd37dd177459cf" gracePeriod=2 Oct 08 18:20:24 crc kubenswrapper[4859]: I1008 18:20:24.191878 4859 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-9ml9h" Oct 08 18:20:24 crc kubenswrapper[4859]: I1008 18:20:24.238791 4859 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-9ml9h" Oct 08 18:20:25 crc kubenswrapper[4859]: I1008 18:20:25.090983 4859 generic.go:334] "Generic (PLEG): container finished" podID="ae751baf-6f53-4a05-b28d-0888f7682db7" containerID="837340b192e24abc2fa902de0e90df4b1d18563db537b98272dd37dd177459cf" exitCode=0 Oct 08 18:20:25 crc kubenswrapper[4859]: I1008 18:20:25.091032 4859 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-q6ktk" event={"ID":"ae751baf-6f53-4a05-b28d-0888f7682db7","Type":"ContainerDied","Data":"837340b192e24abc2fa902de0e90df4b1d18563db537b98272dd37dd177459cf"} Oct 08 18:20:25 crc kubenswrapper[4859]: I1008 18:20:25.092155 4859 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/community-operators-pxq2g" podUID="742016ca-eb57-474a-b4d9-dbfc682628be" containerName="registry-server" containerID="cri-o://1d90316bbffaa725abf95beeb54bcd8eab5e2bc82d8bc3b9c5aa016801b9775e" gracePeriod=2 Oct 08 18:20:25 crc kubenswrapper[4859]: I1008 18:20:25.862550 4859 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-q6ktk" Oct 08 18:20:25 crc kubenswrapper[4859]: I1008 18:20:25.928064 4859 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-t9kgm\" (UniqueName: \"kubernetes.io/projected/ae751baf-6f53-4a05-b28d-0888f7682db7-kube-api-access-t9kgm\") pod \"ae751baf-6f53-4a05-b28d-0888f7682db7\" (UID: \"ae751baf-6f53-4a05-b28d-0888f7682db7\") " Oct 08 18:20:25 crc kubenswrapper[4859]: I1008 18:20:25.928160 4859 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/ae751baf-6f53-4a05-b28d-0888f7682db7-catalog-content\") pod \"ae751baf-6f53-4a05-b28d-0888f7682db7\" (UID: \"ae751baf-6f53-4a05-b28d-0888f7682db7\") " Oct 08 18:20:25 crc kubenswrapper[4859]: I1008 18:20:25.928288 4859 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/ae751baf-6f53-4a05-b28d-0888f7682db7-utilities\") pod \"ae751baf-6f53-4a05-b28d-0888f7682db7\" (UID: \"ae751baf-6f53-4a05-b28d-0888f7682db7\") " Oct 08 18:20:25 crc kubenswrapper[4859]: I1008 18:20:25.929402 4859 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/ae751baf-6f53-4a05-b28d-0888f7682db7-utilities" (OuterVolumeSpecName: "utilities") pod "ae751baf-6f53-4a05-b28d-0888f7682db7" (UID: "ae751baf-6f53-4a05-b28d-0888f7682db7"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 08 18:20:25 crc kubenswrapper[4859]: I1008 18:20:25.934971 4859 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/ae751baf-6f53-4a05-b28d-0888f7682db7-kube-api-access-t9kgm" (OuterVolumeSpecName: "kube-api-access-t9kgm") pod "ae751baf-6f53-4a05-b28d-0888f7682db7" (UID: "ae751baf-6f53-4a05-b28d-0888f7682db7"). InnerVolumeSpecName "kube-api-access-t9kgm". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 08 18:20:25 crc kubenswrapper[4859]: I1008 18:20:25.972708 4859 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-b7vxm"] Oct 08 18:20:25 crc kubenswrapper[4859]: I1008 18:20:25.988233 4859 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/ae751baf-6f53-4a05-b28d-0888f7682db7-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "ae751baf-6f53-4a05-b28d-0888f7682db7" (UID: "ae751baf-6f53-4a05-b28d-0888f7682db7"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 08 18:20:26 crc kubenswrapper[4859]: I1008 18:20:26.030384 4859 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/ae751baf-6f53-4a05-b28d-0888f7682db7-utilities\") on node \"crc\" DevicePath \"\"" Oct 08 18:20:26 crc kubenswrapper[4859]: I1008 18:20:26.030750 4859 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-t9kgm\" (UniqueName: \"kubernetes.io/projected/ae751baf-6f53-4a05-b28d-0888f7682db7-kube-api-access-t9kgm\") on node \"crc\" DevicePath \"\"" Oct 08 18:20:26 crc kubenswrapper[4859]: I1008 18:20:26.030846 4859 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/ae751baf-6f53-4a05-b28d-0888f7682db7-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 08 18:20:26 crc kubenswrapper[4859]: I1008 18:20:26.098808 4859 generic.go:334] "Generic (PLEG): container finished" podID="742016ca-eb57-474a-b4d9-dbfc682628be" containerID="1d90316bbffaa725abf95beeb54bcd8eab5e2bc82d8bc3b9c5aa016801b9775e" exitCode=0 Oct 08 18:20:26 crc kubenswrapper[4859]: I1008 18:20:26.098914 4859 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-pxq2g" event={"ID":"742016ca-eb57-474a-b4d9-dbfc682628be","Type":"ContainerDied","Data":"1d90316bbffaa725abf95beeb54bcd8eab5e2bc82d8bc3b9c5aa016801b9775e"} Oct 08 18:20:26 crc kubenswrapper[4859]: I1008 18:20:26.100867 4859 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-q6ktk" event={"ID":"ae751baf-6f53-4a05-b28d-0888f7682db7","Type":"ContainerDied","Data":"e4c394bb7faf6da21cb8bcdc695d6518a4f4e91eacd37fe96ec9b270dedafb83"} Oct 08 18:20:26 crc kubenswrapper[4859]: I1008 18:20:26.100915 4859 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-q6ktk" Oct 08 18:20:26 crc kubenswrapper[4859]: I1008 18:20:26.100922 4859 scope.go:117] "RemoveContainer" containerID="837340b192e24abc2fa902de0e90df4b1d18563db537b98272dd37dd177459cf" Oct 08 18:20:26 crc kubenswrapper[4859]: I1008 18:20:26.101251 4859 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-marketplace-b7vxm" podUID="eb5e42da-6186-4208-8ceb-348959b9e1da" containerName="registry-server" containerID="cri-o://a8c3827c2b47958d14b87bde8a6b3061aedfb311fd7c96da23198d5b80c3869b" gracePeriod=2 Oct 08 18:20:26 crc kubenswrapper[4859]: I1008 18:20:26.132668 4859 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-q6ktk"] Oct 08 18:20:26 crc kubenswrapper[4859]: I1008 18:20:26.140218 4859 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/certified-operators-q6ktk"] Oct 08 18:20:26 crc kubenswrapper[4859]: I1008 18:20:26.479363 4859 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="ae751baf-6f53-4a05-b28d-0888f7682db7" path="/var/lib/kubelet/pods/ae751baf-6f53-4a05-b28d-0888f7682db7/volumes" Oct 08 18:20:26 crc kubenswrapper[4859]: I1008 18:20:26.871415 4859 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-pxq2g" Oct 08 18:20:26 crc kubenswrapper[4859]: I1008 18:20:26.947091 4859 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/742016ca-eb57-474a-b4d9-dbfc682628be-catalog-content\") pod \"742016ca-eb57-474a-b4d9-dbfc682628be\" (UID: \"742016ca-eb57-474a-b4d9-dbfc682628be\") " Oct 08 18:20:26 crc kubenswrapper[4859]: I1008 18:20:26.947184 4859 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/742016ca-eb57-474a-b4d9-dbfc682628be-utilities\") pod \"742016ca-eb57-474a-b4d9-dbfc682628be\" (UID: \"742016ca-eb57-474a-b4d9-dbfc682628be\") " Oct 08 18:20:26 crc kubenswrapper[4859]: I1008 18:20:26.947240 4859 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-694ds\" (UniqueName: \"kubernetes.io/projected/742016ca-eb57-474a-b4d9-dbfc682628be-kube-api-access-694ds\") pod \"742016ca-eb57-474a-b4d9-dbfc682628be\" (UID: \"742016ca-eb57-474a-b4d9-dbfc682628be\") " Oct 08 18:20:26 crc kubenswrapper[4859]: I1008 18:20:26.948300 4859 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/742016ca-eb57-474a-b4d9-dbfc682628be-utilities" (OuterVolumeSpecName: "utilities") pod "742016ca-eb57-474a-b4d9-dbfc682628be" (UID: "742016ca-eb57-474a-b4d9-dbfc682628be"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 08 18:20:26 crc kubenswrapper[4859]: I1008 18:20:26.953213 4859 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/742016ca-eb57-474a-b4d9-dbfc682628be-kube-api-access-694ds" (OuterVolumeSpecName: "kube-api-access-694ds") pod "742016ca-eb57-474a-b4d9-dbfc682628be" (UID: "742016ca-eb57-474a-b4d9-dbfc682628be"). InnerVolumeSpecName "kube-api-access-694ds". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 08 18:20:27 crc kubenswrapper[4859]: I1008 18:20:27.014563 4859 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/742016ca-eb57-474a-b4d9-dbfc682628be-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "742016ca-eb57-474a-b4d9-dbfc682628be" (UID: "742016ca-eb57-474a-b4d9-dbfc682628be"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 08 18:20:27 crc kubenswrapper[4859]: I1008 18:20:27.057904 4859 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/742016ca-eb57-474a-b4d9-dbfc682628be-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 08 18:20:27 crc kubenswrapper[4859]: I1008 18:20:27.057985 4859 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/742016ca-eb57-474a-b4d9-dbfc682628be-utilities\") on node \"crc\" DevicePath \"\"" Oct 08 18:20:27 crc kubenswrapper[4859]: I1008 18:20:27.058013 4859 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-694ds\" (UniqueName: \"kubernetes.io/projected/742016ca-eb57-474a-b4d9-dbfc682628be-kube-api-access-694ds\") on node \"crc\" DevicePath \"\"" Oct 08 18:20:27 crc kubenswrapper[4859]: I1008 18:20:27.112463 4859 generic.go:334] "Generic (PLEG): container finished" podID="eb5e42da-6186-4208-8ceb-348959b9e1da" containerID="a8c3827c2b47958d14b87bde8a6b3061aedfb311fd7c96da23198d5b80c3869b" exitCode=0 Oct 08 18:20:27 crc kubenswrapper[4859]: I1008 18:20:27.112553 4859 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-b7vxm" event={"ID":"eb5e42da-6186-4208-8ceb-348959b9e1da","Type":"ContainerDied","Data":"a8c3827c2b47958d14b87bde8a6b3061aedfb311fd7c96da23198d5b80c3869b"} Oct 08 18:20:27 crc kubenswrapper[4859]: I1008 18:20:27.116768 4859 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-pxq2g" event={"ID":"742016ca-eb57-474a-b4d9-dbfc682628be","Type":"ContainerDied","Data":"fcb4c394362aaa21fd76341ed436c077abbad9409f43f8d78851b1dc45652d94"} Oct 08 18:20:27 crc kubenswrapper[4859]: I1008 18:20:27.116890 4859 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-pxq2g" Oct 08 18:20:27 crc kubenswrapper[4859]: I1008 18:20:27.152252 4859 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-pxq2g"] Oct 08 18:20:27 crc kubenswrapper[4859]: I1008 18:20:27.154280 4859 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/community-operators-pxq2g"] Oct 08 18:20:27 crc kubenswrapper[4859]: I1008 18:20:27.208055 4859 scope.go:117] "RemoveContainer" containerID="8720d9a552ab4d000922475090494de59806953a5afadcf848cde2c83bb63de3" Oct 08 18:20:27 crc kubenswrapper[4859]: I1008 18:20:27.276605 4859 scope.go:117] "RemoveContainer" containerID="042432da094dcaa29f61e3777c2ac6bcaea4b5a376e5ca646acc8d6ca678a053" Oct 08 18:20:27 crc kubenswrapper[4859]: I1008 18:20:27.340147 4859 scope.go:117] "RemoveContainer" containerID="1d90316bbffaa725abf95beeb54bcd8eab5e2bc82d8bc3b9c5aa016801b9775e" Oct 08 18:20:27 crc kubenswrapper[4859]: I1008 18:20:27.405872 4859 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-b7vxm" Oct 08 18:20:27 crc kubenswrapper[4859]: I1008 18:20:27.426041 4859 scope.go:117] "RemoveContainer" containerID="afba94ebc1a6f3924cbdb193588fcdeeecbdfca1175af7d119e40a1d2224aa4c" Oct 08 18:20:27 crc kubenswrapper[4859]: I1008 18:20:27.458371 4859 scope.go:117] "RemoveContainer" containerID="9deae894e6ab21ee78584acb5fbccdcc8d79c9b4f9fd58f0624cdec092b55ed7" Oct 08 18:20:27 crc kubenswrapper[4859]: I1008 18:20:27.563669 4859 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/eb5e42da-6186-4208-8ceb-348959b9e1da-utilities\") pod \"eb5e42da-6186-4208-8ceb-348959b9e1da\" (UID: \"eb5e42da-6186-4208-8ceb-348959b9e1da\") " Oct 08 18:20:27 crc kubenswrapper[4859]: I1008 18:20:27.563753 4859 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/eb5e42da-6186-4208-8ceb-348959b9e1da-catalog-content\") pod \"eb5e42da-6186-4208-8ceb-348959b9e1da\" (UID: \"eb5e42da-6186-4208-8ceb-348959b9e1da\") " Oct 08 18:20:27 crc kubenswrapper[4859]: I1008 18:20:27.563873 4859 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-7jm2v\" (UniqueName: \"kubernetes.io/projected/eb5e42da-6186-4208-8ceb-348959b9e1da-kube-api-access-7jm2v\") pod \"eb5e42da-6186-4208-8ceb-348959b9e1da\" (UID: \"eb5e42da-6186-4208-8ceb-348959b9e1da\") " Oct 08 18:20:27 crc kubenswrapper[4859]: I1008 18:20:27.565532 4859 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/eb5e42da-6186-4208-8ceb-348959b9e1da-utilities" (OuterVolumeSpecName: "utilities") pod "eb5e42da-6186-4208-8ceb-348959b9e1da" (UID: "eb5e42da-6186-4208-8ceb-348959b9e1da"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 08 18:20:27 crc kubenswrapper[4859]: I1008 18:20:27.572051 4859 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/eb5e42da-6186-4208-8ceb-348959b9e1da-kube-api-access-7jm2v" (OuterVolumeSpecName: "kube-api-access-7jm2v") pod "eb5e42da-6186-4208-8ceb-348959b9e1da" (UID: "eb5e42da-6186-4208-8ceb-348959b9e1da"). InnerVolumeSpecName "kube-api-access-7jm2v". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 08 18:20:27 crc kubenswrapper[4859]: I1008 18:20:27.585715 4859 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/eb5e42da-6186-4208-8ceb-348959b9e1da-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "eb5e42da-6186-4208-8ceb-348959b9e1da" (UID: "eb5e42da-6186-4208-8ceb-348959b9e1da"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 08 18:20:27 crc kubenswrapper[4859]: I1008 18:20:27.665305 4859 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/eb5e42da-6186-4208-8ceb-348959b9e1da-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 08 18:20:27 crc kubenswrapper[4859]: I1008 18:20:27.665358 4859 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/eb5e42da-6186-4208-8ceb-348959b9e1da-utilities\") on node \"crc\" DevicePath \"\"" Oct 08 18:20:27 crc kubenswrapper[4859]: I1008 18:20:27.665373 4859 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-7jm2v\" (UniqueName: \"kubernetes.io/projected/eb5e42da-6186-4208-8ceb-348959b9e1da-kube-api-access-7jm2v\") on node \"crc\" DevicePath \"\"" Oct 08 18:20:28 crc kubenswrapper[4859]: I1008 18:20:28.125913 4859 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-b7vxm" event={"ID":"eb5e42da-6186-4208-8ceb-348959b9e1da","Type":"ContainerDied","Data":"e509d18e7a8e545af82fd281904c29063c4c0510deca7c581e1d73bd4dcda742"} Oct 08 18:20:28 crc kubenswrapper[4859]: I1008 18:20:28.125991 4859 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-b7vxm" Oct 08 18:20:28 crc kubenswrapper[4859]: I1008 18:20:28.126311 4859 scope.go:117] "RemoveContainer" containerID="a8c3827c2b47958d14b87bde8a6b3061aedfb311fd7c96da23198d5b80c3869b" Oct 08 18:20:28 crc kubenswrapper[4859]: I1008 18:20:28.132501 4859 generic.go:334] "Generic (PLEG): container finished" podID="5b68d592-4a46-4fbc-bcec-a8bca2b973aa" containerID="f63c1b5026a47ea0d969d06d88891d99f91b25c2954de93c4619b4e4dc5fe7ba" exitCode=0 Oct 08 18:20:28 crc kubenswrapper[4859]: I1008 18:20:28.132538 4859 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-zb82c" event={"ID":"5b68d592-4a46-4fbc-bcec-a8bca2b973aa","Type":"ContainerDied","Data":"f63c1b5026a47ea0d969d06d88891d99f91b25c2954de93c4619b4e4dc5fe7ba"} Oct 08 18:20:28 crc kubenswrapper[4859]: I1008 18:20:28.138372 4859 generic.go:334] "Generic (PLEG): container finished" podID="aeadfa85-a482-40a8-b947-09d2f6613900" containerID="2eb9a7a4f0c8c6dcb4c312d1f91dbe329a88ed3bf46e411097c4fe7c2595fa4f" exitCode=0 Oct 08 18:20:28 crc kubenswrapper[4859]: I1008 18:20:28.138520 4859 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-xxgmr" event={"ID":"aeadfa85-a482-40a8-b947-09d2f6613900","Type":"ContainerDied","Data":"2eb9a7a4f0c8c6dcb4c312d1f91dbe329a88ed3bf46e411097c4fe7c2595fa4f"} Oct 08 18:20:28 crc kubenswrapper[4859]: I1008 18:20:28.146906 4859 scope.go:117] "RemoveContainer" containerID="1c11ea9059b1720d52e070b6dfc6db266a57ebb6f9f2094bae86f980f45f6bc5" Oct 08 18:20:28 crc kubenswrapper[4859]: I1008 18:20:28.188793 4859 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-b7vxm"] Oct 08 18:20:28 crc kubenswrapper[4859]: I1008 18:20:28.192048 4859 scope.go:117] "RemoveContainer" containerID="5f21941ab9e1b6d9e394e6c1da4946f357475da962ec69127f65dd8e6c344899" Oct 08 18:20:28 crc kubenswrapper[4859]: I1008 18:20:28.192862 4859 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-marketplace-b7vxm"] Oct 08 18:20:28 crc kubenswrapper[4859]: I1008 18:20:28.478323 4859 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="742016ca-eb57-474a-b4d9-dbfc682628be" path="/var/lib/kubelet/pods/742016ca-eb57-474a-b4d9-dbfc682628be/volumes" Oct 08 18:20:28 crc kubenswrapper[4859]: I1008 18:20:28.479166 4859 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="eb5e42da-6186-4208-8ceb-348959b9e1da" path="/var/lib/kubelet/pods/eb5e42da-6186-4208-8ceb-348959b9e1da/volumes" Oct 08 18:20:28 crc kubenswrapper[4859]: I1008 18:20:28.972206 4859 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-9ml9h"] Oct 08 18:20:28 crc kubenswrapper[4859]: I1008 18:20:28.972758 4859 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-operators-9ml9h" podUID="1b84a3b0-464e-4d4a-b87f-2341db3b73cc" containerName="registry-server" containerID="cri-o://8a47ec34fbda7524552f5a47b071ad6395bb48ac13fe587b4fb3e55f085ec428" gracePeriod=2 Oct 08 18:20:29 crc kubenswrapper[4859]: I1008 18:20:29.150161 4859 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-xxgmr" event={"ID":"aeadfa85-a482-40a8-b947-09d2f6613900","Type":"ContainerStarted","Data":"30d7dafc5dc8f0a00b63102ca616e3537a9ab1f2b622798aead42d5f5fe48381"} Oct 08 18:20:29 crc kubenswrapper[4859]: I1008 18:20:29.152326 4859 generic.go:334] "Generic (PLEG): container finished" podID="1b84a3b0-464e-4d4a-b87f-2341db3b73cc" containerID="8a47ec34fbda7524552f5a47b071ad6395bb48ac13fe587b4fb3e55f085ec428" exitCode=0 Oct 08 18:20:29 crc kubenswrapper[4859]: I1008 18:20:29.152448 4859 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-9ml9h" event={"ID":"1b84a3b0-464e-4d4a-b87f-2341db3b73cc","Type":"ContainerDied","Data":"8a47ec34fbda7524552f5a47b071ad6395bb48ac13fe587b4fb3e55f085ec428"} Oct 08 18:20:29 crc kubenswrapper[4859]: I1008 18:20:29.155391 4859 generic.go:334] "Generic (PLEG): container finished" podID="557e8291-7e96-4634-b744-bcf6981d3a05" containerID="012460253b77cfdcb5001d7386ae3a7b69e158658002e1efdac3240841db6fc0" exitCode=0 Oct 08 18:20:29 crc kubenswrapper[4859]: I1008 18:20:29.155521 4859 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-8z5ch" event={"ID":"557e8291-7e96-4634-b744-bcf6981d3a05","Type":"ContainerDied","Data":"012460253b77cfdcb5001d7386ae3a7b69e158658002e1efdac3240841db6fc0"} Oct 08 18:20:29 crc kubenswrapper[4859]: I1008 18:20:29.167970 4859 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-zb82c" event={"ID":"5b68d592-4a46-4fbc-bcec-a8bca2b973aa","Type":"ContainerStarted","Data":"3b5490e97f3c5e1a24d3ba43457bebc4949c6182b88dfe9fdb226ee620caa614"} Oct 08 18:20:29 crc kubenswrapper[4859]: I1008 18:20:29.175722 4859 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-xxgmr" podStartSLOduration=2.261234628 podStartE2EDuration="46.175679759s" podCreationTimestamp="2025-10-08 18:19:43 +0000 UTC" firstStartedPulling="2025-10-08 18:19:44.698567258 +0000 UTC m=+154.945406637" lastFinishedPulling="2025-10-08 18:20:28.613012369 +0000 UTC m=+198.859851768" observedRunningTime="2025-10-08 18:20:29.173455833 +0000 UTC m=+199.420295222" watchObservedRunningTime="2025-10-08 18:20:29.175679759 +0000 UTC m=+199.422519138" Oct 08 18:20:29 crc kubenswrapper[4859]: I1008 18:20:29.218900 4859 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-zb82c" podStartSLOduration=2.244469866 podStartE2EDuration="49.218874908s" podCreationTimestamp="2025-10-08 18:19:40 +0000 UTC" firstStartedPulling="2025-10-08 18:19:41.590825922 +0000 UTC m=+151.837665311" lastFinishedPulling="2025-10-08 18:20:28.565230964 +0000 UTC m=+198.812070353" observedRunningTime="2025-10-08 18:20:29.200093228 +0000 UTC m=+199.446932607" watchObservedRunningTime="2025-10-08 18:20:29.218874908 +0000 UTC m=+199.465714287" Oct 08 18:20:29 crc kubenswrapper[4859]: I1008 18:20:29.362007 4859 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-9ml9h" Oct 08 18:20:29 crc kubenswrapper[4859]: I1008 18:20:29.510718 4859 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/1b84a3b0-464e-4d4a-b87f-2341db3b73cc-utilities\") pod \"1b84a3b0-464e-4d4a-b87f-2341db3b73cc\" (UID: \"1b84a3b0-464e-4d4a-b87f-2341db3b73cc\") " Oct 08 18:20:29 crc kubenswrapper[4859]: I1008 18:20:29.510795 4859 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-4fbtb\" (UniqueName: \"kubernetes.io/projected/1b84a3b0-464e-4d4a-b87f-2341db3b73cc-kube-api-access-4fbtb\") pod \"1b84a3b0-464e-4d4a-b87f-2341db3b73cc\" (UID: \"1b84a3b0-464e-4d4a-b87f-2341db3b73cc\") " Oct 08 18:20:29 crc kubenswrapper[4859]: I1008 18:20:29.510862 4859 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/1b84a3b0-464e-4d4a-b87f-2341db3b73cc-catalog-content\") pod \"1b84a3b0-464e-4d4a-b87f-2341db3b73cc\" (UID: \"1b84a3b0-464e-4d4a-b87f-2341db3b73cc\") " Oct 08 18:20:29 crc kubenswrapper[4859]: I1008 18:20:29.512133 4859 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/1b84a3b0-464e-4d4a-b87f-2341db3b73cc-utilities" (OuterVolumeSpecName: "utilities") pod "1b84a3b0-464e-4d4a-b87f-2341db3b73cc" (UID: "1b84a3b0-464e-4d4a-b87f-2341db3b73cc"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 08 18:20:29 crc kubenswrapper[4859]: I1008 18:20:29.518457 4859 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/1b84a3b0-464e-4d4a-b87f-2341db3b73cc-kube-api-access-4fbtb" (OuterVolumeSpecName: "kube-api-access-4fbtb") pod "1b84a3b0-464e-4d4a-b87f-2341db3b73cc" (UID: "1b84a3b0-464e-4d4a-b87f-2341db3b73cc"). InnerVolumeSpecName "kube-api-access-4fbtb". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 08 18:20:29 crc kubenswrapper[4859]: I1008 18:20:29.592125 4859 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/1b84a3b0-464e-4d4a-b87f-2341db3b73cc-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "1b84a3b0-464e-4d4a-b87f-2341db3b73cc" (UID: "1b84a3b0-464e-4d4a-b87f-2341db3b73cc"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 08 18:20:29 crc kubenswrapper[4859]: I1008 18:20:29.612711 4859 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-4fbtb\" (UniqueName: \"kubernetes.io/projected/1b84a3b0-464e-4d4a-b87f-2341db3b73cc-kube-api-access-4fbtb\") on node \"crc\" DevicePath \"\"" Oct 08 18:20:29 crc kubenswrapper[4859]: I1008 18:20:29.612756 4859 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/1b84a3b0-464e-4d4a-b87f-2341db3b73cc-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 08 18:20:29 crc kubenswrapper[4859]: I1008 18:20:29.612766 4859 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/1b84a3b0-464e-4d4a-b87f-2341db3b73cc-utilities\") on node \"crc\" DevicePath \"\"" Oct 08 18:20:30 crc kubenswrapper[4859]: I1008 18:20:30.178215 4859 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-8z5ch" event={"ID":"557e8291-7e96-4634-b744-bcf6981d3a05","Type":"ContainerStarted","Data":"4444d93f68c351c0aae5349182126b3615263046ce9bb221bee588119677bf6a"} Oct 08 18:20:30 crc kubenswrapper[4859]: I1008 18:20:30.180421 4859 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-9ml9h" event={"ID":"1b84a3b0-464e-4d4a-b87f-2341db3b73cc","Type":"ContainerDied","Data":"934a532c858325e3db47a11b5ffd60c54e62c546dd66e7c6c4157bd333c86765"} Oct 08 18:20:30 crc kubenswrapper[4859]: I1008 18:20:30.180472 4859 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-9ml9h" Oct 08 18:20:30 crc kubenswrapper[4859]: I1008 18:20:30.180476 4859 scope.go:117] "RemoveContainer" containerID="8a47ec34fbda7524552f5a47b071ad6395bb48ac13fe587b4fb3e55f085ec428" Oct 08 18:20:30 crc kubenswrapper[4859]: I1008 18:20:30.202227 4859 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-8z5ch" podStartSLOduration=2.047447679 podStartE2EDuration="48.202204905s" podCreationTimestamp="2025-10-08 18:19:42 +0000 UTC" firstStartedPulling="2025-10-08 18:19:43.656480542 +0000 UTC m=+153.903319921" lastFinishedPulling="2025-10-08 18:20:29.811237768 +0000 UTC m=+200.058077147" observedRunningTime="2025-10-08 18:20:30.198816304 +0000 UTC m=+200.445655703" watchObservedRunningTime="2025-10-08 18:20:30.202204905 +0000 UTC m=+200.449044284" Oct 08 18:20:30 crc kubenswrapper[4859]: I1008 18:20:30.203003 4859 scope.go:117] "RemoveContainer" containerID="ce4a61e3e561228ff7b8c66698b5e93c28edd16df3b59bfc8ccb5c99c0811594" Oct 08 18:20:30 crc kubenswrapper[4859]: I1008 18:20:30.212973 4859 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-9ml9h"] Oct 08 18:20:30 crc kubenswrapper[4859]: I1008 18:20:30.218353 4859 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-operators-9ml9h"] Oct 08 18:20:30 crc kubenswrapper[4859]: I1008 18:20:30.227870 4859 scope.go:117] "RemoveContainer" containerID="ed9287be68925e52499da1bf0708e71fae2564190b4ed03c2c031a97428b5434" Oct 08 18:20:30 crc kubenswrapper[4859]: I1008 18:20:30.478136 4859 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="1b84a3b0-464e-4d4a-b87f-2341db3b73cc" path="/var/lib/kubelet/pods/1b84a3b0-464e-4d4a-b87f-2341db3b73cc/volumes" Oct 08 18:20:30 crc kubenswrapper[4859]: I1008 18:20:30.498118 4859 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-zb82c" Oct 08 18:20:30 crc kubenswrapper[4859]: I1008 18:20:30.498182 4859 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-zb82c" Oct 08 18:20:31 crc kubenswrapper[4859]: I1008 18:20:31.537848 4859 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/community-operators-zb82c" podUID="5b68d592-4a46-4fbc-bcec-a8bca2b973aa" containerName="registry-server" probeResult="failure" output=< Oct 08 18:20:31 crc kubenswrapper[4859]: timeout: failed to connect service ":50051" within 1s Oct 08 18:20:31 crc kubenswrapper[4859]: > Oct 08 18:20:32 crc kubenswrapper[4859]: I1008 18:20:32.693735 4859 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-8z5ch" Oct 08 18:20:32 crc kubenswrapper[4859]: I1008 18:20:32.696159 4859 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-8z5ch" Oct 08 18:20:32 crc kubenswrapper[4859]: I1008 18:20:32.743633 4859 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-8z5ch" Oct 08 18:20:33 crc kubenswrapper[4859]: I1008 18:20:33.719309 4859 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-xxgmr" Oct 08 18:20:33 crc kubenswrapper[4859]: I1008 18:20:33.719366 4859 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-xxgmr" Oct 08 18:20:33 crc kubenswrapper[4859]: I1008 18:20:33.782907 4859 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-xxgmr" Oct 08 18:20:34 crc kubenswrapper[4859]: I1008 18:20:34.279490 4859 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-xxgmr" Oct 08 18:20:40 crc kubenswrapper[4859]: I1008 18:20:40.534698 4859 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-zb82c" Oct 08 18:20:40 crc kubenswrapper[4859]: I1008 18:20:40.581845 4859 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-zb82c" Oct 08 18:20:42 crc kubenswrapper[4859]: I1008 18:20:42.742580 4859 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-8z5ch" Oct 08 18:20:47 crc kubenswrapper[4859]: I1008 18:20:47.925217 4859 patch_prober.go:28] interesting pod/machine-config-daemon-82s52 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 08 18:20:47 crc kubenswrapper[4859]: I1008 18:20:47.925835 4859 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-82s52" podUID="b23a6a6c-9d92-4e7b-840e-55cfda873a2d" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 08 18:20:47 crc kubenswrapper[4859]: I1008 18:20:47.925915 4859 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-82s52" Oct 08 18:20:47 crc kubenswrapper[4859]: I1008 18:20:47.926920 4859 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"c23ea3580998adf90621a0155c20ee4bd590c512629812174e95c28857bcfef9"} pod="openshift-machine-config-operator/machine-config-daemon-82s52" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Oct 08 18:20:47 crc kubenswrapper[4859]: I1008 18:20:47.927044 4859 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-82s52" podUID="b23a6a6c-9d92-4e7b-840e-55cfda873a2d" containerName="machine-config-daemon" containerID="cri-o://c23ea3580998adf90621a0155c20ee4bd590c512629812174e95c28857bcfef9" gracePeriod=600 Oct 08 18:20:48 crc kubenswrapper[4859]: I1008 18:20:48.294162 4859 generic.go:334] "Generic (PLEG): container finished" podID="b23a6a6c-9d92-4e7b-840e-55cfda873a2d" containerID="c23ea3580998adf90621a0155c20ee4bd590c512629812174e95c28857bcfef9" exitCode=0 Oct 08 18:20:48 crc kubenswrapper[4859]: I1008 18:20:48.294254 4859 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-82s52" event={"ID":"b23a6a6c-9d92-4e7b-840e-55cfda873a2d","Type":"ContainerDied","Data":"c23ea3580998adf90621a0155c20ee4bd590c512629812174e95c28857bcfef9"} Oct 08 18:20:48 crc kubenswrapper[4859]: I1008 18:20:48.294780 4859 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-82s52" event={"ID":"b23a6a6c-9d92-4e7b-840e-55cfda873a2d","Type":"ContainerStarted","Data":"c6b208adefaf5eb65ea467fb4313c0f4281605067819f9fbd7f4404606fa52d0"} Oct 08 18:20:51 crc kubenswrapper[4859]: I1008 18:20:51.530141 4859 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-authentication/oauth-openshift-558db77b4-j8gh2"] Oct 08 18:21:16 crc kubenswrapper[4859]: I1008 18:21:16.563283 4859 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-authentication/oauth-openshift-558db77b4-j8gh2" podUID="0902ccb2-78ab-48d0-be02-4c42f03e12fb" containerName="oauth-openshift" containerID="cri-o://7023b5c72b539780ca0fa2a5d77a71496bfa187a6666d8c3d8fd2b6e898c0c70" gracePeriod=15 Oct 08 18:21:16 crc kubenswrapper[4859]: I1008 18:21:16.953573 4859 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-authentication/oauth-openshift-558db77b4-j8gh2" Oct 08 18:21:16 crc kubenswrapper[4859]: I1008 18:21:16.994049 4859 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-authentication/oauth-openshift-cc6d8cb6b-79m29"] Oct 08 18:21:16 crc kubenswrapper[4859]: E1008 18:21:16.994376 4859 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1b84a3b0-464e-4d4a-b87f-2341db3b73cc" containerName="extract-utilities" Oct 08 18:21:16 crc kubenswrapper[4859]: I1008 18:21:16.994394 4859 state_mem.go:107] "Deleted CPUSet assignment" podUID="1b84a3b0-464e-4d4a-b87f-2341db3b73cc" containerName="extract-utilities" Oct 08 18:21:16 crc kubenswrapper[4859]: E1008 18:21:16.994406 4859 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ae751baf-6f53-4a05-b28d-0888f7682db7" containerName="registry-server" Oct 08 18:21:16 crc kubenswrapper[4859]: I1008 18:21:16.994414 4859 state_mem.go:107] "Deleted CPUSet assignment" podUID="ae751baf-6f53-4a05-b28d-0888f7682db7" containerName="registry-server" Oct 08 18:21:16 crc kubenswrapper[4859]: E1008 18:21:16.994424 4859 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1b84a3b0-464e-4d4a-b87f-2341db3b73cc" containerName="registry-server" Oct 08 18:21:16 crc kubenswrapper[4859]: I1008 18:21:16.994433 4859 state_mem.go:107] "Deleted CPUSet assignment" podUID="1b84a3b0-464e-4d4a-b87f-2341db3b73cc" containerName="registry-server" Oct 08 18:21:16 crc kubenswrapper[4859]: E1008 18:21:16.994444 4859 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0902ccb2-78ab-48d0-be02-4c42f03e12fb" containerName="oauth-openshift" Oct 08 18:21:16 crc kubenswrapper[4859]: I1008 18:21:16.994452 4859 state_mem.go:107] "Deleted CPUSet assignment" podUID="0902ccb2-78ab-48d0-be02-4c42f03e12fb" containerName="oauth-openshift" Oct 08 18:21:16 crc kubenswrapper[4859]: E1008 18:21:16.994462 4859 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ae751baf-6f53-4a05-b28d-0888f7682db7" containerName="extract-utilities" Oct 08 18:21:16 crc kubenswrapper[4859]: I1008 18:21:16.994470 4859 state_mem.go:107] "Deleted CPUSet assignment" podUID="ae751baf-6f53-4a05-b28d-0888f7682db7" containerName="extract-utilities" Oct 08 18:21:16 crc kubenswrapper[4859]: E1008 18:21:16.994482 4859 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="eb5e42da-6186-4208-8ceb-348959b9e1da" containerName="extract-utilities" Oct 08 18:21:16 crc kubenswrapper[4859]: I1008 18:21:16.994490 4859 state_mem.go:107] "Deleted CPUSet assignment" podUID="eb5e42da-6186-4208-8ceb-348959b9e1da" containerName="extract-utilities" Oct 08 18:21:16 crc kubenswrapper[4859]: E1008 18:21:16.994505 4859 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="742016ca-eb57-474a-b4d9-dbfc682628be" containerName="registry-server" Oct 08 18:21:16 crc kubenswrapper[4859]: I1008 18:21:16.994513 4859 state_mem.go:107] "Deleted CPUSet assignment" podUID="742016ca-eb57-474a-b4d9-dbfc682628be" containerName="registry-server" Oct 08 18:21:16 crc kubenswrapper[4859]: E1008 18:21:16.994526 4859 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="eb5e42da-6186-4208-8ceb-348959b9e1da" containerName="registry-server" Oct 08 18:21:16 crc kubenswrapper[4859]: I1008 18:21:16.994535 4859 state_mem.go:107] "Deleted CPUSet assignment" podUID="eb5e42da-6186-4208-8ceb-348959b9e1da" containerName="registry-server" Oct 08 18:21:16 crc kubenswrapper[4859]: E1008 18:21:16.994548 4859 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c488e0ea-436c-4766-a311-56b54bf0fb10" containerName="pruner" Oct 08 18:21:16 crc kubenswrapper[4859]: I1008 18:21:16.994556 4859 state_mem.go:107] "Deleted CPUSet assignment" podUID="c488e0ea-436c-4766-a311-56b54bf0fb10" containerName="pruner" Oct 08 18:21:16 crc kubenswrapper[4859]: E1008 18:21:16.994564 4859 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="742016ca-eb57-474a-b4d9-dbfc682628be" containerName="extract-content" Oct 08 18:21:16 crc kubenswrapper[4859]: I1008 18:21:16.994573 4859 state_mem.go:107] "Deleted CPUSet assignment" podUID="742016ca-eb57-474a-b4d9-dbfc682628be" containerName="extract-content" Oct 08 18:21:16 crc kubenswrapper[4859]: E1008 18:21:16.994592 4859 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="742016ca-eb57-474a-b4d9-dbfc682628be" containerName="extract-utilities" Oct 08 18:21:16 crc kubenswrapper[4859]: I1008 18:21:16.994601 4859 state_mem.go:107] "Deleted CPUSet assignment" podUID="742016ca-eb57-474a-b4d9-dbfc682628be" containerName="extract-utilities" Oct 08 18:21:16 crc kubenswrapper[4859]: E1008 18:21:16.994613 4859 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="eb5e42da-6186-4208-8ceb-348959b9e1da" containerName="extract-content" Oct 08 18:21:16 crc kubenswrapper[4859]: I1008 18:21:16.994622 4859 state_mem.go:107] "Deleted CPUSet assignment" podUID="eb5e42da-6186-4208-8ceb-348959b9e1da" containerName="extract-content" Oct 08 18:21:16 crc kubenswrapper[4859]: E1008 18:21:16.994634 4859 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ae751baf-6f53-4a05-b28d-0888f7682db7" containerName="extract-content" Oct 08 18:21:16 crc kubenswrapper[4859]: I1008 18:21:16.994642 4859 state_mem.go:107] "Deleted CPUSet assignment" podUID="ae751baf-6f53-4a05-b28d-0888f7682db7" containerName="extract-content" Oct 08 18:21:16 crc kubenswrapper[4859]: E1008 18:21:16.994652 4859 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1b84a3b0-464e-4d4a-b87f-2341db3b73cc" containerName="extract-content" Oct 08 18:21:16 crc kubenswrapper[4859]: I1008 18:21:16.994659 4859 state_mem.go:107] "Deleted CPUSet assignment" podUID="1b84a3b0-464e-4d4a-b87f-2341db3b73cc" containerName="extract-content" Oct 08 18:21:16 crc kubenswrapper[4859]: I1008 18:21:16.994806 4859 memory_manager.go:354] "RemoveStaleState removing state" podUID="c488e0ea-436c-4766-a311-56b54bf0fb10" containerName="pruner" Oct 08 18:21:16 crc kubenswrapper[4859]: I1008 18:21:16.994827 4859 memory_manager.go:354] "RemoveStaleState removing state" podUID="eb5e42da-6186-4208-8ceb-348959b9e1da" containerName="registry-server" Oct 08 18:21:16 crc kubenswrapper[4859]: I1008 18:21:16.994841 4859 memory_manager.go:354] "RemoveStaleState removing state" podUID="0902ccb2-78ab-48d0-be02-4c42f03e12fb" containerName="oauth-openshift" Oct 08 18:21:16 crc kubenswrapper[4859]: I1008 18:21:16.994850 4859 memory_manager.go:354] "RemoveStaleState removing state" podUID="742016ca-eb57-474a-b4d9-dbfc682628be" containerName="registry-server" Oct 08 18:21:16 crc kubenswrapper[4859]: I1008 18:21:16.994858 4859 memory_manager.go:354] "RemoveStaleState removing state" podUID="ae751baf-6f53-4a05-b28d-0888f7682db7" containerName="registry-server" Oct 08 18:21:16 crc kubenswrapper[4859]: I1008 18:21:16.994869 4859 memory_manager.go:354] "RemoveStaleState removing state" podUID="1b84a3b0-464e-4d4a-b87f-2341db3b73cc" containerName="registry-server" Oct 08 18:21:16 crc kubenswrapper[4859]: I1008 18:21:16.995347 4859 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-authentication/oauth-openshift-cc6d8cb6b-79m29" Oct 08 18:21:17 crc kubenswrapper[4859]: I1008 18:21:17.008123 4859 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-authentication/oauth-openshift-cc6d8cb6b-79m29"] Oct 08 18:21:17 crc kubenswrapper[4859]: I1008 18:21:17.095708 4859 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/0902ccb2-78ab-48d0-be02-4c42f03e12fb-v4-0-config-system-cliconfig\") pod \"0902ccb2-78ab-48d0-be02-4c42f03e12fb\" (UID: \"0902ccb2-78ab-48d0-be02-4c42f03e12fb\") " Oct 08 18:21:17 crc kubenswrapper[4859]: I1008 18:21:17.095866 4859 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/0902ccb2-78ab-48d0-be02-4c42f03e12fb-v4-0-config-system-serving-cert\") pod \"0902ccb2-78ab-48d0-be02-4c42f03e12fb\" (UID: \"0902ccb2-78ab-48d0-be02-4c42f03e12fb\") " Oct 08 18:21:17 crc kubenswrapper[4859]: I1008 18:21:17.095902 4859 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/0902ccb2-78ab-48d0-be02-4c42f03e12fb-audit-dir\") pod \"0902ccb2-78ab-48d0-be02-4c42f03e12fb\" (UID: \"0902ccb2-78ab-48d0-be02-4c42f03e12fb\") " Oct 08 18:21:17 crc kubenswrapper[4859]: I1008 18:21:17.095935 4859 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/0902ccb2-78ab-48d0-be02-4c42f03e12fb-v4-0-config-system-session\") pod \"0902ccb2-78ab-48d0-be02-4c42f03e12fb\" (UID: \"0902ccb2-78ab-48d0-be02-4c42f03e12fb\") " Oct 08 18:21:17 crc kubenswrapper[4859]: I1008 18:21:17.095978 4859 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-npz7k\" (UniqueName: \"kubernetes.io/projected/0902ccb2-78ab-48d0-be02-4c42f03e12fb-kube-api-access-npz7k\") pod \"0902ccb2-78ab-48d0-be02-4c42f03e12fb\" (UID: \"0902ccb2-78ab-48d0-be02-4c42f03e12fb\") " Oct 08 18:21:17 crc kubenswrapper[4859]: I1008 18:21:17.096014 4859 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/0902ccb2-78ab-48d0-be02-4c42f03e12fb-v4-0-config-user-template-login\") pod \"0902ccb2-78ab-48d0-be02-4c42f03e12fb\" (UID: \"0902ccb2-78ab-48d0-be02-4c42f03e12fb\") " Oct 08 18:21:17 crc kubenswrapper[4859]: I1008 18:21:17.096039 4859 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/0902ccb2-78ab-48d0-be02-4c42f03e12fb-v4-0-config-system-ocp-branding-template\") pod \"0902ccb2-78ab-48d0-be02-4c42f03e12fb\" (UID: \"0902ccb2-78ab-48d0-be02-4c42f03e12fb\") " Oct 08 18:21:17 crc kubenswrapper[4859]: I1008 18:21:17.096042 4859 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/0902ccb2-78ab-48d0-be02-4c42f03e12fb-audit-dir" (OuterVolumeSpecName: "audit-dir") pod "0902ccb2-78ab-48d0-be02-4c42f03e12fb" (UID: "0902ccb2-78ab-48d0-be02-4c42f03e12fb"). InnerVolumeSpecName "audit-dir". PluginName "kubernetes.io/host-path", VolumeGidValue "" Oct 08 18:21:17 crc kubenswrapper[4859]: I1008 18:21:17.096113 4859 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/0902ccb2-78ab-48d0-be02-4c42f03e12fb-v4-0-config-system-service-ca\") pod \"0902ccb2-78ab-48d0-be02-4c42f03e12fb\" (UID: \"0902ccb2-78ab-48d0-be02-4c42f03e12fb\") " Oct 08 18:21:17 crc kubenswrapper[4859]: I1008 18:21:17.096165 4859 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/0902ccb2-78ab-48d0-be02-4c42f03e12fb-v4-0-config-user-template-error\") pod \"0902ccb2-78ab-48d0-be02-4c42f03e12fb\" (UID: \"0902ccb2-78ab-48d0-be02-4c42f03e12fb\") " Oct 08 18:21:17 crc kubenswrapper[4859]: I1008 18:21:17.096207 4859 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/0902ccb2-78ab-48d0-be02-4c42f03e12fb-v4-0-config-system-router-certs\") pod \"0902ccb2-78ab-48d0-be02-4c42f03e12fb\" (UID: \"0902ccb2-78ab-48d0-be02-4c42f03e12fb\") " Oct 08 18:21:17 crc kubenswrapper[4859]: I1008 18:21:17.096233 4859 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/0902ccb2-78ab-48d0-be02-4c42f03e12fb-v4-0-config-user-template-provider-selection\") pod \"0902ccb2-78ab-48d0-be02-4c42f03e12fb\" (UID: \"0902ccb2-78ab-48d0-be02-4c42f03e12fb\") " Oct 08 18:21:17 crc kubenswrapper[4859]: I1008 18:21:17.096284 4859 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/0902ccb2-78ab-48d0-be02-4c42f03e12fb-v4-0-config-system-trusted-ca-bundle\") pod \"0902ccb2-78ab-48d0-be02-4c42f03e12fb\" (UID: \"0902ccb2-78ab-48d0-be02-4c42f03e12fb\") " Oct 08 18:21:17 crc kubenswrapper[4859]: I1008 18:21:17.096329 4859 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/0902ccb2-78ab-48d0-be02-4c42f03e12fb-audit-policies\") pod \"0902ccb2-78ab-48d0-be02-4c42f03e12fb\" (UID: \"0902ccb2-78ab-48d0-be02-4c42f03e12fb\") " Oct 08 18:21:17 crc kubenswrapper[4859]: I1008 18:21:17.096363 4859 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/0902ccb2-78ab-48d0-be02-4c42f03e12fb-v4-0-config-user-idp-0-file-data\") pod \"0902ccb2-78ab-48d0-be02-4c42f03e12fb\" (UID: \"0902ccb2-78ab-48d0-be02-4c42f03e12fb\") " Oct 08 18:21:17 crc kubenswrapper[4859]: I1008 18:21:17.096567 4859 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/0902ccb2-78ab-48d0-be02-4c42f03e12fb-v4-0-config-system-cliconfig" (OuterVolumeSpecName: "v4-0-config-system-cliconfig") pod "0902ccb2-78ab-48d0-be02-4c42f03e12fb" (UID: "0902ccb2-78ab-48d0-be02-4c42f03e12fb"). InnerVolumeSpecName "v4-0-config-system-cliconfig". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 08 18:21:17 crc kubenswrapper[4859]: I1008 18:21:17.096591 4859 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/c4a648c4-7e6e-49c0-a5fd-dc9323e9428c-v4-0-config-user-template-login\") pod \"oauth-openshift-cc6d8cb6b-79m29\" (UID: \"c4a648c4-7e6e-49c0-a5fd-dc9323e9428c\") " pod="openshift-authentication/oauth-openshift-cc6d8cb6b-79m29" Oct 08 18:21:17 crc kubenswrapper[4859]: I1008 18:21:17.096642 4859 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/c4a648c4-7e6e-49c0-a5fd-dc9323e9428c-v4-0-config-system-router-certs\") pod \"oauth-openshift-cc6d8cb6b-79m29\" (UID: \"c4a648c4-7e6e-49c0-a5fd-dc9323e9428c\") " pod="openshift-authentication/oauth-openshift-cc6d8cb6b-79m29" Oct 08 18:21:17 crc kubenswrapper[4859]: I1008 18:21:17.096744 4859 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/c4a648c4-7e6e-49c0-a5fd-dc9323e9428c-v4-0-config-user-template-provider-selection\") pod \"oauth-openshift-cc6d8cb6b-79m29\" (UID: \"c4a648c4-7e6e-49c0-a5fd-dc9323e9428c\") " pod="openshift-authentication/oauth-openshift-cc6d8cb6b-79m29" Oct 08 18:21:17 crc kubenswrapper[4859]: I1008 18:21:17.096773 4859 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/c4a648c4-7e6e-49c0-a5fd-dc9323e9428c-v4-0-config-user-template-error\") pod \"oauth-openshift-cc6d8cb6b-79m29\" (UID: \"c4a648c4-7e6e-49c0-a5fd-dc9323e9428c\") " pod="openshift-authentication/oauth-openshift-cc6d8cb6b-79m29" Oct 08 18:21:17 crc kubenswrapper[4859]: I1008 18:21:17.096821 4859 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/c4a648c4-7e6e-49c0-a5fd-dc9323e9428c-audit-policies\") pod \"oauth-openshift-cc6d8cb6b-79m29\" (UID: \"c4a648c4-7e6e-49c0-a5fd-dc9323e9428c\") " pod="openshift-authentication/oauth-openshift-cc6d8cb6b-79m29" Oct 08 18:21:17 crc kubenswrapper[4859]: I1008 18:21:17.096851 4859 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/c4a648c4-7e6e-49c0-a5fd-dc9323e9428c-v4-0-config-system-session\") pod \"oauth-openshift-cc6d8cb6b-79m29\" (UID: \"c4a648c4-7e6e-49c0-a5fd-dc9323e9428c\") " pod="openshift-authentication/oauth-openshift-cc6d8cb6b-79m29" Oct 08 18:21:17 crc kubenswrapper[4859]: I1008 18:21:17.096896 4859 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/c4a648c4-7e6e-49c0-a5fd-dc9323e9428c-v4-0-config-system-cliconfig\") pod \"oauth-openshift-cc6d8cb6b-79m29\" (UID: \"c4a648c4-7e6e-49c0-a5fd-dc9323e9428c\") " pod="openshift-authentication/oauth-openshift-cc6d8cb6b-79m29" Oct 08 18:21:17 crc kubenswrapper[4859]: I1008 18:21:17.096945 4859 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/c4a648c4-7e6e-49c0-a5fd-dc9323e9428c-v4-0-config-system-service-ca\") pod \"oauth-openshift-cc6d8cb6b-79m29\" (UID: \"c4a648c4-7e6e-49c0-a5fd-dc9323e9428c\") " pod="openshift-authentication/oauth-openshift-cc6d8cb6b-79m29" Oct 08 18:21:17 crc kubenswrapper[4859]: I1008 18:21:17.096967 4859 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/c4a648c4-7e6e-49c0-a5fd-dc9323e9428c-v4-0-config-system-trusted-ca-bundle\") pod \"oauth-openshift-cc6d8cb6b-79m29\" (UID: \"c4a648c4-7e6e-49c0-a5fd-dc9323e9428c\") " pod="openshift-authentication/oauth-openshift-cc6d8cb6b-79m29" Oct 08 18:21:17 crc kubenswrapper[4859]: I1008 18:21:17.096990 4859 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/c4a648c4-7e6e-49c0-a5fd-dc9323e9428c-audit-dir\") pod \"oauth-openshift-cc6d8cb6b-79m29\" (UID: \"c4a648c4-7e6e-49c0-a5fd-dc9323e9428c\") " pod="openshift-authentication/oauth-openshift-cc6d8cb6b-79m29" Oct 08 18:21:17 crc kubenswrapper[4859]: I1008 18:21:17.097015 4859 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/c4a648c4-7e6e-49c0-a5fd-dc9323e9428c-v4-0-config-system-serving-cert\") pod \"oauth-openshift-cc6d8cb6b-79m29\" (UID: \"c4a648c4-7e6e-49c0-a5fd-dc9323e9428c\") " pod="openshift-authentication/oauth-openshift-cc6d8cb6b-79m29" Oct 08 18:21:17 crc kubenswrapper[4859]: I1008 18:21:17.097052 4859 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/c4a648c4-7e6e-49c0-a5fd-dc9323e9428c-v4-0-config-user-idp-0-file-data\") pod \"oauth-openshift-cc6d8cb6b-79m29\" (UID: \"c4a648c4-7e6e-49c0-a5fd-dc9323e9428c\") " pod="openshift-authentication/oauth-openshift-cc6d8cb6b-79m29" Oct 08 18:21:17 crc kubenswrapper[4859]: I1008 18:21:17.097091 4859 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-t5b7f\" (UniqueName: \"kubernetes.io/projected/c4a648c4-7e6e-49c0-a5fd-dc9323e9428c-kube-api-access-t5b7f\") pod \"oauth-openshift-cc6d8cb6b-79m29\" (UID: \"c4a648c4-7e6e-49c0-a5fd-dc9323e9428c\") " pod="openshift-authentication/oauth-openshift-cc6d8cb6b-79m29" Oct 08 18:21:17 crc kubenswrapper[4859]: I1008 18:21:17.097137 4859 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/c4a648c4-7e6e-49c0-a5fd-dc9323e9428c-v4-0-config-system-ocp-branding-template\") pod \"oauth-openshift-cc6d8cb6b-79m29\" (UID: \"c4a648c4-7e6e-49c0-a5fd-dc9323e9428c\") " pod="openshift-authentication/oauth-openshift-cc6d8cb6b-79m29" Oct 08 18:21:17 crc kubenswrapper[4859]: I1008 18:21:17.097190 4859 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/0902ccb2-78ab-48d0-be02-4c42f03e12fb-v4-0-config-system-cliconfig\") on node \"crc\" DevicePath \"\"" Oct 08 18:21:17 crc kubenswrapper[4859]: I1008 18:21:17.097205 4859 reconciler_common.go:293] "Volume detached for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/0902ccb2-78ab-48d0-be02-4c42f03e12fb-audit-dir\") on node \"crc\" DevicePath \"\"" Oct 08 18:21:17 crc kubenswrapper[4859]: I1008 18:21:17.098268 4859 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/0902ccb2-78ab-48d0-be02-4c42f03e12fb-v4-0-config-system-service-ca" (OuterVolumeSpecName: "v4-0-config-system-service-ca") pod "0902ccb2-78ab-48d0-be02-4c42f03e12fb" (UID: "0902ccb2-78ab-48d0-be02-4c42f03e12fb"). InnerVolumeSpecName "v4-0-config-system-service-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 08 18:21:17 crc kubenswrapper[4859]: I1008 18:21:17.099231 4859 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/0902ccb2-78ab-48d0-be02-4c42f03e12fb-audit-policies" (OuterVolumeSpecName: "audit-policies") pod "0902ccb2-78ab-48d0-be02-4c42f03e12fb" (UID: "0902ccb2-78ab-48d0-be02-4c42f03e12fb"). InnerVolumeSpecName "audit-policies". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 08 18:21:17 crc kubenswrapper[4859]: I1008 18:21:17.099656 4859 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/0902ccb2-78ab-48d0-be02-4c42f03e12fb-v4-0-config-system-trusted-ca-bundle" (OuterVolumeSpecName: "v4-0-config-system-trusted-ca-bundle") pod "0902ccb2-78ab-48d0-be02-4c42f03e12fb" (UID: "0902ccb2-78ab-48d0-be02-4c42f03e12fb"). InnerVolumeSpecName "v4-0-config-system-trusted-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 08 18:21:17 crc kubenswrapper[4859]: I1008 18:21:17.103766 4859 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/0902ccb2-78ab-48d0-be02-4c42f03e12fb-kube-api-access-npz7k" (OuterVolumeSpecName: "kube-api-access-npz7k") pod "0902ccb2-78ab-48d0-be02-4c42f03e12fb" (UID: "0902ccb2-78ab-48d0-be02-4c42f03e12fb"). InnerVolumeSpecName "kube-api-access-npz7k". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 08 18:21:17 crc kubenswrapper[4859]: I1008 18:21:17.104451 4859 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/0902ccb2-78ab-48d0-be02-4c42f03e12fb-v4-0-config-user-template-error" (OuterVolumeSpecName: "v4-0-config-user-template-error") pod "0902ccb2-78ab-48d0-be02-4c42f03e12fb" (UID: "0902ccb2-78ab-48d0-be02-4c42f03e12fb"). InnerVolumeSpecName "v4-0-config-user-template-error". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 08 18:21:17 crc kubenswrapper[4859]: I1008 18:21:17.104633 4859 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/0902ccb2-78ab-48d0-be02-4c42f03e12fb-v4-0-config-system-ocp-branding-template" (OuterVolumeSpecName: "v4-0-config-system-ocp-branding-template") pod "0902ccb2-78ab-48d0-be02-4c42f03e12fb" (UID: "0902ccb2-78ab-48d0-be02-4c42f03e12fb"). InnerVolumeSpecName "v4-0-config-system-ocp-branding-template". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 08 18:21:17 crc kubenswrapper[4859]: I1008 18:21:17.105503 4859 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/0902ccb2-78ab-48d0-be02-4c42f03e12fb-v4-0-config-user-template-login" (OuterVolumeSpecName: "v4-0-config-user-template-login") pod "0902ccb2-78ab-48d0-be02-4c42f03e12fb" (UID: "0902ccb2-78ab-48d0-be02-4c42f03e12fb"). InnerVolumeSpecName "v4-0-config-user-template-login". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 08 18:21:17 crc kubenswrapper[4859]: I1008 18:21:17.105845 4859 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/0902ccb2-78ab-48d0-be02-4c42f03e12fb-v4-0-config-system-router-certs" (OuterVolumeSpecName: "v4-0-config-system-router-certs") pod "0902ccb2-78ab-48d0-be02-4c42f03e12fb" (UID: "0902ccb2-78ab-48d0-be02-4c42f03e12fb"). InnerVolumeSpecName "v4-0-config-system-router-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 08 18:21:17 crc kubenswrapper[4859]: I1008 18:21:17.106008 4859 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/0902ccb2-78ab-48d0-be02-4c42f03e12fb-v4-0-config-system-session" (OuterVolumeSpecName: "v4-0-config-system-session") pod "0902ccb2-78ab-48d0-be02-4c42f03e12fb" (UID: "0902ccb2-78ab-48d0-be02-4c42f03e12fb"). InnerVolumeSpecName "v4-0-config-system-session". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 08 18:21:17 crc kubenswrapper[4859]: I1008 18:21:17.106228 4859 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/0902ccb2-78ab-48d0-be02-4c42f03e12fb-v4-0-config-user-idp-0-file-data" (OuterVolumeSpecName: "v4-0-config-user-idp-0-file-data") pod "0902ccb2-78ab-48d0-be02-4c42f03e12fb" (UID: "0902ccb2-78ab-48d0-be02-4c42f03e12fb"). InnerVolumeSpecName "v4-0-config-user-idp-0-file-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 08 18:21:17 crc kubenswrapper[4859]: I1008 18:21:17.106736 4859 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/0902ccb2-78ab-48d0-be02-4c42f03e12fb-v4-0-config-system-serving-cert" (OuterVolumeSpecName: "v4-0-config-system-serving-cert") pod "0902ccb2-78ab-48d0-be02-4c42f03e12fb" (UID: "0902ccb2-78ab-48d0-be02-4c42f03e12fb"). InnerVolumeSpecName "v4-0-config-system-serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 08 18:21:17 crc kubenswrapper[4859]: I1008 18:21:17.106818 4859 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/0902ccb2-78ab-48d0-be02-4c42f03e12fb-v4-0-config-user-template-provider-selection" (OuterVolumeSpecName: "v4-0-config-user-template-provider-selection") pod "0902ccb2-78ab-48d0-be02-4c42f03e12fb" (UID: "0902ccb2-78ab-48d0-be02-4c42f03e12fb"). InnerVolumeSpecName "v4-0-config-user-template-provider-selection". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 08 18:21:17 crc kubenswrapper[4859]: I1008 18:21:17.200029 4859 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/c4a648c4-7e6e-49c0-a5fd-dc9323e9428c-v4-0-config-system-cliconfig\") pod \"oauth-openshift-cc6d8cb6b-79m29\" (UID: \"c4a648c4-7e6e-49c0-a5fd-dc9323e9428c\") " pod="openshift-authentication/oauth-openshift-cc6d8cb6b-79m29" Oct 08 18:21:17 crc kubenswrapper[4859]: I1008 18:21:17.200088 4859 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/c4a648c4-7e6e-49c0-a5fd-dc9323e9428c-v4-0-config-system-service-ca\") pod \"oauth-openshift-cc6d8cb6b-79m29\" (UID: \"c4a648c4-7e6e-49c0-a5fd-dc9323e9428c\") " pod="openshift-authentication/oauth-openshift-cc6d8cb6b-79m29" Oct 08 18:21:17 crc kubenswrapper[4859]: I1008 18:21:17.200115 4859 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/c4a648c4-7e6e-49c0-a5fd-dc9323e9428c-v4-0-config-system-trusted-ca-bundle\") pod \"oauth-openshift-cc6d8cb6b-79m29\" (UID: \"c4a648c4-7e6e-49c0-a5fd-dc9323e9428c\") " pod="openshift-authentication/oauth-openshift-cc6d8cb6b-79m29" Oct 08 18:21:17 crc kubenswrapper[4859]: I1008 18:21:17.200145 4859 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/c4a648c4-7e6e-49c0-a5fd-dc9323e9428c-audit-dir\") pod \"oauth-openshift-cc6d8cb6b-79m29\" (UID: \"c4a648c4-7e6e-49c0-a5fd-dc9323e9428c\") " pod="openshift-authentication/oauth-openshift-cc6d8cb6b-79m29" Oct 08 18:21:17 crc kubenswrapper[4859]: I1008 18:21:17.200173 4859 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/c4a648c4-7e6e-49c0-a5fd-dc9323e9428c-v4-0-config-system-serving-cert\") pod \"oauth-openshift-cc6d8cb6b-79m29\" (UID: \"c4a648c4-7e6e-49c0-a5fd-dc9323e9428c\") " pod="openshift-authentication/oauth-openshift-cc6d8cb6b-79m29" Oct 08 18:21:17 crc kubenswrapper[4859]: I1008 18:21:17.200194 4859 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/c4a648c4-7e6e-49c0-a5fd-dc9323e9428c-v4-0-config-user-idp-0-file-data\") pod \"oauth-openshift-cc6d8cb6b-79m29\" (UID: \"c4a648c4-7e6e-49c0-a5fd-dc9323e9428c\") " pod="openshift-authentication/oauth-openshift-cc6d8cb6b-79m29" Oct 08 18:21:17 crc kubenswrapper[4859]: I1008 18:21:17.200222 4859 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-t5b7f\" (UniqueName: \"kubernetes.io/projected/c4a648c4-7e6e-49c0-a5fd-dc9323e9428c-kube-api-access-t5b7f\") pod \"oauth-openshift-cc6d8cb6b-79m29\" (UID: \"c4a648c4-7e6e-49c0-a5fd-dc9323e9428c\") " pod="openshift-authentication/oauth-openshift-cc6d8cb6b-79m29" Oct 08 18:21:17 crc kubenswrapper[4859]: I1008 18:21:17.200259 4859 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/c4a648c4-7e6e-49c0-a5fd-dc9323e9428c-v4-0-config-system-ocp-branding-template\") pod \"oauth-openshift-cc6d8cb6b-79m29\" (UID: \"c4a648c4-7e6e-49c0-a5fd-dc9323e9428c\") " pod="openshift-authentication/oauth-openshift-cc6d8cb6b-79m29" Oct 08 18:21:17 crc kubenswrapper[4859]: I1008 18:21:17.200301 4859 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/c4a648c4-7e6e-49c0-a5fd-dc9323e9428c-v4-0-config-user-template-login\") pod \"oauth-openshift-cc6d8cb6b-79m29\" (UID: \"c4a648c4-7e6e-49c0-a5fd-dc9323e9428c\") " pod="openshift-authentication/oauth-openshift-cc6d8cb6b-79m29" Oct 08 18:21:17 crc kubenswrapper[4859]: I1008 18:21:17.200327 4859 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/c4a648c4-7e6e-49c0-a5fd-dc9323e9428c-v4-0-config-system-router-certs\") pod \"oauth-openshift-cc6d8cb6b-79m29\" (UID: \"c4a648c4-7e6e-49c0-a5fd-dc9323e9428c\") " pod="openshift-authentication/oauth-openshift-cc6d8cb6b-79m29" Oct 08 18:21:17 crc kubenswrapper[4859]: I1008 18:21:17.200350 4859 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/c4a648c4-7e6e-49c0-a5fd-dc9323e9428c-v4-0-config-user-template-provider-selection\") pod \"oauth-openshift-cc6d8cb6b-79m29\" (UID: \"c4a648c4-7e6e-49c0-a5fd-dc9323e9428c\") " pod="openshift-authentication/oauth-openshift-cc6d8cb6b-79m29" Oct 08 18:21:17 crc kubenswrapper[4859]: I1008 18:21:17.200379 4859 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/c4a648c4-7e6e-49c0-a5fd-dc9323e9428c-v4-0-config-user-template-error\") pod \"oauth-openshift-cc6d8cb6b-79m29\" (UID: \"c4a648c4-7e6e-49c0-a5fd-dc9323e9428c\") " pod="openshift-authentication/oauth-openshift-cc6d8cb6b-79m29" Oct 08 18:21:17 crc kubenswrapper[4859]: I1008 18:21:17.200405 4859 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/c4a648c4-7e6e-49c0-a5fd-dc9323e9428c-audit-policies\") pod \"oauth-openshift-cc6d8cb6b-79m29\" (UID: \"c4a648c4-7e6e-49c0-a5fd-dc9323e9428c\") " pod="openshift-authentication/oauth-openshift-cc6d8cb6b-79m29" Oct 08 18:21:17 crc kubenswrapper[4859]: I1008 18:21:17.200426 4859 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/c4a648c4-7e6e-49c0-a5fd-dc9323e9428c-v4-0-config-system-session\") pod \"oauth-openshift-cc6d8cb6b-79m29\" (UID: \"c4a648c4-7e6e-49c0-a5fd-dc9323e9428c\") " pod="openshift-authentication/oauth-openshift-cc6d8cb6b-79m29" Oct 08 18:21:17 crc kubenswrapper[4859]: I1008 18:21:17.200495 4859 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/0902ccb2-78ab-48d0-be02-4c42f03e12fb-v4-0-config-system-serving-cert\") on node \"crc\" DevicePath \"\"" Oct 08 18:21:17 crc kubenswrapper[4859]: I1008 18:21:17.200511 4859 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/0902ccb2-78ab-48d0-be02-4c42f03e12fb-v4-0-config-system-session\") on node \"crc\" DevicePath \"\"" Oct 08 18:21:17 crc kubenswrapper[4859]: I1008 18:21:17.200523 4859 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-npz7k\" (UniqueName: \"kubernetes.io/projected/0902ccb2-78ab-48d0-be02-4c42f03e12fb-kube-api-access-npz7k\") on node \"crc\" DevicePath \"\"" Oct 08 18:21:17 crc kubenswrapper[4859]: I1008 18:21:17.200537 4859 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/0902ccb2-78ab-48d0-be02-4c42f03e12fb-v4-0-config-user-template-login\") on node \"crc\" DevicePath \"\"" Oct 08 18:21:17 crc kubenswrapper[4859]: I1008 18:21:17.200550 4859 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/0902ccb2-78ab-48d0-be02-4c42f03e12fb-v4-0-config-system-ocp-branding-template\") on node \"crc\" DevicePath \"\"" Oct 08 18:21:17 crc kubenswrapper[4859]: I1008 18:21:17.200563 4859 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/0902ccb2-78ab-48d0-be02-4c42f03e12fb-v4-0-config-system-service-ca\") on node \"crc\" DevicePath \"\"" Oct 08 18:21:17 crc kubenswrapper[4859]: I1008 18:21:17.200576 4859 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/0902ccb2-78ab-48d0-be02-4c42f03e12fb-v4-0-config-user-template-error\") on node \"crc\" DevicePath \"\"" Oct 08 18:21:17 crc kubenswrapper[4859]: I1008 18:21:17.200588 4859 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/0902ccb2-78ab-48d0-be02-4c42f03e12fb-v4-0-config-system-router-certs\") on node \"crc\" DevicePath \"\"" Oct 08 18:21:17 crc kubenswrapper[4859]: I1008 18:21:17.200601 4859 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/0902ccb2-78ab-48d0-be02-4c42f03e12fb-v4-0-config-user-template-provider-selection\") on node \"crc\" DevicePath \"\"" Oct 08 18:21:17 crc kubenswrapper[4859]: I1008 18:21:17.200617 4859 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/0902ccb2-78ab-48d0-be02-4c42f03e12fb-v4-0-config-system-trusted-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 08 18:21:17 crc kubenswrapper[4859]: I1008 18:21:17.200631 4859 reconciler_common.go:293] "Volume detached for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/0902ccb2-78ab-48d0-be02-4c42f03e12fb-audit-policies\") on node \"crc\" DevicePath \"\"" Oct 08 18:21:17 crc kubenswrapper[4859]: I1008 18:21:17.200643 4859 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/0902ccb2-78ab-48d0-be02-4c42f03e12fb-v4-0-config-user-idp-0-file-data\") on node \"crc\" DevicePath \"\"" Oct 08 18:21:17 crc kubenswrapper[4859]: I1008 18:21:17.201042 4859 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/c4a648c4-7e6e-49c0-a5fd-dc9323e9428c-audit-dir\") pod \"oauth-openshift-cc6d8cb6b-79m29\" (UID: \"c4a648c4-7e6e-49c0-a5fd-dc9323e9428c\") " pod="openshift-authentication/oauth-openshift-cc6d8cb6b-79m29" Oct 08 18:21:17 crc kubenswrapper[4859]: I1008 18:21:17.201612 4859 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/c4a648c4-7e6e-49c0-a5fd-dc9323e9428c-v4-0-config-system-cliconfig\") pod \"oauth-openshift-cc6d8cb6b-79m29\" (UID: \"c4a648c4-7e6e-49c0-a5fd-dc9323e9428c\") " pod="openshift-authentication/oauth-openshift-cc6d8cb6b-79m29" Oct 08 18:21:17 crc kubenswrapper[4859]: I1008 18:21:17.201779 4859 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/c4a648c4-7e6e-49c0-a5fd-dc9323e9428c-v4-0-config-system-service-ca\") pod \"oauth-openshift-cc6d8cb6b-79m29\" (UID: \"c4a648c4-7e6e-49c0-a5fd-dc9323e9428c\") " pod="openshift-authentication/oauth-openshift-cc6d8cb6b-79m29" Oct 08 18:21:17 crc kubenswrapper[4859]: I1008 18:21:17.202243 4859 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/c4a648c4-7e6e-49c0-a5fd-dc9323e9428c-v4-0-config-system-trusted-ca-bundle\") pod \"oauth-openshift-cc6d8cb6b-79m29\" (UID: \"c4a648c4-7e6e-49c0-a5fd-dc9323e9428c\") " pod="openshift-authentication/oauth-openshift-cc6d8cb6b-79m29" Oct 08 18:21:17 crc kubenswrapper[4859]: I1008 18:21:17.202326 4859 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/c4a648c4-7e6e-49c0-a5fd-dc9323e9428c-audit-policies\") pod \"oauth-openshift-cc6d8cb6b-79m29\" (UID: \"c4a648c4-7e6e-49c0-a5fd-dc9323e9428c\") " pod="openshift-authentication/oauth-openshift-cc6d8cb6b-79m29" Oct 08 18:21:17 crc kubenswrapper[4859]: I1008 18:21:17.204585 4859 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/c4a648c4-7e6e-49c0-a5fd-dc9323e9428c-v4-0-config-system-session\") pod \"oauth-openshift-cc6d8cb6b-79m29\" (UID: \"c4a648c4-7e6e-49c0-a5fd-dc9323e9428c\") " pod="openshift-authentication/oauth-openshift-cc6d8cb6b-79m29" Oct 08 18:21:17 crc kubenswrapper[4859]: I1008 18:21:17.204851 4859 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/c4a648c4-7e6e-49c0-a5fd-dc9323e9428c-v4-0-config-user-template-error\") pod \"oauth-openshift-cc6d8cb6b-79m29\" (UID: \"c4a648c4-7e6e-49c0-a5fd-dc9323e9428c\") " pod="openshift-authentication/oauth-openshift-cc6d8cb6b-79m29" Oct 08 18:21:17 crc kubenswrapper[4859]: I1008 18:21:17.205777 4859 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/c4a648c4-7e6e-49c0-a5fd-dc9323e9428c-v4-0-config-user-idp-0-file-data\") pod \"oauth-openshift-cc6d8cb6b-79m29\" (UID: \"c4a648c4-7e6e-49c0-a5fd-dc9323e9428c\") " pod="openshift-authentication/oauth-openshift-cc6d8cb6b-79m29" Oct 08 18:21:17 crc kubenswrapper[4859]: I1008 18:21:17.206191 4859 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/c4a648c4-7e6e-49c0-a5fd-dc9323e9428c-v4-0-config-system-router-certs\") pod \"oauth-openshift-cc6d8cb6b-79m29\" (UID: \"c4a648c4-7e6e-49c0-a5fd-dc9323e9428c\") " pod="openshift-authentication/oauth-openshift-cc6d8cb6b-79m29" Oct 08 18:21:17 crc kubenswrapper[4859]: I1008 18:21:17.206611 4859 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/c4a648c4-7e6e-49c0-a5fd-dc9323e9428c-v4-0-config-user-template-provider-selection\") pod \"oauth-openshift-cc6d8cb6b-79m29\" (UID: \"c4a648c4-7e6e-49c0-a5fd-dc9323e9428c\") " pod="openshift-authentication/oauth-openshift-cc6d8cb6b-79m29" Oct 08 18:21:17 crc kubenswrapper[4859]: I1008 18:21:17.206963 4859 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/c4a648c4-7e6e-49c0-a5fd-dc9323e9428c-v4-0-config-system-ocp-branding-template\") pod \"oauth-openshift-cc6d8cb6b-79m29\" (UID: \"c4a648c4-7e6e-49c0-a5fd-dc9323e9428c\") " pod="openshift-authentication/oauth-openshift-cc6d8cb6b-79m29" Oct 08 18:21:17 crc kubenswrapper[4859]: I1008 18:21:17.207168 4859 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/c4a648c4-7e6e-49c0-a5fd-dc9323e9428c-v4-0-config-user-template-login\") pod \"oauth-openshift-cc6d8cb6b-79m29\" (UID: \"c4a648c4-7e6e-49c0-a5fd-dc9323e9428c\") " pod="openshift-authentication/oauth-openshift-cc6d8cb6b-79m29" Oct 08 18:21:17 crc kubenswrapper[4859]: I1008 18:21:17.207634 4859 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/c4a648c4-7e6e-49c0-a5fd-dc9323e9428c-v4-0-config-system-serving-cert\") pod \"oauth-openshift-cc6d8cb6b-79m29\" (UID: \"c4a648c4-7e6e-49c0-a5fd-dc9323e9428c\") " pod="openshift-authentication/oauth-openshift-cc6d8cb6b-79m29" Oct 08 18:21:17 crc kubenswrapper[4859]: I1008 18:21:17.218216 4859 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-t5b7f\" (UniqueName: \"kubernetes.io/projected/c4a648c4-7e6e-49c0-a5fd-dc9323e9428c-kube-api-access-t5b7f\") pod \"oauth-openshift-cc6d8cb6b-79m29\" (UID: \"c4a648c4-7e6e-49c0-a5fd-dc9323e9428c\") " pod="openshift-authentication/oauth-openshift-cc6d8cb6b-79m29" Oct 08 18:21:17 crc kubenswrapper[4859]: I1008 18:21:17.328712 4859 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-authentication/oauth-openshift-cc6d8cb6b-79m29" Oct 08 18:21:17 crc kubenswrapper[4859]: I1008 18:21:17.477312 4859 generic.go:334] "Generic (PLEG): container finished" podID="0902ccb2-78ab-48d0-be02-4c42f03e12fb" containerID="7023b5c72b539780ca0fa2a5d77a71496bfa187a6666d8c3d8fd2b6e898c0c70" exitCode=0 Oct 08 18:21:17 crc kubenswrapper[4859]: I1008 18:21:17.477381 4859 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-authentication/oauth-openshift-558db77b4-j8gh2" event={"ID":"0902ccb2-78ab-48d0-be02-4c42f03e12fb","Type":"ContainerDied","Data":"7023b5c72b539780ca0fa2a5d77a71496bfa187a6666d8c3d8fd2b6e898c0c70"} Oct 08 18:21:17 crc kubenswrapper[4859]: I1008 18:21:17.477431 4859 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-authentication/oauth-openshift-558db77b4-j8gh2" event={"ID":"0902ccb2-78ab-48d0-be02-4c42f03e12fb","Type":"ContainerDied","Data":"d787d40d8d2e2389a64edbee99189e2b4ac2bd21b7ae58311a0299d5ea1445a4"} Oct 08 18:21:17 crc kubenswrapper[4859]: I1008 18:21:17.477459 4859 scope.go:117] "RemoveContainer" containerID="7023b5c72b539780ca0fa2a5d77a71496bfa187a6666d8c3d8fd2b6e898c0c70" Oct 08 18:21:17 crc kubenswrapper[4859]: I1008 18:21:17.477664 4859 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-authentication/oauth-openshift-558db77b4-j8gh2" Oct 08 18:21:17 crc kubenswrapper[4859]: I1008 18:21:17.515750 4859 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-authentication/oauth-openshift-558db77b4-j8gh2"] Oct 08 18:21:17 crc kubenswrapper[4859]: I1008 18:21:17.516254 4859 scope.go:117] "RemoveContainer" containerID="7023b5c72b539780ca0fa2a5d77a71496bfa187a6666d8c3d8fd2b6e898c0c70" Oct 08 18:21:17 crc kubenswrapper[4859]: E1008 18:21:17.517040 4859 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"7023b5c72b539780ca0fa2a5d77a71496bfa187a6666d8c3d8fd2b6e898c0c70\": container with ID starting with 7023b5c72b539780ca0fa2a5d77a71496bfa187a6666d8c3d8fd2b6e898c0c70 not found: ID does not exist" containerID="7023b5c72b539780ca0fa2a5d77a71496bfa187a6666d8c3d8fd2b6e898c0c70" Oct 08 18:21:17 crc kubenswrapper[4859]: I1008 18:21:17.517102 4859 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"7023b5c72b539780ca0fa2a5d77a71496bfa187a6666d8c3d8fd2b6e898c0c70"} err="failed to get container status \"7023b5c72b539780ca0fa2a5d77a71496bfa187a6666d8c3d8fd2b6e898c0c70\": rpc error: code = NotFound desc = could not find container \"7023b5c72b539780ca0fa2a5d77a71496bfa187a6666d8c3d8fd2b6e898c0c70\": container with ID starting with 7023b5c72b539780ca0fa2a5d77a71496bfa187a6666d8c3d8fd2b6e898c0c70 not found: ID does not exist" Oct 08 18:21:17 crc kubenswrapper[4859]: I1008 18:21:17.521985 4859 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-authentication/oauth-openshift-558db77b4-j8gh2"] Oct 08 18:21:17 crc kubenswrapper[4859]: I1008 18:21:17.578872 4859 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-authentication/oauth-openshift-cc6d8cb6b-79m29"] Oct 08 18:21:18 crc kubenswrapper[4859]: I1008 18:21:18.487190 4859 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="0902ccb2-78ab-48d0-be02-4c42f03e12fb" path="/var/lib/kubelet/pods/0902ccb2-78ab-48d0-be02-4c42f03e12fb/volumes" Oct 08 18:21:18 crc kubenswrapper[4859]: I1008 18:21:18.488850 4859 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-authentication/oauth-openshift-cc6d8cb6b-79m29" event={"ID":"c4a648c4-7e6e-49c0-a5fd-dc9323e9428c","Type":"ContainerStarted","Data":"4cfd7a57b3d6304642213a252967596219c875751e9c27db0cae12e4b80c54aa"} Oct 08 18:21:18 crc kubenswrapper[4859]: I1008 18:21:18.488938 4859 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-authentication/oauth-openshift-cc6d8cb6b-79m29" event={"ID":"c4a648c4-7e6e-49c0-a5fd-dc9323e9428c","Type":"ContainerStarted","Data":"9e5a10363782f8e4794c7974d450c5efc5a969d113986b82882f69debcd7be1f"} Oct 08 18:21:18 crc kubenswrapper[4859]: I1008 18:21:18.488971 4859 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-authentication/oauth-openshift-cc6d8cb6b-79m29" Oct 08 18:21:18 crc kubenswrapper[4859]: I1008 18:21:18.525957 4859 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-authentication/oauth-openshift-cc6d8cb6b-79m29" Oct 08 18:21:18 crc kubenswrapper[4859]: I1008 18:21:18.527123 4859 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-authentication/oauth-openshift-cc6d8cb6b-79m29" podStartSLOduration=27.527088297 podStartE2EDuration="27.527088297s" podCreationTimestamp="2025-10-08 18:20:51 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-08 18:21:18.520384676 +0000 UTC m=+248.767224095" watchObservedRunningTime="2025-10-08 18:21:18.527088297 +0000 UTC m=+248.773927706" Oct 08 18:21:37 crc kubenswrapper[4859]: I1008 18:21:37.590546 4859 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-c5tcx"] Oct 08 18:21:37 crc kubenswrapper[4859]: I1008 18:21:37.591560 4859 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/certified-operators-c5tcx" podUID="438a5b9a-aaee-4e06-9d22-691862b9207e" containerName="registry-server" containerID="cri-o://8b6f3d58eacd7cf2bce7bfad476971fff5b2db73d55e67cbe99f2fe89dea2ac6" gracePeriod=30 Oct 08 18:21:37 crc kubenswrapper[4859]: I1008 18:21:37.607921 4859 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-zb82c"] Oct 08 18:21:37 crc kubenswrapper[4859]: I1008 18:21:37.608293 4859 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/community-operators-zb82c" podUID="5b68d592-4a46-4fbc-bcec-a8bca2b973aa" containerName="registry-server" containerID="cri-o://3b5490e97f3c5e1a24d3ba43457bebc4949c6182b88dfe9fdb226ee620caa614" gracePeriod=30 Oct 08 18:21:37 crc kubenswrapper[4859]: I1008 18:21:37.615540 4859 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/marketplace-operator-79b997595-bz6v8"] Oct 08 18:21:37 crc kubenswrapper[4859]: I1008 18:21:37.615790 4859 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/marketplace-operator-79b997595-bz6v8" podUID="8d4b5b44-816f-44f6-98ab-c5d66aefaa08" containerName="marketplace-operator" containerID="cri-o://b6f47fb2611d35c2a1a47f8cdcf8133fc94d28820b6d8a1908a935ecfa0d3e2d" gracePeriod=30 Oct 08 18:21:37 crc kubenswrapper[4859]: I1008 18:21:37.629894 4859 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-8z5ch"] Oct 08 18:21:37 crc kubenswrapper[4859]: I1008 18:21:37.630261 4859 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-marketplace-8z5ch" podUID="557e8291-7e96-4634-b744-bcf6981d3a05" containerName="registry-server" containerID="cri-o://4444d93f68c351c0aae5349182126b3615263046ce9bb221bee588119677bf6a" gracePeriod=30 Oct 08 18:21:37 crc kubenswrapper[4859]: I1008 18:21:37.643933 4859 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/marketplace-operator-79b997595-h4tnd"] Oct 08 18:21:37 crc kubenswrapper[4859]: I1008 18:21:37.644881 4859 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/marketplace-operator-79b997595-h4tnd" Oct 08 18:21:37 crc kubenswrapper[4859]: I1008 18:21:37.648650 4859 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-xxgmr"] Oct 08 18:21:37 crc kubenswrapper[4859]: I1008 18:21:37.648957 4859 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-operators-xxgmr" podUID="aeadfa85-a482-40a8-b947-09d2f6613900" containerName="registry-server" containerID="cri-o://30d7dafc5dc8f0a00b63102ca616e3537a9ab1f2b622798aead42d5f5fe48381" gracePeriod=30 Oct 08 18:21:37 crc kubenswrapper[4859]: I1008 18:21:37.663888 4859 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/marketplace-operator-79b997595-h4tnd"] Oct 08 18:21:37 crc kubenswrapper[4859]: I1008 18:21:37.791041 4859 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/8fdf9d2f-74e1-46e7-89ed-1947c3104386-marketplace-trusted-ca\") pod \"marketplace-operator-79b997595-h4tnd\" (UID: \"8fdf9d2f-74e1-46e7-89ed-1947c3104386\") " pod="openshift-marketplace/marketplace-operator-79b997595-h4tnd" Oct 08 18:21:37 crc kubenswrapper[4859]: I1008 18:21:37.791288 4859 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-pcmsz\" (UniqueName: \"kubernetes.io/projected/8fdf9d2f-74e1-46e7-89ed-1947c3104386-kube-api-access-pcmsz\") pod \"marketplace-operator-79b997595-h4tnd\" (UID: \"8fdf9d2f-74e1-46e7-89ed-1947c3104386\") " pod="openshift-marketplace/marketplace-operator-79b997595-h4tnd" Oct 08 18:21:37 crc kubenswrapper[4859]: I1008 18:21:37.791330 4859 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/8fdf9d2f-74e1-46e7-89ed-1947c3104386-marketplace-operator-metrics\") pod \"marketplace-operator-79b997595-h4tnd\" (UID: \"8fdf9d2f-74e1-46e7-89ed-1947c3104386\") " pod="openshift-marketplace/marketplace-operator-79b997595-h4tnd" Oct 08 18:21:37 crc kubenswrapper[4859]: I1008 18:21:37.892565 4859 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/8fdf9d2f-74e1-46e7-89ed-1947c3104386-marketplace-trusted-ca\") pod \"marketplace-operator-79b997595-h4tnd\" (UID: \"8fdf9d2f-74e1-46e7-89ed-1947c3104386\") " pod="openshift-marketplace/marketplace-operator-79b997595-h4tnd" Oct 08 18:21:37 crc kubenswrapper[4859]: I1008 18:21:37.892619 4859 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-pcmsz\" (UniqueName: \"kubernetes.io/projected/8fdf9d2f-74e1-46e7-89ed-1947c3104386-kube-api-access-pcmsz\") pod \"marketplace-operator-79b997595-h4tnd\" (UID: \"8fdf9d2f-74e1-46e7-89ed-1947c3104386\") " pod="openshift-marketplace/marketplace-operator-79b997595-h4tnd" Oct 08 18:21:37 crc kubenswrapper[4859]: I1008 18:21:37.892662 4859 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/8fdf9d2f-74e1-46e7-89ed-1947c3104386-marketplace-operator-metrics\") pod \"marketplace-operator-79b997595-h4tnd\" (UID: \"8fdf9d2f-74e1-46e7-89ed-1947c3104386\") " pod="openshift-marketplace/marketplace-operator-79b997595-h4tnd" Oct 08 18:21:37 crc kubenswrapper[4859]: I1008 18:21:37.894745 4859 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/8fdf9d2f-74e1-46e7-89ed-1947c3104386-marketplace-trusted-ca\") pod \"marketplace-operator-79b997595-h4tnd\" (UID: \"8fdf9d2f-74e1-46e7-89ed-1947c3104386\") " pod="openshift-marketplace/marketplace-operator-79b997595-h4tnd" Oct 08 18:21:37 crc kubenswrapper[4859]: I1008 18:21:37.903497 4859 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/8fdf9d2f-74e1-46e7-89ed-1947c3104386-marketplace-operator-metrics\") pod \"marketplace-operator-79b997595-h4tnd\" (UID: \"8fdf9d2f-74e1-46e7-89ed-1947c3104386\") " pod="openshift-marketplace/marketplace-operator-79b997595-h4tnd" Oct 08 18:21:37 crc kubenswrapper[4859]: I1008 18:21:37.916049 4859 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-pcmsz\" (UniqueName: \"kubernetes.io/projected/8fdf9d2f-74e1-46e7-89ed-1947c3104386-kube-api-access-pcmsz\") pod \"marketplace-operator-79b997595-h4tnd\" (UID: \"8fdf9d2f-74e1-46e7-89ed-1947c3104386\") " pod="openshift-marketplace/marketplace-operator-79b997595-h4tnd" Oct 08 18:21:37 crc kubenswrapper[4859]: I1008 18:21:37.963444 4859 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/marketplace-operator-79b997595-h4tnd" Oct 08 18:21:38 crc kubenswrapper[4859]: I1008 18:21:38.041041 4859 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-zb82c" Oct 08 18:21:38 crc kubenswrapper[4859]: I1008 18:21:38.045861 4859 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-c5tcx" Oct 08 18:21:38 crc kubenswrapper[4859]: I1008 18:21:38.049550 4859 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-8z5ch" Oct 08 18:21:38 crc kubenswrapper[4859]: I1008 18:21:38.056810 4859 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-xxgmr" Oct 08 18:21:38 crc kubenswrapper[4859]: I1008 18:21:38.110609 4859 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/marketplace-operator-79b997595-bz6v8" Oct 08 18:21:38 crc kubenswrapper[4859]: I1008 18:21:38.201776 4859 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/aeadfa85-a482-40a8-b947-09d2f6613900-utilities\") pod \"aeadfa85-a482-40a8-b947-09d2f6613900\" (UID: \"aeadfa85-a482-40a8-b947-09d2f6613900\") " Oct 08 18:21:38 crc kubenswrapper[4859]: I1008 18:21:38.201830 4859 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/5b68d592-4a46-4fbc-bcec-a8bca2b973aa-catalog-content\") pod \"5b68d592-4a46-4fbc-bcec-a8bca2b973aa\" (UID: \"5b68d592-4a46-4fbc-bcec-a8bca2b973aa\") " Oct 08 18:21:38 crc kubenswrapper[4859]: I1008 18:21:38.201859 4859 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-rp9cl\" (UniqueName: \"kubernetes.io/projected/438a5b9a-aaee-4e06-9d22-691862b9207e-kube-api-access-rp9cl\") pod \"438a5b9a-aaee-4e06-9d22-691862b9207e\" (UID: \"438a5b9a-aaee-4e06-9d22-691862b9207e\") " Oct 08 18:21:38 crc kubenswrapper[4859]: I1008 18:21:38.201891 4859 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/8d4b5b44-816f-44f6-98ab-c5d66aefaa08-marketplace-trusted-ca\") pod \"8d4b5b44-816f-44f6-98ab-c5d66aefaa08\" (UID: \"8d4b5b44-816f-44f6-98ab-c5d66aefaa08\") " Oct 08 18:21:38 crc kubenswrapper[4859]: I1008 18:21:38.201921 4859 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-7g9pg\" (UniqueName: \"kubernetes.io/projected/5b68d592-4a46-4fbc-bcec-a8bca2b973aa-kube-api-access-7g9pg\") pod \"5b68d592-4a46-4fbc-bcec-a8bca2b973aa\" (UID: \"5b68d592-4a46-4fbc-bcec-a8bca2b973aa\") " Oct 08 18:21:38 crc kubenswrapper[4859]: I1008 18:21:38.201986 4859 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-r5mcr\" (UniqueName: \"kubernetes.io/projected/557e8291-7e96-4634-b744-bcf6981d3a05-kube-api-access-r5mcr\") pod \"557e8291-7e96-4634-b744-bcf6981d3a05\" (UID: \"557e8291-7e96-4634-b744-bcf6981d3a05\") " Oct 08 18:21:38 crc kubenswrapper[4859]: I1008 18:21:38.202015 4859 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-hv4wh\" (UniqueName: \"kubernetes.io/projected/8d4b5b44-816f-44f6-98ab-c5d66aefaa08-kube-api-access-hv4wh\") pod \"8d4b5b44-816f-44f6-98ab-c5d66aefaa08\" (UID: \"8d4b5b44-816f-44f6-98ab-c5d66aefaa08\") " Oct 08 18:21:38 crc kubenswrapper[4859]: I1008 18:21:38.202091 4859 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/5b68d592-4a46-4fbc-bcec-a8bca2b973aa-utilities\") pod \"5b68d592-4a46-4fbc-bcec-a8bca2b973aa\" (UID: \"5b68d592-4a46-4fbc-bcec-a8bca2b973aa\") " Oct 08 18:21:38 crc kubenswrapper[4859]: I1008 18:21:38.202166 4859 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-9c6gz\" (UniqueName: \"kubernetes.io/projected/aeadfa85-a482-40a8-b947-09d2f6613900-kube-api-access-9c6gz\") pod \"aeadfa85-a482-40a8-b947-09d2f6613900\" (UID: \"aeadfa85-a482-40a8-b947-09d2f6613900\") " Oct 08 18:21:38 crc kubenswrapper[4859]: I1008 18:21:38.202190 4859 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/438a5b9a-aaee-4e06-9d22-691862b9207e-utilities\") pod \"438a5b9a-aaee-4e06-9d22-691862b9207e\" (UID: \"438a5b9a-aaee-4e06-9d22-691862b9207e\") " Oct 08 18:21:38 crc kubenswrapper[4859]: I1008 18:21:38.202255 4859 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/557e8291-7e96-4634-b744-bcf6981d3a05-catalog-content\") pod \"557e8291-7e96-4634-b744-bcf6981d3a05\" (UID: \"557e8291-7e96-4634-b744-bcf6981d3a05\") " Oct 08 18:21:38 crc kubenswrapper[4859]: I1008 18:21:38.202280 4859 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/aeadfa85-a482-40a8-b947-09d2f6613900-catalog-content\") pod \"aeadfa85-a482-40a8-b947-09d2f6613900\" (UID: \"aeadfa85-a482-40a8-b947-09d2f6613900\") " Oct 08 18:21:38 crc kubenswrapper[4859]: I1008 18:21:38.202338 4859 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/557e8291-7e96-4634-b744-bcf6981d3a05-utilities\") pod \"557e8291-7e96-4634-b744-bcf6981d3a05\" (UID: \"557e8291-7e96-4634-b744-bcf6981d3a05\") " Oct 08 18:21:38 crc kubenswrapper[4859]: I1008 18:21:38.202364 4859 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/8d4b5b44-816f-44f6-98ab-c5d66aefaa08-marketplace-operator-metrics\") pod \"8d4b5b44-816f-44f6-98ab-c5d66aefaa08\" (UID: \"8d4b5b44-816f-44f6-98ab-c5d66aefaa08\") " Oct 08 18:21:38 crc kubenswrapper[4859]: I1008 18:21:38.202415 4859 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/438a5b9a-aaee-4e06-9d22-691862b9207e-catalog-content\") pod \"438a5b9a-aaee-4e06-9d22-691862b9207e\" (UID: \"438a5b9a-aaee-4e06-9d22-691862b9207e\") " Oct 08 18:21:38 crc kubenswrapper[4859]: I1008 18:21:38.206168 4859 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/5b68d592-4a46-4fbc-bcec-a8bca2b973aa-utilities" (OuterVolumeSpecName: "utilities") pod "5b68d592-4a46-4fbc-bcec-a8bca2b973aa" (UID: "5b68d592-4a46-4fbc-bcec-a8bca2b973aa"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 08 18:21:38 crc kubenswrapper[4859]: I1008 18:21:38.206570 4859 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/5b68d592-4a46-4fbc-bcec-a8bca2b973aa-utilities\") on node \"crc\" DevicePath \"\"" Oct 08 18:21:38 crc kubenswrapper[4859]: I1008 18:21:38.206829 4859 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/557e8291-7e96-4634-b744-bcf6981d3a05-utilities" (OuterVolumeSpecName: "utilities") pod "557e8291-7e96-4634-b744-bcf6981d3a05" (UID: "557e8291-7e96-4634-b744-bcf6981d3a05"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 08 18:21:38 crc kubenswrapper[4859]: I1008 18:21:38.212775 4859 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/aeadfa85-a482-40a8-b947-09d2f6613900-utilities" (OuterVolumeSpecName: "utilities") pod "aeadfa85-a482-40a8-b947-09d2f6613900" (UID: "aeadfa85-a482-40a8-b947-09d2f6613900"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 08 18:21:38 crc kubenswrapper[4859]: I1008 18:21:38.217151 4859 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/438a5b9a-aaee-4e06-9d22-691862b9207e-utilities" (OuterVolumeSpecName: "utilities") pod "438a5b9a-aaee-4e06-9d22-691862b9207e" (UID: "438a5b9a-aaee-4e06-9d22-691862b9207e"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 08 18:21:38 crc kubenswrapper[4859]: I1008 18:21:38.219362 4859 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/8d4b5b44-816f-44f6-98ab-c5d66aefaa08-marketplace-trusted-ca" (OuterVolumeSpecName: "marketplace-trusted-ca") pod "8d4b5b44-816f-44f6-98ab-c5d66aefaa08" (UID: "8d4b5b44-816f-44f6-98ab-c5d66aefaa08"). InnerVolumeSpecName "marketplace-trusted-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 08 18:21:38 crc kubenswrapper[4859]: I1008 18:21:38.229535 4859 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/557e8291-7e96-4634-b744-bcf6981d3a05-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "557e8291-7e96-4634-b744-bcf6981d3a05" (UID: "557e8291-7e96-4634-b744-bcf6981d3a05"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 08 18:21:38 crc kubenswrapper[4859]: I1008 18:21:38.259263 4859 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/aeadfa85-a482-40a8-b947-09d2f6613900-kube-api-access-9c6gz" (OuterVolumeSpecName: "kube-api-access-9c6gz") pod "aeadfa85-a482-40a8-b947-09d2f6613900" (UID: "aeadfa85-a482-40a8-b947-09d2f6613900"). InnerVolumeSpecName "kube-api-access-9c6gz". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 08 18:21:38 crc kubenswrapper[4859]: I1008 18:21:38.259320 4859 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/557e8291-7e96-4634-b744-bcf6981d3a05-kube-api-access-r5mcr" (OuterVolumeSpecName: "kube-api-access-r5mcr") pod "557e8291-7e96-4634-b744-bcf6981d3a05" (UID: "557e8291-7e96-4634-b744-bcf6981d3a05"). InnerVolumeSpecName "kube-api-access-r5mcr". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 08 18:21:38 crc kubenswrapper[4859]: I1008 18:21:38.259574 4859 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5b68d592-4a46-4fbc-bcec-a8bca2b973aa-kube-api-access-7g9pg" (OuterVolumeSpecName: "kube-api-access-7g9pg") pod "5b68d592-4a46-4fbc-bcec-a8bca2b973aa" (UID: "5b68d592-4a46-4fbc-bcec-a8bca2b973aa"). InnerVolumeSpecName "kube-api-access-7g9pg". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 08 18:21:38 crc kubenswrapper[4859]: I1008 18:21:38.259670 4859 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/438a5b9a-aaee-4e06-9d22-691862b9207e-kube-api-access-rp9cl" (OuterVolumeSpecName: "kube-api-access-rp9cl") pod "438a5b9a-aaee-4e06-9d22-691862b9207e" (UID: "438a5b9a-aaee-4e06-9d22-691862b9207e"). InnerVolumeSpecName "kube-api-access-rp9cl". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 08 18:21:38 crc kubenswrapper[4859]: I1008 18:21:38.259965 4859 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8d4b5b44-816f-44f6-98ab-c5d66aefaa08-kube-api-access-hv4wh" (OuterVolumeSpecName: "kube-api-access-hv4wh") pod "8d4b5b44-816f-44f6-98ab-c5d66aefaa08" (UID: "8d4b5b44-816f-44f6-98ab-c5d66aefaa08"). InnerVolumeSpecName "kube-api-access-hv4wh". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 08 18:21:38 crc kubenswrapper[4859]: I1008 18:21:38.262037 4859 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8d4b5b44-816f-44f6-98ab-c5d66aefaa08-marketplace-operator-metrics" (OuterVolumeSpecName: "marketplace-operator-metrics") pod "8d4b5b44-816f-44f6-98ab-c5d66aefaa08" (UID: "8d4b5b44-816f-44f6-98ab-c5d66aefaa08"). InnerVolumeSpecName "marketplace-operator-metrics". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 08 18:21:38 crc kubenswrapper[4859]: I1008 18:21:38.280326 4859 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/marketplace-operator-79b997595-h4tnd"] Oct 08 18:21:38 crc kubenswrapper[4859]: I1008 18:21:38.309947 4859 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/aeadfa85-a482-40a8-b947-09d2f6613900-utilities\") on node \"crc\" DevicePath \"\"" Oct 08 18:21:38 crc kubenswrapper[4859]: I1008 18:21:38.310088 4859 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-rp9cl\" (UniqueName: \"kubernetes.io/projected/438a5b9a-aaee-4e06-9d22-691862b9207e-kube-api-access-rp9cl\") on node \"crc\" DevicePath \"\"" Oct 08 18:21:38 crc kubenswrapper[4859]: I1008 18:21:38.310150 4859 reconciler_common.go:293] "Volume detached for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/8d4b5b44-816f-44f6-98ab-c5d66aefaa08-marketplace-trusted-ca\") on node \"crc\" DevicePath \"\"" Oct 08 18:21:38 crc kubenswrapper[4859]: I1008 18:21:38.310220 4859 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-7g9pg\" (UniqueName: \"kubernetes.io/projected/5b68d592-4a46-4fbc-bcec-a8bca2b973aa-kube-api-access-7g9pg\") on node \"crc\" DevicePath \"\"" Oct 08 18:21:38 crc kubenswrapper[4859]: I1008 18:21:38.310279 4859 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-r5mcr\" (UniqueName: \"kubernetes.io/projected/557e8291-7e96-4634-b744-bcf6981d3a05-kube-api-access-r5mcr\") on node \"crc\" DevicePath \"\"" Oct 08 18:21:38 crc kubenswrapper[4859]: I1008 18:21:38.310335 4859 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-hv4wh\" (UniqueName: \"kubernetes.io/projected/8d4b5b44-816f-44f6-98ab-c5d66aefaa08-kube-api-access-hv4wh\") on node \"crc\" DevicePath \"\"" Oct 08 18:21:38 crc kubenswrapper[4859]: I1008 18:21:38.310389 4859 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-9c6gz\" (UniqueName: \"kubernetes.io/projected/aeadfa85-a482-40a8-b947-09d2f6613900-kube-api-access-9c6gz\") on node \"crc\" DevicePath \"\"" Oct 08 18:21:38 crc kubenswrapper[4859]: I1008 18:21:38.310445 4859 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/438a5b9a-aaee-4e06-9d22-691862b9207e-utilities\") on node \"crc\" DevicePath \"\"" Oct 08 18:21:38 crc kubenswrapper[4859]: I1008 18:21:38.310504 4859 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/557e8291-7e96-4634-b744-bcf6981d3a05-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 08 18:21:38 crc kubenswrapper[4859]: I1008 18:21:38.310564 4859 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/557e8291-7e96-4634-b744-bcf6981d3a05-utilities\") on node \"crc\" DevicePath \"\"" Oct 08 18:21:38 crc kubenswrapper[4859]: I1008 18:21:38.310620 4859 reconciler_common.go:293] "Volume detached for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/8d4b5b44-816f-44f6-98ab-c5d66aefaa08-marketplace-operator-metrics\") on node \"crc\" DevicePath \"\"" Oct 08 18:21:38 crc kubenswrapper[4859]: I1008 18:21:38.311378 4859 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/5b68d592-4a46-4fbc-bcec-a8bca2b973aa-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "5b68d592-4a46-4fbc-bcec-a8bca2b973aa" (UID: "5b68d592-4a46-4fbc-bcec-a8bca2b973aa"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 08 18:21:38 crc kubenswrapper[4859]: I1008 18:21:38.345740 4859 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/438a5b9a-aaee-4e06-9d22-691862b9207e-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "438a5b9a-aaee-4e06-9d22-691862b9207e" (UID: "438a5b9a-aaee-4e06-9d22-691862b9207e"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 08 18:21:38 crc kubenswrapper[4859]: I1008 18:21:38.376208 4859 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/aeadfa85-a482-40a8-b947-09d2f6613900-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "aeadfa85-a482-40a8-b947-09d2f6613900" (UID: "aeadfa85-a482-40a8-b947-09d2f6613900"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 08 18:21:38 crc kubenswrapper[4859]: I1008 18:21:38.411312 4859 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/aeadfa85-a482-40a8-b947-09d2f6613900-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 08 18:21:38 crc kubenswrapper[4859]: I1008 18:21:38.411347 4859 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/438a5b9a-aaee-4e06-9d22-691862b9207e-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 08 18:21:38 crc kubenswrapper[4859]: I1008 18:21:38.411356 4859 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/5b68d592-4a46-4fbc-bcec-a8bca2b973aa-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 08 18:21:38 crc kubenswrapper[4859]: I1008 18:21:38.626084 4859 generic.go:334] "Generic (PLEG): container finished" podID="5b68d592-4a46-4fbc-bcec-a8bca2b973aa" containerID="3b5490e97f3c5e1a24d3ba43457bebc4949c6182b88dfe9fdb226ee620caa614" exitCode=0 Oct 08 18:21:38 crc kubenswrapper[4859]: I1008 18:21:38.626183 4859 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-zb82c" event={"ID":"5b68d592-4a46-4fbc-bcec-a8bca2b973aa","Type":"ContainerDied","Data":"3b5490e97f3c5e1a24d3ba43457bebc4949c6182b88dfe9fdb226ee620caa614"} Oct 08 18:21:38 crc kubenswrapper[4859]: I1008 18:21:38.626194 4859 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-zb82c" Oct 08 18:21:38 crc kubenswrapper[4859]: I1008 18:21:38.626260 4859 scope.go:117] "RemoveContainer" containerID="3b5490e97f3c5e1a24d3ba43457bebc4949c6182b88dfe9fdb226ee620caa614" Oct 08 18:21:38 crc kubenswrapper[4859]: I1008 18:21:38.626238 4859 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-zb82c" event={"ID":"5b68d592-4a46-4fbc-bcec-a8bca2b973aa","Type":"ContainerDied","Data":"da8c59c1f62a80084e100ca36e84217397827dfa0c3fd9cac0e1b5bc10f153d0"} Oct 08 18:21:38 crc kubenswrapper[4859]: I1008 18:21:38.627344 4859 generic.go:334] "Generic (PLEG): container finished" podID="8d4b5b44-816f-44f6-98ab-c5d66aefaa08" containerID="b6f47fb2611d35c2a1a47f8cdcf8133fc94d28820b6d8a1908a935ecfa0d3e2d" exitCode=0 Oct 08 18:21:38 crc kubenswrapper[4859]: I1008 18:21:38.627384 4859 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/marketplace-operator-79b997595-bz6v8" event={"ID":"8d4b5b44-816f-44f6-98ab-c5d66aefaa08","Type":"ContainerDied","Data":"b6f47fb2611d35c2a1a47f8cdcf8133fc94d28820b6d8a1908a935ecfa0d3e2d"} Oct 08 18:21:38 crc kubenswrapper[4859]: I1008 18:21:38.627405 4859 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/marketplace-operator-79b997595-bz6v8" event={"ID":"8d4b5b44-816f-44f6-98ab-c5d66aefaa08","Type":"ContainerDied","Data":"30e8f4a72ca43a49d25154d0078616d084b696a53c76f98a22049f48584ecae0"} Oct 08 18:21:38 crc kubenswrapper[4859]: I1008 18:21:38.627464 4859 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/marketplace-operator-79b997595-bz6v8" Oct 08 18:21:38 crc kubenswrapper[4859]: I1008 18:21:38.632416 4859 generic.go:334] "Generic (PLEG): container finished" podID="557e8291-7e96-4634-b744-bcf6981d3a05" containerID="4444d93f68c351c0aae5349182126b3615263046ce9bb221bee588119677bf6a" exitCode=0 Oct 08 18:21:38 crc kubenswrapper[4859]: I1008 18:21:38.632554 4859 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-8z5ch" event={"ID":"557e8291-7e96-4634-b744-bcf6981d3a05","Type":"ContainerDied","Data":"4444d93f68c351c0aae5349182126b3615263046ce9bb221bee588119677bf6a"} Oct 08 18:21:38 crc kubenswrapper[4859]: I1008 18:21:38.632866 4859 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-8z5ch" event={"ID":"557e8291-7e96-4634-b744-bcf6981d3a05","Type":"ContainerDied","Data":"bf3646c1c9400204a8151253765e05527f1418f9ec46904a6062307a265566b3"} Oct 08 18:21:38 crc kubenswrapper[4859]: I1008 18:21:38.632660 4859 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-8z5ch" Oct 08 18:21:38 crc kubenswrapper[4859]: I1008 18:21:38.636851 4859 generic.go:334] "Generic (PLEG): container finished" podID="438a5b9a-aaee-4e06-9d22-691862b9207e" containerID="8b6f3d58eacd7cf2bce7bfad476971fff5b2db73d55e67cbe99f2fe89dea2ac6" exitCode=0 Oct 08 18:21:38 crc kubenswrapper[4859]: I1008 18:21:38.636922 4859 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-c5tcx" event={"ID":"438a5b9a-aaee-4e06-9d22-691862b9207e","Type":"ContainerDied","Data":"8b6f3d58eacd7cf2bce7bfad476971fff5b2db73d55e67cbe99f2fe89dea2ac6"} Oct 08 18:21:38 crc kubenswrapper[4859]: I1008 18:21:38.636957 4859 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-c5tcx" event={"ID":"438a5b9a-aaee-4e06-9d22-691862b9207e","Type":"ContainerDied","Data":"018a03ca21455674a1ca2e2f60ca8e66b58ce6fc32a1698043d2788e8c022286"} Oct 08 18:21:38 crc kubenswrapper[4859]: I1008 18:21:38.637043 4859 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-c5tcx" Oct 08 18:21:38 crc kubenswrapper[4859]: I1008 18:21:38.640239 4859 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/marketplace-operator-79b997595-h4tnd" event={"ID":"8fdf9d2f-74e1-46e7-89ed-1947c3104386","Type":"ContainerStarted","Data":"1d3f7dea903a8c0b33235aa910b248a9172601456818b5263a6b2ece7414ec3f"} Oct 08 18:21:38 crc kubenswrapper[4859]: I1008 18:21:38.640285 4859 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/marketplace-operator-79b997595-h4tnd" event={"ID":"8fdf9d2f-74e1-46e7-89ed-1947c3104386","Type":"ContainerStarted","Data":"441b1155e5fd2e724875f2634786c37219a372ff426ab639075e38b0428fa54b"} Oct 08 18:21:38 crc kubenswrapper[4859]: I1008 18:21:38.640639 4859 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/marketplace-operator-79b997595-h4tnd" Oct 08 18:21:38 crc kubenswrapper[4859]: I1008 18:21:38.643723 4859 patch_prober.go:28] interesting pod/marketplace-operator-79b997595-h4tnd container/marketplace-operator namespace/openshift-marketplace: Readiness probe status=failure output="Get \"http://10.217.0.55:8080/healthz\": dial tcp 10.217.0.55:8080: connect: connection refused" start-of-body= Oct 08 18:21:38 crc kubenswrapper[4859]: I1008 18:21:38.643774 4859 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-marketplace/marketplace-operator-79b997595-h4tnd" podUID="8fdf9d2f-74e1-46e7-89ed-1947c3104386" containerName="marketplace-operator" probeResult="failure" output="Get \"http://10.217.0.55:8080/healthz\": dial tcp 10.217.0.55:8080: connect: connection refused" Oct 08 18:21:38 crc kubenswrapper[4859]: I1008 18:21:38.644978 4859 scope.go:117] "RemoveContainer" containerID="f63c1b5026a47ea0d969d06d88891d99f91b25c2954de93c4619b4e4dc5fe7ba" Oct 08 18:21:38 crc kubenswrapper[4859]: I1008 18:21:38.646790 4859 generic.go:334] "Generic (PLEG): container finished" podID="aeadfa85-a482-40a8-b947-09d2f6613900" containerID="30d7dafc5dc8f0a00b63102ca616e3537a9ab1f2b622798aead42d5f5fe48381" exitCode=0 Oct 08 18:21:38 crc kubenswrapper[4859]: I1008 18:21:38.646866 4859 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-xxgmr" event={"ID":"aeadfa85-a482-40a8-b947-09d2f6613900","Type":"ContainerDied","Data":"30d7dafc5dc8f0a00b63102ca616e3537a9ab1f2b622798aead42d5f5fe48381"} Oct 08 18:21:38 crc kubenswrapper[4859]: I1008 18:21:38.646913 4859 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-xxgmr" event={"ID":"aeadfa85-a482-40a8-b947-09d2f6613900","Type":"ContainerDied","Data":"343ad2ab2149bab5b80f2c2b5e34738d0110669b70b2f5af343ad42515c23e78"} Oct 08 18:21:38 crc kubenswrapper[4859]: I1008 18:21:38.647044 4859 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-xxgmr" Oct 08 18:21:38 crc kubenswrapper[4859]: I1008 18:21:38.651607 4859 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/marketplace-operator-79b997595-bz6v8"] Oct 08 18:21:38 crc kubenswrapper[4859]: I1008 18:21:38.654857 4859 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/marketplace-operator-79b997595-bz6v8"] Oct 08 18:21:38 crc kubenswrapper[4859]: I1008 18:21:38.662575 4859 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-zb82c"] Oct 08 18:21:38 crc kubenswrapper[4859]: I1008 18:21:38.666460 4859 scope.go:117] "RemoveContainer" containerID="e38579761dd4da27a2aec9a1a2a6138684ba3f36629898992d39947e3fd12c35" Oct 08 18:21:38 crc kubenswrapper[4859]: I1008 18:21:38.667815 4859 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/community-operators-zb82c"] Oct 08 18:21:38 crc kubenswrapper[4859]: I1008 18:21:38.683047 4859 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-8z5ch"] Oct 08 18:21:38 crc kubenswrapper[4859]: I1008 18:21:38.695329 4859 scope.go:117] "RemoveContainer" containerID="3b5490e97f3c5e1a24d3ba43457bebc4949c6182b88dfe9fdb226ee620caa614" Oct 08 18:21:38 crc kubenswrapper[4859]: E1008 18:21:38.696403 4859 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"3b5490e97f3c5e1a24d3ba43457bebc4949c6182b88dfe9fdb226ee620caa614\": container with ID starting with 3b5490e97f3c5e1a24d3ba43457bebc4949c6182b88dfe9fdb226ee620caa614 not found: ID does not exist" containerID="3b5490e97f3c5e1a24d3ba43457bebc4949c6182b88dfe9fdb226ee620caa614" Oct 08 18:21:38 crc kubenswrapper[4859]: I1008 18:21:38.696449 4859 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"3b5490e97f3c5e1a24d3ba43457bebc4949c6182b88dfe9fdb226ee620caa614"} err="failed to get container status \"3b5490e97f3c5e1a24d3ba43457bebc4949c6182b88dfe9fdb226ee620caa614\": rpc error: code = NotFound desc = could not find container \"3b5490e97f3c5e1a24d3ba43457bebc4949c6182b88dfe9fdb226ee620caa614\": container with ID starting with 3b5490e97f3c5e1a24d3ba43457bebc4949c6182b88dfe9fdb226ee620caa614 not found: ID does not exist" Oct 08 18:21:38 crc kubenswrapper[4859]: I1008 18:21:38.696484 4859 scope.go:117] "RemoveContainer" containerID="f63c1b5026a47ea0d969d06d88891d99f91b25c2954de93c4619b4e4dc5fe7ba" Oct 08 18:21:38 crc kubenswrapper[4859]: I1008 18:21:38.700090 4859 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-marketplace-8z5ch"] Oct 08 18:21:38 crc kubenswrapper[4859]: E1008 18:21:38.700097 4859 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"f63c1b5026a47ea0d969d06d88891d99f91b25c2954de93c4619b4e4dc5fe7ba\": container with ID starting with f63c1b5026a47ea0d969d06d88891d99f91b25c2954de93c4619b4e4dc5fe7ba not found: ID does not exist" containerID="f63c1b5026a47ea0d969d06d88891d99f91b25c2954de93c4619b4e4dc5fe7ba" Oct 08 18:21:38 crc kubenswrapper[4859]: I1008 18:21:38.700143 4859 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"f63c1b5026a47ea0d969d06d88891d99f91b25c2954de93c4619b4e4dc5fe7ba"} err="failed to get container status \"f63c1b5026a47ea0d969d06d88891d99f91b25c2954de93c4619b4e4dc5fe7ba\": rpc error: code = NotFound desc = could not find container \"f63c1b5026a47ea0d969d06d88891d99f91b25c2954de93c4619b4e4dc5fe7ba\": container with ID starting with f63c1b5026a47ea0d969d06d88891d99f91b25c2954de93c4619b4e4dc5fe7ba not found: ID does not exist" Oct 08 18:21:38 crc kubenswrapper[4859]: I1008 18:21:38.700183 4859 scope.go:117] "RemoveContainer" containerID="e38579761dd4da27a2aec9a1a2a6138684ba3f36629898992d39947e3fd12c35" Oct 08 18:21:38 crc kubenswrapper[4859]: E1008 18:21:38.700608 4859 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"e38579761dd4da27a2aec9a1a2a6138684ba3f36629898992d39947e3fd12c35\": container with ID starting with e38579761dd4da27a2aec9a1a2a6138684ba3f36629898992d39947e3fd12c35 not found: ID does not exist" containerID="e38579761dd4da27a2aec9a1a2a6138684ba3f36629898992d39947e3fd12c35" Oct 08 18:21:38 crc kubenswrapper[4859]: I1008 18:21:38.700650 4859 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"e38579761dd4da27a2aec9a1a2a6138684ba3f36629898992d39947e3fd12c35"} err="failed to get container status \"e38579761dd4da27a2aec9a1a2a6138684ba3f36629898992d39947e3fd12c35\": rpc error: code = NotFound desc = could not find container \"e38579761dd4da27a2aec9a1a2a6138684ba3f36629898992d39947e3fd12c35\": container with ID starting with e38579761dd4da27a2aec9a1a2a6138684ba3f36629898992d39947e3fd12c35 not found: ID does not exist" Oct 08 18:21:38 crc kubenswrapper[4859]: I1008 18:21:38.700737 4859 scope.go:117] "RemoveContainer" containerID="b6f47fb2611d35c2a1a47f8cdcf8133fc94d28820b6d8a1908a935ecfa0d3e2d" Oct 08 18:21:38 crc kubenswrapper[4859]: I1008 18:21:38.704483 4859 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-xxgmr"] Oct 08 18:21:38 crc kubenswrapper[4859]: I1008 18:21:38.710383 4859 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-operators-xxgmr"] Oct 08 18:21:38 crc kubenswrapper[4859]: I1008 18:21:38.716834 4859 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-c5tcx"] Oct 08 18:21:38 crc kubenswrapper[4859]: I1008 18:21:38.719922 4859 scope.go:117] "RemoveContainer" containerID="b6f47fb2611d35c2a1a47f8cdcf8133fc94d28820b6d8a1908a935ecfa0d3e2d" Oct 08 18:21:38 crc kubenswrapper[4859]: I1008 18:21:38.721517 4859 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/certified-operators-c5tcx"] Oct 08 18:21:38 crc kubenswrapper[4859]: E1008 18:21:38.721957 4859 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"b6f47fb2611d35c2a1a47f8cdcf8133fc94d28820b6d8a1908a935ecfa0d3e2d\": container with ID starting with b6f47fb2611d35c2a1a47f8cdcf8133fc94d28820b6d8a1908a935ecfa0d3e2d not found: ID does not exist" containerID="b6f47fb2611d35c2a1a47f8cdcf8133fc94d28820b6d8a1908a935ecfa0d3e2d" Oct 08 18:21:38 crc kubenswrapper[4859]: I1008 18:21:38.722005 4859 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"b6f47fb2611d35c2a1a47f8cdcf8133fc94d28820b6d8a1908a935ecfa0d3e2d"} err="failed to get container status \"b6f47fb2611d35c2a1a47f8cdcf8133fc94d28820b6d8a1908a935ecfa0d3e2d\": rpc error: code = NotFound desc = could not find container \"b6f47fb2611d35c2a1a47f8cdcf8133fc94d28820b6d8a1908a935ecfa0d3e2d\": container with ID starting with b6f47fb2611d35c2a1a47f8cdcf8133fc94d28820b6d8a1908a935ecfa0d3e2d not found: ID does not exist" Oct 08 18:21:38 crc kubenswrapper[4859]: I1008 18:21:38.722038 4859 scope.go:117] "RemoveContainer" containerID="4444d93f68c351c0aae5349182126b3615263046ce9bb221bee588119677bf6a" Oct 08 18:21:38 crc kubenswrapper[4859]: I1008 18:21:38.724960 4859 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/marketplace-operator-79b997595-h4tnd" podStartSLOduration=1.724930375 podStartE2EDuration="1.724930375s" podCreationTimestamp="2025-10-08 18:21:37 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-08 18:21:38.720555801 +0000 UTC m=+268.967395180" watchObservedRunningTime="2025-10-08 18:21:38.724930375 +0000 UTC m=+268.971769754" Oct 08 18:21:38 crc kubenswrapper[4859]: I1008 18:21:38.775242 4859 scope.go:117] "RemoveContainer" containerID="012460253b77cfdcb5001d7386ae3a7b69e158658002e1efdac3240841db6fc0" Oct 08 18:21:38 crc kubenswrapper[4859]: I1008 18:21:38.787968 4859 scope.go:117] "RemoveContainer" containerID="af8756f796c45216e3a3536e2f9927a86ac35371e46f68a4b91bb5c60b9e22b3" Oct 08 18:21:38 crc kubenswrapper[4859]: I1008 18:21:38.807307 4859 scope.go:117] "RemoveContainer" containerID="4444d93f68c351c0aae5349182126b3615263046ce9bb221bee588119677bf6a" Oct 08 18:21:38 crc kubenswrapper[4859]: E1008 18:21:38.807889 4859 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"4444d93f68c351c0aae5349182126b3615263046ce9bb221bee588119677bf6a\": container with ID starting with 4444d93f68c351c0aae5349182126b3615263046ce9bb221bee588119677bf6a not found: ID does not exist" containerID="4444d93f68c351c0aae5349182126b3615263046ce9bb221bee588119677bf6a" Oct 08 18:21:38 crc kubenswrapper[4859]: I1008 18:21:38.807929 4859 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"4444d93f68c351c0aae5349182126b3615263046ce9bb221bee588119677bf6a"} err="failed to get container status \"4444d93f68c351c0aae5349182126b3615263046ce9bb221bee588119677bf6a\": rpc error: code = NotFound desc = could not find container \"4444d93f68c351c0aae5349182126b3615263046ce9bb221bee588119677bf6a\": container with ID starting with 4444d93f68c351c0aae5349182126b3615263046ce9bb221bee588119677bf6a not found: ID does not exist" Oct 08 18:21:38 crc kubenswrapper[4859]: I1008 18:21:38.807956 4859 scope.go:117] "RemoveContainer" containerID="012460253b77cfdcb5001d7386ae3a7b69e158658002e1efdac3240841db6fc0" Oct 08 18:21:38 crc kubenswrapper[4859]: E1008 18:21:38.808286 4859 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"012460253b77cfdcb5001d7386ae3a7b69e158658002e1efdac3240841db6fc0\": container with ID starting with 012460253b77cfdcb5001d7386ae3a7b69e158658002e1efdac3240841db6fc0 not found: ID does not exist" containerID="012460253b77cfdcb5001d7386ae3a7b69e158658002e1efdac3240841db6fc0" Oct 08 18:21:38 crc kubenswrapper[4859]: I1008 18:21:38.808329 4859 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"012460253b77cfdcb5001d7386ae3a7b69e158658002e1efdac3240841db6fc0"} err="failed to get container status \"012460253b77cfdcb5001d7386ae3a7b69e158658002e1efdac3240841db6fc0\": rpc error: code = NotFound desc = could not find container \"012460253b77cfdcb5001d7386ae3a7b69e158658002e1efdac3240841db6fc0\": container with ID starting with 012460253b77cfdcb5001d7386ae3a7b69e158658002e1efdac3240841db6fc0 not found: ID does not exist" Oct 08 18:21:38 crc kubenswrapper[4859]: I1008 18:21:38.808360 4859 scope.go:117] "RemoveContainer" containerID="af8756f796c45216e3a3536e2f9927a86ac35371e46f68a4b91bb5c60b9e22b3" Oct 08 18:21:38 crc kubenswrapper[4859]: E1008 18:21:38.808781 4859 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"af8756f796c45216e3a3536e2f9927a86ac35371e46f68a4b91bb5c60b9e22b3\": container with ID starting with af8756f796c45216e3a3536e2f9927a86ac35371e46f68a4b91bb5c60b9e22b3 not found: ID does not exist" containerID="af8756f796c45216e3a3536e2f9927a86ac35371e46f68a4b91bb5c60b9e22b3" Oct 08 18:21:38 crc kubenswrapper[4859]: I1008 18:21:38.808847 4859 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"af8756f796c45216e3a3536e2f9927a86ac35371e46f68a4b91bb5c60b9e22b3"} err="failed to get container status \"af8756f796c45216e3a3536e2f9927a86ac35371e46f68a4b91bb5c60b9e22b3\": rpc error: code = NotFound desc = could not find container \"af8756f796c45216e3a3536e2f9927a86ac35371e46f68a4b91bb5c60b9e22b3\": container with ID starting with af8756f796c45216e3a3536e2f9927a86ac35371e46f68a4b91bb5c60b9e22b3 not found: ID does not exist" Oct 08 18:21:38 crc kubenswrapper[4859]: I1008 18:21:38.808884 4859 scope.go:117] "RemoveContainer" containerID="8b6f3d58eacd7cf2bce7bfad476971fff5b2db73d55e67cbe99f2fe89dea2ac6" Oct 08 18:21:38 crc kubenswrapper[4859]: I1008 18:21:38.831182 4859 scope.go:117] "RemoveContainer" containerID="3a5c39dbc2e6abe1c86f584dff2ab8edde1ccbdd76be2b5d7d17cef9c1396da8" Oct 08 18:21:38 crc kubenswrapper[4859]: I1008 18:21:38.848955 4859 scope.go:117] "RemoveContainer" containerID="223b3d675a30b7c8d3aba21fe8f0a9147dca08dbc807080e4d52889f52dd1da4" Oct 08 18:21:38 crc kubenswrapper[4859]: I1008 18:21:38.870116 4859 scope.go:117] "RemoveContainer" containerID="8b6f3d58eacd7cf2bce7bfad476971fff5b2db73d55e67cbe99f2fe89dea2ac6" Oct 08 18:21:38 crc kubenswrapper[4859]: E1008 18:21:38.870839 4859 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"8b6f3d58eacd7cf2bce7bfad476971fff5b2db73d55e67cbe99f2fe89dea2ac6\": container with ID starting with 8b6f3d58eacd7cf2bce7bfad476971fff5b2db73d55e67cbe99f2fe89dea2ac6 not found: ID does not exist" containerID="8b6f3d58eacd7cf2bce7bfad476971fff5b2db73d55e67cbe99f2fe89dea2ac6" Oct 08 18:21:38 crc kubenswrapper[4859]: I1008 18:21:38.870885 4859 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"8b6f3d58eacd7cf2bce7bfad476971fff5b2db73d55e67cbe99f2fe89dea2ac6"} err="failed to get container status \"8b6f3d58eacd7cf2bce7bfad476971fff5b2db73d55e67cbe99f2fe89dea2ac6\": rpc error: code = NotFound desc = could not find container \"8b6f3d58eacd7cf2bce7bfad476971fff5b2db73d55e67cbe99f2fe89dea2ac6\": container with ID starting with 8b6f3d58eacd7cf2bce7bfad476971fff5b2db73d55e67cbe99f2fe89dea2ac6 not found: ID does not exist" Oct 08 18:21:38 crc kubenswrapper[4859]: I1008 18:21:38.870915 4859 scope.go:117] "RemoveContainer" containerID="3a5c39dbc2e6abe1c86f584dff2ab8edde1ccbdd76be2b5d7d17cef9c1396da8" Oct 08 18:21:38 crc kubenswrapper[4859]: E1008 18:21:38.871473 4859 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"3a5c39dbc2e6abe1c86f584dff2ab8edde1ccbdd76be2b5d7d17cef9c1396da8\": container with ID starting with 3a5c39dbc2e6abe1c86f584dff2ab8edde1ccbdd76be2b5d7d17cef9c1396da8 not found: ID does not exist" containerID="3a5c39dbc2e6abe1c86f584dff2ab8edde1ccbdd76be2b5d7d17cef9c1396da8" Oct 08 18:21:38 crc kubenswrapper[4859]: I1008 18:21:38.871503 4859 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"3a5c39dbc2e6abe1c86f584dff2ab8edde1ccbdd76be2b5d7d17cef9c1396da8"} err="failed to get container status \"3a5c39dbc2e6abe1c86f584dff2ab8edde1ccbdd76be2b5d7d17cef9c1396da8\": rpc error: code = NotFound desc = could not find container \"3a5c39dbc2e6abe1c86f584dff2ab8edde1ccbdd76be2b5d7d17cef9c1396da8\": container with ID starting with 3a5c39dbc2e6abe1c86f584dff2ab8edde1ccbdd76be2b5d7d17cef9c1396da8 not found: ID does not exist" Oct 08 18:21:38 crc kubenswrapper[4859]: I1008 18:21:38.871526 4859 scope.go:117] "RemoveContainer" containerID="223b3d675a30b7c8d3aba21fe8f0a9147dca08dbc807080e4d52889f52dd1da4" Oct 08 18:21:38 crc kubenswrapper[4859]: E1008 18:21:38.871999 4859 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"223b3d675a30b7c8d3aba21fe8f0a9147dca08dbc807080e4d52889f52dd1da4\": container with ID starting with 223b3d675a30b7c8d3aba21fe8f0a9147dca08dbc807080e4d52889f52dd1da4 not found: ID does not exist" containerID="223b3d675a30b7c8d3aba21fe8f0a9147dca08dbc807080e4d52889f52dd1da4" Oct 08 18:21:38 crc kubenswrapper[4859]: I1008 18:21:38.872029 4859 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"223b3d675a30b7c8d3aba21fe8f0a9147dca08dbc807080e4d52889f52dd1da4"} err="failed to get container status \"223b3d675a30b7c8d3aba21fe8f0a9147dca08dbc807080e4d52889f52dd1da4\": rpc error: code = NotFound desc = could not find container \"223b3d675a30b7c8d3aba21fe8f0a9147dca08dbc807080e4d52889f52dd1da4\": container with ID starting with 223b3d675a30b7c8d3aba21fe8f0a9147dca08dbc807080e4d52889f52dd1da4 not found: ID does not exist" Oct 08 18:21:38 crc kubenswrapper[4859]: I1008 18:21:38.872069 4859 scope.go:117] "RemoveContainer" containerID="30d7dafc5dc8f0a00b63102ca616e3537a9ab1f2b622798aead42d5f5fe48381" Oct 08 18:21:38 crc kubenswrapper[4859]: I1008 18:21:38.885818 4859 scope.go:117] "RemoveContainer" containerID="2eb9a7a4f0c8c6dcb4c312d1f91dbe329a88ed3bf46e411097c4fe7c2595fa4f" Oct 08 18:21:38 crc kubenswrapper[4859]: I1008 18:21:38.900378 4859 scope.go:117] "RemoveContainer" containerID="86728cd32e96b87dd733a9b0e8329867e46b270d05e3305e491159fe58b2f03c" Oct 08 18:21:38 crc kubenswrapper[4859]: I1008 18:21:38.917730 4859 scope.go:117] "RemoveContainer" containerID="30d7dafc5dc8f0a00b63102ca616e3537a9ab1f2b622798aead42d5f5fe48381" Oct 08 18:21:38 crc kubenswrapper[4859]: E1008 18:21:38.918480 4859 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"30d7dafc5dc8f0a00b63102ca616e3537a9ab1f2b622798aead42d5f5fe48381\": container with ID starting with 30d7dafc5dc8f0a00b63102ca616e3537a9ab1f2b622798aead42d5f5fe48381 not found: ID does not exist" containerID="30d7dafc5dc8f0a00b63102ca616e3537a9ab1f2b622798aead42d5f5fe48381" Oct 08 18:21:38 crc kubenswrapper[4859]: I1008 18:21:38.918521 4859 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"30d7dafc5dc8f0a00b63102ca616e3537a9ab1f2b622798aead42d5f5fe48381"} err="failed to get container status \"30d7dafc5dc8f0a00b63102ca616e3537a9ab1f2b622798aead42d5f5fe48381\": rpc error: code = NotFound desc = could not find container \"30d7dafc5dc8f0a00b63102ca616e3537a9ab1f2b622798aead42d5f5fe48381\": container with ID starting with 30d7dafc5dc8f0a00b63102ca616e3537a9ab1f2b622798aead42d5f5fe48381 not found: ID does not exist" Oct 08 18:21:38 crc kubenswrapper[4859]: I1008 18:21:38.918549 4859 scope.go:117] "RemoveContainer" containerID="2eb9a7a4f0c8c6dcb4c312d1f91dbe329a88ed3bf46e411097c4fe7c2595fa4f" Oct 08 18:21:38 crc kubenswrapper[4859]: E1008 18:21:38.919276 4859 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"2eb9a7a4f0c8c6dcb4c312d1f91dbe329a88ed3bf46e411097c4fe7c2595fa4f\": container with ID starting with 2eb9a7a4f0c8c6dcb4c312d1f91dbe329a88ed3bf46e411097c4fe7c2595fa4f not found: ID does not exist" containerID="2eb9a7a4f0c8c6dcb4c312d1f91dbe329a88ed3bf46e411097c4fe7c2595fa4f" Oct 08 18:21:38 crc kubenswrapper[4859]: I1008 18:21:38.919316 4859 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"2eb9a7a4f0c8c6dcb4c312d1f91dbe329a88ed3bf46e411097c4fe7c2595fa4f"} err="failed to get container status \"2eb9a7a4f0c8c6dcb4c312d1f91dbe329a88ed3bf46e411097c4fe7c2595fa4f\": rpc error: code = NotFound desc = could not find container \"2eb9a7a4f0c8c6dcb4c312d1f91dbe329a88ed3bf46e411097c4fe7c2595fa4f\": container with ID starting with 2eb9a7a4f0c8c6dcb4c312d1f91dbe329a88ed3bf46e411097c4fe7c2595fa4f not found: ID does not exist" Oct 08 18:21:38 crc kubenswrapper[4859]: I1008 18:21:38.919348 4859 scope.go:117] "RemoveContainer" containerID="86728cd32e96b87dd733a9b0e8329867e46b270d05e3305e491159fe58b2f03c" Oct 08 18:21:38 crc kubenswrapper[4859]: E1008 18:21:38.919643 4859 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"86728cd32e96b87dd733a9b0e8329867e46b270d05e3305e491159fe58b2f03c\": container with ID starting with 86728cd32e96b87dd733a9b0e8329867e46b270d05e3305e491159fe58b2f03c not found: ID does not exist" containerID="86728cd32e96b87dd733a9b0e8329867e46b270d05e3305e491159fe58b2f03c" Oct 08 18:21:38 crc kubenswrapper[4859]: I1008 18:21:38.919671 4859 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"86728cd32e96b87dd733a9b0e8329867e46b270d05e3305e491159fe58b2f03c"} err="failed to get container status \"86728cd32e96b87dd733a9b0e8329867e46b270d05e3305e491159fe58b2f03c\": rpc error: code = NotFound desc = could not find container \"86728cd32e96b87dd733a9b0e8329867e46b270d05e3305e491159fe58b2f03c\": container with ID starting with 86728cd32e96b87dd733a9b0e8329867e46b270d05e3305e491159fe58b2f03c not found: ID does not exist" Oct 08 18:21:39 crc kubenswrapper[4859]: I1008 18:21:39.661325 4859 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/marketplace-operator-79b997595-h4tnd" Oct 08 18:21:39 crc kubenswrapper[4859]: I1008 18:21:39.809487 4859 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-wlwhh"] Oct 08 18:21:39 crc kubenswrapper[4859]: E1008 18:21:39.810054 4859 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="aeadfa85-a482-40a8-b947-09d2f6613900" containerName="extract-utilities" Oct 08 18:21:39 crc kubenswrapper[4859]: I1008 18:21:39.810074 4859 state_mem.go:107] "Deleted CPUSet assignment" podUID="aeadfa85-a482-40a8-b947-09d2f6613900" containerName="extract-utilities" Oct 08 18:21:39 crc kubenswrapper[4859]: E1008 18:21:39.810085 4859 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5b68d592-4a46-4fbc-bcec-a8bca2b973aa" containerName="registry-server" Oct 08 18:21:39 crc kubenswrapper[4859]: I1008 18:21:39.810092 4859 state_mem.go:107] "Deleted CPUSet assignment" podUID="5b68d592-4a46-4fbc-bcec-a8bca2b973aa" containerName="registry-server" Oct 08 18:21:39 crc kubenswrapper[4859]: E1008 18:21:39.810103 4859 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8d4b5b44-816f-44f6-98ab-c5d66aefaa08" containerName="marketplace-operator" Oct 08 18:21:39 crc kubenswrapper[4859]: I1008 18:21:39.810110 4859 state_mem.go:107] "Deleted CPUSet assignment" podUID="8d4b5b44-816f-44f6-98ab-c5d66aefaa08" containerName="marketplace-operator" Oct 08 18:21:39 crc kubenswrapper[4859]: E1008 18:21:39.810120 4859 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5b68d592-4a46-4fbc-bcec-a8bca2b973aa" containerName="extract-utilities" Oct 08 18:21:39 crc kubenswrapper[4859]: I1008 18:21:39.810126 4859 state_mem.go:107] "Deleted CPUSet assignment" podUID="5b68d592-4a46-4fbc-bcec-a8bca2b973aa" containerName="extract-utilities" Oct 08 18:21:39 crc kubenswrapper[4859]: E1008 18:21:39.810135 4859 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="438a5b9a-aaee-4e06-9d22-691862b9207e" containerName="extract-content" Oct 08 18:21:39 crc kubenswrapper[4859]: I1008 18:21:39.810140 4859 state_mem.go:107] "Deleted CPUSet assignment" podUID="438a5b9a-aaee-4e06-9d22-691862b9207e" containerName="extract-content" Oct 08 18:21:39 crc kubenswrapper[4859]: E1008 18:21:39.810149 4859 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="aeadfa85-a482-40a8-b947-09d2f6613900" containerName="registry-server" Oct 08 18:21:39 crc kubenswrapper[4859]: I1008 18:21:39.810158 4859 state_mem.go:107] "Deleted CPUSet assignment" podUID="aeadfa85-a482-40a8-b947-09d2f6613900" containerName="registry-server" Oct 08 18:21:39 crc kubenswrapper[4859]: E1008 18:21:39.810169 4859 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5b68d592-4a46-4fbc-bcec-a8bca2b973aa" containerName="extract-content" Oct 08 18:21:39 crc kubenswrapper[4859]: I1008 18:21:39.810176 4859 state_mem.go:107] "Deleted CPUSet assignment" podUID="5b68d592-4a46-4fbc-bcec-a8bca2b973aa" containerName="extract-content" Oct 08 18:21:39 crc kubenswrapper[4859]: E1008 18:21:39.810183 4859 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="557e8291-7e96-4634-b744-bcf6981d3a05" containerName="registry-server" Oct 08 18:21:39 crc kubenswrapper[4859]: I1008 18:21:39.810189 4859 state_mem.go:107] "Deleted CPUSet assignment" podUID="557e8291-7e96-4634-b744-bcf6981d3a05" containerName="registry-server" Oct 08 18:21:39 crc kubenswrapper[4859]: E1008 18:21:39.810200 4859 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="438a5b9a-aaee-4e06-9d22-691862b9207e" containerName="registry-server" Oct 08 18:21:39 crc kubenswrapper[4859]: I1008 18:21:39.810206 4859 state_mem.go:107] "Deleted CPUSet assignment" podUID="438a5b9a-aaee-4e06-9d22-691862b9207e" containerName="registry-server" Oct 08 18:21:39 crc kubenswrapper[4859]: E1008 18:21:39.810212 4859 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="aeadfa85-a482-40a8-b947-09d2f6613900" containerName="extract-content" Oct 08 18:21:39 crc kubenswrapper[4859]: I1008 18:21:39.810218 4859 state_mem.go:107] "Deleted CPUSet assignment" podUID="aeadfa85-a482-40a8-b947-09d2f6613900" containerName="extract-content" Oct 08 18:21:39 crc kubenswrapper[4859]: E1008 18:21:39.810224 4859 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="557e8291-7e96-4634-b744-bcf6981d3a05" containerName="extract-content" Oct 08 18:21:39 crc kubenswrapper[4859]: I1008 18:21:39.810230 4859 state_mem.go:107] "Deleted CPUSet assignment" podUID="557e8291-7e96-4634-b744-bcf6981d3a05" containerName="extract-content" Oct 08 18:21:39 crc kubenswrapper[4859]: E1008 18:21:39.810240 4859 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="557e8291-7e96-4634-b744-bcf6981d3a05" containerName="extract-utilities" Oct 08 18:21:39 crc kubenswrapper[4859]: I1008 18:21:39.810245 4859 state_mem.go:107] "Deleted CPUSet assignment" podUID="557e8291-7e96-4634-b744-bcf6981d3a05" containerName="extract-utilities" Oct 08 18:21:39 crc kubenswrapper[4859]: E1008 18:21:39.810261 4859 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="438a5b9a-aaee-4e06-9d22-691862b9207e" containerName="extract-utilities" Oct 08 18:21:39 crc kubenswrapper[4859]: I1008 18:21:39.810267 4859 state_mem.go:107] "Deleted CPUSet assignment" podUID="438a5b9a-aaee-4e06-9d22-691862b9207e" containerName="extract-utilities" Oct 08 18:21:39 crc kubenswrapper[4859]: I1008 18:21:39.810354 4859 memory_manager.go:354] "RemoveStaleState removing state" podUID="5b68d592-4a46-4fbc-bcec-a8bca2b973aa" containerName="registry-server" Oct 08 18:21:39 crc kubenswrapper[4859]: I1008 18:21:39.810368 4859 memory_manager.go:354] "RemoveStaleState removing state" podUID="557e8291-7e96-4634-b744-bcf6981d3a05" containerName="registry-server" Oct 08 18:21:39 crc kubenswrapper[4859]: I1008 18:21:39.810375 4859 memory_manager.go:354] "RemoveStaleState removing state" podUID="438a5b9a-aaee-4e06-9d22-691862b9207e" containerName="registry-server" Oct 08 18:21:39 crc kubenswrapper[4859]: I1008 18:21:39.810382 4859 memory_manager.go:354] "RemoveStaleState removing state" podUID="8d4b5b44-816f-44f6-98ab-c5d66aefaa08" containerName="marketplace-operator" Oct 08 18:21:39 crc kubenswrapper[4859]: I1008 18:21:39.810388 4859 memory_manager.go:354] "RemoveStaleState removing state" podUID="aeadfa85-a482-40a8-b947-09d2f6613900" containerName="registry-server" Oct 08 18:21:39 crc kubenswrapper[4859]: I1008 18:21:39.811884 4859 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-wlwhh" Oct 08 18:21:39 crc kubenswrapper[4859]: I1008 18:21:39.813658 4859 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"redhat-marketplace-dockercfg-x2ctb" Oct 08 18:21:39 crc kubenswrapper[4859]: I1008 18:21:39.823204 4859 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-wlwhh"] Oct 08 18:21:39 crc kubenswrapper[4859]: I1008 18:21:39.927252 4859 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/997d41aa-a006-46d1-b488-b6d29186874d-utilities\") pod \"redhat-marketplace-wlwhh\" (UID: \"997d41aa-a006-46d1-b488-b6d29186874d\") " pod="openshift-marketplace/redhat-marketplace-wlwhh" Oct 08 18:21:39 crc kubenswrapper[4859]: I1008 18:21:39.927350 4859 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/997d41aa-a006-46d1-b488-b6d29186874d-catalog-content\") pod \"redhat-marketplace-wlwhh\" (UID: \"997d41aa-a006-46d1-b488-b6d29186874d\") " pod="openshift-marketplace/redhat-marketplace-wlwhh" Oct 08 18:21:39 crc kubenswrapper[4859]: I1008 18:21:39.927606 4859 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-bwdgv\" (UniqueName: \"kubernetes.io/projected/997d41aa-a006-46d1-b488-b6d29186874d-kube-api-access-bwdgv\") pod \"redhat-marketplace-wlwhh\" (UID: \"997d41aa-a006-46d1-b488-b6d29186874d\") " pod="openshift-marketplace/redhat-marketplace-wlwhh" Oct 08 18:21:40 crc kubenswrapper[4859]: I1008 18:21:40.029746 4859 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/997d41aa-a006-46d1-b488-b6d29186874d-utilities\") pod \"redhat-marketplace-wlwhh\" (UID: \"997d41aa-a006-46d1-b488-b6d29186874d\") " pod="openshift-marketplace/redhat-marketplace-wlwhh" Oct 08 18:21:40 crc kubenswrapper[4859]: I1008 18:21:40.029832 4859 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/997d41aa-a006-46d1-b488-b6d29186874d-catalog-content\") pod \"redhat-marketplace-wlwhh\" (UID: \"997d41aa-a006-46d1-b488-b6d29186874d\") " pod="openshift-marketplace/redhat-marketplace-wlwhh" Oct 08 18:21:40 crc kubenswrapper[4859]: I1008 18:21:40.029891 4859 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-bwdgv\" (UniqueName: \"kubernetes.io/projected/997d41aa-a006-46d1-b488-b6d29186874d-kube-api-access-bwdgv\") pod \"redhat-marketplace-wlwhh\" (UID: \"997d41aa-a006-46d1-b488-b6d29186874d\") " pod="openshift-marketplace/redhat-marketplace-wlwhh" Oct 08 18:21:40 crc kubenswrapper[4859]: I1008 18:21:40.030824 4859 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/997d41aa-a006-46d1-b488-b6d29186874d-utilities\") pod \"redhat-marketplace-wlwhh\" (UID: \"997d41aa-a006-46d1-b488-b6d29186874d\") " pod="openshift-marketplace/redhat-marketplace-wlwhh" Oct 08 18:21:40 crc kubenswrapper[4859]: I1008 18:21:40.031228 4859 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/997d41aa-a006-46d1-b488-b6d29186874d-catalog-content\") pod \"redhat-marketplace-wlwhh\" (UID: \"997d41aa-a006-46d1-b488-b6d29186874d\") " pod="openshift-marketplace/redhat-marketplace-wlwhh" Oct 08 18:21:40 crc kubenswrapper[4859]: I1008 18:21:40.031441 4859 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-d8bkm"] Oct 08 18:21:40 crc kubenswrapper[4859]: I1008 18:21:40.034794 4859 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-d8bkm" Oct 08 18:21:40 crc kubenswrapper[4859]: I1008 18:21:40.038168 4859 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"certified-operators-dockercfg-4rs5g" Oct 08 18:21:40 crc kubenswrapper[4859]: I1008 18:21:40.039764 4859 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-d8bkm"] Oct 08 18:21:40 crc kubenswrapper[4859]: I1008 18:21:40.059428 4859 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-bwdgv\" (UniqueName: \"kubernetes.io/projected/997d41aa-a006-46d1-b488-b6d29186874d-kube-api-access-bwdgv\") pod \"redhat-marketplace-wlwhh\" (UID: \"997d41aa-a006-46d1-b488-b6d29186874d\") " pod="openshift-marketplace/redhat-marketplace-wlwhh" Oct 08 18:21:40 crc kubenswrapper[4859]: I1008 18:21:40.131394 4859 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-77jtv\" (UniqueName: \"kubernetes.io/projected/6e7e5636-231d-4788-9277-83549fc6b404-kube-api-access-77jtv\") pod \"certified-operators-d8bkm\" (UID: \"6e7e5636-231d-4788-9277-83549fc6b404\") " pod="openshift-marketplace/certified-operators-d8bkm" Oct 08 18:21:40 crc kubenswrapper[4859]: I1008 18:21:40.131784 4859 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/6e7e5636-231d-4788-9277-83549fc6b404-catalog-content\") pod \"certified-operators-d8bkm\" (UID: \"6e7e5636-231d-4788-9277-83549fc6b404\") " pod="openshift-marketplace/certified-operators-d8bkm" Oct 08 18:21:40 crc kubenswrapper[4859]: I1008 18:21:40.131925 4859 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/6e7e5636-231d-4788-9277-83549fc6b404-utilities\") pod \"certified-operators-d8bkm\" (UID: \"6e7e5636-231d-4788-9277-83549fc6b404\") " pod="openshift-marketplace/certified-operators-d8bkm" Oct 08 18:21:40 crc kubenswrapper[4859]: I1008 18:21:40.136922 4859 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-wlwhh" Oct 08 18:21:40 crc kubenswrapper[4859]: I1008 18:21:40.237318 4859 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/6e7e5636-231d-4788-9277-83549fc6b404-utilities\") pod \"certified-operators-d8bkm\" (UID: \"6e7e5636-231d-4788-9277-83549fc6b404\") " pod="openshift-marketplace/certified-operators-d8bkm" Oct 08 18:21:40 crc kubenswrapper[4859]: I1008 18:21:40.237509 4859 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-77jtv\" (UniqueName: \"kubernetes.io/projected/6e7e5636-231d-4788-9277-83549fc6b404-kube-api-access-77jtv\") pod \"certified-operators-d8bkm\" (UID: \"6e7e5636-231d-4788-9277-83549fc6b404\") " pod="openshift-marketplace/certified-operators-d8bkm" Oct 08 18:21:40 crc kubenswrapper[4859]: I1008 18:21:40.237570 4859 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/6e7e5636-231d-4788-9277-83549fc6b404-catalog-content\") pod \"certified-operators-d8bkm\" (UID: \"6e7e5636-231d-4788-9277-83549fc6b404\") " pod="openshift-marketplace/certified-operators-d8bkm" Oct 08 18:21:40 crc kubenswrapper[4859]: I1008 18:21:40.238202 4859 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/6e7e5636-231d-4788-9277-83549fc6b404-utilities\") pod \"certified-operators-d8bkm\" (UID: \"6e7e5636-231d-4788-9277-83549fc6b404\") " pod="openshift-marketplace/certified-operators-d8bkm" Oct 08 18:21:40 crc kubenswrapper[4859]: I1008 18:21:40.238555 4859 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/6e7e5636-231d-4788-9277-83549fc6b404-catalog-content\") pod \"certified-operators-d8bkm\" (UID: \"6e7e5636-231d-4788-9277-83549fc6b404\") " pod="openshift-marketplace/certified-operators-d8bkm" Oct 08 18:21:40 crc kubenswrapper[4859]: I1008 18:21:40.257603 4859 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-77jtv\" (UniqueName: \"kubernetes.io/projected/6e7e5636-231d-4788-9277-83549fc6b404-kube-api-access-77jtv\") pod \"certified-operators-d8bkm\" (UID: \"6e7e5636-231d-4788-9277-83549fc6b404\") " pod="openshift-marketplace/certified-operators-d8bkm" Oct 08 18:21:40 crc kubenswrapper[4859]: I1008 18:21:40.339998 4859 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-wlwhh"] Oct 08 18:21:40 crc kubenswrapper[4859]: I1008 18:21:40.406005 4859 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-d8bkm" Oct 08 18:21:40 crc kubenswrapper[4859]: I1008 18:21:40.476676 4859 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="438a5b9a-aaee-4e06-9d22-691862b9207e" path="/var/lib/kubelet/pods/438a5b9a-aaee-4e06-9d22-691862b9207e/volumes" Oct 08 18:21:40 crc kubenswrapper[4859]: I1008 18:21:40.478181 4859 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="557e8291-7e96-4634-b744-bcf6981d3a05" path="/var/lib/kubelet/pods/557e8291-7e96-4634-b744-bcf6981d3a05/volumes" Oct 08 18:21:40 crc kubenswrapper[4859]: I1008 18:21:40.478963 4859 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="5b68d592-4a46-4fbc-bcec-a8bca2b973aa" path="/var/lib/kubelet/pods/5b68d592-4a46-4fbc-bcec-a8bca2b973aa/volumes" Oct 08 18:21:40 crc kubenswrapper[4859]: I1008 18:21:40.480459 4859 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="8d4b5b44-816f-44f6-98ab-c5d66aefaa08" path="/var/lib/kubelet/pods/8d4b5b44-816f-44f6-98ab-c5d66aefaa08/volumes" Oct 08 18:21:40 crc kubenswrapper[4859]: I1008 18:21:40.480925 4859 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="aeadfa85-a482-40a8-b947-09d2f6613900" path="/var/lib/kubelet/pods/aeadfa85-a482-40a8-b947-09d2f6613900/volumes" Oct 08 18:21:40 crc kubenswrapper[4859]: I1008 18:21:40.643544 4859 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-d8bkm"] Oct 08 18:21:40 crc kubenswrapper[4859]: W1008 18:21:40.653459 4859 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod6e7e5636_231d_4788_9277_83549fc6b404.slice/crio-936d84e5057fdc9a8dd2b7c0f737180a06c1291be6ac02e33311cda0c86a7b8f WatchSource:0}: Error finding container 936d84e5057fdc9a8dd2b7c0f737180a06c1291be6ac02e33311cda0c86a7b8f: Status 404 returned error can't find the container with id 936d84e5057fdc9a8dd2b7c0f737180a06c1291be6ac02e33311cda0c86a7b8f Oct 08 18:21:40 crc kubenswrapper[4859]: I1008 18:21:40.662776 4859 generic.go:334] "Generic (PLEG): container finished" podID="997d41aa-a006-46d1-b488-b6d29186874d" containerID="5dff21af6517f36a662449b209a4df4b953bde8931bb3bef7f452928f8b34fbd" exitCode=0 Oct 08 18:21:40 crc kubenswrapper[4859]: I1008 18:21:40.662841 4859 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-wlwhh" event={"ID":"997d41aa-a006-46d1-b488-b6d29186874d","Type":"ContainerDied","Data":"5dff21af6517f36a662449b209a4df4b953bde8931bb3bef7f452928f8b34fbd"} Oct 08 18:21:40 crc kubenswrapper[4859]: I1008 18:21:40.662871 4859 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-wlwhh" event={"ID":"997d41aa-a006-46d1-b488-b6d29186874d","Type":"ContainerStarted","Data":"4b354ca15d3ea024c857fa892df20342b94d80c7917689d824203de3890e6966"} Oct 08 18:21:40 crc kubenswrapper[4859]: I1008 18:21:40.664908 4859 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-d8bkm" event={"ID":"6e7e5636-231d-4788-9277-83549fc6b404","Type":"ContainerStarted","Data":"936d84e5057fdc9a8dd2b7c0f737180a06c1291be6ac02e33311cda0c86a7b8f"} Oct 08 18:21:41 crc kubenswrapper[4859]: I1008 18:21:41.673639 4859 generic.go:334] "Generic (PLEG): container finished" podID="6e7e5636-231d-4788-9277-83549fc6b404" containerID="6ab5d0f7ad6ae783f669061db7dff5588e3b47f761bc17643bd81abc4e9ad04e" exitCode=0 Oct 08 18:21:41 crc kubenswrapper[4859]: I1008 18:21:41.673739 4859 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-d8bkm" event={"ID":"6e7e5636-231d-4788-9277-83549fc6b404","Type":"ContainerDied","Data":"6ab5d0f7ad6ae783f669061db7dff5588e3b47f761bc17643bd81abc4e9ad04e"} Oct 08 18:21:42 crc kubenswrapper[4859]: I1008 18:21:42.214545 4859 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-hlqjs"] Oct 08 18:21:42 crc kubenswrapper[4859]: I1008 18:21:42.216951 4859 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-hlqjs" Oct 08 18:21:42 crc kubenswrapper[4859]: I1008 18:21:42.220092 4859 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"redhat-operators-dockercfg-ct8rh" Oct 08 18:21:42 crc kubenswrapper[4859]: I1008 18:21:42.232159 4859 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-hlqjs"] Oct 08 18:21:42 crc kubenswrapper[4859]: I1008 18:21:42.374259 4859 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/748a4b84-e115-444a-b0f8-24029e8d1896-catalog-content\") pod \"redhat-operators-hlqjs\" (UID: \"748a4b84-e115-444a-b0f8-24029e8d1896\") " pod="openshift-marketplace/redhat-operators-hlqjs" Oct 08 18:21:42 crc kubenswrapper[4859]: I1008 18:21:42.374544 4859 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-dt5n7\" (UniqueName: \"kubernetes.io/projected/748a4b84-e115-444a-b0f8-24029e8d1896-kube-api-access-dt5n7\") pod \"redhat-operators-hlqjs\" (UID: \"748a4b84-e115-444a-b0f8-24029e8d1896\") " pod="openshift-marketplace/redhat-operators-hlqjs" Oct 08 18:21:42 crc kubenswrapper[4859]: I1008 18:21:42.374628 4859 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/748a4b84-e115-444a-b0f8-24029e8d1896-utilities\") pod \"redhat-operators-hlqjs\" (UID: \"748a4b84-e115-444a-b0f8-24029e8d1896\") " pod="openshift-marketplace/redhat-operators-hlqjs" Oct 08 18:21:42 crc kubenswrapper[4859]: I1008 18:21:42.412415 4859 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-p7bgm"] Oct 08 18:21:42 crc kubenswrapper[4859]: I1008 18:21:42.413728 4859 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-p7bgm" Oct 08 18:21:42 crc kubenswrapper[4859]: I1008 18:21:42.419139 4859 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"community-operators-dockercfg-dmngl" Oct 08 18:21:42 crc kubenswrapper[4859]: I1008 18:21:42.426185 4859 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-p7bgm"] Oct 08 18:21:42 crc kubenswrapper[4859]: I1008 18:21:42.475365 4859 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/748a4b84-e115-444a-b0f8-24029e8d1896-catalog-content\") pod \"redhat-operators-hlqjs\" (UID: \"748a4b84-e115-444a-b0f8-24029e8d1896\") " pod="openshift-marketplace/redhat-operators-hlqjs" Oct 08 18:21:42 crc kubenswrapper[4859]: I1008 18:21:42.475575 4859 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-dt5n7\" (UniqueName: \"kubernetes.io/projected/748a4b84-e115-444a-b0f8-24029e8d1896-kube-api-access-dt5n7\") pod \"redhat-operators-hlqjs\" (UID: \"748a4b84-e115-444a-b0f8-24029e8d1896\") " pod="openshift-marketplace/redhat-operators-hlqjs" Oct 08 18:21:42 crc kubenswrapper[4859]: I1008 18:21:42.475956 4859 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/748a4b84-e115-444a-b0f8-24029e8d1896-utilities\") pod \"redhat-operators-hlqjs\" (UID: \"748a4b84-e115-444a-b0f8-24029e8d1896\") " pod="openshift-marketplace/redhat-operators-hlqjs" Oct 08 18:21:42 crc kubenswrapper[4859]: I1008 18:21:42.475851 4859 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/748a4b84-e115-444a-b0f8-24029e8d1896-catalog-content\") pod \"redhat-operators-hlqjs\" (UID: \"748a4b84-e115-444a-b0f8-24029e8d1896\") " pod="openshift-marketplace/redhat-operators-hlqjs" Oct 08 18:21:42 crc kubenswrapper[4859]: I1008 18:21:42.476321 4859 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/748a4b84-e115-444a-b0f8-24029e8d1896-utilities\") pod \"redhat-operators-hlqjs\" (UID: \"748a4b84-e115-444a-b0f8-24029e8d1896\") " pod="openshift-marketplace/redhat-operators-hlqjs" Oct 08 18:21:42 crc kubenswrapper[4859]: I1008 18:21:42.497926 4859 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-dt5n7\" (UniqueName: \"kubernetes.io/projected/748a4b84-e115-444a-b0f8-24029e8d1896-kube-api-access-dt5n7\") pod \"redhat-operators-hlqjs\" (UID: \"748a4b84-e115-444a-b0f8-24029e8d1896\") " pod="openshift-marketplace/redhat-operators-hlqjs" Oct 08 18:21:42 crc kubenswrapper[4859]: I1008 18:21:42.537889 4859 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-hlqjs" Oct 08 18:21:42 crc kubenswrapper[4859]: I1008 18:21:42.577823 4859 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-2cdq2\" (UniqueName: \"kubernetes.io/projected/ad354068-21f9-40aa-a2ae-739f84e32469-kube-api-access-2cdq2\") pod \"community-operators-p7bgm\" (UID: \"ad354068-21f9-40aa-a2ae-739f84e32469\") " pod="openshift-marketplace/community-operators-p7bgm" Oct 08 18:21:42 crc kubenswrapper[4859]: I1008 18:21:42.577936 4859 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/ad354068-21f9-40aa-a2ae-739f84e32469-utilities\") pod \"community-operators-p7bgm\" (UID: \"ad354068-21f9-40aa-a2ae-739f84e32469\") " pod="openshift-marketplace/community-operators-p7bgm" Oct 08 18:21:42 crc kubenswrapper[4859]: I1008 18:21:42.578016 4859 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/ad354068-21f9-40aa-a2ae-739f84e32469-catalog-content\") pod \"community-operators-p7bgm\" (UID: \"ad354068-21f9-40aa-a2ae-739f84e32469\") " pod="openshift-marketplace/community-operators-p7bgm" Oct 08 18:21:42 crc kubenswrapper[4859]: I1008 18:21:42.697034 4859 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/ad354068-21f9-40aa-a2ae-739f84e32469-utilities\") pod \"community-operators-p7bgm\" (UID: \"ad354068-21f9-40aa-a2ae-739f84e32469\") " pod="openshift-marketplace/community-operators-p7bgm" Oct 08 18:21:42 crc kubenswrapper[4859]: I1008 18:21:42.697581 4859 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/ad354068-21f9-40aa-a2ae-739f84e32469-catalog-content\") pod \"community-operators-p7bgm\" (UID: \"ad354068-21f9-40aa-a2ae-739f84e32469\") " pod="openshift-marketplace/community-operators-p7bgm" Oct 08 18:21:42 crc kubenswrapper[4859]: I1008 18:21:42.697678 4859 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-2cdq2\" (UniqueName: \"kubernetes.io/projected/ad354068-21f9-40aa-a2ae-739f84e32469-kube-api-access-2cdq2\") pod \"community-operators-p7bgm\" (UID: \"ad354068-21f9-40aa-a2ae-739f84e32469\") " pod="openshift-marketplace/community-operators-p7bgm" Oct 08 18:21:42 crc kubenswrapper[4859]: I1008 18:21:42.697895 4859 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/ad354068-21f9-40aa-a2ae-739f84e32469-utilities\") pod \"community-operators-p7bgm\" (UID: \"ad354068-21f9-40aa-a2ae-739f84e32469\") " pod="openshift-marketplace/community-operators-p7bgm" Oct 08 18:21:42 crc kubenswrapper[4859]: I1008 18:21:42.697970 4859 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/ad354068-21f9-40aa-a2ae-739f84e32469-catalog-content\") pod \"community-operators-p7bgm\" (UID: \"ad354068-21f9-40aa-a2ae-739f84e32469\") " pod="openshift-marketplace/community-operators-p7bgm" Oct 08 18:21:42 crc kubenswrapper[4859]: I1008 18:21:42.700228 4859 generic.go:334] "Generic (PLEG): container finished" podID="997d41aa-a006-46d1-b488-b6d29186874d" containerID="c7a1d033cb16e90f3740ff9cbfe50c3d20f459a5e29c05b3822aabb07a59b272" exitCode=0 Oct 08 18:21:42 crc kubenswrapper[4859]: I1008 18:21:42.700597 4859 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-wlwhh" event={"ID":"997d41aa-a006-46d1-b488-b6d29186874d","Type":"ContainerDied","Data":"c7a1d033cb16e90f3740ff9cbfe50c3d20f459a5e29c05b3822aabb07a59b272"} Oct 08 18:21:42 crc kubenswrapper[4859]: I1008 18:21:42.713102 4859 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-d8bkm" event={"ID":"6e7e5636-231d-4788-9277-83549fc6b404","Type":"ContainerStarted","Data":"1c86c0f4e566c3320dc26f8fcfce719ec38669dddd685a982529c7885c384e83"} Oct 08 18:21:42 crc kubenswrapper[4859]: I1008 18:21:42.727676 4859 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-2cdq2\" (UniqueName: \"kubernetes.io/projected/ad354068-21f9-40aa-a2ae-739f84e32469-kube-api-access-2cdq2\") pod \"community-operators-p7bgm\" (UID: \"ad354068-21f9-40aa-a2ae-739f84e32469\") " pod="openshift-marketplace/community-operators-p7bgm" Oct 08 18:21:42 crc kubenswrapper[4859]: I1008 18:21:42.730700 4859 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-p7bgm" Oct 08 18:21:42 crc kubenswrapper[4859]: I1008 18:21:42.769878 4859 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-hlqjs"] Oct 08 18:21:42 crc kubenswrapper[4859]: W1008 18:21:42.777200 4859 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod748a4b84_e115_444a_b0f8_24029e8d1896.slice/crio-dc547304710591ee37c52b9b6363d72e12ea46aee21e48b975889092967234ae WatchSource:0}: Error finding container dc547304710591ee37c52b9b6363d72e12ea46aee21e48b975889092967234ae: Status 404 returned error can't find the container with id dc547304710591ee37c52b9b6363d72e12ea46aee21e48b975889092967234ae Oct 08 18:21:42 crc kubenswrapper[4859]: I1008 18:21:42.938349 4859 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-p7bgm"] Oct 08 18:21:42 crc kubenswrapper[4859]: W1008 18:21:42.948195 4859 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podad354068_21f9_40aa_a2ae_739f84e32469.slice/crio-51f81cca9ae9970df7d9a2a8e72c8d924ba1a867d65dea921fac5f8220297f56 WatchSource:0}: Error finding container 51f81cca9ae9970df7d9a2a8e72c8d924ba1a867d65dea921fac5f8220297f56: Status 404 returned error can't find the container with id 51f81cca9ae9970df7d9a2a8e72c8d924ba1a867d65dea921fac5f8220297f56 Oct 08 18:21:43 crc kubenswrapper[4859]: E1008 18:21:43.121229 4859 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod748a4b84_e115_444a_b0f8_24029e8d1896.slice/crio-conmon-b307190deb9033186860551b4bcb369baab214578419b6e6b1093458c8641517.scope\": RecentStats: unable to find data in memory cache]" Oct 08 18:21:43 crc kubenswrapper[4859]: I1008 18:21:43.720454 4859 generic.go:334] "Generic (PLEG): container finished" podID="6e7e5636-231d-4788-9277-83549fc6b404" containerID="1c86c0f4e566c3320dc26f8fcfce719ec38669dddd685a982529c7885c384e83" exitCode=0 Oct 08 18:21:43 crc kubenswrapper[4859]: I1008 18:21:43.721196 4859 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-d8bkm" event={"ID":"6e7e5636-231d-4788-9277-83549fc6b404","Type":"ContainerDied","Data":"1c86c0f4e566c3320dc26f8fcfce719ec38669dddd685a982529c7885c384e83"} Oct 08 18:21:43 crc kubenswrapper[4859]: I1008 18:21:43.726272 4859 generic.go:334] "Generic (PLEG): container finished" podID="ad354068-21f9-40aa-a2ae-739f84e32469" containerID="925d34b56e3fa96c4a2415fdaeda4a44b5cc43b1138ff186637c877733c5cf89" exitCode=0 Oct 08 18:21:43 crc kubenswrapper[4859]: I1008 18:21:43.726399 4859 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-p7bgm" event={"ID":"ad354068-21f9-40aa-a2ae-739f84e32469","Type":"ContainerDied","Data":"925d34b56e3fa96c4a2415fdaeda4a44b5cc43b1138ff186637c877733c5cf89"} Oct 08 18:21:43 crc kubenswrapper[4859]: I1008 18:21:43.726451 4859 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-p7bgm" event={"ID":"ad354068-21f9-40aa-a2ae-739f84e32469","Type":"ContainerStarted","Data":"51f81cca9ae9970df7d9a2a8e72c8d924ba1a867d65dea921fac5f8220297f56"} Oct 08 18:21:43 crc kubenswrapper[4859]: I1008 18:21:43.729890 4859 generic.go:334] "Generic (PLEG): container finished" podID="748a4b84-e115-444a-b0f8-24029e8d1896" containerID="b307190deb9033186860551b4bcb369baab214578419b6e6b1093458c8641517" exitCode=0 Oct 08 18:21:43 crc kubenswrapper[4859]: I1008 18:21:43.729952 4859 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-hlqjs" event={"ID":"748a4b84-e115-444a-b0f8-24029e8d1896","Type":"ContainerDied","Data":"b307190deb9033186860551b4bcb369baab214578419b6e6b1093458c8641517"} Oct 08 18:21:43 crc kubenswrapper[4859]: I1008 18:21:43.730039 4859 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-hlqjs" event={"ID":"748a4b84-e115-444a-b0f8-24029e8d1896","Type":"ContainerStarted","Data":"dc547304710591ee37c52b9b6363d72e12ea46aee21e48b975889092967234ae"} Oct 08 18:21:43 crc kubenswrapper[4859]: I1008 18:21:43.732446 4859 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-wlwhh" event={"ID":"997d41aa-a006-46d1-b488-b6d29186874d","Type":"ContainerStarted","Data":"0140ed75fd8e787997ce0b4ee99de4e858cede779695e2b80537c5fc54a7adb7"} Oct 08 18:21:43 crc kubenswrapper[4859]: I1008 18:21:43.807065 4859 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-wlwhh" podStartSLOduration=2.332413138 podStartE2EDuration="4.807039228s" podCreationTimestamp="2025-10-08 18:21:39 +0000 UTC" firstStartedPulling="2025-10-08 18:21:40.666565469 +0000 UTC m=+270.913404848" lastFinishedPulling="2025-10-08 18:21:43.141191559 +0000 UTC m=+273.388030938" observedRunningTime="2025-10-08 18:21:43.806331867 +0000 UTC m=+274.053171266" watchObservedRunningTime="2025-10-08 18:21:43.807039228 +0000 UTC m=+274.053878607" Oct 08 18:21:44 crc kubenswrapper[4859]: I1008 18:21:44.740267 4859 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-d8bkm" event={"ID":"6e7e5636-231d-4788-9277-83549fc6b404","Type":"ContainerStarted","Data":"6b47fe95e9a78503dd73b320ba45787498384e6a8c7c847cbf9cb55a73fed78b"} Oct 08 18:21:44 crc kubenswrapper[4859]: I1008 18:21:44.743576 4859 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-p7bgm" event={"ID":"ad354068-21f9-40aa-a2ae-739f84e32469","Type":"ContainerStarted","Data":"02113f0210ddb55b50d872c6fbb0c3bb5b62cce65087a03978397da276932ce9"} Oct 08 18:21:44 crc kubenswrapper[4859]: I1008 18:21:44.746183 4859 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-hlqjs" event={"ID":"748a4b84-e115-444a-b0f8-24029e8d1896","Type":"ContainerStarted","Data":"bfe46ad44f72ab4f3692a43292f8054496005e88e75c5500f72a8a12015a4b37"} Oct 08 18:21:44 crc kubenswrapper[4859]: I1008 18:21:44.800168 4859 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-d8bkm" podStartSLOduration=3.32302933 podStartE2EDuration="5.800137747s" podCreationTimestamp="2025-10-08 18:21:39 +0000 UTC" firstStartedPulling="2025-10-08 18:21:41.675549485 +0000 UTC m=+271.922388874" lastFinishedPulling="2025-10-08 18:21:44.152657912 +0000 UTC m=+274.399497291" observedRunningTime="2025-10-08 18:21:44.770743056 +0000 UTC m=+275.017582485" watchObservedRunningTime="2025-10-08 18:21:44.800137747 +0000 UTC m=+275.046977126" Oct 08 18:21:45 crc kubenswrapper[4859]: I1008 18:21:45.753860 4859 generic.go:334] "Generic (PLEG): container finished" podID="ad354068-21f9-40aa-a2ae-739f84e32469" containerID="02113f0210ddb55b50d872c6fbb0c3bb5b62cce65087a03978397da276932ce9" exitCode=0 Oct 08 18:21:45 crc kubenswrapper[4859]: I1008 18:21:45.753972 4859 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-p7bgm" event={"ID":"ad354068-21f9-40aa-a2ae-739f84e32469","Type":"ContainerDied","Data":"02113f0210ddb55b50d872c6fbb0c3bb5b62cce65087a03978397da276932ce9"} Oct 08 18:21:45 crc kubenswrapper[4859]: I1008 18:21:45.756874 4859 generic.go:334] "Generic (PLEG): container finished" podID="748a4b84-e115-444a-b0f8-24029e8d1896" containerID="bfe46ad44f72ab4f3692a43292f8054496005e88e75c5500f72a8a12015a4b37" exitCode=0 Oct 08 18:21:45 crc kubenswrapper[4859]: I1008 18:21:45.756991 4859 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-hlqjs" event={"ID":"748a4b84-e115-444a-b0f8-24029e8d1896","Type":"ContainerDied","Data":"bfe46ad44f72ab4f3692a43292f8054496005e88e75c5500f72a8a12015a4b37"} Oct 08 18:21:46 crc kubenswrapper[4859]: I1008 18:21:46.765197 4859 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-p7bgm" event={"ID":"ad354068-21f9-40aa-a2ae-739f84e32469","Type":"ContainerStarted","Data":"21f824902bbc5a057fc94a3730ed38c40cbe8c2f2d3f225ba62ab1844de07dd1"} Oct 08 18:21:46 crc kubenswrapper[4859]: I1008 18:21:46.772642 4859 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-hlqjs" event={"ID":"748a4b84-e115-444a-b0f8-24029e8d1896","Type":"ContainerStarted","Data":"91de141eca1297582a7867af5acf64d6ef4acae66f5c87c1333084cbd5f646d8"} Oct 08 18:21:46 crc kubenswrapper[4859]: I1008 18:21:46.788099 4859 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-p7bgm" podStartSLOduration=2.334747044 podStartE2EDuration="4.788079831s" podCreationTimestamp="2025-10-08 18:21:42 +0000 UTC" firstStartedPulling="2025-10-08 18:21:43.728264804 +0000 UTC m=+273.975104223" lastFinishedPulling="2025-10-08 18:21:46.181597631 +0000 UTC m=+276.428437010" observedRunningTime="2025-10-08 18:21:46.784230523 +0000 UTC m=+277.031069952" watchObservedRunningTime="2025-10-08 18:21:46.788079831 +0000 UTC m=+277.034919210" Oct 08 18:21:46 crc kubenswrapper[4859]: I1008 18:21:46.813286 4859 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-hlqjs" podStartSLOduration=2.145372948 podStartE2EDuration="4.813263372s" podCreationTimestamp="2025-10-08 18:21:42 +0000 UTC" firstStartedPulling="2025-10-08 18:21:43.731571995 +0000 UTC m=+273.978411374" lastFinishedPulling="2025-10-08 18:21:46.399462419 +0000 UTC m=+276.646301798" observedRunningTime="2025-10-08 18:21:46.812972533 +0000 UTC m=+277.059811922" watchObservedRunningTime="2025-10-08 18:21:46.813263372 +0000 UTC m=+277.060102751" Oct 08 18:21:50 crc kubenswrapper[4859]: I1008 18:21:50.138056 4859 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-wlwhh" Oct 08 18:21:50 crc kubenswrapper[4859]: I1008 18:21:50.138727 4859 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-wlwhh" Oct 08 18:21:50 crc kubenswrapper[4859]: I1008 18:21:50.201471 4859 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-wlwhh" Oct 08 18:21:50 crc kubenswrapper[4859]: I1008 18:21:50.407398 4859 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-d8bkm" Oct 08 18:21:50 crc kubenswrapper[4859]: I1008 18:21:50.407495 4859 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-d8bkm" Oct 08 18:21:50 crc kubenswrapper[4859]: I1008 18:21:50.456962 4859 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-d8bkm" Oct 08 18:21:50 crc kubenswrapper[4859]: I1008 18:21:50.845978 4859 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-d8bkm" Oct 08 18:21:50 crc kubenswrapper[4859]: I1008 18:21:50.847245 4859 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-wlwhh" Oct 08 18:21:52 crc kubenswrapper[4859]: I1008 18:21:52.539016 4859 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-hlqjs" Oct 08 18:21:52 crc kubenswrapper[4859]: I1008 18:21:52.539881 4859 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-hlqjs" Oct 08 18:21:52 crc kubenswrapper[4859]: I1008 18:21:52.577629 4859 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-hlqjs" Oct 08 18:21:52 crc kubenswrapper[4859]: I1008 18:21:52.731926 4859 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-p7bgm" Oct 08 18:21:52 crc kubenswrapper[4859]: I1008 18:21:52.732660 4859 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-p7bgm" Oct 08 18:21:52 crc kubenswrapper[4859]: I1008 18:21:52.772716 4859 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-p7bgm" Oct 08 18:21:52 crc kubenswrapper[4859]: I1008 18:21:52.849556 4859 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-p7bgm" Oct 08 18:21:52 crc kubenswrapper[4859]: I1008 18:21:52.861121 4859 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-hlqjs" Oct 08 18:23:17 crc kubenswrapper[4859]: I1008 18:23:17.924629 4859 patch_prober.go:28] interesting pod/machine-config-daemon-82s52 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 08 18:23:17 crc kubenswrapper[4859]: I1008 18:23:17.925579 4859 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-82s52" podUID="b23a6a6c-9d92-4e7b-840e-55cfda873a2d" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 08 18:23:47 crc kubenswrapper[4859]: I1008 18:23:47.925391 4859 patch_prober.go:28] interesting pod/machine-config-daemon-82s52 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 08 18:23:47 crc kubenswrapper[4859]: I1008 18:23:47.926417 4859 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-82s52" podUID="b23a6a6c-9d92-4e7b-840e-55cfda873a2d" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 08 18:24:17 crc kubenswrapper[4859]: I1008 18:24:17.924965 4859 patch_prober.go:28] interesting pod/machine-config-daemon-82s52 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 08 18:24:17 crc kubenswrapper[4859]: I1008 18:24:17.925629 4859 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-82s52" podUID="b23a6a6c-9d92-4e7b-840e-55cfda873a2d" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 08 18:24:17 crc kubenswrapper[4859]: I1008 18:24:17.925743 4859 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-82s52" Oct 08 18:24:17 crc kubenswrapper[4859]: I1008 18:24:17.926810 4859 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"c6b208adefaf5eb65ea467fb4313c0f4281605067819f9fbd7f4404606fa52d0"} pod="openshift-machine-config-operator/machine-config-daemon-82s52" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Oct 08 18:24:17 crc kubenswrapper[4859]: I1008 18:24:17.926909 4859 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-82s52" podUID="b23a6a6c-9d92-4e7b-840e-55cfda873a2d" containerName="machine-config-daemon" containerID="cri-o://c6b208adefaf5eb65ea467fb4313c0f4281605067819f9fbd7f4404606fa52d0" gracePeriod=600 Oct 08 18:24:18 crc kubenswrapper[4859]: I1008 18:24:18.795456 4859 generic.go:334] "Generic (PLEG): container finished" podID="b23a6a6c-9d92-4e7b-840e-55cfda873a2d" containerID="c6b208adefaf5eb65ea467fb4313c0f4281605067819f9fbd7f4404606fa52d0" exitCode=0 Oct 08 18:24:18 crc kubenswrapper[4859]: I1008 18:24:18.795541 4859 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-82s52" event={"ID":"b23a6a6c-9d92-4e7b-840e-55cfda873a2d","Type":"ContainerDied","Data":"c6b208adefaf5eb65ea467fb4313c0f4281605067819f9fbd7f4404606fa52d0"} Oct 08 18:24:18 crc kubenswrapper[4859]: I1008 18:24:18.795919 4859 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-82s52" event={"ID":"b23a6a6c-9d92-4e7b-840e-55cfda873a2d","Type":"ContainerStarted","Data":"487a3a1048ef98faa7383425fe83462232329f063579bf1e501fc8c04727fcea"} Oct 08 18:24:18 crc kubenswrapper[4859]: I1008 18:24:18.795952 4859 scope.go:117] "RemoveContainer" containerID="c23ea3580998adf90621a0155c20ee4bd590c512629812174e95c28857bcfef9" Oct 08 18:24:19 crc kubenswrapper[4859]: I1008 18:24:19.855853 4859 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-image-registry/image-registry-66df7c8f76-8gddl"] Oct 08 18:24:19 crc kubenswrapper[4859]: I1008 18:24:19.858166 4859 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/image-registry-66df7c8f76-8gddl" Oct 08 18:24:19 crc kubenswrapper[4859]: I1008 18:24:19.871509 4859 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-image-registry/image-registry-66df7c8f76-8gddl"] Oct 08 18:24:20 crc kubenswrapper[4859]: I1008 18:24:20.037223 4859 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-66df7c8f76-8gddl\" (UID: \"34684e60-133d-4439-b0ee-d363f143858a\") " pod="openshift-image-registry/image-registry-66df7c8f76-8gddl" Oct 08 18:24:20 crc kubenswrapper[4859]: I1008 18:24:20.037309 4859 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/34684e60-133d-4439-b0ee-d363f143858a-bound-sa-token\") pod \"image-registry-66df7c8f76-8gddl\" (UID: \"34684e60-133d-4439-b0ee-d363f143858a\") " pod="openshift-image-registry/image-registry-66df7c8f76-8gddl" Oct 08 18:24:20 crc kubenswrapper[4859]: I1008 18:24:20.037372 4859 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/34684e60-133d-4439-b0ee-d363f143858a-ca-trust-extracted\") pod \"image-registry-66df7c8f76-8gddl\" (UID: \"34684e60-133d-4439-b0ee-d363f143858a\") " pod="openshift-image-registry/image-registry-66df7c8f76-8gddl" Oct 08 18:24:20 crc kubenswrapper[4859]: I1008 18:24:20.037393 4859 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-8mbmn\" (UniqueName: \"kubernetes.io/projected/34684e60-133d-4439-b0ee-d363f143858a-kube-api-access-8mbmn\") pod \"image-registry-66df7c8f76-8gddl\" (UID: \"34684e60-133d-4439-b0ee-d363f143858a\") " pod="openshift-image-registry/image-registry-66df7c8f76-8gddl" Oct 08 18:24:20 crc kubenswrapper[4859]: I1008 18:24:20.037419 4859 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/34684e60-133d-4439-b0ee-d363f143858a-trusted-ca\") pod \"image-registry-66df7c8f76-8gddl\" (UID: \"34684e60-133d-4439-b0ee-d363f143858a\") " pod="openshift-image-registry/image-registry-66df7c8f76-8gddl" Oct 08 18:24:20 crc kubenswrapper[4859]: I1008 18:24:20.037449 4859 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/34684e60-133d-4439-b0ee-d363f143858a-installation-pull-secrets\") pod \"image-registry-66df7c8f76-8gddl\" (UID: \"34684e60-133d-4439-b0ee-d363f143858a\") " pod="openshift-image-registry/image-registry-66df7c8f76-8gddl" Oct 08 18:24:20 crc kubenswrapper[4859]: I1008 18:24:20.037498 4859 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/34684e60-133d-4439-b0ee-d363f143858a-registry-tls\") pod \"image-registry-66df7c8f76-8gddl\" (UID: \"34684e60-133d-4439-b0ee-d363f143858a\") " pod="openshift-image-registry/image-registry-66df7c8f76-8gddl" Oct 08 18:24:20 crc kubenswrapper[4859]: I1008 18:24:20.037540 4859 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/34684e60-133d-4439-b0ee-d363f143858a-registry-certificates\") pod \"image-registry-66df7c8f76-8gddl\" (UID: \"34684e60-133d-4439-b0ee-d363f143858a\") " pod="openshift-image-registry/image-registry-66df7c8f76-8gddl" Oct 08 18:24:20 crc kubenswrapper[4859]: I1008 18:24:20.057914 4859 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-66df7c8f76-8gddl\" (UID: \"34684e60-133d-4439-b0ee-d363f143858a\") " pod="openshift-image-registry/image-registry-66df7c8f76-8gddl" Oct 08 18:24:20 crc kubenswrapper[4859]: I1008 18:24:20.139083 4859 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/34684e60-133d-4439-b0ee-d363f143858a-installation-pull-secrets\") pod \"image-registry-66df7c8f76-8gddl\" (UID: \"34684e60-133d-4439-b0ee-d363f143858a\") " pod="openshift-image-registry/image-registry-66df7c8f76-8gddl" Oct 08 18:24:20 crc kubenswrapper[4859]: I1008 18:24:20.139166 4859 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/34684e60-133d-4439-b0ee-d363f143858a-registry-tls\") pod \"image-registry-66df7c8f76-8gddl\" (UID: \"34684e60-133d-4439-b0ee-d363f143858a\") " pod="openshift-image-registry/image-registry-66df7c8f76-8gddl" Oct 08 18:24:20 crc kubenswrapper[4859]: I1008 18:24:20.139211 4859 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/34684e60-133d-4439-b0ee-d363f143858a-registry-certificates\") pod \"image-registry-66df7c8f76-8gddl\" (UID: \"34684e60-133d-4439-b0ee-d363f143858a\") " pod="openshift-image-registry/image-registry-66df7c8f76-8gddl" Oct 08 18:24:20 crc kubenswrapper[4859]: I1008 18:24:20.139243 4859 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/34684e60-133d-4439-b0ee-d363f143858a-bound-sa-token\") pod \"image-registry-66df7c8f76-8gddl\" (UID: \"34684e60-133d-4439-b0ee-d363f143858a\") " pod="openshift-image-registry/image-registry-66df7c8f76-8gddl" Oct 08 18:24:20 crc kubenswrapper[4859]: I1008 18:24:20.139284 4859 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/34684e60-133d-4439-b0ee-d363f143858a-ca-trust-extracted\") pod \"image-registry-66df7c8f76-8gddl\" (UID: \"34684e60-133d-4439-b0ee-d363f143858a\") " pod="openshift-image-registry/image-registry-66df7c8f76-8gddl" Oct 08 18:24:20 crc kubenswrapper[4859]: I1008 18:24:20.139300 4859 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-8mbmn\" (UniqueName: \"kubernetes.io/projected/34684e60-133d-4439-b0ee-d363f143858a-kube-api-access-8mbmn\") pod \"image-registry-66df7c8f76-8gddl\" (UID: \"34684e60-133d-4439-b0ee-d363f143858a\") " pod="openshift-image-registry/image-registry-66df7c8f76-8gddl" Oct 08 18:24:20 crc kubenswrapper[4859]: I1008 18:24:20.139323 4859 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/34684e60-133d-4439-b0ee-d363f143858a-trusted-ca\") pod \"image-registry-66df7c8f76-8gddl\" (UID: \"34684e60-133d-4439-b0ee-d363f143858a\") " pod="openshift-image-registry/image-registry-66df7c8f76-8gddl" Oct 08 18:24:20 crc kubenswrapper[4859]: I1008 18:24:20.140070 4859 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/34684e60-133d-4439-b0ee-d363f143858a-ca-trust-extracted\") pod \"image-registry-66df7c8f76-8gddl\" (UID: \"34684e60-133d-4439-b0ee-d363f143858a\") " pod="openshift-image-registry/image-registry-66df7c8f76-8gddl" Oct 08 18:24:20 crc kubenswrapper[4859]: I1008 18:24:20.140741 4859 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/34684e60-133d-4439-b0ee-d363f143858a-trusted-ca\") pod \"image-registry-66df7c8f76-8gddl\" (UID: \"34684e60-133d-4439-b0ee-d363f143858a\") " pod="openshift-image-registry/image-registry-66df7c8f76-8gddl" Oct 08 18:24:20 crc kubenswrapper[4859]: I1008 18:24:20.141296 4859 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/34684e60-133d-4439-b0ee-d363f143858a-registry-certificates\") pod \"image-registry-66df7c8f76-8gddl\" (UID: \"34684e60-133d-4439-b0ee-d363f143858a\") " pod="openshift-image-registry/image-registry-66df7c8f76-8gddl" Oct 08 18:24:20 crc kubenswrapper[4859]: I1008 18:24:20.146490 4859 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/34684e60-133d-4439-b0ee-d363f143858a-installation-pull-secrets\") pod \"image-registry-66df7c8f76-8gddl\" (UID: \"34684e60-133d-4439-b0ee-d363f143858a\") " pod="openshift-image-registry/image-registry-66df7c8f76-8gddl" Oct 08 18:24:20 crc kubenswrapper[4859]: I1008 18:24:20.148805 4859 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/34684e60-133d-4439-b0ee-d363f143858a-registry-tls\") pod \"image-registry-66df7c8f76-8gddl\" (UID: \"34684e60-133d-4439-b0ee-d363f143858a\") " pod="openshift-image-registry/image-registry-66df7c8f76-8gddl" Oct 08 18:24:20 crc kubenswrapper[4859]: I1008 18:24:20.156416 4859 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/34684e60-133d-4439-b0ee-d363f143858a-bound-sa-token\") pod \"image-registry-66df7c8f76-8gddl\" (UID: \"34684e60-133d-4439-b0ee-d363f143858a\") " pod="openshift-image-registry/image-registry-66df7c8f76-8gddl" Oct 08 18:24:20 crc kubenswrapper[4859]: I1008 18:24:20.159132 4859 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-8mbmn\" (UniqueName: \"kubernetes.io/projected/34684e60-133d-4439-b0ee-d363f143858a-kube-api-access-8mbmn\") pod \"image-registry-66df7c8f76-8gddl\" (UID: \"34684e60-133d-4439-b0ee-d363f143858a\") " pod="openshift-image-registry/image-registry-66df7c8f76-8gddl" Oct 08 18:24:20 crc kubenswrapper[4859]: I1008 18:24:20.205655 4859 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/image-registry-66df7c8f76-8gddl" Oct 08 18:24:20 crc kubenswrapper[4859]: I1008 18:24:20.605359 4859 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-image-registry/image-registry-66df7c8f76-8gddl"] Oct 08 18:24:20 crc kubenswrapper[4859]: W1008 18:24:20.610594 4859 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod34684e60_133d_4439_b0ee_d363f143858a.slice/crio-1fb8b7768f5ab9ab9e111a277f4225c01015d96cda4ecfe61880f4774b69814c WatchSource:0}: Error finding container 1fb8b7768f5ab9ab9e111a277f4225c01015d96cda4ecfe61880f4774b69814c: Status 404 returned error can't find the container with id 1fb8b7768f5ab9ab9e111a277f4225c01015d96cda4ecfe61880f4774b69814c Oct 08 18:24:20 crc kubenswrapper[4859]: I1008 18:24:20.812342 4859 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/image-registry-66df7c8f76-8gddl" event={"ID":"34684e60-133d-4439-b0ee-d363f143858a","Type":"ContainerStarted","Data":"02898d6d36966506623ad84b9aa9ddc8da8562c1fdc379e063e3da81e1f19ec8"} Oct 08 18:24:20 crc kubenswrapper[4859]: I1008 18:24:20.812422 4859 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/image-registry-66df7c8f76-8gddl" event={"ID":"34684e60-133d-4439-b0ee-d363f143858a","Type":"ContainerStarted","Data":"1fb8b7768f5ab9ab9e111a277f4225c01015d96cda4ecfe61880f4774b69814c"} Oct 08 18:24:20 crc kubenswrapper[4859]: I1008 18:24:20.812488 4859 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-image-registry/image-registry-66df7c8f76-8gddl" Oct 08 18:24:40 crc kubenswrapper[4859]: I1008 18:24:40.214352 4859 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-image-registry/image-registry-66df7c8f76-8gddl" Oct 08 18:24:40 crc kubenswrapper[4859]: I1008 18:24:40.250155 4859 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-image-registry/image-registry-66df7c8f76-8gddl" podStartSLOduration=21.250125689 podStartE2EDuration="21.250125689s" podCreationTimestamp="2025-10-08 18:24:19 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-08 18:24:20.838903545 +0000 UTC m=+431.085742914" watchObservedRunningTime="2025-10-08 18:24:40.250125689 +0000 UTC m=+450.496965108" Oct 08 18:24:40 crc kubenswrapper[4859]: I1008 18:24:40.324594 4859 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-image-registry/image-registry-697d97f7c8-kg44m"] Oct 08 18:25:05 crc kubenswrapper[4859]: I1008 18:25:05.391023 4859 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-image-registry/image-registry-697d97f7c8-kg44m" podUID="4afad99b-9bf7-43c9-8dcf-7a3217136be5" containerName="registry" containerID="cri-o://c4e4ac2065ed20944a9de0ebaa0ce1d4d3c970b586c0a4da7abe8fb14445e685" gracePeriod=30 Oct 08 18:25:05 crc kubenswrapper[4859]: I1008 18:25:05.745826 4859 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/image-registry-697d97f7c8-kg44m" Oct 08 18:25:05 crc kubenswrapper[4859]: I1008 18:25:05.780165 4859 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/4afad99b-9bf7-43c9-8dcf-7a3217136be5-trusted-ca\") pod \"4afad99b-9bf7-43c9-8dcf-7a3217136be5\" (UID: \"4afad99b-9bf7-43c9-8dcf-7a3217136be5\") " Oct 08 18:25:05 crc kubenswrapper[4859]: I1008 18:25:05.780466 4859 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"registry-storage\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"4afad99b-9bf7-43c9-8dcf-7a3217136be5\" (UID: \"4afad99b-9bf7-43c9-8dcf-7a3217136be5\") " Oct 08 18:25:05 crc kubenswrapper[4859]: I1008 18:25:05.780520 4859 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/4afad99b-9bf7-43c9-8dcf-7a3217136be5-registry-certificates\") pod \"4afad99b-9bf7-43c9-8dcf-7a3217136be5\" (UID: \"4afad99b-9bf7-43c9-8dcf-7a3217136be5\") " Oct 08 18:25:05 crc kubenswrapper[4859]: I1008 18:25:05.780545 4859 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/4afad99b-9bf7-43c9-8dcf-7a3217136be5-ca-trust-extracted\") pod \"4afad99b-9bf7-43c9-8dcf-7a3217136be5\" (UID: \"4afad99b-9bf7-43c9-8dcf-7a3217136be5\") " Oct 08 18:25:05 crc kubenswrapper[4859]: I1008 18:25:05.780563 4859 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/4afad99b-9bf7-43c9-8dcf-7a3217136be5-installation-pull-secrets\") pod \"4afad99b-9bf7-43c9-8dcf-7a3217136be5\" (UID: \"4afad99b-9bf7-43c9-8dcf-7a3217136be5\") " Oct 08 18:25:05 crc kubenswrapper[4859]: I1008 18:25:05.780582 4859 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/4afad99b-9bf7-43c9-8dcf-7a3217136be5-registry-tls\") pod \"4afad99b-9bf7-43c9-8dcf-7a3217136be5\" (UID: \"4afad99b-9bf7-43c9-8dcf-7a3217136be5\") " Oct 08 18:25:05 crc kubenswrapper[4859]: I1008 18:25:05.780629 4859 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-cs2l8\" (UniqueName: \"kubernetes.io/projected/4afad99b-9bf7-43c9-8dcf-7a3217136be5-kube-api-access-cs2l8\") pod \"4afad99b-9bf7-43c9-8dcf-7a3217136be5\" (UID: \"4afad99b-9bf7-43c9-8dcf-7a3217136be5\") " Oct 08 18:25:05 crc kubenswrapper[4859]: I1008 18:25:05.780665 4859 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/4afad99b-9bf7-43c9-8dcf-7a3217136be5-bound-sa-token\") pod \"4afad99b-9bf7-43c9-8dcf-7a3217136be5\" (UID: \"4afad99b-9bf7-43c9-8dcf-7a3217136be5\") " Oct 08 18:25:05 crc kubenswrapper[4859]: I1008 18:25:05.781347 4859 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/4afad99b-9bf7-43c9-8dcf-7a3217136be5-trusted-ca" (OuterVolumeSpecName: "trusted-ca") pod "4afad99b-9bf7-43c9-8dcf-7a3217136be5" (UID: "4afad99b-9bf7-43c9-8dcf-7a3217136be5"). InnerVolumeSpecName "trusted-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 08 18:25:05 crc kubenswrapper[4859]: I1008 18:25:05.784524 4859 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/4afad99b-9bf7-43c9-8dcf-7a3217136be5-registry-certificates" (OuterVolumeSpecName: "registry-certificates") pod "4afad99b-9bf7-43c9-8dcf-7a3217136be5" (UID: "4afad99b-9bf7-43c9-8dcf-7a3217136be5"). InnerVolumeSpecName "registry-certificates". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 08 18:25:05 crc kubenswrapper[4859]: I1008 18:25:05.786979 4859 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/4afad99b-9bf7-43c9-8dcf-7a3217136be5-registry-tls" (OuterVolumeSpecName: "registry-tls") pod "4afad99b-9bf7-43c9-8dcf-7a3217136be5" (UID: "4afad99b-9bf7-43c9-8dcf-7a3217136be5"). InnerVolumeSpecName "registry-tls". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 08 18:25:05 crc kubenswrapper[4859]: I1008 18:25:05.788017 4859 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/4afad99b-9bf7-43c9-8dcf-7a3217136be5-kube-api-access-cs2l8" (OuterVolumeSpecName: "kube-api-access-cs2l8") pod "4afad99b-9bf7-43c9-8dcf-7a3217136be5" (UID: "4afad99b-9bf7-43c9-8dcf-7a3217136be5"). InnerVolumeSpecName "kube-api-access-cs2l8". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 08 18:25:05 crc kubenswrapper[4859]: I1008 18:25:05.788188 4859 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/4afad99b-9bf7-43c9-8dcf-7a3217136be5-installation-pull-secrets" (OuterVolumeSpecName: "installation-pull-secrets") pod "4afad99b-9bf7-43c9-8dcf-7a3217136be5" (UID: "4afad99b-9bf7-43c9-8dcf-7a3217136be5"). InnerVolumeSpecName "installation-pull-secrets". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 08 18:25:05 crc kubenswrapper[4859]: I1008 18:25:05.788711 4859 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/4afad99b-9bf7-43c9-8dcf-7a3217136be5-bound-sa-token" (OuterVolumeSpecName: "bound-sa-token") pod "4afad99b-9bf7-43c9-8dcf-7a3217136be5" (UID: "4afad99b-9bf7-43c9-8dcf-7a3217136be5"). InnerVolumeSpecName "bound-sa-token". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 08 18:25:05 crc kubenswrapper[4859]: I1008 18:25:05.792181 4859 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (OuterVolumeSpecName: "registry-storage") pod "4afad99b-9bf7-43c9-8dcf-7a3217136be5" (UID: "4afad99b-9bf7-43c9-8dcf-7a3217136be5"). InnerVolumeSpecName "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8". PluginName "kubernetes.io/csi", VolumeGidValue "" Oct 08 18:25:05 crc kubenswrapper[4859]: I1008 18:25:05.798360 4859 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/4afad99b-9bf7-43c9-8dcf-7a3217136be5-ca-trust-extracted" (OuterVolumeSpecName: "ca-trust-extracted") pod "4afad99b-9bf7-43c9-8dcf-7a3217136be5" (UID: "4afad99b-9bf7-43c9-8dcf-7a3217136be5"). InnerVolumeSpecName "ca-trust-extracted". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 08 18:25:05 crc kubenswrapper[4859]: I1008 18:25:05.881430 4859 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-cs2l8\" (UniqueName: \"kubernetes.io/projected/4afad99b-9bf7-43c9-8dcf-7a3217136be5-kube-api-access-cs2l8\") on node \"crc\" DevicePath \"\"" Oct 08 18:25:05 crc kubenswrapper[4859]: I1008 18:25:05.881464 4859 reconciler_common.go:293] "Volume detached for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/4afad99b-9bf7-43c9-8dcf-7a3217136be5-bound-sa-token\") on node \"crc\" DevicePath \"\"" Oct 08 18:25:05 crc kubenswrapper[4859]: I1008 18:25:05.881474 4859 reconciler_common.go:293] "Volume detached for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/4afad99b-9bf7-43c9-8dcf-7a3217136be5-trusted-ca\") on node \"crc\" DevicePath \"\"" Oct 08 18:25:05 crc kubenswrapper[4859]: I1008 18:25:05.881484 4859 reconciler_common.go:293] "Volume detached for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/4afad99b-9bf7-43c9-8dcf-7a3217136be5-registry-certificates\") on node \"crc\" DevicePath \"\"" Oct 08 18:25:05 crc kubenswrapper[4859]: I1008 18:25:05.881492 4859 reconciler_common.go:293] "Volume detached for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/4afad99b-9bf7-43c9-8dcf-7a3217136be5-installation-pull-secrets\") on node \"crc\" DevicePath \"\"" Oct 08 18:25:05 crc kubenswrapper[4859]: I1008 18:25:05.881500 4859 reconciler_common.go:293] "Volume detached for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/4afad99b-9bf7-43c9-8dcf-7a3217136be5-ca-trust-extracted\") on node \"crc\" DevicePath \"\"" Oct 08 18:25:05 crc kubenswrapper[4859]: I1008 18:25:05.881507 4859 reconciler_common.go:293] "Volume detached for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/4afad99b-9bf7-43c9-8dcf-7a3217136be5-registry-tls\") on node \"crc\" DevicePath \"\"" Oct 08 18:25:06 crc kubenswrapper[4859]: I1008 18:25:06.117098 4859 generic.go:334] "Generic (PLEG): container finished" podID="4afad99b-9bf7-43c9-8dcf-7a3217136be5" containerID="c4e4ac2065ed20944a9de0ebaa0ce1d4d3c970b586c0a4da7abe8fb14445e685" exitCode=0 Oct 08 18:25:06 crc kubenswrapper[4859]: I1008 18:25:06.117164 4859 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/image-registry-697d97f7c8-kg44m" event={"ID":"4afad99b-9bf7-43c9-8dcf-7a3217136be5","Type":"ContainerDied","Data":"c4e4ac2065ed20944a9de0ebaa0ce1d4d3c970b586c0a4da7abe8fb14445e685"} Oct 08 18:25:06 crc kubenswrapper[4859]: I1008 18:25:06.117204 4859 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/image-registry-697d97f7c8-kg44m" event={"ID":"4afad99b-9bf7-43c9-8dcf-7a3217136be5","Type":"ContainerDied","Data":"e936f0856e9768e37d0620582beb93df7b58bedfc12fd790f00e445203415533"} Oct 08 18:25:06 crc kubenswrapper[4859]: I1008 18:25:06.117231 4859 scope.go:117] "RemoveContainer" containerID="c4e4ac2065ed20944a9de0ebaa0ce1d4d3c970b586c0a4da7abe8fb14445e685" Oct 08 18:25:06 crc kubenswrapper[4859]: I1008 18:25:06.117735 4859 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/image-registry-697d97f7c8-kg44m" Oct 08 18:25:06 crc kubenswrapper[4859]: I1008 18:25:06.151068 4859 scope.go:117] "RemoveContainer" containerID="c4e4ac2065ed20944a9de0ebaa0ce1d4d3c970b586c0a4da7abe8fb14445e685" Oct 08 18:25:06 crc kubenswrapper[4859]: E1008 18:25:06.153869 4859 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"c4e4ac2065ed20944a9de0ebaa0ce1d4d3c970b586c0a4da7abe8fb14445e685\": container with ID starting with c4e4ac2065ed20944a9de0ebaa0ce1d4d3c970b586c0a4da7abe8fb14445e685 not found: ID does not exist" containerID="c4e4ac2065ed20944a9de0ebaa0ce1d4d3c970b586c0a4da7abe8fb14445e685" Oct 08 18:25:06 crc kubenswrapper[4859]: I1008 18:25:06.153926 4859 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"c4e4ac2065ed20944a9de0ebaa0ce1d4d3c970b586c0a4da7abe8fb14445e685"} err="failed to get container status \"c4e4ac2065ed20944a9de0ebaa0ce1d4d3c970b586c0a4da7abe8fb14445e685\": rpc error: code = NotFound desc = could not find container \"c4e4ac2065ed20944a9de0ebaa0ce1d4d3c970b586c0a4da7abe8fb14445e685\": container with ID starting with c4e4ac2065ed20944a9de0ebaa0ce1d4d3c970b586c0a4da7abe8fb14445e685 not found: ID does not exist" Oct 08 18:25:06 crc kubenswrapper[4859]: I1008 18:25:06.155759 4859 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-image-registry/image-registry-697d97f7c8-kg44m"] Oct 08 18:25:06 crc kubenswrapper[4859]: I1008 18:25:06.161844 4859 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-image-registry/image-registry-697d97f7c8-kg44m"] Oct 08 18:25:06 crc kubenswrapper[4859]: I1008 18:25:06.487242 4859 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="4afad99b-9bf7-43c9-8dcf-7a3217136be5" path="/var/lib/kubelet/pods/4afad99b-9bf7-43c9-8dcf-7a3217136be5/volumes" Oct 08 18:26:47 crc kubenswrapper[4859]: I1008 18:26:47.924739 4859 patch_prober.go:28] interesting pod/machine-config-daemon-82s52 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 08 18:26:47 crc kubenswrapper[4859]: I1008 18:26:47.926220 4859 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-82s52" podUID="b23a6a6c-9d92-4e7b-840e-55cfda873a2d" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 08 18:26:56 crc kubenswrapper[4859]: I1008 18:26:56.963848 4859 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["cert-manager/cert-manager-cainjector-7f985d654d-lqldn"] Oct 08 18:26:56 crc kubenswrapper[4859]: E1008 18:26:56.964508 4859 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="4afad99b-9bf7-43c9-8dcf-7a3217136be5" containerName="registry" Oct 08 18:26:56 crc kubenswrapper[4859]: I1008 18:26:56.964520 4859 state_mem.go:107] "Deleted CPUSet assignment" podUID="4afad99b-9bf7-43c9-8dcf-7a3217136be5" containerName="registry" Oct 08 18:26:56 crc kubenswrapper[4859]: I1008 18:26:56.964626 4859 memory_manager.go:354] "RemoveStaleState removing state" podUID="4afad99b-9bf7-43c9-8dcf-7a3217136be5" containerName="registry" Oct 08 18:26:56 crc kubenswrapper[4859]: I1008 18:26:56.965032 4859 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="cert-manager/cert-manager-cainjector-7f985d654d-lqldn" Oct 08 18:26:56 crc kubenswrapper[4859]: I1008 18:26:56.969664 4859 reflector.go:368] Caches populated for *v1.ConfigMap from object-"cert-manager"/"kube-root-ca.crt" Oct 08 18:26:56 crc kubenswrapper[4859]: I1008 18:26:56.971236 4859 reflector.go:368] Caches populated for *v1.ConfigMap from object-"cert-manager"/"openshift-service-ca.crt" Oct 08 18:26:56 crc kubenswrapper[4859]: I1008 18:26:56.971422 4859 reflector.go:368] Caches populated for *v1.Secret from object-"cert-manager"/"cert-manager-cainjector-dockercfg-qmtxj" Oct 08 18:26:56 crc kubenswrapper[4859]: I1008 18:26:56.982906 4859 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["cert-manager/cert-manager-5b446d88c5-w95bk"] Oct 08 18:26:56 crc kubenswrapper[4859]: I1008 18:26:56.984005 4859 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="cert-manager/cert-manager-5b446d88c5-w95bk" Oct 08 18:26:56 crc kubenswrapper[4859]: I1008 18:26:56.986325 4859 reflector.go:368] Caches populated for *v1.Secret from object-"cert-manager"/"cert-manager-dockercfg-h4rws" Oct 08 18:26:56 crc kubenswrapper[4859]: I1008 18:26:56.989647 4859 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["cert-manager/cert-manager-cainjector-7f985d654d-lqldn"] Oct 08 18:26:57 crc kubenswrapper[4859]: I1008 18:26:57.003578 4859 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["cert-manager/cert-manager-webhook-5655c58dd6-jkfxx"] Oct 08 18:26:57 crc kubenswrapper[4859]: I1008 18:26:57.004674 4859 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="cert-manager/cert-manager-webhook-5655c58dd6-jkfxx" Oct 08 18:26:57 crc kubenswrapper[4859]: I1008 18:26:57.006537 4859 reflector.go:368] Caches populated for *v1.Secret from object-"cert-manager"/"cert-manager-webhook-dockercfg-j8k2s" Oct 08 18:26:57 crc kubenswrapper[4859]: I1008 18:26:57.017525 4859 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["cert-manager/cert-manager-webhook-5655c58dd6-jkfxx"] Oct 08 18:26:57 crc kubenswrapper[4859]: I1008 18:26:57.018503 4859 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-757pb\" (UniqueName: \"kubernetes.io/projected/ccaaab95-a4e1-43b8-a962-0b833beec038-kube-api-access-757pb\") pod \"cert-manager-cainjector-7f985d654d-lqldn\" (UID: \"ccaaab95-a4e1-43b8-a962-0b833beec038\") " pod="cert-manager/cert-manager-cainjector-7f985d654d-lqldn" Oct 08 18:26:57 crc kubenswrapper[4859]: I1008 18:26:57.018567 4859 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-pql6x\" (UniqueName: \"kubernetes.io/projected/2fa6fe28-fd3a-4611-ad50-2a4fc67ca563-kube-api-access-pql6x\") pod \"cert-manager-5b446d88c5-w95bk\" (UID: \"2fa6fe28-fd3a-4611-ad50-2a4fc67ca563\") " pod="cert-manager/cert-manager-5b446d88c5-w95bk" Oct 08 18:26:57 crc kubenswrapper[4859]: I1008 18:26:57.039008 4859 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["cert-manager/cert-manager-5b446d88c5-w95bk"] Oct 08 18:26:57 crc kubenswrapper[4859]: I1008 18:26:57.120467 4859 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-22xbr\" (UniqueName: \"kubernetes.io/projected/24b1f415-e420-420a-b6c1-7fde05b080b6-kube-api-access-22xbr\") pod \"cert-manager-webhook-5655c58dd6-jkfxx\" (UID: \"24b1f415-e420-420a-b6c1-7fde05b080b6\") " pod="cert-manager/cert-manager-webhook-5655c58dd6-jkfxx" Oct 08 18:26:57 crc kubenswrapper[4859]: I1008 18:26:57.120555 4859 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-757pb\" (UniqueName: \"kubernetes.io/projected/ccaaab95-a4e1-43b8-a962-0b833beec038-kube-api-access-757pb\") pod \"cert-manager-cainjector-7f985d654d-lqldn\" (UID: \"ccaaab95-a4e1-43b8-a962-0b833beec038\") " pod="cert-manager/cert-manager-cainjector-7f985d654d-lqldn" Oct 08 18:26:57 crc kubenswrapper[4859]: I1008 18:26:57.120583 4859 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-pql6x\" (UniqueName: \"kubernetes.io/projected/2fa6fe28-fd3a-4611-ad50-2a4fc67ca563-kube-api-access-pql6x\") pod \"cert-manager-5b446d88c5-w95bk\" (UID: \"2fa6fe28-fd3a-4611-ad50-2a4fc67ca563\") " pod="cert-manager/cert-manager-5b446d88c5-w95bk" Oct 08 18:26:57 crc kubenswrapper[4859]: I1008 18:26:57.142143 4859 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-757pb\" (UniqueName: \"kubernetes.io/projected/ccaaab95-a4e1-43b8-a962-0b833beec038-kube-api-access-757pb\") pod \"cert-manager-cainjector-7f985d654d-lqldn\" (UID: \"ccaaab95-a4e1-43b8-a962-0b833beec038\") " pod="cert-manager/cert-manager-cainjector-7f985d654d-lqldn" Oct 08 18:26:57 crc kubenswrapper[4859]: I1008 18:26:57.156078 4859 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-pql6x\" (UniqueName: \"kubernetes.io/projected/2fa6fe28-fd3a-4611-ad50-2a4fc67ca563-kube-api-access-pql6x\") pod \"cert-manager-5b446d88c5-w95bk\" (UID: \"2fa6fe28-fd3a-4611-ad50-2a4fc67ca563\") " pod="cert-manager/cert-manager-5b446d88c5-w95bk" Oct 08 18:26:57 crc kubenswrapper[4859]: I1008 18:26:57.222860 4859 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-22xbr\" (UniqueName: \"kubernetes.io/projected/24b1f415-e420-420a-b6c1-7fde05b080b6-kube-api-access-22xbr\") pod \"cert-manager-webhook-5655c58dd6-jkfxx\" (UID: \"24b1f415-e420-420a-b6c1-7fde05b080b6\") " pod="cert-manager/cert-manager-webhook-5655c58dd6-jkfxx" Oct 08 18:26:57 crc kubenswrapper[4859]: I1008 18:26:57.241423 4859 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-22xbr\" (UniqueName: \"kubernetes.io/projected/24b1f415-e420-420a-b6c1-7fde05b080b6-kube-api-access-22xbr\") pod \"cert-manager-webhook-5655c58dd6-jkfxx\" (UID: \"24b1f415-e420-420a-b6c1-7fde05b080b6\") " pod="cert-manager/cert-manager-webhook-5655c58dd6-jkfxx" Oct 08 18:26:57 crc kubenswrapper[4859]: I1008 18:26:57.281327 4859 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="cert-manager/cert-manager-cainjector-7f985d654d-lqldn" Oct 08 18:26:57 crc kubenswrapper[4859]: I1008 18:26:57.296365 4859 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="cert-manager/cert-manager-5b446d88c5-w95bk" Oct 08 18:26:57 crc kubenswrapper[4859]: I1008 18:26:57.321346 4859 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="cert-manager/cert-manager-webhook-5655c58dd6-jkfxx" Oct 08 18:26:57 crc kubenswrapper[4859]: I1008 18:26:57.498793 4859 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["cert-manager/cert-manager-cainjector-7f985d654d-lqldn"] Oct 08 18:26:57 crc kubenswrapper[4859]: I1008 18:26:57.510756 4859 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Oct 08 18:26:57 crc kubenswrapper[4859]: I1008 18:26:57.543150 4859 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["cert-manager/cert-manager-5b446d88c5-w95bk"] Oct 08 18:26:57 crc kubenswrapper[4859]: W1008 18:26:57.551161 4859 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod2fa6fe28_fd3a_4611_ad50_2a4fc67ca563.slice/crio-b481ed7423889cbaa4cff30990630d6e5fbe7ffae852f84ddf0a4b4d2e043363 WatchSource:0}: Error finding container b481ed7423889cbaa4cff30990630d6e5fbe7ffae852f84ddf0a4b4d2e043363: Status 404 returned error can't find the container with id b481ed7423889cbaa4cff30990630d6e5fbe7ffae852f84ddf0a4b4d2e043363 Oct 08 18:26:57 crc kubenswrapper[4859]: I1008 18:26:57.579967 4859 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["cert-manager/cert-manager-webhook-5655c58dd6-jkfxx"] Oct 08 18:26:57 crc kubenswrapper[4859]: W1008 18:26:57.584947 4859 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod24b1f415_e420_420a_b6c1_7fde05b080b6.slice/crio-74740da0ea34e8a30f95c3b6576b046c1111ac1f10c4f41f4bbe90e4a238801c WatchSource:0}: Error finding container 74740da0ea34e8a30f95c3b6576b046c1111ac1f10c4f41f4bbe90e4a238801c: Status 404 returned error can't find the container with id 74740da0ea34e8a30f95c3b6576b046c1111ac1f10c4f41f4bbe90e4a238801c Oct 08 18:26:57 crc kubenswrapper[4859]: I1008 18:26:57.857957 4859 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cert-manager/cert-manager-5b446d88c5-w95bk" event={"ID":"2fa6fe28-fd3a-4611-ad50-2a4fc67ca563","Type":"ContainerStarted","Data":"b481ed7423889cbaa4cff30990630d6e5fbe7ffae852f84ddf0a4b4d2e043363"} Oct 08 18:26:57 crc kubenswrapper[4859]: I1008 18:26:57.859227 4859 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cert-manager/cert-manager-webhook-5655c58dd6-jkfxx" event={"ID":"24b1f415-e420-420a-b6c1-7fde05b080b6","Type":"ContainerStarted","Data":"74740da0ea34e8a30f95c3b6576b046c1111ac1f10c4f41f4bbe90e4a238801c"} Oct 08 18:26:57 crc kubenswrapper[4859]: I1008 18:26:57.860871 4859 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cert-manager/cert-manager-cainjector-7f985d654d-lqldn" event={"ID":"ccaaab95-a4e1-43b8-a962-0b833beec038","Type":"ContainerStarted","Data":"f0d4cb2fc65c383bbcc3e32a821db2a0d31e839c2567b8cecd41e4282a0f0e9d"} Oct 08 18:27:01 crc kubenswrapper[4859]: I1008 18:27:01.885330 4859 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cert-manager/cert-manager-5b446d88c5-w95bk" event={"ID":"2fa6fe28-fd3a-4611-ad50-2a4fc67ca563","Type":"ContainerStarted","Data":"6f297fcd116e520d58bffaf51d4542ef572cdcbb94c9066fe52d6e4614e2958f"} Oct 08 18:27:01 crc kubenswrapper[4859]: I1008 18:27:01.888666 4859 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cert-manager/cert-manager-webhook-5655c58dd6-jkfxx" event={"ID":"24b1f415-e420-420a-b6c1-7fde05b080b6","Type":"ContainerStarted","Data":"bea73b2f669f612fefe9577074314a2d90f088ee797c48c7b3d0332f2d008e3a"} Oct 08 18:27:01 crc kubenswrapper[4859]: I1008 18:27:01.888920 4859 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="cert-manager/cert-manager-webhook-5655c58dd6-jkfxx" Oct 08 18:27:01 crc kubenswrapper[4859]: I1008 18:27:01.914298 4859 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="cert-manager/cert-manager-5b446d88c5-w95bk" podStartSLOduration=2.310896174 podStartE2EDuration="5.914271354s" podCreationTimestamp="2025-10-08 18:26:56 +0000 UTC" firstStartedPulling="2025-10-08 18:26:57.556048363 +0000 UTC m=+587.802887742" lastFinishedPulling="2025-10-08 18:27:01.159423513 +0000 UTC m=+591.406262922" observedRunningTime="2025-10-08 18:27:01.906872967 +0000 UTC m=+592.153712356" watchObservedRunningTime="2025-10-08 18:27:01.914271354 +0000 UTC m=+592.161110733" Oct 08 18:27:01 crc kubenswrapper[4859]: I1008 18:27:01.925902 4859 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="cert-manager/cert-manager-webhook-5655c58dd6-jkfxx" podStartSLOduration=2.297340907 podStartE2EDuration="5.925881091s" podCreationTimestamp="2025-10-08 18:26:56 +0000 UTC" firstStartedPulling="2025-10-08 18:26:57.587634104 +0000 UTC m=+587.834473483" lastFinishedPulling="2025-10-08 18:27:01.216174288 +0000 UTC m=+591.463013667" observedRunningTime="2025-10-08 18:27:01.92391721 +0000 UTC m=+592.170756589" watchObservedRunningTime="2025-10-08 18:27:01.925881091 +0000 UTC m=+592.172720480" Oct 08 18:27:02 crc kubenswrapper[4859]: I1008 18:27:02.904820 4859 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cert-manager/cert-manager-cainjector-7f985d654d-lqldn" event={"ID":"ccaaab95-a4e1-43b8-a962-0b833beec038","Type":"ContainerStarted","Data":"7422996a137507ff615a96dab2d3eab9835bde55e8c38049828b55a2ba956ce3"} Oct 08 18:27:02 crc kubenswrapper[4859]: I1008 18:27:02.923997 4859 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="cert-manager/cert-manager-cainjector-7f985d654d-lqldn" podStartSLOduration=2.4831325570000002 podStartE2EDuration="6.923967627s" podCreationTimestamp="2025-10-08 18:26:56 +0000 UTC" firstStartedPulling="2025-10-08 18:26:57.510529784 +0000 UTC m=+587.757369163" lastFinishedPulling="2025-10-08 18:27:01.951364854 +0000 UTC m=+592.198204233" observedRunningTime="2025-10-08 18:27:02.920380337 +0000 UTC m=+593.167219766" watchObservedRunningTime="2025-10-08 18:27:02.923967627 +0000 UTC m=+593.170807006" Oct 08 18:27:07 crc kubenswrapper[4859]: I1008 18:27:07.328505 4859 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="cert-manager/cert-manager-webhook-5655c58dd6-jkfxx" Oct 08 18:27:07 crc kubenswrapper[4859]: I1008 18:27:07.589144 4859 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-ovn-kubernetes/ovnkube-node-4g8gf"] Oct 08 18:27:07 crc kubenswrapper[4859]: I1008 18:27:07.589504 4859 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-ovn-kubernetes/ovnkube-node-4g8gf" podUID="1dff864e-b75d-4e0b-b182-75f710eac8df" containerName="ovn-controller" containerID="cri-o://6ddc0e842bb8f81663406070023b0e5ed9cc99c98db0123b1c84f20da59fb5b5" gracePeriod=30 Oct 08 18:27:07 crc kubenswrapper[4859]: I1008 18:27:07.589583 4859 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-ovn-kubernetes/ovnkube-node-4g8gf" podUID="1dff864e-b75d-4e0b-b182-75f710eac8df" containerName="nbdb" containerID="cri-o://b8009a42cc4c9a816457a0678bdb2f7b7c53d0c8604e007a6fb141f66d06a516" gracePeriod=30 Oct 08 18:27:07 crc kubenswrapper[4859]: I1008 18:27:07.589648 4859 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-ovn-kubernetes/ovnkube-node-4g8gf" podUID="1dff864e-b75d-4e0b-b182-75f710eac8df" containerName="kube-rbac-proxy-ovn-metrics" containerID="cri-o://669f5f3c6d3c1ec0c4b4870e090b0a853adc01727c23383e595715b476c32668" gracePeriod=30 Oct 08 18:27:07 crc kubenswrapper[4859]: I1008 18:27:07.589731 4859 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-ovn-kubernetes/ovnkube-node-4g8gf" podUID="1dff864e-b75d-4e0b-b182-75f710eac8df" containerName="kube-rbac-proxy-node" containerID="cri-o://768ca274dedcc6847fc21d15234d3665fa9ccae371b5b6a882cfcb9b182f527d" gracePeriod=30 Oct 08 18:27:07 crc kubenswrapper[4859]: I1008 18:27:07.589769 4859 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-ovn-kubernetes/ovnkube-node-4g8gf" podUID="1dff864e-b75d-4e0b-b182-75f710eac8df" containerName="ovn-acl-logging" containerID="cri-o://4a74c1f0d21c90cdd7e776ee14e1f48757a04e2220fbd1f884449181ec6d1c3a" gracePeriod=30 Oct 08 18:27:07 crc kubenswrapper[4859]: I1008 18:27:07.589635 4859 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-ovn-kubernetes/ovnkube-node-4g8gf" podUID="1dff864e-b75d-4e0b-b182-75f710eac8df" containerName="northd" containerID="cri-o://676511a11ed4102084b57ebdd5f3fda4969741b2bc8443b415c4bf94f39ae4c9" gracePeriod=30 Oct 08 18:27:07 crc kubenswrapper[4859]: I1008 18:27:07.590072 4859 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-ovn-kubernetes/ovnkube-node-4g8gf" podUID="1dff864e-b75d-4e0b-b182-75f710eac8df" containerName="sbdb" containerID="cri-o://2c67d2aed7126d76989e057d8e2d8e188a8902b2f3fe26ed6d1d45d3b013d897" gracePeriod=30 Oct 08 18:27:07 crc kubenswrapper[4859]: I1008 18:27:07.640461 4859 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-ovn-kubernetes/ovnkube-node-4g8gf" podUID="1dff864e-b75d-4e0b-b182-75f710eac8df" containerName="ovnkube-controller" containerID="cri-o://b737cb013e2877d970d7e301048f95e0ba0ade0b8db42b2607a78fdf2580659f" gracePeriod=30 Oct 08 18:27:07 crc kubenswrapper[4859]: I1008 18:27:07.937424 4859 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-8plkg_92b2d7de-31cc-4a91-95a5-ed5ea964b028/kube-multus/2.log" Oct 08 18:27:07 crc kubenswrapper[4859]: I1008 18:27:07.938154 4859 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-8plkg_92b2d7de-31cc-4a91-95a5-ed5ea964b028/kube-multus/1.log" Oct 08 18:27:07 crc kubenswrapper[4859]: I1008 18:27:07.938199 4859 generic.go:334] "Generic (PLEG): container finished" podID="92b2d7de-31cc-4a91-95a5-ed5ea964b028" containerID="d18a8ef0ab4fc3ed687ccf7e94753da8a4a29dae40a39aae2b64a854761f90d0" exitCode=2 Oct 08 18:27:07 crc kubenswrapper[4859]: I1008 18:27:07.938272 4859 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-8plkg" event={"ID":"92b2d7de-31cc-4a91-95a5-ed5ea964b028","Type":"ContainerDied","Data":"d18a8ef0ab4fc3ed687ccf7e94753da8a4a29dae40a39aae2b64a854761f90d0"} Oct 08 18:27:07 crc kubenswrapper[4859]: I1008 18:27:07.938316 4859 scope.go:117] "RemoveContainer" containerID="729bd32fd01dcab495bffa64f2f5e44990fc1d78b05e7f131868009f92305831" Oct 08 18:27:07 crc kubenswrapper[4859]: I1008 18:27:07.939228 4859 scope.go:117] "RemoveContainer" containerID="d18a8ef0ab4fc3ed687ccf7e94753da8a4a29dae40a39aae2b64a854761f90d0" Oct 08 18:27:07 crc kubenswrapper[4859]: E1008 18:27:07.939666 4859 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"kube-multus\" with CrashLoopBackOff: \"back-off 20s restarting failed container=kube-multus pod=multus-8plkg_openshift-multus(92b2d7de-31cc-4a91-95a5-ed5ea964b028)\"" pod="openshift-multus/multus-8plkg" podUID="92b2d7de-31cc-4a91-95a5-ed5ea964b028" Oct 08 18:27:07 crc kubenswrapper[4859]: I1008 18:27:07.941285 4859 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-4g8gf_1dff864e-b75d-4e0b-b182-75f710eac8df/ovnkube-controller/3.log" Oct 08 18:27:07 crc kubenswrapper[4859]: I1008 18:27:07.950995 4859 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-4g8gf_1dff864e-b75d-4e0b-b182-75f710eac8df/ovn-acl-logging/0.log" Oct 08 18:27:07 crc kubenswrapper[4859]: I1008 18:27:07.951899 4859 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-4g8gf_1dff864e-b75d-4e0b-b182-75f710eac8df/ovn-controller/0.log" Oct 08 18:27:07 crc kubenswrapper[4859]: I1008 18:27:07.954772 4859 generic.go:334] "Generic (PLEG): container finished" podID="1dff864e-b75d-4e0b-b182-75f710eac8df" containerID="b737cb013e2877d970d7e301048f95e0ba0ade0b8db42b2607a78fdf2580659f" exitCode=0 Oct 08 18:27:07 crc kubenswrapper[4859]: I1008 18:27:07.954804 4859 generic.go:334] "Generic (PLEG): container finished" podID="1dff864e-b75d-4e0b-b182-75f710eac8df" containerID="2c67d2aed7126d76989e057d8e2d8e188a8902b2f3fe26ed6d1d45d3b013d897" exitCode=0 Oct 08 18:27:07 crc kubenswrapper[4859]: I1008 18:27:07.954811 4859 generic.go:334] "Generic (PLEG): container finished" podID="1dff864e-b75d-4e0b-b182-75f710eac8df" containerID="b8009a42cc4c9a816457a0678bdb2f7b7c53d0c8604e007a6fb141f66d06a516" exitCode=0 Oct 08 18:27:07 crc kubenswrapper[4859]: I1008 18:27:07.954819 4859 generic.go:334] "Generic (PLEG): container finished" podID="1dff864e-b75d-4e0b-b182-75f710eac8df" containerID="676511a11ed4102084b57ebdd5f3fda4969741b2bc8443b415c4bf94f39ae4c9" exitCode=0 Oct 08 18:27:07 crc kubenswrapper[4859]: I1008 18:27:07.954827 4859 generic.go:334] "Generic (PLEG): container finished" podID="1dff864e-b75d-4e0b-b182-75f710eac8df" containerID="669f5f3c6d3c1ec0c4b4870e090b0a853adc01727c23383e595715b476c32668" exitCode=0 Oct 08 18:27:07 crc kubenswrapper[4859]: I1008 18:27:07.954834 4859 generic.go:334] "Generic (PLEG): container finished" podID="1dff864e-b75d-4e0b-b182-75f710eac8df" containerID="768ca274dedcc6847fc21d15234d3665fa9ccae371b5b6a882cfcb9b182f527d" exitCode=0 Oct 08 18:27:07 crc kubenswrapper[4859]: I1008 18:27:07.954843 4859 generic.go:334] "Generic (PLEG): container finished" podID="1dff864e-b75d-4e0b-b182-75f710eac8df" containerID="4a74c1f0d21c90cdd7e776ee14e1f48757a04e2220fbd1f884449181ec6d1c3a" exitCode=143 Oct 08 18:27:07 crc kubenswrapper[4859]: I1008 18:27:07.954850 4859 generic.go:334] "Generic (PLEG): container finished" podID="1dff864e-b75d-4e0b-b182-75f710eac8df" containerID="6ddc0e842bb8f81663406070023b0e5ed9cc99c98db0123b1c84f20da59fb5b5" exitCode=143 Oct 08 18:27:07 crc kubenswrapper[4859]: I1008 18:27:07.954861 4859 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-4g8gf" event={"ID":"1dff864e-b75d-4e0b-b182-75f710eac8df","Type":"ContainerDied","Data":"b737cb013e2877d970d7e301048f95e0ba0ade0b8db42b2607a78fdf2580659f"} Oct 08 18:27:07 crc kubenswrapper[4859]: I1008 18:27:07.954913 4859 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-4g8gf" event={"ID":"1dff864e-b75d-4e0b-b182-75f710eac8df","Type":"ContainerDied","Data":"2c67d2aed7126d76989e057d8e2d8e188a8902b2f3fe26ed6d1d45d3b013d897"} Oct 08 18:27:07 crc kubenswrapper[4859]: I1008 18:27:07.954923 4859 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-4g8gf" event={"ID":"1dff864e-b75d-4e0b-b182-75f710eac8df","Type":"ContainerDied","Data":"b8009a42cc4c9a816457a0678bdb2f7b7c53d0c8604e007a6fb141f66d06a516"} Oct 08 18:27:07 crc kubenswrapper[4859]: I1008 18:27:07.954933 4859 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-4g8gf" event={"ID":"1dff864e-b75d-4e0b-b182-75f710eac8df","Type":"ContainerDied","Data":"676511a11ed4102084b57ebdd5f3fda4969741b2bc8443b415c4bf94f39ae4c9"} Oct 08 18:27:07 crc kubenswrapper[4859]: I1008 18:27:07.954945 4859 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-4g8gf" event={"ID":"1dff864e-b75d-4e0b-b182-75f710eac8df","Type":"ContainerDied","Data":"669f5f3c6d3c1ec0c4b4870e090b0a853adc01727c23383e595715b476c32668"} Oct 08 18:27:07 crc kubenswrapper[4859]: I1008 18:27:07.954957 4859 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-4g8gf" event={"ID":"1dff864e-b75d-4e0b-b182-75f710eac8df","Type":"ContainerDied","Data":"768ca274dedcc6847fc21d15234d3665fa9ccae371b5b6a882cfcb9b182f527d"} Oct 08 18:27:07 crc kubenswrapper[4859]: I1008 18:27:07.954967 4859 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-4g8gf" event={"ID":"1dff864e-b75d-4e0b-b182-75f710eac8df","Type":"ContainerDied","Data":"4a74c1f0d21c90cdd7e776ee14e1f48757a04e2220fbd1f884449181ec6d1c3a"} Oct 08 18:27:07 crc kubenswrapper[4859]: I1008 18:27:07.954977 4859 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-4g8gf" event={"ID":"1dff864e-b75d-4e0b-b182-75f710eac8df","Type":"ContainerDied","Data":"6ddc0e842bb8f81663406070023b0e5ed9cc99c98db0123b1c84f20da59fb5b5"} Oct 08 18:27:07 crc kubenswrapper[4859]: I1008 18:27:07.954987 4859 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-4g8gf" event={"ID":"1dff864e-b75d-4e0b-b182-75f710eac8df","Type":"ContainerDied","Data":"9cde21bcc626ceb4fd1a09cfe04cba4c327c7e4ad650edc972e0765907465fba"} Oct 08 18:27:07 crc kubenswrapper[4859]: I1008 18:27:07.954998 4859 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="9cde21bcc626ceb4fd1a09cfe04cba4c327c7e4ad650edc972e0765907465fba" Oct 08 18:27:07 crc kubenswrapper[4859]: I1008 18:27:07.971497 4859 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-4g8gf_1dff864e-b75d-4e0b-b182-75f710eac8df/ovnkube-controller/3.log" Oct 08 18:27:07 crc kubenswrapper[4859]: I1008 18:27:07.977393 4859 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-4g8gf_1dff864e-b75d-4e0b-b182-75f710eac8df/ovn-acl-logging/0.log" Oct 08 18:27:07 crc kubenswrapper[4859]: I1008 18:27:07.977399 4859 scope.go:117] "RemoveContainer" containerID="f3a760c557ef21fbc6b54e78c3f6c5b325f296f93a022425cca7b1399eed8661" Oct 08 18:27:07 crc kubenswrapper[4859]: I1008 18:27:07.977828 4859 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-4g8gf_1dff864e-b75d-4e0b-b182-75f710eac8df/ovn-controller/0.log" Oct 08 18:27:07 crc kubenswrapper[4859]: I1008 18:27:07.978381 4859 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-ovn-kubernetes/ovnkube-node-4g8gf" Oct 08 18:27:08 crc kubenswrapper[4859]: I1008 18:27:08.070971 4859 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"node-log\" (UniqueName: \"kubernetes.io/host-path/1dff864e-b75d-4e0b-b182-75f710eac8df-node-log\") pod \"1dff864e-b75d-4e0b-b182-75f710eac8df\" (UID: \"1dff864e-b75d-4e0b-b182-75f710eac8df\") " Oct 08 18:27:08 crc kubenswrapper[4859]: I1008 18:27:08.071034 4859 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/1dff864e-b75d-4e0b-b182-75f710eac8df-env-overrides\") pod \"1dff864e-b75d-4e0b-b182-75f710eac8df\" (UID: \"1dff864e-b75d-4e0b-b182-75f710eac8df\") " Oct 08 18:27:08 crc kubenswrapper[4859]: I1008 18:27:08.071071 4859 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host-kubelet\" (UniqueName: \"kubernetes.io/host-path/1dff864e-b75d-4e0b-b182-75f710eac8df-host-kubelet\") pod \"1dff864e-b75d-4e0b-b182-75f710eac8df\" (UID: \"1dff864e-b75d-4e0b-b182-75f710eac8df\") " Oct 08 18:27:08 crc kubenswrapper[4859]: I1008 18:27:08.071098 4859 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/1dff864e-b75d-4e0b-b182-75f710eac8df-ovnkube-script-lib\") pod \"1dff864e-b75d-4e0b-b182-75f710eac8df\" (UID: \"1dff864e-b75d-4e0b-b182-75f710eac8df\") " Oct 08 18:27:08 crc kubenswrapper[4859]: I1008 18:27:08.071131 4859 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"log-socket\" (UniqueName: \"kubernetes.io/host-path/1dff864e-b75d-4e0b-b182-75f710eac8df-log-socket\") pod \"1dff864e-b75d-4e0b-b182-75f710eac8df\" (UID: \"1dff864e-b75d-4e0b-b182-75f710eac8df\") " Oct 08 18:27:08 crc kubenswrapper[4859]: I1008 18:27:08.071117 4859 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/1dff864e-b75d-4e0b-b182-75f710eac8df-node-log" (OuterVolumeSpecName: "node-log") pod "1dff864e-b75d-4e0b-b182-75f710eac8df" (UID: "1dff864e-b75d-4e0b-b182-75f710eac8df"). InnerVolumeSpecName "node-log". PluginName "kubernetes.io/host-path", VolumeGidValue "" Oct 08 18:27:08 crc kubenswrapper[4859]: I1008 18:27:08.071199 4859 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host-run-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/1dff864e-b75d-4e0b-b182-75f710eac8df-host-run-ovn-kubernetes\") pod \"1dff864e-b75d-4e0b-b182-75f710eac8df\" (UID: \"1dff864e-b75d-4e0b-b182-75f710eac8df\") " Oct 08 18:27:08 crc kubenswrapper[4859]: I1008 18:27:08.071223 4859 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/1dff864e-b75d-4e0b-b182-75f710eac8df-host-slash\") pod \"1dff864e-b75d-4e0b-b182-75f710eac8df\" (UID: \"1dff864e-b75d-4e0b-b182-75f710eac8df\") " Oct 08 18:27:08 crc kubenswrapper[4859]: I1008 18:27:08.071248 4859 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-systemd\" (UniqueName: \"kubernetes.io/host-path/1dff864e-b75d-4e0b-b182-75f710eac8df-run-systemd\") pod \"1dff864e-b75d-4e0b-b182-75f710eac8df\" (UID: \"1dff864e-b75d-4e0b-b182-75f710eac8df\") " Oct 08 18:27:08 crc kubenswrapper[4859]: I1008 18:27:08.071271 4859 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-openvswitch\" (UniqueName: \"kubernetes.io/host-path/1dff864e-b75d-4e0b-b182-75f710eac8df-run-openvswitch\") pod \"1dff864e-b75d-4e0b-b182-75f710eac8df\" (UID: \"1dff864e-b75d-4e0b-b182-75f710eac8df\") " Oct 08 18:27:08 crc kubenswrapper[4859]: I1008 18:27:08.071237 4859 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/1dff864e-b75d-4e0b-b182-75f710eac8df-log-socket" (OuterVolumeSpecName: "log-socket") pod "1dff864e-b75d-4e0b-b182-75f710eac8df" (UID: "1dff864e-b75d-4e0b-b182-75f710eac8df"). InnerVolumeSpecName "log-socket". PluginName "kubernetes.io/host-path", VolumeGidValue "" Oct 08 18:27:08 crc kubenswrapper[4859]: I1008 18:27:08.071296 4859 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/1dff864e-b75d-4e0b-b182-75f710eac8df-host-slash" (OuterVolumeSpecName: "host-slash") pod "1dff864e-b75d-4e0b-b182-75f710eac8df" (UID: "1dff864e-b75d-4e0b-b182-75f710eac8df"). InnerVolumeSpecName "host-slash". PluginName "kubernetes.io/host-path", VolumeGidValue "" Oct 08 18:27:08 crc kubenswrapper[4859]: I1008 18:27:08.071295 4859 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/1dff864e-b75d-4e0b-b182-75f710eac8df-host-run-netns\") pod \"1dff864e-b75d-4e0b-b182-75f710eac8df\" (UID: \"1dff864e-b75d-4e0b-b182-75f710eac8df\") " Oct 08 18:27:08 crc kubenswrapper[4859]: I1008 18:27:08.071282 4859 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/1dff864e-b75d-4e0b-b182-75f710eac8df-host-run-ovn-kubernetes" (OuterVolumeSpecName: "host-run-ovn-kubernetes") pod "1dff864e-b75d-4e0b-b182-75f710eac8df" (UID: "1dff864e-b75d-4e0b-b182-75f710eac8df"). InnerVolumeSpecName "host-run-ovn-kubernetes". PluginName "kubernetes.io/host-path", VolumeGidValue "" Oct 08 18:27:08 crc kubenswrapper[4859]: I1008 18:27:08.071342 4859 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/1dff864e-b75d-4e0b-b182-75f710eac8df-run-openvswitch" (OuterVolumeSpecName: "run-openvswitch") pod "1dff864e-b75d-4e0b-b182-75f710eac8df" (UID: "1dff864e-b75d-4e0b-b182-75f710eac8df"). InnerVolumeSpecName "run-openvswitch". PluginName "kubernetes.io/host-path", VolumeGidValue "" Oct 08 18:27:08 crc kubenswrapper[4859]: I1008 18:27:08.071372 4859 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/1dff864e-b75d-4e0b-b182-75f710eac8df-host-run-netns" (OuterVolumeSpecName: "host-run-netns") pod "1dff864e-b75d-4e0b-b182-75f710eac8df" (UID: "1dff864e-b75d-4e0b-b182-75f710eac8df"). InnerVolumeSpecName "host-run-netns". PluginName "kubernetes.io/host-path", VolumeGidValue "" Oct 08 18:27:08 crc kubenswrapper[4859]: I1008 18:27:08.071484 4859 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host-var-lib-cni-networks-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/1dff864e-b75d-4e0b-b182-75f710eac8df-host-var-lib-cni-networks-ovn-kubernetes\") pod \"1dff864e-b75d-4e0b-b182-75f710eac8df\" (UID: \"1dff864e-b75d-4e0b-b182-75f710eac8df\") " Oct 08 18:27:08 crc kubenswrapper[4859]: I1008 18:27:08.071518 4859 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host-cni-netd\" (UniqueName: \"kubernetes.io/host-path/1dff864e-b75d-4e0b-b182-75f710eac8df-host-cni-netd\") pod \"1dff864e-b75d-4e0b-b182-75f710eac8df\" (UID: \"1dff864e-b75d-4e0b-b182-75f710eac8df\") " Oct 08 18:27:08 crc kubenswrapper[4859]: I1008 18:27:08.071529 4859 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/1dff864e-b75d-4e0b-b182-75f710eac8df-host-var-lib-cni-networks-ovn-kubernetes" (OuterVolumeSpecName: "host-var-lib-cni-networks-ovn-kubernetes") pod "1dff864e-b75d-4e0b-b182-75f710eac8df" (UID: "1dff864e-b75d-4e0b-b182-75f710eac8df"). InnerVolumeSpecName "host-var-lib-cni-networks-ovn-kubernetes". PluginName "kubernetes.io/host-path", VolumeGidValue "" Oct 08 18:27:08 crc kubenswrapper[4859]: I1008 18:27:08.071551 4859 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"var-lib-openvswitch\" (UniqueName: \"kubernetes.io/host-path/1dff864e-b75d-4e0b-b182-75f710eac8df-var-lib-openvswitch\") pod \"1dff864e-b75d-4e0b-b182-75f710eac8df\" (UID: \"1dff864e-b75d-4e0b-b182-75f710eac8df\") " Oct 08 18:27:08 crc kubenswrapper[4859]: I1008 18:27:08.071565 4859 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/1dff864e-b75d-4e0b-b182-75f710eac8df-host-cni-netd" (OuterVolumeSpecName: "host-cni-netd") pod "1dff864e-b75d-4e0b-b182-75f710eac8df" (UID: "1dff864e-b75d-4e0b-b182-75f710eac8df"). InnerVolumeSpecName "host-cni-netd". PluginName "kubernetes.io/host-path", VolumeGidValue "" Oct 08 18:27:08 crc kubenswrapper[4859]: I1008 18:27:08.071575 4859 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-ovn\" (UniqueName: \"kubernetes.io/host-path/1dff864e-b75d-4e0b-b182-75f710eac8df-run-ovn\") pod \"1dff864e-b75d-4e0b-b182-75f710eac8df\" (UID: \"1dff864e-b75d-4e0b-b182-75f710eac8df\") " Oct 08 18:27:08 crc kubenswrapper[4859]: I1008 18:27:08.071596 4859 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/1dff864e-b75d-4e0b-b182-75f710eac8df-var-lib-openvswitch" (OuterVolumeSpecName: "var-lib-openvswitch") pod "1dff864e-b75d-4e0b-b182-75f710eac8df" (UID: "1dff864e-b75d-4e0b-b182-75f710eac8df"). InnerVolumeSpecName "var-lib-openvswitch". PluginName "kubernetes.io/host-path", VolumeGidValue "" Oct 08 18:27:08 crc kubenswrapper[4859]: I1008 18:27:08.071614 4859 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/1dff864e-b75d-4e0b-b182-75f710eac8df-run-ovn" (OuterVolumeSpecName: "run-ovn") pod "1dff864e-b75d-4e0b-b182-75f710eac8df" (UID: "1dff864e-b75d-4e0b-b182-75f710eac8df"). InnerVolumeSpecName "run-ovn". PluginName "kubernetes.io/host-path", VolumeGidValue "" Oct 08 18:27:08 crc kubenswrapper[4859]: I1008 18:27:08.071623 4859 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1dff864e-b75d-4e0b-b182-75f710eac8df-ovnkube-script-lib" (OuterVolumeSpecName: "ovnkube-script-lib") pod "1dff864e-b75d-4e0b-b182-75f710eac8df" (UID: "1dff864e-b75d-4e0b-b182-75f710eac8df"). InnerVolumeSpecName "ovnkube-script-lib". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 08 18:27:08 crc kubenswrapper[4859]: I1008 18:27:08.071624 4859 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/1dff864e-b75d-4e0b-b182-75f710eac8df-ovnkube-config\") pod \"1dff864e-b75d-4e0b-b182-75f710eac8df\" (UID: \"1dff864e-b75d-4e0b-b182-75f710eac8df\") " Oct 08 18:27:08 crc kubenswrapper[4859]: I1008 18:27:08.071710 4859 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-openvswitch\" (UniqueName: \"kubernetes.io/host-path/1dff864e-b75d-4e0b-b182-75f710eac8df-etc-openvswitch\") pod \"1dff864e-b75d-4e0b-b182-75f710eac8df\" (UID: \"1dff864e-b75d-4e0b-b182-75f710eac8df\") " Oct 08 18:27:08 crc kubenswrapper[4859]: I1008 18:27:08.071739 4859 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"systemd-units\" (UniqueName: \"kubernetes.io/host-path/1dff864e-b75d-4e0b-b182-75f710eac8df-systemd-units\") pod \"1dff864e-b75d-4e0b-b182-75f710eac8df\" (UID: \"1dff864e-b75d-4e0b-b182-75f710eac8df\") " Oct 08 18:27:08 crc kubenswrapper[4859]: I1008 18:27:08.071764 4859 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host-cni-bin\" (UniqueName: \"kubernetes.io/host-path/1dff864e-b75d-4e0b-b182-75f710eac8df-host-cni-bin\") pod \"1dff864e-b75d-4e0b-b182-75f710eac8df\" (UID: \"1dff864e-b75d-4e0b-b182-75f710eac8df\") " Oct 08 18:27:08 crc kubenswrapper[4859]: I1008 18:27:08.071758 4859 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1dff864e-b75d-4e0b-b182-75f710eac8df-env-overrides" (OuterVolumeSpecName: "env-overrides") pod "1dff864e-b75d-4e0b-b182-75f710eac8df" (UID: "1dff864e-b75d-4e0b-b182-75f710eac8df"). InnerVolumeSpecName "env-overrides". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 08 18:27:08 crc kubenswrapper[4859]: I1008 18:27:08.071795 4859 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/1dff864e-b75d-4e0b-b182-75f710eac8df-systemd-units" (OuterVolumeSpecName: "systemd-units") pod "1dff864e-b75d-4e0b-b182-75f710eac8df" (UID: "1dff864e-b75d-4e0b-b182-75f710eac8df"). InnerVolumeSpecName "systemd-units". PluginName "kubernetes.io/host-path", VolumeGidValue "" Oct 08 18:27:08 crc kubenswrapper[4859]: I1008 18:27:08.071796 4859 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/1dff864e-b75d-4e0b-b182-75f710eac8df-etc-openvswitch" (OuterVolumeSpecName: "etc-openvswitch") pod "1dff864e-b75d-4e0b-b182-75f710eac8df" (UID: "1dff864e-b75d-4e0b-b182-75f710eac8df"). InnerVolumeSpecName "etc-openvswitch". PluginName "kubernetes.io/host-path", VolumeGidValue "" Oct 08 18:27:08 crc kubenswrapper[4859]: I1008 18:27:08.071810 4859 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/1dff864e-b75d-4e0b-b182-75f710eac8df-host-cni-bin" (OuterVolumeSpecName: "host-cni-bin") pod "1dff864e-b75d-4e0b-b182-75f710eac8df" (UID: "1dff864e-b75d-4e0b-b182-75f710eac8df"). InnerVolumeSpecName "host-cni-bin". PluginName "kubernetes.io/host-path", VolumeGidValue "" Oct 08 18:27:08 crc kubenswrapper[4859]: I1008 18:27:08.071818 4859 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/1dff864e-b75d-4e0b-b182-75f710eac8df-ovn-node-metrics-cert\") pod \"1dff864e-b75d-4e0b-b182-75f710eac8df\" (UID: \"1dff864e-b75d-4e0b-b182-75f710eac8df\") " Oct 08 18:27:08 crc kubenswrapper[4859]: I1008 18:27:08.071923 4859 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-s8jnn\" (UniqueName: \"kubernetes.io/projected/1dff864e-b75d-4e0b-b182-75f710eac8df-kube-api-access-s8jnn\") pod \"1dff864e-b75d-4e0b-b182-75f710eac8df\" (UID: \"1dff864e-b75d-4e0b-b182-75f710eac8df\") " Oct 08 18:27:08 crc kubenswrapper[4859]: I1008 18:27:08.072102 4859 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1dff864e-b75d-4e0b-b182-75f710eac8df-ovnkube-config" (OuterVolumeSpecName: "ovnkube-config") pod "1dff864e-b75d-4e0b-b182-75f710eac8df" (UID: "1dff864e-b75d-4e0b-b182-75f710eac8df"). InnerVolumeSpecName "ovnkube-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 08 18:27:08 crc kubenswrapper[4859]: I1008 18:27:08.072368 4859 reconciler_common.go:293] "Volume detached for volume \"log-socket\" (UniqueName: \"kubernetes.io/host-path/1dff864e-b75d-4e0b-b182-75f710eac8df-log-socket\") on node \"crc\" DevicePath \"\"" Oct 08 18:27:08 crc kubenswrapper[4859]: I1008 18:27:08.072383 4859 reconciler_common.go:293] "Volume detached for volume \"host-run-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/1dff864e-b75d-4e0b-b182-75f710eac8df-host-run-ovn-kubernetes\") on node \"crc\" DevicePath \"\"" Oct 08 18:27:08 crc kubenswrapper[4859]: I1008 18:27:08.072394 4859 reconciler_common.go:293] "Volume detached for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/1dff864e-b75d-4e0b-b182-75f710eac8df-host-slash\") on node \"crc\" DevicePath \"\"" Oct 08 18:27:08 crc kubenswrapper[4859]: I1008 18:27:08.072393 4859 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/1dff864e-b75d-4e0b-b182-75f710eac8df-host-kubelet" (OuterVolumeSpecName: "host-kubelet") pod "1dff864e-b75d-4e0b-b182-75f710eac8df" (UID: "1dff864e-b75d-4e0b-b182-75f710eac8df"). InnerVolumeSpecName "host-kubelet". PluginName "kubernetes.io/host-path", VolumeGidValue "" Oct 08 18:27:08 crc kubenswrapper[4859]: I1008 18:27:08.072402 4859 reconciler_common.go:293] "Volume detached for volume \"run-openvswitch\" (UniqueName: \"kubernetes.io/host-path/1dff864e-b75d-4e0b-b182-75f710eac8df-run-openvswitch\") on node \"crc\" DevicePath \"\"" Oct 08 18:27:08 crc kubenswrapper[4859]: I1008 18:27:08.072449 4859 reconciler_common.go:293] "Volume detached for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/1dff864e-b75d-4e0b-b182-75f710eac8df-host-run-netns\") on node \"crc\" DevicePath \"\"" Oct 08 18:27:08 crc kubenswrapper[4859]: I1008 18:27:08.072463 4859 reconciler_common.go:293] "Volume detached for volume \"host-var-lib-cni-networks-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/1dff864e-b75d-4e0b-b182-75f710eac8df-host-var-lib-cni-networks-ovn-kubernetes\") on node \"crc\" DevicePath \"\"" Oct 08 18:27:08 crc kubenswrapper[4859]: I1008 18:27:08.072479 4859 reconciler_common.go:293] "Volume detached for volume \"host-cni-netd\" (UniqueName: \"kubernetes.io/host-path/1dff864e-b75d-4e0b-b182-75f710eac8df-host-cni-netd\") on node \"crc\" DevicePath \"\"" Oct 08 18:27:08 crc kubenswrapper[4859]: I1008 18:27:08.072503 4859 reconciler_common.go:293] "Volume detached for volume \"var-lib-openvswitch\" (UniqueName: \"kubernetes.io/host-path/1dff864e-b75d-4e0b-b182-75f710eac8df-var-lib-openvswitch\") on node \"crc\" DevicePath \"\"" Oct 08 18:27:08 crc kubenswrapper[4859]: I1008 18:27:08.072515 4859 reconciler_common.go:293] "Volume detached for volume \"run-ovn\" (UniqueName: \"kubernetes.io/host-path/1dff864e-b75d-4e0b-b182-75f710eac8df-run-ovn\") on node \"crc\" DevicePath \"\"" Oct 08 18:27:08 crc kubenswrapper[4859]: I1008 18:27:08.072526 4859 reconciler_common.go:293] "Volume detached for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/1dff864e-b75d-4e0b-b182-75f710eac8df-ovnkube-config\") on node \"crc\" DevicePath \"\"" Oct 08 18:27:08 crc kubenswrapper[4859]: I1008 18:27:08.072537 4859 reconciler_common.go:293] "Volume detached for volume \"systemd-units\" (UniqueName: \"kubernetes.io/host-path/1dff864e-b75d-4e0b-b182-75f710eac8df-systemd-units\") on node \"crc\" DevicePath \"\"" Oct 08 18:27:08 crc kubenswrapper[4859]: I1008 18:27:08.072548 4859 reconciler_common.go:293] "Volume detached for volume \"etc-openvswitch\" (UniqueName: \"kubernetes.io/host-path/1dff864e-b75d-4e0b-b182-75f710eac8df-etc-openvswitch\") on node \"crc\" DevicePath \"\"" Oct 08 18:27:08 crc kubenswrapper[4859]: I1008 18:27:08.072558 4859 reconciler_common.go:293] "Volume detached for volume \"host-cni-bin\" (UniqueName: \"kubernetes.io/host-path/1dff864e-b75d-4e0b-b182-75f710eac8df-host-cni-bin\") on node \"crc\" DevicePath \"\"" Oct 08 18:27:08 crc kubenswrapper[4859]: I1008 18:27:08.072569 4859 reconciler_common.go:293] "Volume detached for volume \"node-log\" (UniqueName: \"kubernetes.io/host-path/1dff864e-b75d-4e0b-b182-75f710eac8df-node-log\") on node \"crc\" DevicePath \"\"" Oct 08 18:27:08 crc kubenswrapper[4859]: I1008 18:27:08.072581 4859 reconciler_common.go:293] "Volume detached for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/1dff864e-b75d-4e0b-b182-75f710eac8df-env-overrides\") on node \"crc\" DevicePath \"\"" Oct 08 18:27:08 crc kubenswrapper[4859]: I1008 18:27:08.072592 4859 reconciler_common.go:293] "Volume detached for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/1dff864e-b75d-4e0b-b182-75f710eac8df-ovnkube-script-lib\") on node \"crc\" DevicePath \"\"" Oct 08 18:27:08 crc kubenswrapper[4859]: I1008 18:27:08.081356 4859 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1dff864e-b75d-4e0b-b182-75f710eac8df-ovn-node-metrics-cert" (OuterVolumeSpecName: "ovn-node-metrics-cert") pod "1dff864e-b75d-4e0b-b182-75f710eac8df" (UID: "1dff864e-b75d-4e0b-b182-75f710eac8df"). InnerVolumeSpecName "ovn-node-metrics-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 08 18:27:08 crc kubenswrapper[4859]: I1008 18:27:08.087127 4859 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-ovn-kubernetes/ovnkube-node-mt6h5"] Oct 08 18:27:08 crc kubenswrapper[4859]: E1008 18:27:08.087377 4859 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1dff864e-b75d-4e0b-b182-75f710eac8df" containerName="nbdb" Oct 08 18:27:08 crc kubenswrapper[4859]: I1008 18:27:08.087393 4859 state_mem.go:107] "Deleted CPUSet assignment" podUID="1dff864e-b75d-4e0b-b182-75f710eac8df" containerName="nbdb" Oct 08 18:27:08 crc kubenswrapper[4859]: E1008 18:27:08.087414 4859 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1dff864e-b75d-4e0b-b182-75f710eac8df" containerName="ovnkube-controller" Oct 08 18:27:08 crc kubenswrapper[4859]: I1008 18:27:08.087423 4859 state_mem.go:107] "Deleted CPUSet assignment" podUID="1dff864e-b75d-4e0b-b182-75f710eac8df" containerName="ovnkube-controller" Oct 08 18:27:08 crc kubenswrapper[4859]: E1008 18:27:08.087433 4859 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1dff864e-b75d-4e0b-b182-75f710eac8df" containerName="ovnkube-controller" Oct 08 18:27:08 crc kubenswrapper[4859]: I1008 18:27:08.087441 4859 state_mem.go:107] "Deleted CPUSet assignment" podUID="1dff864e-b75d-4e0b-b182-75f710eac8df" containerName="ovnkube-controller" Oct 08 18:27:08 crc kubenswrapper[4859]: E1008 18:27:08.087452 4859 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1dff864e-b75d-4e0b-b182-75f710eac8df" containerName="kubecfg-setup" Oct 08 18:27:08 crc kubenswrapper[4859]: I1008 18:27:08.087460 4859 state_mem.go:107] "Deleted CPUSet assignment" podUID="1dff864e-b75d-4e0b-b182-75f710eac8df" containerName="kubecfg-setup" Oct 08 18:27:08 crc kubenswrapper[4859]: E1008 18:27:08.087472 4859 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1dff864e-b75d-4e0b-b182-75f710eac8df" containerName="ovnkube-controller" Oct 08 18:27:08 crc kubenswrapper[4859]: I1008 18:27:08.087479 4859 state_mem.go:107] "Deleted CPUSet assignment" podUID="1dff864e-b75d-4e0b-b182-75f710eac8df" containerName="ovnkube-controller" Oct 08 18:27:08 crc kubenswrapper[4859]: E1008 18:27:08.087487 4859 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1dff864e-b75d-4e0b-b182-75f710eac8df" containerName="ovn-acl-logging" Oct 08 18:27:08 crc kubenswrapper[4859]: I1008 18:27:08.087497 4859 state_mem.go:107] "Deleted CPUSet assignment" podUID="1dff864e-b75d-4e0b-b182-75f710eac8df" containerName="ovn-acl-logging" Oct 08 18:27:08 crc kubenswrapper[4859]: E1008 18:27:08.087509 4859 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1dff864e-b75d-4e0b-b182-75f710eac8df" containerName="ovn-controller" Oct 08 18:27:08 crc kubenswrapper[4859]: I1008 18:27:08.087518 4859 state_mem.go:107] "Deleted CPUSet assignment" podUID="1dff864e-b75d-4e0b-b182-75f710eac8df" containerName="ovn-controller" Oct 08 18:27:08 crc kubenswrapper[4859]: E1008 18:27:08.087529 4859 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1dff864e-b75d-4e0b-b182-75f710eac8df" containerName="northd" Oct 08 18:27:08 crc kubenswrapper[4859]: I1008 18:27:08.087536 4859 state_mem.go:107] "Deleted CPUSet assignment" podUID="1dff864e-b75d-4e0b-b182-75f710eac8df" containerName="northd" Oct 08 18:27:08 crc kubenswrapper[4859]: E1008 18:27:08.087547 4859 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1dff864e-b75d-4e0b-b182-75f710eac8df" containerName="ovnkube-controller" Oct 08 18:27:08 crc kubenswrapper[4859]: I1008 18:27:08.087553 4859 state_mem.go:107] "Deleted CPUSet assignment" podUID="1dff864e-b75d-4e0b-b182-75f710eac8df" containerName="ovnkube-controller" Oct 08 18:27:08 crc kubenswrapper[4859]: E1008 18:27:08.087563 4859 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1dff864e-b75d-4e0b-b182-75f710eac8df" containerName="sbdb" Oct 08 18:27:08 crc kubenswrapper[4859]: I1008 18:27:08.087570 4859 state_mem.go:107] "Deleted CPUSet assignment" podUID="1dff864e-b75d-4e0b-b182-75f710eac8df" containerName="sbdb" Oct 08 18:27:08 crc kubenswrapper[4859]: E1008 18:27:08.087580 4859 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1dff864e-b75d-4e0b-b182-75f710eac8df" containerName="kube-rbac-proxy-node" Oct 08 18:27:08 crc kubenswrapper[4859]: I1008 18:27:08.087587 4859 state_mem.go:107] "Deleted CPUSet assignment" podUID="1dff864e-b75d-4e0b-b182-75f710eac8df" containerName="kube-rbac-proxy-node" Oct 08 18:27:08 crc kubenswrapper[4859]: E1008 18:27:08.087595 4859 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1dff864e-b75d-4e0b-b182-75f710eac8df" containerName="kube-rbac-proxy-ovn-metrics" Oct 08 18:27:08 crc kubenswrapper[4859]: I1008 18:27:08.087602 4859 state_mem.go:107] "Deleted CPUSet assignment" podUID="1dff864e-b75d-4e0b-b182-75f710eac8df" containerName="kube-rbac-proxy-ovn-metrics" Oct 08 18:27:08 crc kubenswrapper[4859]: I1008 18:27:08.087729 4859 memory_manager.go:354] "RemoveStaleState removing state" podUID="1dff864e-b75d-4e0b-b182-75f710eac8df" containerName="ovnkube-controller" Oct 08 18:27:08 crc kubenswrapper[4859]: I1008 18:27:08.087740 4859 memory_manager.go:354] "RemoveStaleState removing state" podUID="1dff864e-b75d-4e0b-b182-75f710eac8df" containerName="ovnkube-controller" Oct 08 18:27:08 crc kubenswrapper[4859]: I1008 18:27:08.087753 4859 memory_manager.go:354] "RemoveStaleState removing state" podUID="1dff864e-b75d-4e0b-b182-75f710eac8df" containerName="kube-rbac-proxy-node" Oct 08 18:27:08 crc kubenswrapper[4859]: I1008 18:27:08.087761 4859 memory_manager.go:354] "RemoveStaleState removing state" podUID="1dff864e-b75d-4e0b-b182-75f710eac8df" containerName="ovn-acl-logging" Oct 08 18:27:08 crc kubenswrapper[4859]: I1008 18:27:08.087771 4859 memory_manager.go:354] "RemoveStaleState removing state" podUID="1dff864e-b75d-4e0b-b182-75f710eac8df" containerName="northd" Oct 08 18:27:08 crc kubenswrapper[4859]: I1008 18:27:08.087780 4859 memory_manager.go:354] "RemoveStaleState removing state" podUID="1dff864e-b75d-4e0b-b182-75f710eac8df" containerName="nbdb" Oct 08 18:27:08 crc kubenswrapper[4859]: I1008 18:27:08.087791 4859 memory_manager.go:354] "RemoveStaleState removing state" podUID="1dff864e-b75d-4e0b-b182-75f710eac8df" containerName="sbdb" Oct 08 18:27:08 crc kubenswrapper[4859]: I1008 18:27:08.087800 4859 memory_manager.go:354] "RemoveStaleState removing state" podUID="1dff864e-b75d-4e0b-b182-75f710eac8df" containerName="ovnkube-controller" Oct 08 18:27:08 crc kubenswrapper[4859]: I1008 18:27:08.087808 4859 memory_manager.go:354] "RemoveStaleState removing state" podUID="1dff864e-b75d-4e0b-b182-75f710eac8df" containerName="ovn-controller" Oct 08 18:27:08 crc kubenswrapper[4859]: I1008 18:27:08.087817 4859 memory_manager.go:354] "RemoveStaleState removing state" podUID="1dff864e-b75d-4e0b-b182-75f710eac8df" containerName="kube-rbac-proxy-ovn-metrics" Oct 08 18:27:08 crc kubenswrapper[4859]: E1008 18:27:08.087945 4859 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1dff864e-b75d-4e0b-b182-75f710eac8df" containerName="ovnkube-controller" Oct 08 18:27:08 crc kubenswrapper[4859]: I1008 18:27:08.087954 4859 state_mem.go:107] "Deleted CPUSet assignment" podUID="1dff864e-b75d-4e0b-b182-75f710eac8df" containerName="ovnkube-controller" Oct 08 18:27:08 crc kubenswrapper[4859]: I1008 18:27:08.088059 4859 memory_manager.go:354] "RemoveStaleState removing state" podUID="1dff864e-b75d-4e0b-b182-75f710eac8df" containerName="ovnkube-controller" Oct 08 18:27:08 crc kubenswrapper[4859]: I1008 18:27:08.088282 4859 memory_manager.go:354] "RemoveStaleState removing state" podUID="1dff864e-b75d-4e0b-b182-75f710eac8df" containerName="ovnkube-controller" Oct 08 18:27:08 crc kubenswrapper[4859]: I1008 18:27:08.094037 4859 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/1dff864e-b75d-4e0b-b182-75f710eac8df-kube-api-access-s8jnn" (OuterVolumeSpecName: "kube-api-access-s8jnn") pod "1dff864e-b75d-4e0b-b182-75f710eac8df" (UID: "1dff864e-b75d-4e0b-b182-75f710eac8df"). InnerVolumeSpecName "kube-api-access-s8jnn". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 08 18:27:08 crc kubenswrapper[4859]: I1008 18:27:08.094278 4859 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ovn-kubernetes/ovnkube-node-mt6h5" Oct 08 18:27:08 crc kubenswrapper[4859]: I1008 18:27:08.102721 4859 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/1dff864e-b75d-4e0b-b182-75f710eac8df-run-systemd" (OuterVolumeSpecName: "run-systemd") pod "1dff864e-b75d-4e0b-b182-75f710eac8df" (UID: "1dff864e-b75d-4e0b-b182-75f710eac8df"). InnerVolumeSpecName "run-systemd". PluginName "kubernetes.io/host-path", VolumeGidValue "" Oct 08 18:27:08 crc kubenswrapper[4859]: I1008 18:27:08.173568 4859 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-openvswitch\" (UniqueName: \"kubernetes.io/host-path/372dca22-ecaf-4e04-83b8-888a03d0eab2-run-openvswitch\") pod \"ovnkube-node-mt6h5\" (UID: \"372dca22-ecaf-4e04-83b8-888a03d0eab2\") " pod="openshift-ovn-kubernetes/ovnkube-node-mt6h5" Oct 08 18:27:08 crc kubenswrapper[4859]: I1008 18:27:08.173628 4859 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/372dca22-ecaf-4e04-83b8-888a03d0eab2-ovnkube-script-lib\") pod \"ovnkube-node-mt6h5\" (UID: \"372dca22-ecaf-4e04-83b8-888a03d0eab2\") " pod="openshift-ovn-kubernetes/ovnkube-node-mt6h5" Oct 08 18:27:08 crc kubenswrapper[4859]: I1008 18:27:08.173654 4859 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-openvswitch\" (UniqueName: \"kubernetes.io/host-path/372dca22-ecaf-4e04-83b8-888a03d0eab2-etc-openvswitch\") pod \"ovnkube-node-mt6h5\" (UID: \"372dca22-ecaf-4e04-83b8-888a03d0eab2\") " pod="openshift-ovn-kubernetes/ovnkube-node-mt6h5" Oct 08 18:27:08 crc kubenswrapper[4859]: I1008 18:27:08.173670 4859 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-l6p7l\" (UniqueName: \"kubernetes.io/projected/372dca22-ecaf-4e04-83b8-888a03d0eab2-kube-api-access-l6p7l\") pod \"ovnkube-node-mt6h5\" (UID: \"372dca22-ecaf-4e04-83b8-888a03d0eab2\") " pod="openshift-ovn-kubernetes/ovnkube-node-mt6h5" Oct 08 18:27:08 crc kubenswrapper[4859]: I1008 18:27:08.173740 4859 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-systemd\" (UniqueName: \"kubernetes.io/host-path/372dca22-ecaf-4e04-83b8-888a03d0eab2-run-systemd\") pod \"ovnkube-node-mt6h5\" (UID: \"372dca22-ecaf-4e04-83b8-888a03d0eab2\") " pod="openshift-ovn-kubernetes/ovnkube-node-mt6h5" Oct 08 18:27:08 crc kubenswrapper[4859]: I1008 18:27:08.173759 4859 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/372dca22-ecaf-4e04-83b8-888a03d0eab2-ovn-node-metrics-cert\") pod \"ovnkube-node-mt6h5\" (UID: \"372dca22-ecaf-4e04-83b8-888a03d0eab2\") " pod="openshift-ovn-kubernetes/ovnkube-node-mt6h5" Oct 08 18:27:08 crc kubenswrapper[4859]: I1008 18:27:08.173872 4859 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-run-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/372dca22-ecaf-4e04-83b8-888a03d0eab2-host-run-ovn-kubernetes\") pod \"ovnkube-node-mt6h5\" (UID: \"372dca22-ecaf-4e04-83b8-888a03d0eab2\") " pod="openshift-ovn-kubernetes/ovnkube-node-mt6h5" Oct 08 18:27:08 crc kubenswrapper[4859]: I1008 18:27:08.173959 4859 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/372dca22-ecaf-4e04-83b8-888a03d0eab2-host-slash\") pod \"ovnkube-node-mt6h5\" (UID: \"372dca22-ecaf-4e04-83b8-888a03d0eab2\") " pod="openshift-ovn-kubernetes/ovnkube-node-mt6h5" Oct 08 18:27:08 crc kubenswrapper[4859]: I1008 18:27:08.173996 4859 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-cni-bin\" (UniqueName: \"kubernetes.io/host-path/372dca22-ecaf-4e04-83b8-888a03d0eab2-host-cni-bin\") pod \"ovnkube-node-mt6h5\" (UID: \"372dca22-ecaf-4e04-83b8-888a03d0eab2\") " pod="openshift-ovn-kubernetes/ovnkube-node-mt6h5" Oct 08 18:27:08 crc kubenswrapper[4859]: I1008 18:27:08.174017 4859 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-lib-openvswitch\" (UniqueName: \"kubernetes.io/host-path/372dca22-ecaf-4e04-83b8-888a03d0eab2-var-lib-openvswitch\") pod \"ovnkube-node-mt6h5\" (UID: \"372dca22-ecaf-4e04-83b8-888a03d0eab2\") " pod="openshift-ovn-kubernetes/ovnkube-node-mt6h5" Oct 08 18:27:08 crc kubenswrapper[4859]: I1008 18:27:08.174049 4859 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/372dca22-ecaf-4e04-83b8-888a03d0eab2-env-overrides\") pod \"ovnkube-node-mt6h5\" (UID: \"372dca22-ecaf-4e04-83b8-888a03d0eab2\") " pod="openshift-ovn-kubernetes/ovnkube-node-mt6h5" Oct 08 18:27:08 crc kubenswrapper[4859]: I1008 18:27:08.174066 4859 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-cni-netd\" (UniqueName: \"kubernetes.io/host-path/372dca22-ecaf-4e04-83b8-888a03d0eab2-host-cni-netd\") pod \"ovnkube-node-mt6h5\" (UID: \"372dca22-ecaf-4e04-83b8-888a03d0eab2\") " pod="openshift-ovn-kubernetes/ovnkube-node-mt6h5" Oct 08 18:27:08 crc kubenswrapper[4859]: I1008 18:27:08.174088 4859 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-kubelet\" (UniqueName: \"kubernetes.io/host-path/372dca22-ecaf-4e04-83b8-888a03d0eab2-host-kubelet\") pod \"ovnkube-node-mt6h5\" (UID: \"372dca22-ecaf-4e04-83b8-888a03d0eab2\") " pod="openshift-ovn-kubernetes/ovnkube-node-mt6h5" Oct 08 18:27:08 crc kubenswrapper[4859]: I1008 18:27:08.174109 4859 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-ovn\" (UniqueName: \"kubernetes.io/host-path/372dca22-ecaf-4e04-83b8-888a03d0eab2-run-ovn\") pod \"ovnkube-node-mt6h5\" (UID: \"372dca22-ecaf-4e04-83b8-888a03d0eab2\") " pod="openshift-ovn-kubernetes/ovnkube-node-mt6h5" Oct 08 18:27:08 crc kubenswrapper[4859]: I1008 18:27:08.174130 4859 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-socket\" (UniqueName: \"kubernetes.io/host-path/372dca22-ecaf-4e04-83b8-888a03d0eab2-log-socket\") pod \"ovnkube-node-mt6h5\" (UID: \"372dca22-ecaf-4e04-83b8-888a03d0eab2\") " pod="openshift-ovn-kubernetes/ovnkube-node-mt6h5" Oct 08 18:27:08 crc kubenswrapper[4859]: I1008 18:27:08.174166 4859 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-var-lib-cni-networks-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/372dca22-ecaf-4e04-83b8-888a03d0eab2-host-var-lib-cni-networks-ovn-kubernetes\") pod \"ovnkube-node-mt6h5\" (UID: \"372dca22-ecaf-4e04-83b8-888a03d0eab2\") " pod="openshift-ovn-kubernetes/ovnkube-node-mt6h5" Oct 08 18:27:08 crc kubenswrapper[4859]: I1008 18:27:08.174188 4859 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/372dca22-ecaf-4e04-83b8-888a03d0eab2-ovnkube-config\") pod \"ovnkube-node-mt6h5\" (UID: \"372dca22-ecaf-4e04-83b8-888a03d0eab2\") " pod="openshift-ovn-kubernetes/ovnkube-node-mt6h5" Oct 08 18:27:08 crc kubenswrapper[4859]: I1008 18:27:08.174215 4859 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/372dca22-ecaf-4e04-83b8-888a03d0eab2-host-run-netns\") pod \"ovnkube-node-mt6h5\" (UID: \"372dca22-ecaf-4e04-83b8-888a03d0eab2\") " pod="openshift-ovn-kubernetes/ovnkube-node-mt6h5" Oct 08 18:27:08 crc kubenswrapper[4859]: I1008 18:27:08.174236 4859 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"node-log\" (UniqueName: \"kubernetes.io/host-path/372dca22-ecaf-4e04-83b8-888a03d0eab2-node-log\") pod \"ovnkube-node-mt6h5\" (UID: \"372dca22-ecaf-4e04-83b8-888a03d0eab2\") " pod="openshift-ovn-kubernetes/ovnkube-node-mt6h5" Oct 08 18:27:08 crc kubenswrapper[4859]: I1008 18:27:08.174318 4859 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"systemd-units\" (UniqueName: \"kubernetes.io/host-path/372dca22-ecaf-4e04-83b8-888a03d0eab2-systemd-units\") pod \"ovnkube-node-mt6h5\" (UID: \"372dca22-ecaf-4e04-83b8-888a03d0eab2\") " pod="openshift-ovn-kubernetes/ovnkube-node-mt6h5" Oct 08 18:27:08 crc kubenswrapper[4859]: I1008 18:27:08.174384 4859 reconciler_common.go:293] "Volume detached for volume \"host-kubelet\" (UniqueName: \"kubernetes.io/host-path/1dff864e-b75d-4e0b-b182-75f710eac8df-host-kubelet\") on node \"crc\" DevicePath \"\"" Oct 08 18:27:08 crc kubenswrapper[4859]: I1008 18:27:08.174412 4859 reconciler_common.go:293] "Volume detached for volume \"run-systemd\" (UniqueName: \"kubernetes.io/host-path/1dff864e-b75d-4e0b-b182-75f710eac8df-run-systemd\") on node \"crc\" DevicePath \"\"" Oct 08 18:27:08 crc kubenswrapper[4859]: I1008 18:27:08.174424 4859 reconciler_common.go:293] "Volume detached for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/1dff864e-b75d-4e0b-b182-75f710eac8df-ovn-node-metrics-cert\") on node \"crc\" DevicePath \"\"" Oct 08 18:27:08 crc kubenswrapper[4859]: I1008 18:27:08.174434 4859 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-s8jnn\" (UniqueName: \"kubernetes.io/projected/1dff864e-b75d-4e0b-b182-75f710eac8df-kube-api-access-s8jnn\") on node \"crc\" DevicePath \"\"" Oct 08 18:27:08 crc kubenswrapper[4859]: I1008 18:27:08.275245 4859 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/372dca22-ecaf-4e04-83b8-888a03d0eab2-host-run-netns\") pod \"ovnkube-node-mt6h5\" (UID: \"372dca22-ecaf-4e04-83b8-888a03d0eab2\") " pod="openshift-ovn-kubernetes/ovnkube-node-mt6h5" Oct 08 18:27:08 crc kubenswrapper[4859]: I1008 18:27:08.275300 4859 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"node-log\" (UniqueName: \"kubernetes.io/host-path/372dca22-ecaf-4e04-83b8-888a03d0eab2-node-log\") pod \"ovnkube-node-mt6h5\" (UID: \"372dca22-ecaf-4e04-83b8-888a03d0eab2\") " pod="openshift-ovn-kubernetes/ovnkube-node-mt6h5" Oct 08 18:27:08 crc kubenswrapper[4859]: I1008 18:27:08.275323 4859 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"systemd-units\" (UniqueName: \"kubernetes.io/host-path/372dca22-ecaf-4e04-83b8-888a03d0eab2-systemd-units\") pod \"ovnkube-node-mt6h5\" (UID: \"372dca22-ecaf-4e04-83b8-888a03d0eab2\") " pod="openshift-ovn-kubernetes/ovnkube-node-mt6h5" Oct 08 18:27:08 crc kubenswrapper[4859]: I1008 18:27:08.275370 4859 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-openvswitch\" (UniqueName: \"kubernetes.io/host-path/372dca22-ecaf-4e04-83b8-888a03d0eab2-run-openvswitch\") pod \"ovnkube-node-mt6h5\" (UID: \"372dca22-ecaf-4e04-83b8-888a03d0eab2\") " pod="openshift-ovn-kubernetes/ovnkube-node-mt6h5" Oct 08 18:27:08 crc kubenswrapper[4859]: I1008 18:27:08.275397 4859 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/372dca22-ecaf-4e04-83b8-888a03d0eab2-host-run-netns\") pod \"ovnkube-node-mt6h5\" (UID: \"372dca22-ecaf-4e04-83b8-888a03d0eab2\") " pod="openshift-ovn-kubernetes/ovnkube-node-mt6h5" Oct 08 18:27:08 crc kubenswrapper[4859]: I1008 18:27:08.275407 4859 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/372dca22-ecaf-4e04-83b8-888a03d0eab2-ovnkube-script-lib\") pod \"ovnkube-node-mt6h5\" (UID: \"372dca22-ecaf-4e04-83b8-888a03d0eab2\") " pod="openshift-ovn-kubernetes/ovnkube-node-mt6h5" Oct 08 18:27:08 crc kubenswrapper[4859]: I1008 18:27:08.275472 4859 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-openvswitch\" (UniqueName: \"kubernetes.io/host-path/372dca22-ecaf-4e04-83b8-888a03d0eab2-etc-openvswitch\") pod \"ovnkube-node-mt6h5\" (UID: \"372dca22-ecaf-4e04-83b8-888a03d0eab2\") " pod="openshift-ovn-kubernetes/ovnkube-node-mt6h5" Oct 08 18:27:08 crc kubenswrapper[4859]: I1008 18:27:08.275486 4859 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-openvswitch\" (UniqueName: \"kubernetes.io/host-path/372dca22-ecaf-4e04-83b8-888a03d0eab2-run-openvswitch\") pod \"ovnkube-node-mt6h5\" (UID: \"372dca22-ecaf-4e04-83b8-888a03d0eab2\") " pod="openshift-ovn-kubernetes/ovnkube-node-mt6h5" Oct 08 18:27:08 crc kubenswrapper[4859]: I1008 18:27:08.275499 4859 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-l6p7l\" (UniqueName: \"kubernetes.io/projected/372dca22-ecaf-4e04-83b8-888a03d0eab2-kube-api-access-l6p7l\") pod \"ovnkube-node-mt6h5\" (UID: \"372dca22-ecaf-4e04-83b8-888a03d0eab2\") " pod="openshift-ovn-kubernetes/ovnkube-node-mt6h5" Oct 08 18:27:08 crc kubenswrapper[4859]: I1008 18:27:08.275528 4859 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"systemd-units\" (UniqueName: \"kubernetes.io/host-path/372dca22-ecaf-4e04-83b8-888a03d0eab2-systemd-units\") pod \"ovnkube-node-mt6h5\" (UID: \"372dca22-ecaf-4e04-83b8-888a03d0eab2\") " pod="openshift-ovn-kubernetes/ovnkube-node-mt6h5" Oct 08 18:27:08 crc kubenswrapper[4859]: I1008 18:27:08.275579 4859 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-systemd\" (UniqueName: \"kubernetes.io/host-path/372dca22-ecaf-4e04-83b8-888a03d0eab2-run-systemd\") pod \"ovnkube-node-mt6h5\" (UID: \"372dca22-ecaf-4e04-83b8-888a03d0eab2\") " pod="openshift-ovn-kubernetes/ovnkube-node-mt6h5" Oct 08 18:27:08 crc kubenswrapper[4859]: I1008 18:27:08.275619 4859 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"node-log\" (UniqueName: \"kubernetes.io/host-path/372dca22-ecaf-4e04-83b8-888a03d0eab2-node-log\") pod \"ovnkube-node-mt6h5\" (UID: \"372dca22-ecaf-4e04-83b8-888a03d0eab2\") " pod="openshift-ovn-kubernetes/ovnkube-node-mt6h5" Oct 08 18:27:08 crc kubenswrapper[4859]: I1008 18:27:08.275629 4859 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-openvswitch\" (UniqueName: \"kubernetes.io/host-path/372dca22-ecaf-4e04-83b8-888a03d0eab2-etc-openvswitch\") pod \"ovnkube-node-mt6h5\" (UID: \"372dca22-ecaf-4e04-83b8-888a03d0eab2\") " pod="openshift-ovn-kubernetes/ovnkube-node-mt6h5" Oct 08 18:27:08 crc kubenswrapper[4859]: I1008 18:27:08.275555 4859 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-systemd\" (UniqueName: \"kubernetes.io/host-path/372dca22-ecaf-4e04-83b8-888a03d0eab2-run-systemd\") pod \"ovnkube-node-mt6h5\" (UID: \"372dca22-ecaf-4e04-83b8-888a03d0eab2\") " pod="openshift-ovn-kubernetes/ovnkube-node-mt6h5" Oct 08 18:27:08 crc kubenswrapper[4859]: I1008 18:27:08.275779 4859 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/372dca22-ecaf-4e04-83b8-888a03d0eab2-ovn-node-metrics-cert\") pod \"ovnkube-node-mt6h5\" (UID: \"372dca22-ecaf-4e04-83b8-888a03d0eab2\") " pod="openshift-ovn-kubernetes/ovnkube-node-mt6h5" Oct 08 18:27:08 crc kubenswrapper[4859]: I1008 18:27:08.275880 4859 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-run-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/372dca22-ecaf-4e04-83b8-888a03d0eab2-host-run-ovn-kubernetes\") pod \"ovnkube-node-mt6h5\" (UID: \"372dca22-ecaf-4e04-83b8-888a03d0eab2\") " pod="openshift-ovn-kubernetes/ovnkube-node-mt6h5" Oct 08 18:27:08 crc kubenswrapper[4859]: I1008 18:27:08.275930 4859 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/372dca22-ecaf-4e04-83b8-888a03d0eab2-host-slash\") pod \"ovnkube-node-mt6h5\" (UID: \"372dca22-ecaf-4e04-83b8-888a03d0eab2\") " pod="openshift-ovn-kubernetes/ovnkube-node-mt6h5" Oct 08 18:27:08 crc kubenswrapper[4859]: I1008 18:27:08.275982 4859 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-cni-bin\" (UniqueName: \"kubernetes.io/host-path/372dca22-ecaf-4e04-83b8-888a03d0eab2-host-cni-bin\") pod \"ovnkube-node-mt6h5\" (UID: \"372dca22-ecaf-4e04-83b8-888a03d0eab2\") " pod="openshift-ovn-kubernetes/ovnkube-node-mt6h5" Oct 08 18:27:08 crc kubenswrapper[4859]: I1008 18:27:08.276005 4859 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-run-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/372dca22-ecaf-4e04-83b8-888a03d0eab2-host-run-ovn-kubernetes\") pod \"ovnkube-node-mt6h5\" (UID: \"372dca22-ecaf-4e04-83b8-888a03d0eab2\") " pod="openshift-ovn-kubernetes/ovnkube-node-mt6h5" Oct 08 18:27:08 crc kubenswrapper[4859]: I1008 18:27:08.276017 4859 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-lib-openvswitch\" (UniqueName: \"kubernetes.io/host-path/372dca22-ecaf-4e04-83b8-888a03d0eab2-var-lib-openvswitch\") pod \"ovnkube-node-mt6h5\" (UID: \"372dca22-ecaf-4e04-83b8-888a03d0eab2\") " pod="openshift-ovn-kubernetes/ovnkube-node-mt6h5" Oct 08 18:27:08 crc kubenswrapper[4859]: I1008 18:27:08.276053 4859 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/372dca22-ecaf-4e04-83b8-888a03d0eab2-host-slash\") pod \"ovnkube-node-mt6h5\" (UID: \"372dca22-ecaf-4e04-83b8-888a03d0eab2\") " pod="openshift-ovn-kubernetes/ovnkube-node-mt6h5" Oct 08 18:27:08 crc kubenswrapper[4859]: I1008 18:27:08.276084 4859 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/372dca22-ecaf-4e04-83b8-888a03d0eab2-env-overrides\") pod \"ovnkube-node-mt6h5\" (UID: \"372dca22-ecaf-4e04-83b8-888a03d0eab2\") " pod="openshift-ovn-kubernetes/ovnkube-node-mt6h5" Oct 08 18:27:08 crc kubenswrapper[4859]: I1008 18:27:08.276113 4859 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-lib-openvswitch\" (UniqueName: \"kubernetes.io/host-path/372dca22-ecaf-4e04-83b8-888a03d0eab2-var-lib-openvswitch\") pod \"ovnkube-node-mt6h5\" (UID: \"372dca22-ecaf-4e04-83b8-888a03d0eab2\") " pod="openshift-ovn-kubernetes/ovnkube-node-mt6h5" Oct 08 18:27:08 crc kubenswrapper[4859]: I1008 18:27:08.276118 4859 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-cni-netd\" (UniqueName: \"kubernetes.io/host-path/372dca22-ecaf-4e04-83b8-888a03d0eab2-host-cni-netd\") pod \"ovnkube-node-mt6h5\" (UID: \"372dca22-ecaf-4e04-83b8-888a03d0eab2\") " pod="openshift-ovn-kubernetes/ovnkube-node-mt6h5" Oct 08 18:27:08 crc kubenswrapper[4859]: I1008 18:27:08.276173 4859 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-kubelet\" (UniqueName: \"kubernetes.io/host-path/372dca22-ecaf-4e04-83b8-888a03d0eab2-host-kubelet\") pod \"ovnkube-node-mt6h5\" (UID: \"372dca22-ecaf-4e04-83b8-888a03d0eab2\") " pod="openshift-ovn-kubernetes/ovnkube-node-mt6h5" Oct 08 18:27:08 crc kubenswrapper[4859]: I1008 18:27:08.276206 4859 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-ovn\" (UniqueName: \"kubernetes.io/host-path/372dca22-ecaf-4e04-83b8-888a03d0eab2-run-ovn\") pod \"ovnkube-node-mt6h5\" (UID: \"372dca22-ecaf-4e04-83b8-888a03d0eab2\") " pod="openshift-ovn-kubernetes/ovnkube-node-mt6h5" Oct 08 18:27:08 crc kubenswrapper[4859]: I1008 18:27:08.276230 4859 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-socket\" (UniqueName: \"kubernetes.io/host-path/372dca22-ecaf-4e04-83b8-888a03d0eab2-log-socket\") pod \"ovnkube-node-mt6h5\" (UID: \"372dca22-ecaf-4e04-83b8-888a03d0eab2\") " pod="openshift-ovn-kubernetes/ovnkube-node-mt6h5" Oct 08 18:27:08 crc kubenswrapper[4859]: I1008 18:27:08.276272 4859 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-var-lib-cni-networks-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/372dca22-ecaf-4e04-83b8-888a03d0eab2-host-var-lib-cni-networks-ovn-kubernetes\") pod \"ovnkube-node-mt6h5\" (UID: \"372dca22-ecaf-4e04-83b8-888a03d0eab2\") " pod="openshift-ovn-kubernetes/ovnkube-node-mt6h5" Oct 08 18:27:08 crc kubenswrapper[4859]: I1008 18:27:08.276308 4859 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/372dca22-ecaf-4e04-83b8-888a03d0eab2-ovnkube-script-lib\") pod \"ovnkube-node-mt6h5\" (UID: \"372dca22-ecaf-4e04-83b8-888a03d0eab2\") " pod="openshift-ovn-kubernetes/ovnkube-node-mt6h5" Oct 08 18:27:08 crc kubenswrapper[4859]: I1008 18:27:08.276311 4859 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/372dca22-ecaf-4e04-83b8-888a03d0eab2-ovnkube-config\") pod \"ovnkube-node-mt6h5\" (UID: \"372dca22-ecaf-4e04-83b8-888a03d0eab2\") " pod="openshift-ovn-kubernetes/ovnkube-node-mt6h5" Oct 08 18:27:08 crc kubenswrapper[4859]: I1008 18:27:08.276403 4859 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-kubelet\" (UniqueName: \"kubernetes.io/host-path/372dca22-ecaf-4e04-83b8-888a03d0eab2-host-kubelet\") pod \"ovnkube-node-mt6h5\" (UID: \"372dca22-ecaf-4e04-83b8-888a03d0eab2\") " pod="openshift-ovn-kubernetes/ovnkube-node-mt6h5" Oct 08 18:27:08 crc kubenswrapper[4859]: I1008 18:27:08.276088 4859 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-cni-bin\" (UniqueName: \"kubernetes.io/host-path/372dca22-ecaf-4e04-83b8-888a03d0eab2-host-cni-bin\") pod \"ovnkube-node-mt6h5\" (UID: \"372dca22-ecaf-4e04-83b8-888a03d0eab2\") " pod="openshift-ovn-kubernetes/ovnkube-node-mt6h5" Oct 08 18:27:08 crc kubenswrapper[4859]: I1008 18:27:08.276448 4859 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-cni-netd\" (UniqueName: \"kubernetes.io/host-path/372dca22-ecaf-4e04-83b8-888a03d0eab2-host-cni-netd\") pod \"ovnkube-node-mt6h5\" (UID: \"372dca22-ecaf-4e04-83b8-888a03d0eab2\") " pod="openshift-ovn-kubernetes/ovnkube-node-mt6h5" Oct 08 18:27:08 crc kubenswrapper[4859]: I1008 18:27:08.276478 4859 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-socket\" (UniqueName: \"kubernetes.io/host-path/372dca22-ecaf-4e04-83b8-888a03d0eab2-log-socket\") pod \"ovnkube-node-mt6h5\" (UID: \"372dca22-ecaf-4e04-83b8-888a03d0eab2\") " pod="openshift-ovn-kubernetes/ovnkube-node-mt6h5" Oct 08 18:27:08 crc kubenswrapper[4859]: I1008 18:27:08.276511 4859 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-ovn\" (UniqueName: \"kubernetes.io/host-path/372dca22-ecaf-4e04-83b8-888a03d0eab2-run-ovn\") pod \"ovnkube-node-mt6h5\" (UID: \"372dca22-ecaf-4e04-83b8-888a03d0eab2\") " pod="openshift-ovn-kubernetes/ovnkube-node-mt6h5" Oct 08 18:27:08 crc kubenswrapper[4859]: I1008 18:27:08.276542 4859 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-var-lib-cni-networks-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/372dca22-ecaf-4e04-83b8-888a03d0eab2-host-var-lib-cni-networks-ovn-kubernetes\") pod \"ovnkube-node-mt6h5\" (UID: \"372dca22-ecaf-4e04-83b8-888a03d0eab2\") " pod="openshift-ovn-kubernetes/ovnkube-node-mt6h5" Oct 08 18:27:08 crc kubenswrapper[4859]: I1008 18:27:08.276550 4859 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/372dca22-ecaf-4e04-83b8-888a03d0eab2-env-overrides\") pod \"ovnkube-node-mt6h5\" (UID: \"372dca22-ecaf-4e04-83b8-888a03d0eab2\") " pod="openshift-ovn-kubernetes/ovnkube-node-mt6h5" Oct 08 18:27:08 crc kubenswrapper[4859]: I1008 18:27:08.277323 4859 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/372dca22-ecaf-4e04-83b8-888a03d0eab2-ovnkube-config\") pod \"ovnkube-node-mt6h5\" (UID: \"372dca22-ecaf-4e04-83b8-888a03d0eab2\") " pod="openshift-ovn-kubernetes/ovnkube-node-mt6h5" Oct 08 18:27:08 crc kubenswrapper[4859]: I1008 18:27:08.280121 4859 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/372dca22-ecaf-4e04-83b8-888a03d0eab2-ovn-node-metrics-cert\") pod \"ovnkube-node-mt6h5\" (UID: \"372dca22-ecaf-4e04-83b8-888a03d0eab2\") " pod="openshift-ovn-kubernetes/ovnkube-node-mt6h5" Oct 08 18:27:08 crc kubenswrapper[4859]: I1008 18:27:08.300883 4859 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-l6p7l\" (UniqueName: \"kubernetes.io/projected/372dca22-ecaf-4e04-83b8-888a03d0eab2-kube-api-access-l6p7l\") pod \"ovnkube-node-mt6h5\" (UID: \"372dca22-ecaf-4e04-83b8-888a03d0eab2\") " pod="openshift-ovn-kubernetes/ovnkube-node-mt6h5" Oct 08 18:27:08 crc kubenswrapper[4859]: I1008 18:27:08.419860 4859 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ovn-kubernetes/ovnkube-node-mt6h5" Oct 08 18:27:08 crc kubenswrapper[4859]: I1008 18:27:08.964404 4859 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-8plkg_92b2d7de-31cc-4a91-95a5-ed5ea964b028/kube-multus/2.log" Oct 08 18:27:08 crc kubenswrapper[4859]: I1008 18:27:08.971269 4859 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-4g8gf_1dff864e-b75d-4e0b-b182-75f710eac8df/ovn-acl-logging/0.log" Oct 08 18:27:08 crc kubenswrapper[4859]: I1008 18:27:08.972053 4859 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-4g8gf_1dff864e-b75d-4e0b-b182-75f710eac8df/ovn-controller/0.log" Oct 08 18:27:08 crc kubenswrapper[4859]: I1008 18:27:08.973461 4859 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-ovn-kubernetes/ovnkube-node-4g8gf" Oct 08 18:27:08 crc kubenswrapper[4859]: I1008 18:27:08.990348 4859 generic.go:334] "Generic (PLEG): container finished" podID="372dca22-ecaf-4e04-83b8-888a03d0eab2" containerID="a39d4baf4d0ae7f3b80d9cda828e5058c205a368dfbea487ddb66fd1495339db" exitCode=0 Oct 08 18:27:08 crc kubenswrapper[4859]: I1008 18:27:08.990403 4859 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-mt6h5" event={"ID":"372dca22-ecaf-4e04-83b8-888a03d0eab2","Type":"ContainerDied","Data":"a39d4baf4d0ae7f3b80d9cda828e5058c205a368dfbea487ddb66fd1495339db"} Oct 08 18:27:08 crc kubenswrapper[4859]: I1008 18:27:08.990439 4859 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-mt6h5" event={"ID":"372dca22-ecaf-4e04-83b8-888a03d0eab2","Type":"ContainerStarted","Data":"7a611bb2e6ec59c2dfd280c5dbe6bdc51ba69409a8e2932afc9a7c0b233206aa"} Oct 08 18:27:09 crc kubenswrapper[4859]: I1008 18:27:09.084834 4859 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-ovn-kubernetes/ovnkube-node-4g8gf"] Oct 08 18:27:09 crc kubenswrapper[4859]: I1008 18:27:09.098146 4859 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-ovn-kubernetes/ovnkube-node-4g8gf"] Oct 08 18:27:10 crc kubenswrapper[4859]: I1008 18:27:10.002491 4859 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-mt6h5" event={"ID":"372dca22-ecaf-4e04-83b8-888a03d0eab2","Type":"ContainerStarted","Data":"6089df0e878e96d83fbae581827444bb557d825ee3cba78ae7e1ec17a94311b4"} Oct 08 18:27:10 crc kubenswrapper[4859]: I1008 18:27:10.002548 4859 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-mt6h5" event={"ID":"372dca22-ecaf-4e04-83b8-888a03d0eab2","Type":"ContainerStarted","Data":"3417cb420d33b421c9719b6a98fefd1ec17d225887e00c264decc8fd9a919416"} Oct 08 18:27:10 crc kubenswrapper[4859]: I1008 18:27:10.002570 4859 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-mt6h5" event={"ID":"372dca22-ecaf-4e04-83b8-888a03d0eab2","Type":"ContainerStarted","Data":"6af9fd599465fc6d32e5f659fca813e9d5a503a417f30519239a7c46b2f5fdae"} Oct 08 18:27:10 crc kubenswrapper[4859]: I1008 18:27:10.002587 4859 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-mt6h5" event={"ID":"372dca22-ecaf-4e04-83b8-888a03d0eab2","Type":"ContainerStarted","Data":"108aa849541884d90750120e687e1c4aec8a97ec38528626de82c77052a44cb9"} Oct 08 18:27:10 crc kubenswrapper[4859]: I1008 18:27:10.002600 4859 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-mt6h5" event={"ID":"372dca22-ecaf-4e04-83b8-888a03d0eab2","Type":"ContainerStarted","Data":"97f43324928062daa5a2de3cf58faeb2c39492b8223fcadcb64d4bc8323642c8"} Oct 08 18:27:10 crc kubenswrapper[4859]: I1008 18:27:10.002617 4859 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-mt6h5" event={"ID":"372dca22-ecaf-4e04-83b8-888a03d0eab2","Type":"ContainerStarted","Data":"f56b4cb4480620c92008dfda83320f8ac7507b0eaa581f84ccaf5e3551f7a1f7"} Oct 08 18:27:10 crc kubenswrapper[4859]: I1008 18:27:10.492657 4859 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="1dff864e-b75d-4e0b-b182-75f710eac8df" path="/var/lib/kubelet/pods/1dff864e-b75d-4e0b-b182-75f710eac8df/volumes" Oct 08 18:27:10 crc kubenswrapper[4859]: I1008 18:27:10.697669 4859 scope.go:117] "RemoveContainer" containerID="b737cb013e2877d970d7e301048f95e0ba0ade0b8db42b2607a78fdf2580659f" Oct 08 18:27:10 crc kubenswrapper[4859]: I1008 18:27:10.722798 4859 scope.go:117] "RemoveContainer" containerID="669f5f3c6d3c1ec0c4b4870e090b0a853adc01727c23383e595715b476c32668" Oct 08 18:27:10 crc kubenswrapper[4859]: I1008 18:27:10.742642 4859 scope.go:117] "RemoveContainer" containerID="2c67d2aed7126d76989e057d8e2d8e188a8902b2f3fe26ed6d1d45d3b013d897" Oct 08 18:27:10 crc kubenswrapper[4859]: I1008 18:27:10.758423 4859 scope.go:117] "RemoveContainer" containerID="676511a11ed4102084b57ebdd5f3fda4969741b2bc8443b415c4bf94f39ae4c9" Oct 08 18:27:10 crc kubenswrapper[4859]: I1008 18:27:10.774037 4859 scope.go:117] "RemoveContainer" containerID="6ddc0e842bb8f81663406070023b0e5ed9cc99c98db0123b1c84f20da59fb5b5" Oct 08 18:27:10 crc kubenswrapper[4859]: I1008 18:27:10.791506 4859 scope.go:117] "RemoveContainer" containerID="b8009a42cc4c9a816457a0678bdb2f7b7c53d0c8604e007a6fb141f66d06a516" Oct 08 18:27:10 crc kubenswrapper[4859]: I1008 18:27:10.810680 4859 scope.go:117] "RemoveContainer" containerID="768ca274dedcc6847fc21d15234d3665fa9ccae371b5b6a882cfcb9b182f527d" Oct 08 18:27:10 crc kubenswrapper[4859]: I1008 18:27:10.828647 4859 scope.go:117] "RemoveContainer" containerID="4a74c1f0d21c90cdd7e776ee14e1f48757a04e2220fbd1f884449181ec6d1c3a" Oct 08 18:27:10 crc kubenswrapper[4859]: I1008 18:27:10.844935 4859 scope.go:117] "RemoveContainer" containerID="7a211d3da02b173867ff3a912eecd3b379483c7cc77e6b91122edd1ffd30fd53" Oct 08 18:27:13 crc kubenswrapper[4859]: I1008 18:27:13.031964 4859 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-mt6h5" event={"ID":"372dca22-ecaf-4e04-83b8-888a03d0eab2","Type":"ContainerStarted","Data":"772345e93eddae944c23372aa8c9f129c15ea13b6866f27aead198efab838920"} Oct 08 18:27:15 crc kubenswrapper[4859]: I1008 18:27:15.050461 4859 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-mt6h5" event={"ID":"372dca22-ecaf-4e04-83b8-888a03d0eab2","Type":"ContainerStarted","Data":"2cffa5335fddd9a0f54f46be68031854f167acf9c148fc390abdecb50e95eb45"} Oct 08 18:27:15 crc kubenswrapper[4859]: I1008 18:27:15.050872 4859 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ovn-kubernetes/ovnkube-node-mt6h5" Oct 08 18:27:15 crc kubenswrapper[4859]: I1008 18:27:15.050893 4859 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ovn-kubernetes/ovnkube-node-mt6h5" Oct 08 18:27:15 crc kubenswrapper[4859]: I1008 18:27:15.050904 4859 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ovn-kubernetes/ovnkube-node-mt6h5" Oct 08 18:27:15 crc kubenswrapper[4859]: I1008 18:27:15.081129 4859 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-ovn-kubernetes/ovnkube-node-mt6h5" Oct 08 18:27:15 crc kubenswrapper[4859]: I1008 18:27:15.081569 4859 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-ovn-kubernetes/ovnkube-node-mt6h5" Oct 08 18:27:15 crc kubenswrapper[4859]: I1008 18:27:15.088264 4859 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-ovn-kubernetes/ovnkube-node-mt6h5" podStartSLOduration=7.088243636 podStartE2EDuration="7.088243636s" podCreationTimestamp="2025-10-08 18:27:08 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-08 18:27:15.084249303 +0000 UTC m=+605.331088682" watchObservedRunningTime="2025-10-08 18:27:15.088243636 +0000 UTC m=+605.335083015" Oct 08 18:27:17 crc kubenswrapper[4859]: I1008 18:27:17.924737 4859 patch_prober.go:28] interesting pod/machine-config-daemon-82s52 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 08 18:27:17 crc kubenswrapper[4859]: I1008 18:27:17.925245 4859 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-82s52" podUID="b23a6a6c-9d92-4e7b-840e-55cfda873a2d" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 08 18:27:20 crc kubenswrapper[4859]: I1008 18:27:20.477048 4859 scope.go:117] "RemoveContainer" containerID="d18a8ef0ab4fc3ed687ccf7e94753da8a4a29dae40a39aae2b64a854761f90d0" Oct 08 18:27:20 crc kubenswrapper[4859]: E1008 18:27:20.477667 4859 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"kube-multus\" with CrashLoopBackOff: \"back-off 20s restarting failed container=kube-multus pod=multus-8plkg_openshift-multus(92b2d7de-31cc-4a91-95a5-ed5ea964b028)\"" pod="openshift-multus/multus-8plkg" podUID="92b2d7de-31cc-4a91-95a5-ed5ea964b028" Oct 08 18:27:34 crc kubenswrapper[4859]: I1008 18:27:34.470569 4859 scope.go:117] "RemoveContainer" containerID="d18a8ef0ab4fc3ed687ccf7e94753da8a4a29dae40a39aae2b64a854761f90d0" Oct 08 18:27:35 crc kubenswrapper[4859]: I1008 18:27:35.184597 4859 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-8plkg_92b2d7de-31cc-4a91-95a5-ed5ea964b028/kube-multus/2.log" Oct 08 18:27:35 crc kubenswrapper[4859]: I1008 18:27:35.185174 4859 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-8plkg" event={"ID":"92b2d7de-31cc-4a91-95a5-ed5ea964b028","Type":"ContainerStarted","Data":"a89447ff287eba59d6fc40c3d61fa602798ec7970e190cb503a6e0295183941c"} Oct 08 18:27:38 crc kubenswrapper[4859]: I1008 18:27:38.463807 4859 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-ovn-kubernetes/ovnkube-node-mt6h5" Oct 08 18:27:47 crc kubenswrapper[4859]: I1008 18:27:47.010714 4859 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/fa9831ede5d93c33d525b70ce6ddf94e500d80992af75a3305fe98835cdxmbx"] Oct 08 18:27:47 crc kubenswrapper[4859]: I1008 18:27:47.012833 4859 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/fa9831ede5d93c33d525b70ce6ddf94e500d80992af75a3305fe98835cdxmbx" Oct 08 18:27:47 crc kubenswrapper[4859]: I1008 18:27:47.016332 4859 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"default-dockercfg-vmwhc" Oct 08 18:27:47 crc kubenswrapper[4859]: I1008 18:27:47.023059 4859 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/fa9831ede5d93c33d525b70ce6ddf94e500d80992af75a3305fe98835cdxmbx"] Oct 08 18:27:47 crc kubenswrapper[4859]: I1008 18:27:47.092906 4859 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-vkjbd\" (UniqueName: \"kubernetes.io/projected/cc21f1af-ca05-4cf4-b66f-f7fcde5026ee-kube-api-access-vkjbd\") pod \"fa9831ede5d93c33d525b70ce6ddf94e500d80992af75a3305fe98835cdxmbx\" (UID: \"cc21f1af-ca05-4cf4-b66f-f7fcde5026ee\") " pod="openshift-marketplace/fa9831ede5d93c33d525b70ce6ddf94e500d80992af75a3305fe98835cdxmbx" Oct 08 18:27:47 crc kubenswrapper[4859]: I1008 18:27:47.092994 4859 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/cc21f1af-ca05-4cf4-b66f-f7fcde5026ee-bundle\") pod \"fa9831ede5d93c33d525b70ce6ddf94e500d80992af75a3305fe98835cdxmbx\" (UID: \"cc21f1af-ca05-4cf4-b66f-f7fcde5026ee\") " pod="openshift-marketplace/fa9831ede5d93c33d525b70ce6ddf94e500d80992af75a3305fe98835cdxmbx" Oct 08 18:27:47 crc kubenswrapper[4859]: I1008 18:27:47.093044 4859 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/cc21f1af-ca05-4cf4-b66f-f7fcde5026ee-util\") pod \"fa9831ede5d93c33d525b70ce6ddf94e500d80992af75a3305fe98835cdxmbx\" (UID: \"cc21f1af-ca05-4cf4-b66f-f7fcde5026ee\") " pod="openshift-marketplace/fa9831ede5d93c33d525b70ce6ddf94e500d80992af75a3305fe98835cdxmbx" Oct 08 18:27:47 crc kubenswrapper[4859]: I1008 18:27:47.195122 4859 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-vkjbd\" (UniqueName: \"kubernetes.io/projected/cc21f1af-ca05-4cf4-b66f-f7fcde5026ee-kube-api-access-vkjbd\") pod \"fa9831ede5d93c33d525b70ce6ddf94e500d80992af75a3305fe98835cdxmbx\" (UID: \"cc21f1af-ca05-4cf4-b66f-f7fcde5026ee\") " pod="openshift-marketplace/fa9831ede5d93c33d525b70ce6ddf94e500d80992af75a3305fe98835cdxmbx" Oct 08 18:27:47 crc kubenswrapper[4859]: I1008 18:27:47.195279 4859 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/cc21f1af-ca05-4cf4-b66f-f7fcde5026ee-bundle\") pod \"fa9831ede5d93c33d525b70ce6ddf94e500d80992af75a3305fe98835cdxmbx\" (UID: \"cc21f1af-ca05-4cf4-b66f-f7fcde5026ee\") " pod="openshift-marketplace/fa9831ede5d93c33d525b70ce6ddf94e500d80992af75a3305fe98835cdxmbx" Oct 08 18:27:47 crc kubenswrapper[4859]: I1008 18:27:47.195342 4859 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/cc21f1af-ca05-4cf4-b66f-f7fcde5026ee-util\") pod \"fa9831ede5d93c33d525b70ce6ddf94e500d80992af75a3305fe98835cdxmbx\" (UID: \"cc21f1af-ca05-4cf4-b66f-f7fcde5026ee\") " pod="openshift-marketplace/fa9831ede5d93c33d525b70ce6ddf94e500d80992af75a3305fe98835cdxmbx" Oct 08 18:27:47 crc kubenswrapper[4859]: I1008 18:27:47.196078 4859 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/cc21f1af-ca05-4cf4-b66f-f7fcde5026ee-bundle\") pod \"fa9831ede5d93c33d525b70ce6ddf94e500d80992af75a3305fe98835cdxmbx\" (UID: \"cc21f1af-ca05-4cf4-b66f-f7fcde5026ee\") " pod="openshift-marketplace/fa9831ede5d93c33d525b70ce6ddf94e500d80992af75a3305fe98835cdxmbx" Oct 08 18:27:47 crc kubenswrapper[4859]: I1008 18:27:47.196191 4859 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/cc21f1af-ca05-4cf4-b66f-f7fcde5026ee-util\") pod \"fa9831ede5d93c33d525b70ce6ddf94e500d80992af75a3305fe98835cdxmbx\" (UID: \"cc21f1af-ca05-4cf4-b66f-f7fcde5026ee\") " pod="openshift-marketplace/fa9831ede5d93c33d525b70ce6ddf94e500d80992af75a3305fe98835cdxmbx" Oct 08 18:27:47 crc kubenswrapper[4859]: I1008 18:27:47.216342 4859 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-vkjbd\" (UniqueName: \"kubernetes.io/projected/cc21f1af-ca05-4cf4-b66f-f7fcde5026ee-kube-api-access-vkjbd\") pod \"fa9831ede5d93c33d525b70ce6ddf94e500d80992af75a3305fe98835cdxmbx\" (UID: \"cc21f1af-ca05-4cf4-b66f-f7fcde5026ee\") " pod="openshift-marketplace/fa9831ede5d93c33d525b70ce6ddf94e500d80992af75a3305fe98835cdxmbx" Oct 08 18:27:47 crc kubenswrapper[4859]: I1008 18:27:47.332514 4859 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/fa9831ede5d93c33d525b70ce6ddf94e500d80992af75a3305fe98835cdxmbx" Oct 08 18:27:47 crc kubenswrapper[4859]: I1008 18:27:47.752006 4859 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/fa9831ede5d93c33d525b70ce6ddf94e500d80992af75a3305fe98835cdxmbx"] Oct 08 18:27:47 crc kubenswrapper[4859]: I1008 18:27:47.924943 4859 patch_prober.go:28] interesting pod/machine-config-daemon-82s52 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 08 18:27:47 crc kubenswrapper[4859]: I1008 18:27:47.925333 4859 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-82s52" podUID="b23a6a6c-9d92-4e7b-840e-55cfda873a2d" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 08 18:27:47 crc kubenswrapper[4859]: I1008 18:27:47.925401 4859 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-82s52" Oct 08 18:27:47 crc kubenswrapper[4859]: I1008 18:27:47.926067 4859 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"487a3a1048ef98faa7383425fe83462232329f063579bf1e501fc8c04727fcea"} pod="openshift-machine-config-operator/machine-config-daemon-82s52" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Oct 08 18:27:47 crc kubenswrapper[4859]: I1008 18:27:47.926140 4859 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-82s52" podUID="b23a6a6c-9d92-4e7b-840e-55cfda873a2d" containerName="machine-config-daemon" containerID="cri-o://487a3a1048ef98faa7383425fe83462232329f063579bf1e501fc8c04727fcea" gracePeriod=600 Oct 08 18:27:48 crc kubenswrapper[4859]: I1008 18:27:48.280997 4859 generic.go:334] "Generic (PLEG): container finished" podID="cc21f1af-ca05-4cf4-b66f-f7fcde5026ee" containerID="7859e2bfb4ff6585f7e21bdd25d20ff67b1c6064b20bcba8cf071e4ee51e3cdf" exitCode=0 Oct 08 18:27:48 crc kubenswrapper[4859]: I1008 18:27:48.281133 4859 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/fa9831ede5d93c33d525b70ce6ddf94e500d80992af75a3305fe98835cdxmbx" event={"ID":"cc21f1af-ca05-4cf4-b66f-f7fcde5026ee","Type":"ContainerDied","Data":"7859e2bfb4ff6585f7e21bdd25d20ff67b1c6064b20bcba8cf071e4ee51e3cdf"} Oct 08 18:27:48 crc kubenswrapper[4859]: I1008 18:27:48.281212 4859 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/fa9831ede5d93c33d525b70ce6ddf94e500d80992af75a3305fe98835cdxmbx" event={"ID":"cc21f1af-ca05-4cf4-b66f-f7fcde5026ee","Type":"ContainerStarted","Data":"37606c93575ceafd5f27583358253a1a385c85ec9f2e640c43b016b55c85f970"} Oct 08 18:27:48 crc kubenswrapper[4859]: I1008 18:27:48.296178 4859 generic.go:334] "Generic (PLEG): container finished" podID="b23a6a6c-9d92-4e7b-840e-55cfda873a2d" containerID="487a3a1048ef98faa7383425fe83462232329f063579bf1e501fc8c04727fcea" exitCode=0 Oct 08 18:27:48 crc kubenswrapper[4859]: I1008 18:27:48.296240 4859 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-82s52" event={"ID":"b23a6a6c-9d92-4e7b-840e-55cfda873a2d","Type":"ContainerDied","Data":"487a3a1048ef98faa7383425fe83462232329f063579bf1e501fc8c04727fcea"} Oct 08 18:27:48 crc kubenswrapper[4859]: I1008 18:27:48.296280 4859 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-82s52" event={"ID":"b23a6a6c-9d92-4e7b-840e-55cfda873a2d","Type":"ContainerStarted","Data":"6ca72342b251d747b10d72d072b3e566097d846a52f53943f7da38fe2a34e100"} Oct 08 18:27:48 crc kubenswrapper[4859]: I1008 18:27:48.296298 4859 scope.go:117] "RemoveContainer" containerID="c6b208adefaf5eb65ea467fb4313c0f4281605067819f9fbd7f4404606fa52d0" Oct 08 18:27:50 crc kubenswrapper[4859]: I1008 18:27:50.312286 4859 generic.go:334] "Generic (PLEG): container finished" podID="cc21f1af-ca05-4cf4-b66f-f7fcde5026ee" containerID="e06d650a029a1411434c63f198cb68e5e32b88980d99a69a8ffd4d3b3704e652" exitCode=0 Oct 08 18:27:50 crc kubenswrapper[4859]: I1008 18:27:50.312378 4859 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/fa9831ede5d93c33d525b70ce6ddf94e500d80992af75a3305fe98835cdxmbx" event={"ID":"cc21f1af-ca05-4cf4-b66f-f7fcde5026ee","Type":"ContainerDied","Data":"e06d650a029a1411434c63f198cb68e5e32b88980d99a69a8ffd4d3b3704e652"} Oct 08 18:27:51 crc kubenswrapper[4859]: I1008 18:27:51.322891 4859 generic.go:334] "Generic (PLEG): container finished" podID="cc21f1af-ca05-4cf4-b66f-f7fcde5026ee" containerID="e2bd968228147763a810b9788ed56103209a91912b4126a7ef12de95b8bdd0bd" exitCode=0 Oct 08 18:27:51 crc kubenswrapper[4859]: I1008 18:27:51.322946 4859 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/fa9831ede5d93c33d525b70ce6ddf94e500d80992af75a3305fe98835cdxmbx" event={"ID":"cc21f1af-ca05-4cf4-b66f-f7fcde5026ee","Type":"ContainerDied","Data":"e2bd968228147763a810b9788ed56103209a91912b4126a7ef12de95b8bdd0bd"} Oct 08 18:27:52 crc kubenswrapper[4859]: I1008 18:27:52.567492 4859 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/fa9831ede5d93c33d525b70ce6ddf94e500d80992af75a3305fe98835cdxmbx" Oct 08 18:27:52 crc kubenswrapper[4859]: I1008 18:27:52.670996 4859 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/cc21f1af-ca05-4cf4-b66f-f7fcde5026ee-util\") pod \"cc21f1af-ca05-4cf4-b66f-f7fcde5026ee\" (UID: \"cc21f1af-ca05-4cf4-b66f-f7fcde5026ee\") " Oct 08 18:27:52 crc kubenswrapper[4859]: I1008 18:27:52.671058 4859 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-vkjbd\" (UniqueName: \"kubernetes.io/projected/cc21f1af-ca05-4cf4-b66f-f7fcde5026ee-kube-api-access-vkjbd\") pod \"cc21f1af-ca05-4cf4-b66f-f7fcde5026ee\" (UID: \"cc21f1af-ca05-4cf4-b66f-f7fcde5026ee\") " Oct 08 18:27:52 crc kubenswrapper[4859]: I1008 18:27:52.671103 4859 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/cc21f1af-ca05-4cf4-b66f-f7fcde5026ee-bundle\") pod \"cc21f1af-ca05-4cf4-b66f-f7fcde5026ee\" (UID: \"cc21f1af-ca05-4cf4-b66f-f7fcde5026ee\") " Oct 08 18:27:52 crc kubenswrapper[4859]: I1008 18:27:52.671916 4859 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/cc21f1af-ca05-4cf4-b66f-f7fcde5026ee-bundle" (OuterVolumeSpecName: "bundle") pod "cc21f1af-ca05-4cf4-b66f-f7fcde5026ee" (UID: "cc21f1af-ca05-4cf4-b66f-f7fcde5026ee"). InnerVolumeSpecName "bundle". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 08 18:27:52 crc kubenswrapper[4859]: I1008 18:27:52.676606 4859 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/cc21f1af-ca05-4cf4-b66f-f7fcde5026ee-kube-api-access-vkjbd" (OuterVolumeSpecName: "kube-api-access-vkjbd") pod "cc21f1af-ca05-4cf4-b66f-f7fcde5026ee" (UID: "cc21f1af-ca05-4cf4-b66f-f7fcde5026ee"). InnerVolumeSpecName "kube-api-access-vkjbd". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 08 18:27:52 crc kubenswrapper[4859]: I1008 18:27:52.690417 4859 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/cc21f1af-ca05-4cf4-b66f-f7fcde5026ee-util" (OuterVolumeSpecName: "util") pod "cc21f1af-ca05-4cf4-b66f-f7fcde5026ee" (UID: "cc21f1af-ca05-4cf4-b66f-f7fcde5026ee"). InnerVolumeSpecName "util". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 08 18:27:52 crc kubenswrapper[4859]: I1008 18:27:52.773050 4859 reconciler_common.go:293] "Volume detached for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/cc21f1af-ca05-4cf4-b66f-f7fcde5026ee-util\") on node \"crc\" DevicePath \"\"" Oct 08 18:27:52 crc kubenswrapper[4859]: I1008 18:27:52.773309 4859 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-vkjbd\" (UniqueName: \"kubernetes.io/projected/cc21f1af-ca05-4cf4-b66f-f7fcde5026ee-kube-api-access-vkjbd\") on node \"crc\" DevicePath \"\"" Oct 08 18:27:52 crc kubenswrapper[4859]: I1008 18:27:52.773369 4859 reconciler_common.go:293] "Volume detached for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/cc21f1af-ca05-4cf4-b66f-f7fcde5026ee-bundle\") on node \"crc\" DevicePath \"\"" Oct 08 18:27:53 crc kubenswrapper[4859]: I1008 18:27:53.337468 4859 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/fa9831ede5d93c33d525b70ce6ddf94e500d80992af75a3305fe98835cdxmbx" event={"ID":"cc21f1af-ca05-4cf4-b66f-f7fcde5026ee","Type":"ContainerDied","Data":"37606c93575ceafd5f27583358253a1a385c85ec9f2e640c43b016b55c85f970"} Oct 08 18:27:53 crc kubenswrapper[4859]: I1008 18:27:53.337506 4859 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="37606c93575ceafd5f27583358253a1a385c85ec9f2e640c43b016b55c85f970" Oct 08 18:27:53 crc kubenswrapper[4859]: I1008 18:27:53.338013 4859 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/fa9831ede5d93c33d525b70ce6ddf94e500d80992af75a3305fe98835cdxmbx" Oct 08 18:27:58 crc kubenswrapper[4859]: I1008 18:27:58.387099 4859 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-nmstate/nmstate-operator-858ddd8f98-vbswp"] Oct 08 18:27:58 crc kubenswrapper[4859]: E1008 18:27:58.388124 4859 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="cc21f1af-ca05-4cf4-b66f-f7fcde5026ee" containerName="util" Oct 08 18:27:58 crc kubenswrapper[4859]: I1008 18:27:58.388144 4859 state_mem.go:107] "Deleted CPUSet assignment" podUID="cc21f1af-ca05-4cf4-b66f-f7fcde5026ee" containerName="util" Oct 08 18:27:58 crc kubenswrapper[4859]: E1008 18:27:58.388176 4859 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="cc21f1af-ca05-4cf4-b66f-f7fcde5026ee" containerName="pull" Oct 08 18:27:58 crc kubenswrapper[4859]: I1008 18:27:58.388189 4859 state_mem.go:107] "Deleted CPUSet assignment" podUID="cc21f1af-ca05-4cf4-b66f-f7fcde5026ee" containerName="pull" Oct 08 18:27:58 crc kubenswrapper[4859]: E1008 18:27:58.388200 4859 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="cc21f1af-ca05-4cf4-b66f-f7fcde5026ee" containerName="extract" Oct 08 18:27:58 crc kubenswrapper[4859]: I1008 18:27:58.388211 4859 state_mem.go:107] "Deleted CPUSet assignment" podUID="cc21f1af-ca05-4cf4-b66f-f7fcde5026ee" containerName="extract" Oct 08 18:27:58 crc kubenswrapper[4859]: I1008 18:27:58.388359 4859 memory_manager.go:354] "RemoveStaleState removing state" podUID="cc21f1af-ca05-4cf4-b66f-f7fcde5026ee" containerName="extract" Oct 08 18:27:58 crc kubenswrapper[4859]: I1008 18:27:58.388993 4859 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-nmstate/nmstate-operator-858ddd8f98-vbswp" Oct 08 18:27:58 crc kubenswrapper[4859]: I1008 18:27:58.391680 4859 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-nmstate"/"kube-root-ca.crt" Oct 08 18:27:58 crc kubenswrapper[4859]: I1008 18:27:58.391757 4859 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-nmstate"/"nmstate-operator-dockercfg-vjfxp" Oct 08 18:27:58 crc kubenswrapper[4859]: I1008 18:27:58.391895 4859 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-nmstate"/"openshift-service-ca.crt" Oct 08 18:27:58 crc kubenswrapper[4859]: I1008 18:27:58.424641 4859 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-nmstate/nmstate-operator-858ddd8f98-vbswp"] Oct 08 18:27:58 crc kubenswrapper[4859]: I1008 18:27:58.454456 4859 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-gbkg9\" (UniqueName: \"kubernetes.io/projected/3a114f4c-72fa-4640-86e4-cb4326069481-kube-api-access-gbkg9\") pod \"nmstate-operator-858ddd8f98-vbswp\" (UID: \"3a114f4c-72fa-4640-86e4-cb4326069481\") " pod="openshift-nmstate/nmstate-operator-858ddd8f98-vbswp" Oct 08 18:27:58 crc kubenswrapper[4859]: I1008 18:27:58.555184 4859 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-gbkg9\" (UniqueName: \"kubernetes.io/projected/3a114f4c-72fa-4640-86e4-cb4326069481-kube-api-access-gbkg9\") pod \"nmstate-operator-858ddd8f98-vbswp\" (UID: \"3a114f4c-72fa-4640-86e4-cb4326069481\") " pod="openshift-nmstate/nmstate-operator-858ddd8f98-vbswp" Oct 08 18:27:58 crc kubenswrapper[4859]: I1008 18:27:58.582365 4859 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-gbkg9\" (UniqueName: \"kubernetes.io/projected/3a114f4c-72fa-4640-86e4-cb4326069481-kube-api-access-gbkg9\") pod \"nmstate-operator-858ddd8f98-vbswp\" (UID: \"3a114f4c-72fa-4640-86e4-cb4326069481\") " pod="openshift-nmstate/nmstate-operator-858ddd8f98-vbswp" Oct 08 18:27:58 crc kubenswrapper[4859]: I1008 18:27:58.714597 4859 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-nmstate/nmstate-operator-858ddd8f98-vbswp" Oct 08 18:27:58 crc kubenswrapper[4859]: I1008 18:27:58.924069 4859 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-nmstate/nmstate-operator-858ddd8f98-vbswp"] Oct 08 18:27:59 crc kubenswrapper[4859]: I1008 18:27:59.374120 4859 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-nmstate/nmstate-operator-858ddd8f98-vbswp" event={"ID":"3a114f4c-72fa-4640-86e4-cb4326069481","Type":"ContainerStarted","Data":"fd45c1d16807ab9ac823f1ae14817bb7097b5efa7923c0d587113d1ce6793aa2"} Oct 08 18:28:02 crc kubenswrapper[4859]: I1008 18:28:02.400575 4859 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-nmstate/nmstate-operator-858ddd8f98-vbswp" event={"ID":"3a114f4c-72fa-4640-86e4-cb4326069481","Type":"ContainerStarted","Data":"9a07b0d425fae66b0c7b450b2166a5d30a5bda12815ed5e97e8059ec49243811"} Oct 08 18:28:02 crc kubenswrapper[4859]: I1008 18:28:02.425646 4859 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-nmstate/nmstate-operator-858ddd8f98-vbswp" podStartSLOduration=1.76193483 podStartE2EDuration="4.425607354s" podCreationTimestamp="2025-10-08 18:27:58 +0000 UTC" firstStartedPulling="2025-10-08 18:27:58.931421233 +0000 UTC m=+649.178260612" lastFinishedPulling="2025-10-08 18:28:01.595093757 +0000 UTC m=+651.841933136" observedRunningTime="2025-10-08 18:28:02.417845092 +0000 UTC m=+652.664684481" watchObservedRunningTime="2025-10-08 18:28:02.425607354 +0000 UTC m=+652.672446733" Oct 08 18:28:07 crc kubenswrapper[4859]: I1008 18:28:07.650433 4859 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-nmstate/nmstate-metrics-fdff9cb8d-8vdx7"] Oct 08 18:28:07 crc kubenswrapper[4859]: I1008 18:28:07.651945 4859 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-nmstate/nmstate-metrics-fdff9cb8d-8vdx7" Oct 08 18:28:07 crc kubenswrapper[4859]: I1008 18:28:07.654245 4859 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-nmstate"/"nmstate-handler-dockercfg-4wdzg" Oct 08 18:28:07 crc kubenswrapper[4859]: I1008 18:28:07.662601 4859 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-nmstate/nmstate-webhook-6cdbc54649-728jz"] Oct 08 18:28:07 crc kubenswrapper[4859]: I1008 18:28:07.663959 4859 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-nmstate/nmstate-webhook-6cdbc54649-728jz" Oct 08 18:28:07 crc kubenswrapper[4859]: I1008 18:28:07.671421 4859 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-nmstate"/"openshift-nmstate-webhook" Oct 08 18:28:07 crc kubenswrapper[4859]: I1008 18:28:07.683675 4859 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-nmstate/nmstate-metrics-fdff9cb8d-8vdx7"] Oct 08 18:28:07 crc kubenswrapper[4859]: I1008 18:28:07.692423 4859 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-nmstate/nmstate-handler-htrgb"] Oct 08 18:28:07 crc kubenswrapper[4859]: I1008 18:28:07.693415 4859 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-nmstate/nmstate-handler-htrgb" Oct 08 18:28:07 crc kubenswrapper[4859]: I1008 18:28:07.696544 4859 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-nmstate/nmstate-webhook-6cdbc54649-728jz"] Oct 08 18:28:07 crc kubenswrapper[4859]: I1008 18:28:07.774102 4859 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovs-socket\" (UniqueName: \"kubernetes.io/host-path/692be5e4-be66-4010-9547-610a2d3d85fd-ovs-socket\") pod \"nmstate-handler-htrgb\" (UID: \"692be5e4-be66-4010-9547-610a2d3d85fd\") " pod="openshift-nmstate/nmstate-handler-htrgb" Oct 08 18:28:07 crc kubenswrapper[4859]: I1008 18:28:07.774162 4859 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nmstate-lock\" (UniqueName: \"kubernetes.io/host-path/692be5e4-be66-4010-9547-610a2d3d85fd-nmstate-lock\") pod \"nmstate-handler-htrgb\" (UID: \"692be5e4-be66-4010-9547-610a2d3d85fd\") " pod="openshift-nmstate/nmstate-handler-htrgb" Oct 08 18:28:07 crc kubenswrapper[4859]: I1008 18:28:07.774195 4859 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-2rzdw\" (UniqueName: \"kubernetes.io/projected/0beb8934-6616-4771-b93d-5b833f87600d-kube-api-access-2rzdw\") pod \"nmstate-metrics-fdff9cb8d-8vdx7\" (UID: \"0beb8934-6616-4771-b93d-5b833f87600d\") " pod="openshift-nmstate/nmstate-metrics-fdff9cb8d-8vdx7" Oct 08 18:28:07 crc kubenswrapper[4859]: I1008 18:28:07.774253 4859 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-rkv6k\" (UniqueName: \"kubernetes.io/projected/5af71d73-77d0-409b-8b48-cca9164c529e-kube-api-access-rkv6k\") pod \"nmstate-webhook-6cdbc54649-728jz\" (UID: \"5af71d73-77d0-409b-8b48-cca9164c529e\") " pod="openshift-nmstate/nmstate-webhook-6cdbc54649-728jz" Oct 08 18:28:07 crc kubenswrapper[4859]: I1008 18:28:07.774271 4859 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"tls-key-pair\" (UniqueName: \"kubernetes.io/secret/5af71d73-77d0-409b-8b48-cca9164c529e-tls-key-pair\") pod \"nmstate-webhook-6cdbc54649-728jz\" (UID: \"5af71d73-77d0-409b-8b48-cca9164c529e\") " pod="openshift-nmstate/nmstate-webhook-6cdbc54649-728jz" Oct 08 18:28:07 crc kubenswrapper[4859]: I1008 18:28:07.774306 4859 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dbus-socket\" (UniqueName: \"kubernetes.io/host-path/692be5e4-be66-4010-9547-610a2d3d85fd-dbus-socket\") pod \"nmstate-handler-htrgb\" (UID: \"692be5e4-be66-4010-9547-610a2d3d85fd\") " pod="openshift-nmstate/nmstate-handler-htrgb" Oct 08 18:28:07 crc kubenswrapper[4859]: I1008 18:28:07.774322 4859 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-pvrj7\" (UniqueName: \"kubernetes.io/projected/692be5e4-be66-4010-9547-610a2d3d85fd-kube-api-access-pvrj7\") pod \"nmstate-handler-htrgb\" (UID: \"692be5e4-be66-4010-9547-610a2d3d85fd\") " pod="openshift-nmstate/nmstate-handler-htrgb" Oct 08 18:28:07 crc kubenswrapper[4859]: I1008 18:28:07.813495 4859 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-nmstate/nmstate-console-plugin-6b874cbd85-r5l52"] Oct 08 18:28:07 crc kubenswrapper[4859]: I1008 18:28:07.814269 4859 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-nmstate/nmstate-console-plugin-6b874cbd85-r5l52" Oct 08 18:28:07 crc kubenswrapper[4859]: I1008 18:28:07.821331 4859 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-nmstate"/"plugin-serving-cert" Oct 08 18:28:07 crc kubenswrapper[4859]: I1008 18:28:07.821408 4859 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-nmstate"/"nginx-conf" Oct 08 18:28:07 crc kubenswrapper[4859]: I1008 18:28:07.821363 4859 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-nmstate"/"default-dockercfg-9rftj" Oct 08 18:28:07 crc kubenswrapper[4859]: I1008 18:28:07.839930 4859 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-nmstate/nmstate-console-plugin-6b874cbd85-r5l52"] Oct 08 18:28:07 crc kubenswrapper[4859]: I1008 18:28:07.876082 4859 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dbus-socket\" (UniqueName: \"kubernetes.io/host-path/692be5e4-be66-4010-9547-610a2d3d85fd-dbus-socket\") pod \"nmstate-handler-htrgb\" (UID: \"692be5e4-be66-4010-9547-610a2d3d85fd\") " pod="openshift-nmstate/nmstate-handler-htrgb" Oct 08 18:28:07 crc kubenswrapper[4859]: I1008 18:28:07.876151 4859 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-pvrj7\" (UniqueName: \"kubernetes.io/projected/692be5e4-be66-4010-9547-610a2d3d85fd-kube-api-access-pvrj7\") pod \"nmstate-handler-htrgb\" (UID: \"692be5e4-be66-4010-9547-610a2d3d85fd\") " pod="openshift-nmstate/nmstate-handler-htrgb" Oct 08 18:28:07 crc kubenswrapper[4859]: I1008 18:28:07.876229 4859 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/a8be1e40-aedc-40bd-b47d-e381b5746080-nginx-conf\") pod \"nmstate-console-plugin-6b874cbd85-r5l52\" (UID: \"a8be1e40-aedc-40bd-b47d-e381b5746080\") " pod="openshift-nmstate/nmstate-console-plugin-6b874cbd85-r5l52" Oct 08 18:28:07 crc kubenswrapper[4859]: I1008 18:28:07.876266 4859 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"plugin-serving-cert\" (UniqueName: \"kubernetes.io/secret/a8be1e40-aedc-40bd-b47d-e381b5746080-plugin-serving-cert\") pod \"nmstate-console-plugin-6b874cbd85-r5l52\" (UID: \"a8be1e40-aedc-40bd-b47d-e381b5746080\") " pod="openshift-nmstate/nmstate-console-plugin-6b874cbd85-r5l52" Oct 08 18:28:07 crc kubenswrapper[4859]: I1008 18:28:07.876289 4859 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-4q9d7\" (UniqueName: \"kubernetes.io/projected/a8be1e40-aedc-40bd-b47d-e381b5746080-kube-api-access-4q9d7\") pod \"nmstate-console-plugin-6b874cbd85-r5l52\" (UID: \"a8be1e40-aedc-40bd-b47d-e381b5746080\") " pod="openshift-nmstate/nmstate-console-plugin-6b874cbd85-r5l52" Oct 08 18:28:07 crc kubenswrapper[4859]: I1008 18:28:07.876314 4859 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovs-socket\" (UniqueName: \"kubernetes.io/host-path/692be5e4-be66-4010-9547-610a2d3d85fd-ovs-socket\") pod \"nmstate-handler-htrgb\" (UID: \"692be5e4-be66-4010-9547-610a2d3d85fd\") " pod="openshift-nmstate/nmstate-handler-htrgb" Oct 08 18:28:07 crc kubenswrapper[4859]: I1008 18:28:07.876348 4859 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nmstate-lock\" (UniqueName: \"kubernetes.io/host-path/692be5e4-be66-4010-9547-610a2d3d85fd-nmstate-lock\") pod \"nmstate-handler-htrgb\" (UID: \"692be5e4-be66-4010-9547-610a2d3d85fd\") " pod="openshift-nmstate/nmstate-handler-htrgb" Oct 08 18:28:07 crc kubenswrapper[4859]: I1008 18:28:07.876374 4859 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-2rzdw\" (UniqueName: \"kubernetes.io/projected/0beb8934-6616-4771-b93d-5b833f87600d-kube-api-access-2rzdw\") pod \"nmstate-metrics-fdff9cb8d-8vdx7\" (UID: \"0beb8934-6616-4771-b93d-5b833f87600d\") " pod="openshift-nmstate/nmstate-metrics-fdff9cb8d-8vdx7" Oct 08 18:28:07 crc kubenswrapper[4859]: I1008 18:28:07.876403 4859 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rkv6k\" (UniqueName: \"kubernetes.io/projected/5af71d73-77d0-409b-8b48-cca9164c529e-kube-api-access-rkv6k\") pod \"nmstate-webhook-6cdbc54649-728jz\" (UID: \"5af71d73-77d0-409b-8b48-cca9164c529e\") " pod="openshift-nmstate/nmstate-webhook-6cdbc54649-728jz" Oct 08 18:28:07 crc kubenswrapper[4859]: I1008 18:28:07.876420 4859 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"tls-key-pair\" (UniqueName: \"kubernetes.io/secret/5af71d73-77d0-409b-8b48-cca9164c529e-tls-key-pair\") pod \"nmstate-webhook-6cdbc54649-728jz\" (UID: \"5af71d73-77d0-409b-8b48-cca9164c529e\") " pod="openshift-nmstate/nmstate-webhook-6cdbc54649-728jz" Oct 08 18:28:07 crc kubenswrapper[4859]: E1008 18:28:07.876564 4859 secret.go:188] Couldn't get secret openshift-nmstate/openshift-nmstate-webhook: secret "openshift-nmstate-webhook" not found Oct 08 18:28:07 crc kubenswrapper[4859]: E1008 18:28:07.876647 4859 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/5af71d73-77d0-409b-8b48-cca9164c529e-tls-key-pair podName:5af71d73-77d0-409b-8b48-cca9164c529e nodeName:}" failed. No retries permitted until 2025-10-08 18:28:08.376618436 +0000 UTC m=+658.623457805 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "tls-key-pair" (UniqueName: "kubernetes.io/secret/5af71d73-77d0-409b-8b48-cca9164c529e-tls-key-pair") pod "nmstate-webhook-6cdbc54649-728jz" (UID: "5af71d73-77d0-409b-8b48-cca9164c529e") : secret "openshift-nmstate-webhook" not found Oct 08 18:28:07 crc kubenswrapper[4859]: I1008 18:28:07.876810 4859 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovs-socket\" (UniqueName: \"kubernetes.io/host-path/692be5e4-be66-4010-9547-610a2d3d85fd-ovs-socket\") pod \"nmstate-handler-htrgb\" (UID: \"692be5e4-be66-4010-9547-610a2d3d85fd\") " pod="openshift-nmstate/nmstate-handler-htrgb" Oct 08 18:28:07 crc kubenswrapper[4859]: I1008 18:28:07.876988 4859 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nmstate-lock\" (UniqueName: \"kubernetes.io/host-path/692be5e4-be66-4010-9547-610a2d3d85fd-nmstate-lock\") pod \"nmstate-handler-htrgb\" (UID: \"692be5e4-be66-4010-9547-610a2d3d85fd\") " pod="openshift-nmstate/nmstate-handler-htrgb" Oct 08 18:28:07 crc kubenswrapper[4859]: I1008 18:28:07.877456 4859 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dbus-socket\" (UniqueName: \"kubernetes.io/host-path/692be5e4-be66-4010-9547-610a2d3d85fd-dbus-socket\") pod \"nmstate-handler-htrgb\" (UID: \"692be5e4-be66-4010-9547-610a2d3d85fd\") " pod="openshift-nmstate/nmstate-handler-htrgb" Oct 08 18:28:07 crc kubenswrapper[4859]: I1008 18:28:07.899644 4859 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-2rzdw\" (UniqueName: \"kubernetes.io/projected/0beb8934-6616-4771-b93d-5b833f87600d-kube-api-access-2rzdw\") pod \"nmstate-metrics-fdff9cb8d-8vdx7\" (UID: \"0beb8934-6616-4771-b93d-5b833f87600d\") " pod="openshift-nmstate/nmstate-metrics-fdff9cb8d-8vdx7" Oct 08 18:28:07 crc kubenswrapper[4859]: I1008 18:28:07.899666 4859 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rkv6k\" (UniqueName: \"kubernetes.io/projected/5af71d73-77d0-409b-8b48-cca9164c529e-kube-api-access-rkv6k\") pod \"nmstate-webhook-6cdbc54649-728jz\" (UID: \"5af71d73-77d0-409b-8b48-cca9164c529e\") " pod="openshift-nmstate/nmstate-webhook-6cdbc54649-728jz" Oct 08 18:28:07 crc kubenswrapper[4859]: I1008 18:28:07.903490 4859 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-pvrj7\" (UniqueName: \"kubernetes.io/projected/692be5e4-be66-4010-9547-610a2d3d85fd-kube-api-access-pvrj7\") pod \"nmstate-handler-htrgb\" (UID: \"692be5e4-be66-4010-9547-610a2d3d85fd\") " pod="openshift-nmstate/nmstate-handler-htrgb" Oct 08 18:28:07 crc kubenswrapper[4859]: I1008 18:28:07.977572 4859 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/a8be1e40-aedc-40bd-b47d-e381b5746080-nginx-conf\") pod \"nmstate-console-plugin-6b874cbd85-r5l52\" (UID: \"a8be1e40-aedc-40bd-b47d-e381b5746080\") " pod="openshift-nmstate/nmstate-console-plugin-6b874cbd85-r5l52" Oct 08 18:28:07 crc kubenswrapper[4859]: I1008 18:28:07.977639 4859 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"plugin-serving-cert\" (UniqueName: \"kubernetes.io/secret/a8be1e40-aedc-40bd-b47d-e381b5746080-plugin-serving-cert\") pod \"nmstate-console-plugin-6b874cbd85-r5l52\" (UID: \"a8be1e40-aedc-40bd-b47d-e381b5746080\") " pod="openshift-nmstate/nmstate-console-plugin-6b874cbd85-r5l52" Oct 08 18:28:07 crc kubenswrapper[4859]: I1008 18:28:07.977662 4859 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-4q9d7\" (UniqueName: \"kubernetes.io/projected/a8be1e40-aedc-40bd-b47d-e381b5746080-kube-api-access-4q9d7\") pod \"nmstate-console-plugin-6b874cbd85-r5l52\" (UID: \"a8be1e40-aedc-40bd-b47d-e381b5746080\") " pod="openshift-nmstate/nmstate-console-plugin-6b874cbd85-r5l52" Oct 08 18:28:07 crc kubenswrapper[4859]: I1008 18:28:07.979592 4859 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/a8be1e40-aedc-40bd-b47d-e381b5746080-nginx-conf\") pod \"nmstate-console-plugin-6b874cbd85-r5l52\" (UID: \"a8be1e40-aedc-40bd-b47d-e381b5746080\") " pod="openshift-nmstate/nmstate-console-plugin-6b874cbd85-r5l52" Oct 08 18:28:07 crc kubenswrapper[4859]: I1008 18:28:07.981659 4859 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"plugin-serving-cert\" (UniqueName: \"kubernetes.io/secret/a8be1e40-aedc-40bd-b47d-e381b5746080-plugin-serving-cert\") pod \"nmstate-console-plugin-6b874cbd85-r5l52\" (UID: \"a8be1e40-aedc-40bd-b47d-e381b5746080\") " pod="openshift-nmstate/nmstate-console-plugin-6b874cbd85-r5l52" Oct 08 18:28:07 crc kubenswrapper[4859]: I1008 18:28:07.988670 4859 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-nmstate/nmstate-metrics-fdff9cb8d-8vdx7" Oct 08 18:28:08 crc kubenswrapper[4859]: I1008 18:28:08.003450 4859 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-4q9d7\" (UniqueName: \"kubernetes.io/projected/a8be1e40-aedc-40bd-b47d-e381b5746080-kube-api-access-4q9d7\") pod \"nmstate-console-plugin-6b874cbd85-r5l52\" (UID: \"a8be1e40-aedc-40bd-b47d-e381b5746080\") " pod="openshift-nmstate/nmstate-console-plugin-6b874cbd85-r5l52" Oct 08 18:28:08 crc kubenswrapper[4859]: I1008 18:28:08.021265 4859 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-nmstate/nmstate-handler-htrgb" Oct 08 18:28:08 crc kubenswrapper[4859]: I1008 18:28:08.028782 4859 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-console/console-5fc69bdff-x57fv"] Oct 08 18:28:08 crc kubenswrapper[4859]: I1008 18:28:08.029529 4859 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-console/console-5fc69bdff-x57fv" Oct 08 18:28:08 crc kubenswrapper[4859]: I1008 18:28:08.046925 4859 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-console/console-5fc69bdff-x57fv"] Oct 08 18:28:08 crc kubenswrapper[4859]: I1008 18:28:08.079118 4859 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/f9c07424-efe0-49d5-8de0-8e74405dfd48-service-ca\") pod \"console-5fc69bdff-x57fv\" (UID: \"f9c07424-efe0-49d5-8de0-8e74405dfd48\") " pod="openshift-console/console-5fc69bdff-x57fv" Oct 08 18:28:08 crc kubenswrapper[4859]: I1008 18:28:08.079175 4859 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/f9c07424-efe0-49d5-8de0-8e74405dfd48-console-oauth-config\") pod \"console-5fc69bdff-x57fv\" (UID: \"f9c07424-efe0-49d5-8de0-8e74405dfd48\") " pod="openshift-console/console-5fc69bdff-x57fv" Oct 08 18:28:08 crc kubenswrapper[4859]: I1008 18:28:08.079196 4859 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/f9c07424-efe0-49d5-8de0-8e74405dfd48-console-config\") pod \"console-5fc69bdff-x57fv\" (UID: \"f9c07424-efe0-49d5-8de0-8e74405dfd48\") " pod="openshift-console/console-5fc69bdff-x57fv" Oct 08 18:28:08 crc kubenswrapper[4859]: I1008 18:28:08.079223 4859 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/f9c07424-efe0-49d5-8de0-8e74405dfd48-console-serving-cert\") pod \"console-5fc69bdff-x57fv\" (UID: \"f9c07424-efe0-49d5-8de0-8e74405dfd48\") " pod="openshift-console/console-5fc69bdff-x57fv" Oct 08 18:28:08 crc kubenswrapper[4859]: I1008 18:28:08.079284 4859 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/f9c07424-efe0-49d5-8de0-8e74405dfd48-oauth-serving-cert\") pod \"console-5fc69bdff-x57fv\" (UID: \"f9c07424-efe0-49d5-8de0-8e74405dfd48\") " pod="openshift-console/console-5fc69bdff-x57fv" Oct 08 18:28:08 crc kubenswrapper[4859]: I1008 18:28:08.079657 4859 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-w6ckn\" (UniqueName: \"kubernetes.io/projected/f9c07424-efe0-49d5-8de0-8e74405dfd48-kube-api-access-w6ckn\") pod \"console-5fc69bdff-x57fv\" (UID: \"f9c07424-efe0-49d5-8de0-8e74405dfd48\") " pod="openshift-console/console-5fc69bdff-x57fv" Oct 08 18:28:08 crc kubenswrapper[4859]: I1008 18:28:08.079713 4859 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/f9c07424-efe0-49d5-8de0-8e74405dfd48-trusted-ca-bundle\") pod \"console-5fc69bdff-x57fv\" (UID: \"f9c07424-efe0-49d5-8de0-8e74405dfd48\") " pod="openshift-console/console-5fc69bdff-x57fv" Oct 08 18:28:08 crc kubenswrapper[4859]: I1008 18:28:08.147879 4859 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-nmstate/nmstate-console-plugin-6b874cbd85-r5l52" Oct 08 18:28:08 crc kubenswrapper[4859]: I1008 18:28:08.185315 4859 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/f9c07424-efe0-49d5-8de0-8e74405dfd48-console-oauth-config\") pod \"console-5fc69bdff-x57fv\" (UID: \"f9c07424-efe0-49d5-8de0-8e74405dfd48\") " pod="openshift-console/console-5fc69bdff-x57fv" Oct 08 18:28:08 crc kubenswrapper[4859]: I1008 18:28:08.185355 4859 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/f9c07424-efe0-49d5-8de0-8e74405dfd48-console-config\") pod \"console-5fc69bdff-x57fv\" (UID: \"f9c07424-efe0-49d5-8de0-8e74405dfd48\") " pod="openshift-console/console-5fc69bdff-x57fv" Oct 08 18:28:08 crc kubenswrapper[4859]: I1008 18:28:08.185378 4859 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/f9c07424-efe0-49d5-8de0-8e74405dfd48-console-serving-cert\") pod \"console-5fc69bdff-x57fv\" (UID: \"f9c07424-efe0-49d5-8de0-8e74405dfd48\") " pod="openshift-console/console-5fc69bdff-x57fv" Oct 08 18:28:08 crc kubenswrapper[4859]: I1008 18:28:08.185438 4859 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/f9c07424-efe0-49d5-8de0-8e74405dfd48-oauth-serving-cert\") pod \"console-5fc69bdff-x57fv\" (UID: \"f9c07424-efe0-49d5-8de0-8e74405dfd48\") " pod="openshift-console/console-5fc69bdff-x57fv" Oct 08 18:28:08 crc kubenswrapper[4859]: I1008 18:28:08.185474 4859 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-w6ckn\" (UniqueName: \"kubernetes.io/projected/f9c07424-efe0-49d5-8de0-8e74405dfd48-kube-api-access-w6ckn\") pod \"console-5fc69bdff-x57fv\" (UID: \"f9c07424-efe0-49d5-8de0-8e74405dfd48\") " pod="openshift-console/console-5fc69bdff-x57fv" Oct 08 18:28:08 crc kubenswrapper[4859]: I1008 18:28:08.185502 4859 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/f9c07424-efe0-49d5-8de0-8e74405dfd48-trusted-ca-bundle\") pod \"console-5fc69bdff-x57fv\" (UID: \"f9c07424-efe0-49d5-8de0-8e74405dfd48\") " pod="openshift-console/console-5fc69bdff-x57fv" Oct 08 18:28:08 crc kubenswrapper[4859]: I1008 18:28:08.185519 4859 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/f9c07424-efe0-49d5-8de0-8e74405dfd48-service-ca\") pod \"console-5fc69bdff-x57fv\" (UID: \"f9c07424-efe0-49d5-8de0-8e74405dfd48\") " pod="openshift-console/console-5fc69bdff-x57fv" Oct 08 18:28:08 crc kubenswrapper[4859]: I1008 18:28:08.188943 4859 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/f9c07424-efe0-49d5-8de0-8e74405dfd48-service-ca\") pod \"console-5fc69bdff-x57fv\" (UID: \"f9c07424-efe0-49d5-8de0-8e74405dfd48\") " pod="openshift-console/console-5fc69bdff-x57fv" Oct 08 18:28:08 crc kubenswrapper[4859]: I1008 18:28:08.188974 4859 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/f9c07424-efe0-49d5-8de0-8e74405dfd48-oauth-serving-cert\") pod \"console-5fc69bdff-x57fv\" (UID: \"f9c07424-efe0-49d5-8de0-8e74405dfd48\") " pod="openshift-console/console-5fc69bdff-x57fv" Oct 08 18:28:08 crc kubenswrapper[4859]: I1008 18:28:08.189067 4859 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/f9c07424-efe0-49d5-8de0-8e74405dfd48-console-config\") pod \"console-5fc69bdff-x57fv\" (UID: \"f9c07424-efe0-49d5-8de0-8e74405dfd48\") " pod="openshift-console/console-5fc69bdff-x57fv" Oct 08 18:28:08 crc kubenswrapper[4859]: I1008 18:28:08.189207 4859 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/f9c07424-efe0-49d5-8de0-8e74405dfd48-trusted-ca-bundle\") pod \"console-5fc69bdff-x57fv\" (UID: \"f9c07424-efe0-49d5-8de0-8e74405dfd48\") " pod="openshift-console/console-5fc69bdff-x57fv" Oct 08 18:28:08 crc kubenswrapper[4859]: I1008 18:28:08.195106 4859 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/f9c07424-efe0-49d5-8de0-8e74405dfd48-console-oauth-config\") pod \"console-5fc69bdff-x57fv\" (UID: \"f9c07424-efe0-49d5-8de0-8e74405dfd48\") " pod="openshift-console/console-5fc69bdff-x57fv" Oct 08 18:28:08 crc kubenswrapper[4859]: I1008 18:28:08.195616 4859 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/f9c07424-efe0-49d5-8de0-8e74405dfd48-console-serving-cert\") pod \"console-5fc69bdff-x57fv\" (UID: \"f9c07424-efe0-49d5-8de0-8e74405dfd48\") " pod="openshift-console/console-5fc69bdff-x57fv" Oct 08 18:28:08 crc kubenswrapper[4859]: I1008 18:28:08.208111 4859 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-w6ckn\" (UniqueName: \"kubernetes.io/projected/f9c07424-efe0-49d5-8de0-8e74405dfd48-kube-api-access-w6ckn\") pod \"console-5fc69bdff-x57fv\" (UID: \"f9c07424-efe0-49d5-8de0-8e74405dfd48\") " pod="openshift-console/console-5fc69bdff-x57fv" Oct 08 18:28:08 crc kubenswrapper[4859]: I1008 18:28:08.350710 4859 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-nmstate/nmstate-console-plugin-6b874cbd85-r5l52"] Oct 08 18:28:08 crc kubenswrapper[4859]: W1008 18:28:08.356182 4859 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-poda8be1e40_aedc_40bd_b47d_e381b5746080.slice/crio-a84c82d8c4a6f06153b3246e68bbc420b5efce689cd45a4c72b8348ed3bea779 WatchSource:0}: Error finding container a84c82d8c4a6f06153b3246e68bbc420b5efce689cd45a4c72b8348ed3bea779: Status 404 returned error can't find the container with id a84c82d8c4a6f06153b3246e68bbc420b5efce689cd45a4c72b8348ed3bea779 Oct 08 18:28:08 crc kubenswrapper[4859]: I1008 18:28:08.382148 4859 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-console/console-5fc69bdff-x57fv" Oct 08 18:28:08 crc kubenswrapper[4859]: I1008 18:28:08.388310 4859 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"tls-key-pair\" (UniqueName: \"kubernetes.io/secret/5af71d73-77d0-409b-8b48-cca9164c529e-tls-key-pair\") pod \"nmstate-webhook-6cdbc54649-728jz\" (UID: \"5af71d73-77d0-409b-8b48-cca9164c529e\") " pod="openshift-nmstate/nmstate-webhook-6cdbc54649-728jz" Oct 08 18:28:08 crc kubenswrapper[4859]: I1008 18:28:08.393233 4859 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"tls-key-pair\" (UniqueName: \"kubernetes.io/secret/5af71d73-77d0-409b-8b48-cca9164c529e-tls-key-pair\") pod \"nmstate-webhook-6cdbc54649-728jz\" (UID: \"5af71d73-77d0-409b-8b48-cca9164c529e\") " pod="openshift-nmstate/nmstate-webhook-6cdbc54649-728jz" Oct 08 18:28:08 crc kubenswrapper[4859]: I1008 18:28:08.436166 4859 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-nmstate/nmstate-console-plugin-6b874cbd85-r5l52" event={"ID":"a8be1e40-aedc-40bd-b47d-e381b5746080","Type":"ContainerStarted","Data":"a84c82d8c4a6f06153b3246e68bbc420b5efce689cd45a4c72b8348ed3bea779"} Oct 08 18:28:08 crc kubenswrapper[4859]: I1008 18:28:08.450564 4859 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-nmstate/nmstate-handler-htrgb" event={"ID":"692be5e4-be66-4010-9547-610a2d3d85fd","Type":"ContainerStarted","Data":"58c6c2e7ec039950be8dbd5a0359815044cd9f9c05e1eb995dd63dba9ad2fce1"} Oct 08 18:28:08 crc kubenswrapper[4859]: I1008 18:28:08.460344 4859 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-nmstate/nmstate-metrics-fdff9cb8d-8vdx7"] Oct 08 18:28:08 crc kubenswrapper[4859]: W1008 18:28:08.482122 4859 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod0beb8934_6616_4771_b93d_5b833f87600d.slice/crio-cd438476f69c6a37f2d4cb2c12d2e860f341dce84548ab0ff865ff1dcccf1f88 WatchSource:0}: Error finding container cd438476f69c6a37f2d4cb2c12d2e860f341dce84548ab0ff865ff1dcccf1f88: Status 404 returned error can't find the container with id cd438476f69c6a37f2d4cb2c12d2e860f341dce84548ab0ff865ff1dcccf1f88 Oct 08 18:28:08 crc kubenswrapper[4859]: I1008 18:28:08.559730 4859 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-console/console-5fc69bdff-x57fv"] Oct 08 18:28:08 crc kubenswrapper[4859]: W1008 18:28:08.562574 4859 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podf9c07424_efe0_49d5_8de0_8e74405dfd48.slice/crio-7d681e29337af3db082c65e40c873b059de95ea6595a37b45589eba19390e7f6 WatchSource:0}: Error finding container 7d681e29337af3db082c65e40c873b059de95ea6595a37b45589eba19390e7f6: Status 404 returned error can't find the container with id 7d681e29337af3db082c65e40c873b059de95ea6595a37b45589eba19390e7f6 Oct 08 18:28:08 crc kubenswrapper[4859]: I1008 18:28:08.602530 4859 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-nmstate/nmstate-webhook-6cdbc54649-728jz" Oct 08 18:28:08 crc kubenswrapper[4859]: I1008 18:28:08.819002 4859 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-nmstate/nmstate-webhook-6cdbc54649-728jz"] Oct 08 18:28:09 crc kubenswrapper[4859]: I1008 18:28:09.458964 4859 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console/console-5fc69bdff-x57fv" event={"ID":"f9c07424-efe0-49d5-8de0-8e74405dfd48","Type":"ContainerStarted","Data":"547f8f243e8783530ba9ea35d14ceef898bb47f72ccf3400038ce601ce572c97"} Oct 08 18:28:09 crc kubenswrapper[4859]: I1008 18:28:09.459042 4859 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console/console-5fc69bdff-x57fv" event={"ID":"f9c07424-efe0-49d5-8de0-8e74405dfd48","Type":"ContainerStarted","Data":"7d681e29337af3db082c65e40c873b059de95ea6595a37b45589eba19390e7f6"} Oct 08 18:28:09 crc kubenswrapper[4859]: I1008 18:28:09.460840 4859 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-nmstate/nmstate-metrics-fdff9cb8d-8vdx7" event={"ID":"0beb8934-6616-4771-b93d-5b833f87600d","Type":"ContainerStarted","Data":"cd438476f69c6a37f2d4cb2c12d2e860f341dce84548ab0ff865ff1dcccf1f88"} Oct 08 18:28:09 crc kubenswrapper[4859]: I1008 18:28:09.462088 4859 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-nmstate/nmstate-webhook-6cdbc54649-728jz" event={"ID":"5af71d73-77d0-409b-8b48-cca9164c529e","Type":"ContainerStarted","Data":"dce98c3256648495f13b07701819dd98cd1a697d42922fef8136c54e17964f0c"} Oct 08 18:28:09 crc kubenswrapper[4859]: I1008 18:28:09.478743 4859 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-console/console-5fc69bdff-x57fv" podStartSLOduration=1.478720342 podStartE2EDuration="1.478720342s" podCreationTimestamp="2025-10-08 18:28:08 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-08 18:28:09.47665395 +0000 UTC m=+659.723493329" watchObservedRunningTime="2025-10-08 18:28:09.478720342 +0000 UTC m=+659.725559721" Oct 08 18:28:11 crc kubenswrapper[4859]: I1008 18:28:11.475352 4859 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-nmstate/nmstate-metrics-fdff9cb8d-8vdx7" event={"ID":"0beb8934-6616-4771-b93d-5b833f87600d","Type":"ContainerStarted","Data":"fcfcb9f28bff375350c72d73082cc306cadee7f5adfa041ee5109ef5c6d59901"} Oct 08 18:28:11 crc kubenswrapper[4859]: I1008 18:28:11.476867 4859 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-nmstate/nmstate-console-plugin-6b874cbd85-r5l52" event={"ID":"a8be1e40-aedc-40bd-b47d-e381b5746080","Type":"ContainerStarted","Data":"c9ca9c85c7fb1d9b011c9885bbfd6286cf81dcef643147251968d8ee1f4ca177"} Oct 08 18:28:11 crc kubenswrapper[4859]: I1008 18:28:11.478230 4859 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-nmstate/nmstate-webhook-6cdbc54649-728jz" event={"ID":"5af71d73-77d0-409b-8b48-cca9164c529e","Type":"ContainerStarted","Data":"fae632c95cec9c4dd28f35e7bf636d6ec57c8056e550e2a2ae05d36c6629b76c"} Oct 08 18:28:11 crc kubenswrapper[4859]: I1008 18:28:11.478363 4859 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-nmstate/nmstate-webhook-6cdbc54649-728jz" Oct 08 18:28:11 crc kubenswrapper[4859]: I1008 18:28:11.480260 4859 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-nmstate/nmstate-handler-htrgb" event={"ID":"692be5e4-be66-4010-9547-610a2d3d85fd","Type":"ContainerStarted","Data":"7312709c6206d4527913415bd0ce45275647c849c660cd9eeaea18f81f47b60e"} Oct 08 18:28:11 crc kubenswrapper[4859]: I1008 18:28:11.480436 4859 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-nmstate/nmstate-handler-htrgb" Oct 08 18:28:11 crc kubenswrapper[4859]: I1008 18:28:11.490856 4859 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-nmstate/nmstate-console-plugin-6b874cbd85-r5l52" podStartSLOduration=1.95557479 podStartE2EDuration="4.490836333s" podCreationTimestamp="2025-10-08 18:28:07 +0000 UTC" firstStartedPulling="2025-10-08 18:28:08.358578168 +0000 UTC m=+658.605417547" lastFinishedPulling="2025-10-08 18:28:10.893839711 +0000 UTC m=+661.140679090" observedRunningTime="2025-10-08 18:28:11.49039745 +0000 UTC m=+661.737236829" watchObservedRunningTime="2025-10-08 18:28:11.490836333 +0000 UTC m=+661.737675712" Oct 08 18:28:11 crc kubenswrapper[4859]: I1008 18:28:11.523870 4859 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-nmstate/nmstate-webhook-6cdbc54649-728jz" podStartSLOduration=2.451072005 podStartE2EDuration="4.523843897s" podCreationTimestamp="2025-10-08 18:28:07 +0000 UTC" firstStartedPulling="2025-10-08 18:28:08.841473368 +0000 UTC m=+659.088312747" lastFinishedPulling="2025-10-08 18:28:10.91424525 +0000 UTC m=+661.161084639" observedRunningTime="2025-10-08 18:28:11.521979772 +0000 UTC m=+661.768819151" watchObservedRunningTime="2025-10-08 18:28:11.523843897 +0000 UTC m=+661.770683276" Oct 08 18:28:11 crc kubenswrapper[4859]: I1008 18:28:11.524004 4859 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-nmstate/nmstate-handler-htrgb" podStartSLOduration=1.693125773 podStartE2EDuration="4.523998382s" podCreationTimestamp="2025-10-08 18:28:07 +0000 UTC" firstStartedPulling="2025-10-08 18:28:08.078579438 +0000 UTC m=+658.325418817" lastFinishedPulling="2025-10-08 18:28:10.909452007 +0000 UTC m=+661.156291426" observedRunningTime="2025-10-08 18:28:11.50915634 +0000 UTC m=+661.755995769" watchObservedRunningTime="2025-10-08 18:28:11.523998382 +0000 UTC m=+661.770837761" Oct 08 18:28:14 crc kubenswrapper[4859]: I1008 18:28:14.498202 4859 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-nmstate/nmstate-metrics-fdff9cb8d-8vdx7" event={"ID":"0beb8934-6616-4771-b93d-5b833f87600d","Type":"ContainerStarted","Data":"59662597be3d79cf030d3c3dd27e38af7f2bfd5779b436c30eee83ac565a7c93"} Oct 08 18:28:14 crc kubenswrapper[4859]: I1008 18:28:14.513228 4859 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-nmstate/nmstate-metrics-fdff9cb8d-8vdx7" podStartSLOduration=2.480779333 podStartE2EDuration="7.513202952s" podCreationTimestamp="2025-10-08 18:28:07 +0000 UTC" firstStartedPulling="2025-10-08 18:28:08.485158353 +0000 UTC m=+658.731997732" lastFinishedPulling="2025-10-08 18:28:13.517581972 +0000 UTC m=+663.764421351" observedRunningTime="2025-10-08 18:28:14.512887933 +0000 UTC m=+664.759727332" watchObservedRunningTime="2025-10-08 18:28:14.513202952 +0000 UTC m=+664.760042331" Oct 08 18:28:18 crc kubenswrapper[4859]: I1008 18:28:18.055038 4859 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-nmstate/nmstate-handler-htrgb" Oct 08 18:28:18 crc kubenswrapper[4859]: I1008 18:28:18.383180 4859 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-console/console-5fc69bdff-x57fv" Oct 08 18:28:18 crc kubenswrapper[4859]: I1008 18:28:18.383606 4859 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-console/console-5fc69bdff-x57fv" Oct 08 18:28:18 crc kubenswrapper[4859]: I1008 18:28:18.390714 4859 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-console/console-5fc69bdff-x57fv" Oct 08 18:28:18 crc kubenswrapper[4859]: I1008 18:28:18.536668 4859 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-console/console-5fc69bdff-x57fv" Oct 08 18:28:18 crc kubenswrapper[4859]: I1008 18:28:18.606609 4859 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-console/console-f9d7485db-m2xmp"] Oct 08 18:28:28 crc kubenswrapper[4859]: I1008 18:28:28.611094 4859 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-nmstate/nmstate-webhook-6cdbc54649-728jz" Oct 08 18:28:41 crc kubenswrapper[4859]: I1008 18:28:41.293273 4859 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/8f2f4ee801e5826a37d84a7b1fc4ccbf6b79de668302737d0f1152d8d2lfzpq"] Oct 08 18:28:41 crc kubenswrapper[4859]: I1008 18:28:41.294939 4859 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/8f2f4ee801e5826a37d84a7b1fc4ccbf6b79de668302737d0f1152d8d2lfzpq" Oct 08 18:28:41 crc kubenswrapper[4859]: I1008 18:28:41.296935 4859 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"default-dockercfg-vmwhc" Oct 08 18:28:41 crc kubenswrapper[4859]: I1008 18:28:41.304312 4859 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/8f2f4ee801e5826a37d84a7b1fc4ccbf6b79de668302737d0f1152d8d2lfzpq"] Oct 08 18:28:41 crc kubenswrapper[4859]: I1008 18:28:41.336046 4859 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/a200c80e-a0f7-40ee-b697-acf6d429d14f-util\") pod \"8f2f4ee801e5826a37d84a7b1fc4ccbf6b79de668302737d0f1152d8d2lfzpq\" (UID: \"a200c80e-a0f7-40ee-b697-acf6d429d14f\") " pod="openshift-marketplace/8f2f4ee801e5826a37d84a7b1fc4ccbf6b79de668302737d0f1152d8d2lfzpq" Oct 08 18:28:41 crc kubenswrapper[4859]: I1008 18:28:41.336116 4859 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-svb6t\" (UniqueName: \"kubernetes.io/projected/a200c80e-a0f7-40ee-b697-acf6d429d14f-kube-api-access-svb6t\") pod \"8f2f4ee801e5826a37d84a7b1fc4ccbf6b79de668302737d0f1152d8d2lfzpq\" (UID: \"a200c80e-a0f7-40ee-b697-acf6d429d14f\") " pod="openshift-marketplace/8f2f4ee801e5826a37d84a7b1fc4ccbf6b79de668302737d0f1152d8d2lfzpq" Oct 08 18:28:41 crc kubenswrapper[4859]: I1008 18:28:41.336171 4859 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/a200c80e-a0f7-40ee-b697-acf6d429d14f-bundle\") pod \"8f2f4ee801e5826a37d84a7b1fc4ccbf6b79de668302737d0f1152d8d2lfzpq\" (UID: \"a200c80e-a0f7-40ee-b697-acf6d429d14f\") " pod="openshift-marketplace/8f2f4ee801e5826a37d84a7b1fc4ccbf6b79de668302737d0f1152d8d2lfzpq" Oct 08 18:28:41 crc kubenswrapper[4859]: I1008 18:28:41.438070 4859 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/a200c80e-a0f7-40ee-b697-acf6d429d14f-util\") pod \"8f2f4ee801e5826a37d84a7b1fc4ccbf6b79de668302737d0f1152d8d2lfzpq\" (UID: \"a200c80e-a0f7-40ee-b697-acf6d429d14f\") " pod="openshift-marketplace/8f2f4ee801e5826a37d84a7b1fc4ccbf6b79de668302737d0f1152d8d2lfzpq" Oct 08 18:28:41 crc kubenswrapper[4859]: I1008 18:28:41.438146 4859 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-svb6t\" (UniqueName: \"kubernetes.io/projected/a200c80e-a0f7-40ee-b697-acf6d429d14f-kube-api-access-svb6t\") pod \"8f2f4ee801e5826a37d84a7b1fc4ccbf6b79de668302737d0f1152d8d2lfzpq\" (UID: \"a200c80e-a0f7-40ee-b697-acf6d429d14f\") " pod="openshift-marketplace/8f2f4ee801e5826a37d84a7b1fc4ccbf6b79de668302737d0f1152d8d2lfzpq" Oct 08 18:28:41 crc kubenswrapper[4859]: I1008 18:28:41.438192 4859 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/a200c80e-a0f7-40ee-b697-acf6d429d14f-bundle\") pod \"8f2f4ee801e5826a37d84a7b1fc4ccbf6b79de668302737d0f1152d8d2lfzpq\" (UID: \"a200c80e-a0f7-40ee-b697-acf6d429d14f\") " pod="openshift-marketplace/8f2f4ee801e5826a37d84a7b1fc4ccbf6b79de668302737d0f1152d8d2lfzpq" Oct 08 18:28:41 crc kubenswrapper[4859]: I1008 18:28:41.438805 4859 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/a200c80e-a0f7-40ee-b697-acf6d429d14f-bundle\") pod \"8f2f4ee801e5826a37d84a7b1fc4ccbf6b79de668302737d0f1152d8d2lfzpq\" (UID: \"a200c80e-a0f7-40ee-b697-acf6d429d14f\") " pod="openshift-marketplace/8f2f4ee801e5826a37d84a7b1fc4ccbf6b79de668302737d0f1152d8d2lfzpq" Oct 08 18:28:41 crc kubenswrapper[4859]: I1008 18:28:41.439029 4859 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/a200c80e-a0f7-40ee-b697-acf6d429d14f-util\") pod \"8f2f4ee801e5826a37d84a7b1fc4ccbf6b79de668302737d0f1152d8d2lfzpq\" (UID: \"a200c80e-a0f7-40ee-b697-acf6d429d14f\") " pod="openshift-marketplace/8f2f4ee801e5826a37d84a7b1fc4ccbf6b79de668302737d0f1152d8d2lfzpq" Oct 08 18:28:41 crc kubenswrapper[4859]: I1008 18:28:41.458841 4859 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-svb6t\" (UniqueName: \"kubernetes.io/projected/a200c80e-a0f7-40ee-b697-acf6d429d14f-kube-api-access-svb6t\") pod \"8f2f4ee801e5826a37d84a7b1fc4ccbf6b79de668302737d0f1152d8d2lfzpq\" (UID: \"a200c80e-a0f7-40ee-b697-acf6d429d14f\") " pod="openshift-marketplace/8f2f4ee801e5826a37d84a7b1fc4ccbf6b79de668302737d0f1152d8d2lfzpq" Oct 08 18:28:41 crc kubenswrapper[4859]: I1008 18:28:41.622976 4859 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/8f2f4ee801e5826a37d84a7b1fc4ccbf6b79de668302737d0f1152d8d2lfzpq" Oct 08 18:28:42 crc kubenswrapper[4859]: I1008 18:28:42.063266 4859 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/8f2f4ee801e5826a37d84a7b1fc4ccbf6b79de668302737d0f1152d8d2lfzpq"] Oct 08 18:28:42 crc kubenswrapper[4859]: I1008 18:28:42.707059 4859 generic.go:334] "Generic (PLEG): container finished" podID="a200c80e-a0f7-40ee-b697-acf6d429d14f" containerID="788bd9e00a7eadb83d788b133d0e83061914bfcf3c89f53e7d9842084c009f16" exitCode=0 Oct 08 18:28:42 crc kubenswrapper[4859]: I1008 18:28:42.707241 4859 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/8f2f4ee801e5826a37d84a7b1fc4ccbf6b79de668302737d0f1152d8d2lfzpq" event={"ID":"a200c80e-a0f7-40ee-b697-acf6d429d14f","Type":"ContainerDied","Data":"788bd9e00a7eadb83d788b133d0e83061914bfcf3c89f53e7d9842084c009f16"} Oct 08 18:28:42 crc kubenswrapper[4859]: I1008 18:28:42.708383 4859 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/8f2f4ee801e5826a37d84a7b1fc4ccbf6b79de668302737d0f1152d8d2lfzpq" event={"ID":"a200c80e-a0f7-40ee-b697-acf6d429d14f","Type":"ContainerStarted","Data":"d814eb7e8aa93b9337bd7d7f2a1d33fe8174aebc98549adbc27d65953d432ed4"} Oct 08 18:28:43 crc kubenswrapper[4859]: I1008 18:28:43.661054 4859 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-console/console-f9d7485db-m2xmp" podUID="9a829106-59b1-4389-95bf-2ee4d56e317c" containerName="console" containerID="cri-o://18ddf7cb370db54cf37e40e1f580d3e3a64ea5cacb148bb385217940676b8053" gracePeriod=15 Oct 08 18:28:44 crc kubenswrapper[4859]: I1008 18:28:44.110798 4859 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-console_console-f9d7485db-m2xmp_9a829106-59b1-4389-95bf-2ee4d56e317c/console/0.log" Oct 08 18:28:44 crc kubenswrapper[4859]: I1008 18:28:44.111106 4859 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-console/console-f9d7485db-m2xmp" Oct 08 18:28:44 crc kubenswrapper[4859]: I1008 18:28:44.176835 4859 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/9a829106-59b1-4389-95bf-2ee4d56e317c-trusted-ca-bundle\") pod \"9a829106-59b1-4389-95bf-2ee4d56e317c\" (UID: \"9a829106-59b1-4389-95bf-2ee4d56e317c\") " Oct 08 18:28:44 crc kubenswrapper[4859]: I1008 18:28:44.176910 4859 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/9a829106-59b1-4389-95bf-2ee4d56e317c-console-oauth-config\") pod \"9a829106-59b1-4389-95bf-2ee4d56e317c\" (UID: \"9a829106-59b1-4389-95bf-2ee4d56e317c\") " Oct 08 18:28:44 crc kubenswrapper[4859]: I1008 18:28:44.176953 4859 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-df4rr\" (UniqueName: \"kubernetes.io/projected/9a829106-59b1-4389-95bf-2ee4d56e317c-kube-api-access-df4rr\") pod \"9a829106-59b1-4389-95bf-2ee4d56e317c\" (UID: \"9a829106-59b1-4389-95bf-2ee4d56e317c\") " Oct 08 18:28:44 crc kubenswrapper[4859]: I1008 18:28:44.176988 4859 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/9a829106-59b1-4389-95bf-2ee4d56e317c-service-ca\") pod \"9a829106-59b1-4389-95bf-2ee4d56e317c\" (UID: \"9a829106-59b1-4389-95bf-2ee4d56e317c\") " Oct 08 18:28:44 crc kubenswrapper[4859]: I1008 18:28:44.177012 4859 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/9a829106-59b1-4389-95bf-2ee4d56e317c-console-config\") pod \"9a829106-59b1-4389-95bf-2ee4d56e317c\" (UID: \"9a829106-59b1-4389-95bf-2ee4d56e317c\") " Oct 08 18:28:44 crc kubenswrapper[4859]: I1008 18:28:44.177035 4859 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/9a829106-59b1-4389-95bf-2ee4d56e317c-console-serving-cert\") pod \"9a829106-59b1-4389-95bf-2ee4d56e317c\" (UID: \"9a829106-59b1-4389-95bf-2ee4d56e317c\") " Oct 08 18:28:44 crc kubenswrapper[4859]: I1008 18:28:44.177074 4859 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/9a829106-59b1-4389-95bf-2ee4d56e317c-oauth-serving-cert\") pod \"9a829106-59b1-4389-95bf-2ee4d56e317c\" (UID: \"9a829106-59b1-4389-95bf-2ee4d56e317c\") " Oct 08 18:28:44 crc kubenswrapper[4859]: I1008 18:28:44.178371 4859 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/9a829106-59b1-4389-95bf-2ee4d56e317c-service-ca" (OuterVolumeSpecName: "service-ca") pod "9a829106-59b1-4389-95bf-2ee4d56e317c" (UID: "9a829106-59b1-4389-95bf-2ee4d56e317c"). InnerVolumeSpecName "service-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 08 18:28:44 crc kubenswrapper[4859]: I1008 18:28:44.178578 4859 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/9a829106-59b1-4389-95bf-2ee4d56e317c-console-config" (OuterVolumeSpecName: "console-config") pod "9a829106-59b1-4389-95bf-2ee4d56e317c" (UID: "9a829106-59b1-4389-95bf-2ee4d56e317c"). InnerVolumeSpecName "console-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 08 18:28:44 crc kubenswrapper[4859]: I1008 18:28:44.178596 4859 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/9a829106-59b1-4389-95bf-2ee4d56e317c-oauth-serving-cert" (OuterVolumeSpecName: "oauth-serving-cert") pod "9a829106-59b1-4389-95bf-2ee4d56e317c" (UID: "9a829106-59b1-4389-95bf-2ee4d56e317c"). InnerVolumeSpecName "oauth-serving-cert". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 08 18:28:44 crc kubenswrapper[4859]: I1008 18:28:44.178980 4859 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/9a829106-59b1-4389-95bf-2ee4d56e317c-trusted-ca-bundle" (OuterVolumeSpecName: "trusted-ca-bundle") pod "9a829106-59b1-4389-95bf-2ee4d56e317c" (UID: "9a829106-59b1-4389-95bf-2ee4d56e317c"). InnerVolumeSpecName "trusted-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 08 18:28:44 crc kubenswrapper[4859]: I1008 18:28:44.183789 4859 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/9a829106-59b1-4389-95bf-2ee4d56e317c-console-oauth-config" (OuterVolumeSpecName: "console-oauth-config") pod "9a829106-59b1-4389-95bf-2ee4d56e317c" (UID: "9a829106-59b1-4389-95bf-2ee4d56e317c"). InnerVolumeSpecName "console-oauth-config". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 08 18:28:44 crc kubenswrapper[4859]: I1008 18:28:44.184704 4859 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/9a829106-59b1-4389-95bf-2ee4d56e317c-kube-api-access-df4rr" (OuterVolumeSpecName: "kube-api-access-df4rr") pod "9a829106-59b1-4389-95bf-2ee4d56e317c" (UID: "9a829106-59b1-4389-95bf-2ee4d56e317c"). InnerVolumeSpecName "kube-api-access-df4rr". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 08 18:28:44 crc kubenswrapper[4859]: I1008 18:28:44.185996 4859 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/9a829106-59b1-4389-95bf-2ee4d56e317c-console-serving-cert" (OuterVolumeSpecName: "console-serving-cert") pod "9a829106-59b1-4389-95bf-2ee4d56e317c" (UID: "9a829106-59b1-4389-95bf-2ee4d56e317c"). InnerVolumeSpecName "console-serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 08 18:28:44 crc kubenswrapper[4859]: I1008 18:28:44.278680 4859 reconciler_common.go:293] "Volume detached for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/9a829106-59b1-4389-95bf-2ee4d56e317c-oauth-serving-cert\") on node \"crc\" DevicePath \"\"" Oct 08 18:28:44 crc kubenswrapper[4859]: I1008 18:28:44.278753 4859 reconciler_common.go:293] "Volume detached for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/9a829106-59b1-4389-95bf-2ee4d56e317c-trusted-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 08 18:28:44 crc kubenswrapper[4859]: I1008 18:28:44.278765 4859 reconciler_common.go:293] "Volume detached for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/9a829106-59b1-4389-95bf-2ee4d56e317c-console-oauth-config\") on node \"crc\" DevicePath \"\"" Oct 08 18:28:44 crc kubenswrapper[4859]: I1008 18:28:44.278782 4859 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-df4rr\" (UniqueName: \"kubernetes.io/projected/9a829106-59b1-4389-95bf-2ee4d56e317c-kube-api-access-df4rr\") on node \"crc\" DevicePath \"\"" Oct 08 18:28:44 crc kubenswrapper[4859]: I1008 18:28:44.278798 4859 reconciler_common.go:293] "Volume detached for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/9a829106-59b1-4389-95bf-2ee4d56e317c-service-ca\") on node \"crc\" DevicePath \"\"" Oct 08 18:28:44 crc kubenswrapper[4859]: I1008 18:28:44.278809 4859 reconciler_common.go:293] "Volume detached for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/9a829106-59b1-4389-95bf-2ee4d56e317c-console-config\") on node \"crc\" DevicePath \"\"" Oct 08 18:28:44 crc kubenswrapper[4859]: I1008 18:28:44.278820 4859 reconciler_common.go:293] "Volume detached for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/9a829106-59b1-4389-95bf-2ee4d56e317c-console-serving-cert\") on node \"crc\" DevicePath \"\"" Oct 08 18:28:44 crc kubenswrapper[4859]: I1008 18:28:44.728769 4859 generic.go:334] "Generic (PLEG): container finished" podID="a200c80e-a0f7-40ee-b697-acf6d429d14f" containerID="1a28fc1248bfd179796ab0c70ce56d127cf943dd0d7e4e4d7d7c2d8f35cd4948" exitCode=0 Oct 08 18:28:44 crc kubenswrapper[4859]: I1008 18:28:44.728919 4859 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/8f2f4ee801e5826a37d84a7b1fc4ccbf6b79de668302737d0f1152d8d2lfzpq" event={"ID":"a200c80e-a0f7-40ee-b697-acf6d429d14f","Type":"ContainerDied","Data":"1a28fc1248bfd179796ab0c70ce56d127cf943dd0d7e4e4d7d7c2d8f35cd4948"} Oct 08 18:28:44 crc kubenswrapper[4859]: I1008 18:28:44.735886 4859 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-console_console-f9d7485db-m2xmp_9a829106-59b1-4389-95bf-2ee4d56e317c/console/0.log" Oct 08 18:28:44 crc kubenswrapper[4859]: I1008 18:28:44.735951 4859 generic.go:334] "Generic (PLEG): container finished" podID="9a829106-59b1-4389-95bf-2ee4d56e317c" containerID="18ddf7cb370db54cf37e40e1f580d3e3a64ea5cacb148bb385217940676b8053" exitCode=2 Oct 08 18:28:44 crc kubenswrapper[4859]: I1008 18:28:44.735991 4859 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console/console-f9d7485db-m2xmp" event={"ID":"9a829106-59b1-4389-95bf-2ee4d56e317c","Type":"ContainerDied","Data":"18ddf7cb370db54cf37e40e1f580d3e3a64ea5cacb148bb385217940676b8053"} Oct 08 18:28:44 crc kubenswrapper[4859]: I1008 18:28:44.736025 4859 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console/console-f9d7485db-m2xmp" event={"ID":"9a829106-59b1-4389-95bf-2ee4d56e317c","Type":"ContainerDied","Data":"d678c5240a9fe6aca79f30a7a28b1340b3c082b8b6cdeee98fbb95d197a14f0a"} Oct 08 18:28:44 crc kubenswrapper[4859]: I1008 18:28:44.736046 4859 scope.go:117] "RemoveContainer" containerID="18ddf7cb370db54cf37e40e1f580d3e3a64ea5cacb148bb385217940676b8053" Oct 08 18:28:44 crc kubenswrapper[4859]: I1008 18:28:44.736081 4859 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-console/console-f9d7485db-m2xmp" Oct 08 18:28:44 crc kubenswrapper[4859]: I1008 18:28:44.777909 4859 scope.go:117] "RemoveContainer" containerID="18ddf7cb370db54cf37e40e1f580d3e3a64ea5cacb148bb385217940676b8053" Oct 08 18:28:44 crc kubenswrapper[4859]: E1008 18:28:44.778656 4859 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"18ddf7cb370db54cf37e40e1f580d3e3a64ea5cacb148bb385217940676b8053\": container with ID starting with 18ddf7cb370db54cf37e40e1f580d3e3a64ea5cacb148bb385217940676b8053 not found: ID does not exist" containerID="18ddf7cb370db54cf37e40e1f580d3e3a64ea5cacb148bb385217940676b8053" Oct 08 18:28:44 crc kubenswrapper[4859]: I1008 18:28:44.778765 4859 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"18ddf7cb370db54cf37e40e1f580d3e3a64ea5cacb148bb385217940676b8053"} err="failed to get container status \"18ddf7cb370db54cf37e40e1f580d3e3a64ea5cacb148bb385217940676b8053\": rpc error: code = NotFound desc = could not find container \"18ddf7cb370db54cf37e40e1f580d3e3a64ea5cacb148bb385217940676b8053\": container with ID starting with 18ddf7cb370db54cf37e40e1f580d3e3a64ea5cacb148bb385217940676b8053 not found: ID does not exist" Oct 08 18:28:44 crc kubenswrapper[4859]: I1008 18:28:44.786183 4859 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-console/console-f9d7485db-m2xmp"] Oct 08 18:28:44 crc kubenswrapper[4859]: I1008 18:28:44.793267 4859 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-console/console-f9d7485db-m2xmp"] Oct 08 18:28:45 crc kubenswrapper[4859]: I1008 18:28:45.752982 4859 generic.go:334] "Generic (PLEG): container finished" podID="a200c80e-a0f7-40ee-b697-acf6d429d14f" containerID="92d931f9b1756d1e04c199d5322a22ae7c7e1ef1b908b3c59ed82143697c2e5d" exitCode=0 Oct 08 18:28:45 crc kubenswrapper[4859]: I1008 18:28:45.753301 4859 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/8f2f4ee801e5826a37d84a7b1fc4ccbf6b79de668302737d0f1152d8d2lfzpq" event={"ID":"a200c80e-a0f7-40ee-b697-acf6d429d14f","Type":"ContainerDied","Data":"92d931f9b1756d1e04c199d5322a22ae7c7e1ef1b908b3c59ed82143697c2e5d"} Oct 08 18:28:46 crc kubenswrapper[4859]: I1008 18:28:46.483015 4859 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="9a829106-59b1-4389-95bf-2ee4d56e317c" path="/var/lib/kubelet/pods/9a829106-59b1-4389-95bf-2ee4d56e317c/volumes" Oct 08 18:28:47 crc kubenswrapper[4859]: I1008 18:28:47.023904 4859 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/8f2f4ee801e5826a37d84a7b1fc4ccbf6b79de668302737d0f1152d8d2lfzpq" Oct 08 18:28:47 crc kubenswrapper[4859]: I1008 18:28:47.113069 4859 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-svb6t\" (UniqueName: \"kubernetes.io/projected/a200c80e-a0f7-40ee-b697-acf6d429d14f-kube-api-access-svb6t\") pod \"a200c80e-a0f7-40ee-b697-acf6d429d14f\" (UID: \"a200c80e-a0f7-40ee-b697-acf6d429d14f\") " Oct 08 18:28:47 crc kubenswrapper[4859]: I1008 18:28:47.113453 4859 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/a200c80e-a0f7-40ee-b697-acf6d429d14f-bundle\") pod \"a200c80e-a0f7-40ee-b697-acf6d429d14f\" (UID: \"a200c80e-a0f7-40ee-b697-acf6d429d14f\") " Oct 08 18:28:47 crc kubenswrapper[4859]: I1008 18:28:47.113602 4859 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/a200c80e-a0f7-40ee-b697-acf6d429d14f-util\") pod \"a200c80e-a0f7-40ee-b697-acf6d429d14f\" (UID: \"a200c80e-a0f7-40ee-b697-acf6d429d14f\") " Oct 08 18:28:47 crc kubenswrapper[4859]: I1008 18:28:47.114542 4859 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/a200c80e-a0f7-40ee-b697-acf6d429d14f-bundle" (OuterVolumeSpecName: "bundle") pod "a200c80e-a0f7-40ee-b697-acf6d429d14f" (UID: "a200c80e-a0f7-40ee-b697-acf6d429d14f"). InnerVolumeSpecName "bundle". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 08 18:28:47 crc kubenswrapper[4859]: I1008 18:28:47.120470 4859 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a200c80e-a0f7-40ee-b697-acf6d429d14f-kube-api-access-svb6t" (OuterVolumeSpecName: "kube-api-access-svb6t") pod "a200c80e-a0f7-40ee-b697-acf6d429d14f" (UID: "a200c80e-a0f7-40ee-b697-acf6d429d14f"). InnerVolumeSpecName "kube-api-access-svb6t". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 08 18:28:47 crc kubenswrapper[4859]: I1008 18:28:47.127091 4859 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/a200c80e-a0f7-40ee-b697-acf6d429d14f-util" (OuterVolumeSpecName: "util") pod "a200c80e-a0f7-40ee-b697-acf6d429d14f" (UID: "a200c80e-a0f7-40ee-b697-acf6d429d14f"). InnerVolumeSpecName "util". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 08 18:28:47 crc kubenswrapper[4859]: I1008 18:28:47.216457 4859 reconciler_common.go:293] "Volume detached for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/a200c80e-a0f7-40ee-b697-acf6d429d14f-util\") on node \"crc\" DevicePath \"\"" Oct 08 18:28:47 crc kubenswrapper[4859]: I1008 18:28:47.216514 4859 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-svb6t\" (UniqueName: \"kubernetes.io/projected/a200c80e-a0f7-40ee-b697-acf6d429d14f-kube-api-access-svb6t\") on node \"crc\" DevicePath \"\"" Oct 08 18:28:47 crc kubenswrapper[4859]: I1008 18:28:47.216534 4859 reconciler_common.go:293] "Volume detached for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/a200c80e-a0f7-40ee-b697-acf6d429d14f-bundle\") on node \"crc\" DevicePath \"\"" Oct 08 18:28:47 crc kubenswrapper[4859]: I1008 18:28:47.769607 4859 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/8f2f4ee801e5826a37d84a7b1fc4ccbf6b79de668302737d0f1152d8d2lfzpq" event={"ID":"a200c80e-a0f7-40ee-b697-acf6d429d14f","Type":"ContainerDied","Data":"d814eb7e8aa93b9337bd7d7f2a1d33fe8174aebc98549adbc27d65953d432ed4"} Oct 08 18:28:47 crc kubenswrapper[4859]: I1008 18:28:47.769659 4859 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="d814eb7e8aa93b9337bd7d7f2a1d33fe8174aebc98549adbc27d65953d432ed4" Oct 08 18:28:47 crc kubenswrapper[4859]: I1008 18:28:47.769658 4859 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/8f2f4ee801e5826a37d84a7b1fc4ccbf6b79de668302737d0f1152d8d2lfzpq" Oct 08 18:28:56 crc kubenswrapper[4859]: I1008 18:28:56.456935 4859 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["metallb-system/metallb-operator-controller-manager-6789c5bc7c-k7fl4"] Oct 08 18:28:56 crc kubenswrapper[4859]: E1008 18:28:56.457868 4859 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a200c80e-a0f7-40ee-b697-acf6d429d14f" containerName="pull" Oct 08 18:28:56 crc kubenswrapper[4859]: I1008 18:28:56.457888 4859 state_mem.go:107] "Deleted CPUSet assignment" podUID="a200c80e-a0f7-40ee-b697-acf6d429d14f" containerName="pull" Oct 08 18:28:56 crc kubenswrapper[4859]: E1008 18:28:56.457902 4859 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a200c80e-a0f7-40ee-b697-acf6d429d14f" containerName="util" Oct 08 18:28:56 crc kubenswrapper[4859]: I1008 18:28:56.457909 4859 state_mem.go:107] "Deleted CPUSet assignment" podUID="a200c80e-a0f7-40ee-b697-acf6d429d14f" containerName="util" Oct 08 18:28:56 crc kubenswrapper[4859]: E1008 18:28:56.457919 4859 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a200c80e-a0f7-40ee-b697-acf6d429d14f" containerName="extract" Oct 08 18:28:56 crc kubenswrapper[4859]: I1008 18:28:56.457929 4859 state_mem.go:107] "Deleted CPUSet assignment" podUID="a200c80e-a0f7-40ee-b697-acf6d429d14f" containerName="extract" Oct 08 18:28:56 crc kubenswrapper[4859]: E1008 18:28:56.457949 4859 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9a829106-59b1-4389-95bf-2ee4d56e317c" containerName="console" Oct 08 18:28:56 crc kubenswrapper[4859]: I1008 18:28:56.457956 4859 state_mem.go:107] "Deleted CPUSet assignment" podUID="9a829106-59b1-4389-95bf-2ee4d56e317c" containerName="console" Oct 08 18:28:56 crc kubenswrapper[4859]: I1008 18:28:56.458092 4859 memory_manager.go:354] "RemoveStaleState removing state" podUID="9a829106-59b1-4389-95bf-2ee4d56e317c" containerName="console" Oct 08 18:28:56 crc kubenswrapper[4859]: I1008 18:28:56.458107 4859 memory_manager.go:354] "RemoveStaleState removing state" podUID="a200c80e-a0f7-40ee-b697-acf6d429d14f" containerName="extract" Oct 08 18:28:56 crc kubenswrapper[4859]: I1008 18:28:56.458616 4859 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/metallb-operator-controller-manager-6789c5bc7c-k7fl4" Oct 08 18:28:56 crc kubenswrapper[4859]: W1008 18:28:56.461117 4859 reflector.go:561] object-"metallb-system"/"metallb-operator-controller-manager-service-cert": failed to list *v1.Secret: secrets "metallb-operator-controller-manager-service-cert" is forbidden: User "system:node:crc" cannot list resource "secrets" in API group "" in the namespace "metallb-system": no relationship found between node 'crc' and this object Oct 08 18:28:56 crc kubenswrapper[4859]: E1008 18:28:56.461197 4859 reflector.go:158] "Unhandled Error" err="object-\"metallb-system\"/\"metallb-operator-controller-manager-service-cert\": Failed to watch *v1.Secret: failed to list *v1.Secret: secrets \"metallb-operator-controller-manager-service-cert\" is forbidden: User \"system:node:crc\" cannot list resource \"secrets\" in API group \"\" in the namespace \"metallb-system\": no relationship found between node 'crc' and this object" logger="UnhandledError" Oct 08 18:28:56 crc kubenswrapper[4859]: W1008 18:28:56.461300 4859 reflector.go:561] object-"metallb-system"/"metallb-operator-webhook-server-cert": failed to list *v1.Secret: secrets "metallb-operator-webhook-server-cert" is forbidden: User "system:node:crc" cannot list resource "secrets" in API group "" in the namespace "metallb-system": no relationship found between node 'crc' and this object Oct 08 18:28:56 crc kubenswrapper[4859]: E1008 18:28:56.461361 4859 reflector.go:158] "Unhandled Error" err="object-\"metallb-system\"/\"metallb-operator-webhook-server-cert\": Failed to watch *v1.Secret: failed to list *v1.Secret: secrets \"metallb-operator-webhook-server-cert\" is forbidden: User \"system:node:crc\" cannot list resource \"secrets\" in API group \"\" in the namespace \"metallb-system\": no relationship found between node 'crc' and this object" logger="UnhandledError" Oct 08 18:28:56 crc kubenswrapper[4859]: W1008 18:28:56.461367 4859 reflector.go:561] object-"metallb-system"/"kube-root-ca.crt": failed to list *v1.ConfigMap: configmaps "kube-root-ca.crt" is forbidden: User "system:node:crc" cannot list resource "configmaps" in API group "" in the namespace "metallb-system": no relationship found between node 'crc' and this object Oct 08 18:28:56 crc kubenswrapper[4859]: E1008 18:28:56.461419 4859 reflector.go:158] "Unhandled Error" err="object-\"metallb-system\"/\"kube-root-ca.crt\": Failed to watch *v1.ConfigMap: failed to list *v1.ConfigMap: configmaps \"kube-root-ca.crt\" is forbidden: User \"system:node:crc\" cannot list resource \"configmaps\" in API group \"\" in the namespace \"metallb-system\": no relationship found between node 'crc' and this object" logger="UnhandledError" Oct 08 18:28:56 crc kubenswrapper[4859]: I1008 18:28:56.461376 4859 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"manager-account-dockercfg-chb8m" Oct 08 18:28:56 crc kubenswrapper[4859]: I1008 18:28:56.461606 4859 reflector.go:368] Caches populated for *v1.ConfigMap from object-"metallb-system"/"openshift-service-ca.crt" Oct 08 18:28:56 crc kubenswrapper[4859]: I1008 18:28:56.477315 4859 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["metallb-system/metallb-operator-controller-manager-6789c5bc7c-k7fl4"] Oct 08 18:28:56 crc kubenswrapper[4859]: I1008 18:28:56.537007 4859 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-jczfj\" (UniqueName: \"kubernetes.io/projected/47c9b7b8-561d-489b-a5d2-34ef1f6f4be3-kube-api-access-jczfj\") pod \"metallb-operator-controller-manager-6789c5bc7c-k7fl4\" (UID: \"47c9b7b8-561d-489b-a5d2-34ef1f6f4be3\") " pod="metallb-system/metallb-operator-controller-manager-6789c5bc7c-k7fl4" Oct 08 18:28:56 crc kubenswrapper[4859]: I1008 18:28:56.537630 4859 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/47c9b7b8-561d-489b-a5d2-34ef1f6f4be3-webhook-cert\") pod \"metallb-operator-controller-manager-6789c5bc7c-k7fl4\" (UID: \"47c9b7b8-561d-489b-a5d2-34ef1f6f4be3\") " pod="metallb-system/metallb-operator-controller-manager-6789c5bc7c-k7fl4" Oct 08 18:28:56 crc kubenswrapper[4859]: I1008 18:28:56.537786 4859 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/47c9b7b8-561d-489b-a5d2-34ef1f6f4be3-apiservice-cert\") pod \"metallb-operator-controller-manager-6789c5bc7c-k7fl4\" (UID: \"47c9b7b8-561d-489b-a5d2-34ef1f6f4be3\") " pod="metallb-system/metallb-operator-controller-manager-6789c5bc7c-k7fl4" Oct 08 18:28:56 crc kubenswrapper[4859]: I1008 18:28:56.639046 4859 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/47c9b7b8-561d-489b-a5d2-34ef1f6f4be3-webhook-cert\") pod \"metallb-operator-controller-manager-6789c5bc7c-k7fl4\" (UID: \"47c9b7b8-561d-489b-a5d2-34ef1f6f4be3\") " pod="metallb-system/metallb-operator-controller-manager-6789c5bc7c-k7fl4" Oct 08 18:28:56 crc kubenswrapper[4859]: I1008 18:28:56.639093 4859 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/47c9b7b8-561d-489b-a5d2-34ef1f6f4be3-apiservice-cert\") pod \"metallb-operator-controller-manager-6789c5bc7c-k7fl4\" (UID: \"47c9b7b8-561d-489b-a5d2-34ef1f6f4be3\") " pod="metallb-system/metallb-operator-controller-manager-6789c5bc7c-k7fl4" Oct 08 18:28:56 crc kubenswrapper[4859]: I1008 18:28:56.639130 4859 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-jczfj\" (UniqueName: \"kubernetes.io/projected/47c9b7b8-561d-489b-a5d2-34ef1f6f4be3-kube-api-access-jczfj\") pod \"metallb-operator-controller-manager-6789c5bc7c-k7fl4\" (UID: \"47c9b7b8-561d-489b-a5d2-34ef1f6f4be3\") " pod="metallb-system/metallb-operator-controller-manager-6789c5bc7c-k7fl4" Oct 08 18:28:56 crc kubenswrapper[4859]: I1008 18:28:56.811636 4859 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["metallb-system/metallb-operator-webhook-server-c4d794b9c-6lnpl"] Oct 08 18:28:56 crc kubenswrapper[4859]: I1008 18:28:56.812508 4859 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/metallb-operator-webhook-server-c4d794b9c-6lnpl" Oct 08 18:28:56 crc kubenswrapper[4859]: I1008 18:28:56.818001 4859 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"metallb-webhook-cert" Oct 08 18:28:56 crc kubenswrapper[4859]: I1008 18:28:56.818435 4859 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"controller-dockercfg-fxq77" Oct 08 18:28:56 crc kubenswrapper[4859]: I1008 18:28:56.822240 4859 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"metallb-operator-webhook-server-service-cert" Oct 08 18:28:56 crc kubenswrapper[4859]: I1008 18:28:56.841623 4859 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["metallb-system/metallb-operator-webhook-server-c4d794b9c-6lnpl"] Oct 08 18:28:56 crc kubenswrapper[4859]: I1008 18:28:56.842091 4859 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/c5192e5b-dc81-4df1-8ba2-f103a6b869eb-apiservice-cert\") pod \"metallb-operator-webhook-server-c4d794b9c-6lnpl\" (UID: \"c5192e5b-dc81-4df1-8ba2-f103a6b869eb\") " pod="metallb-system/metallb-operator-webhook-server-c4d794b9c-6lnpl" Oct 08 18:28:56 crc kubenswrapper[4859]: I1008 18:28:56.842158 4859 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-hhhmz\" (UniqueName: \"kubernetes.io/projected/c5192e5b-dc81-4df1-8ba2-f103a6b869eb-kube-api-access-hhhmz\") pod \"metallb-operator-webhook-server-c4d794b9c-6lnpl\" (UID: \"c5192e5b-dc81-4df1-8ba2-f103a6b869eb\") " pod="metallb-system/metallb-operator-webhook-server-c4d794b9c-6lnpl" Oct 08 18:28:56 crc kubenswrapper[4859]: I1008 18:28:56.842238 4859 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/c5192e5b-dc81-4df1-8ba2-f103a6b869eb-webhook-cert\") pod \"metallb-operator-webhook-server-c4d794b9c-6lnpl\" (UID: \"c5192e5b-dc81-4df1-8ba2-f103a6b869eb\") " pod="metallb-system/metallb-operator-webhook-server-c4d794b9c-6lnpl" Oct 08 18:28:56 crc kubenswrapper[4859]: I1008 18:28:56.943307 4859 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/c5192e5b-dc81-4df1-8ba2-f103a6b869eb-apiservice-cert\") pod \"metallb-operator-webhook-server-c4d794b9c-6lnpl\" (UID: \"c5192e5b-dc81-4df1-8ba2-f103a6b869eb\") " pod="metallb-system/metallb-operator-webhook-server-c4d794b9c-6lnpl" Oct 08 18:28:56 crc kubenswrapper[4859]: I1008 18:28:56.943384 4859 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-hhhmz\" (UniqueName: \"kubernetes.io/projected/c5192e5b-dc81-4df1-8ba2-f103a6b869eb-kube-api-access-hhhmz\") pod \"metallb-operator-webhook-server-c4d794b9c-6lnpl\" (UID: \"c5192e5b-dc81-4df1-8ba2-f103a6b869eb\") " pod="metallb-system/metallb-operator-webhook-server-c4d794b9c-6lnpl" Oct 08 18:28:56 crc kubenswrapper[4859]: I1008 18:28:56.943475 4859 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/c5192e5b-dc81-4df1-8ba2-f103a6b869eb-webhook-cert\") pod \"metallb-operator-webhook-server-c4d794b9c-6lnpl\" (UID: \"c5192e5b-dc81-4df1-8ba2-f103a6b869eb\") " pod="metallb-system/metallb-operator-webhook-server-c4d794b9c-6lnpl" Oct 08 18:28:56 crc kubenswrapper[4859]: I1008 18:28:56.950230 4859 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/c5192e5b-dc81-4df1-8ba2-f103a6b869eb-apiservice-cert\") pod \"metallb-operator-webhook-server-c4d794b9c-6lnpl\" (UID: \"c5192e5b-dc81-4df1-8ba2-f103a6b869eb\") " pod="metallb-system/metallb-operator-webhook-server-c4d794b9c-6lnpl" Oct 08 18:28:56 crc kubenswrapper[4859]: I1008 18:28:56.952923 4859 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/c5192e5b-dc81-4df1-8ba2-f103a6b869eb-webhook-cert\") pod \"metallb-operator-webhook-server-c4d794b9c-6lnpl\" (UID: \"c5192e5b-dc81-4df1-8ba2-f103a6b869eb\") " pod="metallb-system/metallb-operator-webhook-server-c4d794b9c-6lnpl" Oct 08 18:28:57 crc kubenswrapper[4859]: I1008 18:28:57.482543 4859 reflector.go:368] Caches populated for *v1.ConfigMap from object-"metallb-system"/"kube-root-ca.crt" Oct 08 18:28:57 crc kubenswrapper[4859]: I1008 18:28:57.492171 4859 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-jczfj\" (UniqueName: \"kubernetes.io/projected/47c9b7b8-561d-489b-a5d2-34ef1f6f4be3-kube-api-access-jczfj\") pod \"metallb-operator-controller-manager-6789c5bc7c-k7fl4\" (UID: \"47c9b7b8-561d-489b-a5d2-34ef1f6f4be3\") " pod="metallb-system/metallb-operator-controller-manager-6789c5bc7c-k7fl4" Oct 08 18:28:57 crc kubenswrapper[4859]: I1008 18:28:57.493129 4859 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-hhhmz\" (UniqueName: \"kubernetes.io/projected/c5192e5b-dc81-4df1-8ba2-f103a6b869eb-kube-api-access-hhhmz\") pod \"metallb-operator-webhook-server-c4d794b9c-6lnpl\" (UID: \"c5192e5b-dc81-4df1-8ba2-f103a6b869eb\") " pod="metallb-system/metallb-operator-webhook-server-c4d794b9c-6lnpl" Oct 08 18:28:57 crc kubenswrapper[4859]: I1008 18:28:57.526475 4859 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"metallb-operator-webhook-server-cert" Oct 08 18:28:57 crc kubenswrapper[4859]: E1008 18:28:57.639794 4859 secret.go:188] Couldn't get secret metallb-system/metallb-operator-controller-manager-service-cert: failed to sync secret cache: timed out waiting for the condition Oct 08 18:28:57 crc kubenswrapper[4859]: E1008 18:28:57.640084 4859 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/47c9b7b8-561d-489b-a5d2-34ef1f6f4be3-webhook-cert podName:47c9b7b8-561d-489b-a5d2-34ef1f6f4be3 nodeName:}" failed. No retries permitted until 2025-10-08 18:28:58.140063032 +0000 UTC m=+708.386902411 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "webhook-cert" (UniqueName: "kubernetes.io/secret/47c9b7b8-561d-489b-a5d2-34ef1f6f4be3-webhook-cert") pod "metallb-operator-controller-manager-6789c5bc7c-k7fl4" (UID: "47c9b7b8-561d-489b-a5d2-34ef1f6f4be3") : failed to sync secret cache: timed out waiting for the condition Oct 08 18:28:57 crc kubenswrapper[4859]: E1008 18:28:57.639793 4859 secret.go:188] Couldn't get secret metallb-system/metallb-operator-controller-manager-service-cert: failed to sync secret cache: timed out waiting for the condition Oct 08 18:28:57 crc kubenswrapper[4859]: E1008 18:28:57.640270 4859 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/47c9b7b8-561d-489b-a5d2-34ef1f6f4be3-apiservice-cert podName:47c9b7b8-561d-489b-a5d2-34ef1f6f4be3 nodeName:}" failed. No retries permitted until 2025-10-08 18:28:58.140261758 +0000 UTC m=+708.387101137 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "apiservice-cert" (UniqueName: "kubernetes.io/secret/47c9b7b8-561d-489b-a5d2-34ef1f6f4be3-apiservice-cert") pod "metallb-operator-controller-manager-6789c5bc7c-k7fl4" (UID: "47c9b7b8-561d-489b-a5d2-34ef1f6f4be3") : failed to sync secret cache: timed out waiting for the condition Oct 08 18:28:57 crc kubenswrapper[4859]: I1008 18:28:57.726036 4859 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/metallb-operator-webhook-server-c4d794b9c-6lnpl" Oct 08 18:28:57 crc kubenswrapper[4859]: I1008 18:28:57.751494 4859 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"metallb-operator-controller-manager-service-cert" Oct 08 18:28:58 crc kubenswrapper[4859]: I1008 18:28:58.146750 4859 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["metallb-system/metallb-operator-webhook-server-c4d794b9c-6lnpl"] Oct 08 18:28:58 crc kubenswrapper[4859]: I1008 18:28:58.160633 4859 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/47c9b7b8-561d-489b-a5d2-34ef1f6f4be3-webhook-cert\") pod \"metallb-operator-controller-manager-6789c5bc7c-k7fl4\" (UID: \"47c9b7b8-561d-489b-a5d2-34ef1f6f4be3\") " pod="metallb-system/metallb-operator-controller-manager-6789c5bc7c-k7fl4" Oct 08 18:28:58 crc kubenswrapper[4859]: I1008 18:28:58.160671 4859 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/47c9b7b8-561d-489b-a5d2-34ef1f6f4be3-apiservice-cert\") pod \"metallb-operator-controller-manager-6789c5bc7c-k7fl4\" (UID: \"47c9b7b8-561d-489b-a5d2-34ef1f6f4be3\") " pod="metallb-system/metallb-operator-controller-manager-6789c5bc7c-k7fl4" Oct 08 18:28:58 crc kubenswrapper[4859]: I1008 18:28:58.165975 4859 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/47c9b7b8-561d-489b-a5d2-34ef1f6f4be3-webhook-cert\") pod \"metallb-operator-controller-manager-6789c5bc7c-k7fl4\" (UID: \"47c9b7b8-561d-489b-a5d2-34ef1f6f4be3\") " pod="metallb-system/metallb-operator-controller-manager-6789c5bc7c-k7fl4" Oct 08 18:28:58 crc kubenswrapper[4859]: I1008 18:28:58.166372 4859 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/47c9b7b8-561d-489b-a5d2-34ef1f6f4be3-apiservice-cert\") pod \"metallb-operator-controller-manager-6789c5bc7c-k7fl4\" (UID: \"47c9b7b8-561d-489b-a5d2-34ef1f6f4be3\") " pod="metallb-system/metallb-operator-controller-manager-6789c5bc7c-k7fl4" Oct 08 18:28:58 crc kubenswrapper[4859]: I1008 18:28:58.276569 4859 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/metallb-operator-controller-manager-6789c5bc7c-k7fl4" Oct 08 18:28:58 crc kubenswrapper[4859]: I1008 18:28:58.501563 4859 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["metallb-system/metallb-operator-controller-manager-6789c5bc7c-k7fl4"] Oct 08 18:28:58 crc kubenswrapper[4859]: W1008 18:28:58.506679 4859 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod47c9b7b8_561d_489b_a5d2_34ef1f6f4be3.slice/crio-4c14912ae02386e9c881733784d3dae1b9c074ef21b8eb437e04a8b347f73d84 WatchSource:0}: Error finding container 4c14912ae02386e9c881733784d3dae1b9c074ef21b8eb437e04a8b347f73d84: Status 404 returned error can't find the container with id 4c14912ae02386e9c881733784d3dae1b9c074ef21b8eb437e04a8b347f73d84 Oct 08 18:28:58 crc kubenswrapper[4859]: I1008 18:28:58.841441 4859 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/metallb-operator-controller-manager-6789c5bc7c-k7fl4" event={"ID":"47c9b7b8-561d-489b-a5d2-34ef1f6f4be3","Type":"ContainerStarted","Data":"4c14912ae02386e9c881733784d3dae1b9c074ef21b8eb437e04a8b347f73d84"} Oct 08 18:28:58 crc kubenswrapper[4859]: I1008 18:28:58.842762 4859 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/metallb-operator-webhook-server-c4d794b9c-6lnpl" event={"ID":"c5192e5b-dc81-4df1-8ba2-f103a6b869eb","Type":"ContainerStarted","Data":"7c0131c6e8e6f8b9abe5ec097791bb9867763c29bbe87c13b5d900ffe56fe138"} Oct 08 18:29:03 crc kubenswrapper[4859]: I1008 18:29:03.888463 4859 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/metallb-operator-controller-manager-6789c5bc7c-k7fl4" event={"ID":"47c9b7b8-561d-489b-a5d2-34ef1f6f4be3","Type":"ContainerStarted","Data":"1cb22b1edc710dc6111a5ff358814e27afb8cf8b79ff2b74eb94e069abef6902"} Oct 08 18:29:03 crc kubenswrapper[4859]: I1008 18:29:03.889151 4859 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="metallb-system/metallb-operator-controller-manager-6789c5bc7c-k7fl4" Oct 08 18:29:03 crc kubenswrapper[4859]: I1008 18:29:03.891067 4859 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/metallb-operator-webhook-server-c4d794b9c-6lnpl" event={"ID":"c5192e5b-dc81-4df1-8ba2-f103a6b869eb","Type":"ContainerStarted","Data":"53d29a5e35d32121d23a25efdf28439dd534201482fae3ceb523473875079210"} Oct 08 18:29:03 crc kubenswrapper[4859]: I1008 18:29:03.891215 4859 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="metallb-system/metallb-operator-webhook-server-c4d794b9c-6lnpl" Oct 08 18:29:03 crc kubenswrapper[4859]: I1008 18:29:03.912521 4859 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="metallb-system/metallb-operator-controller-manager-6789c5bc7c-k7fl4" podStartSLOduration=2.808520575 podStartE2EDuration="7.912501039s" podCreationTimestamp="2025-10-08 18:28:56 +0000 UTC" firstStartedPulling="2025-10-08 18:28:58.509769887 +0000 UTC m=+708.756609266" lastFinishedPulling="2025-10-08 18:29:03.613750351 +0000 UTC m=+713.860589730" observedRunningTime="2025-10-08 18:29:03.910981074 +0000 UTC m=+714.157820463" watchObservedRunningTime="2025-10-08 18:29:03.912501039 +0000 UTC m=+714.159340418" Oct 08 18:29:03 crc kubenswrapper[4859]: I1008 18:29:03.934511 4859 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="metallb-system/metallb-operator-webhook-server-c4d794b9c-6lnpl" podStartSLOduration=2.486824132 podStartE2EDuration="7.934491715s" podCreationTimestamp="2025-10-08 18:28:56 +0000 UTC" firstStartedPulling="2025-10-08 18:28:58.159840251 +0000 UTC m=+708.406679640" lastFinishedPulling="2025-10-08 18:29:03.607507854 +0000 UTC m=+713.854347223" observedRunningTime="2025-10-08 18:29:03.931471665 +0000 UTC m=+714.178311064" watchObservedRunningTime="2025-10-08 18:29:03.934491715 +0000 UTC m=+714.181331094" Oct 08 18:29:17 crc kubenswrapper[4859]: I1008 18:29:17.732982 4859 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="metallb-system/metallb-operator-webhook-server-c4d794b9c-6lnpl" Oct 08 18:29:37 crc kubenswrapper[4859]: I1008 18:29:37.320381 4859 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-controller-manager/controller-manager-879f6c89f-9gpzj"] Oct 08 18:29:37 crc kubenswrapper[4859]: I1008 18:29:37.321196 4859 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-controller-manager/controller-manager-879f6c89f-9gpzj" podUID="460ad6e7-859d-4111-a555-b55c5b7228e0" containerName="controller-manager" containerID="cri-o://46ee8cb842f339e6b4ba065cab54a4e76c9bac684c499161345db073e276bd89" gracePeriod=30 Oct 08 18:29:37 crc kubenswrapper[4859]: I1008 18:29:37.473832 4859 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-6576b87f9c-xlfdd"] Oct 08 18:29:37 crc kubenswrapper[4859]: I1008 18:29:37.474353 4859 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-xlfdd" podUID="ca64d874-5d86-4c49-8953-631a75d59788" containerName="route-controller-manager" containerID="cri-o://44d292882e24bdc475f139a3b68beb09c235ab83c773cff48466e5d53d3b8a19" gracePeriod=30 Oct 08 18:29:37 crc kubenswrapper[4859]: I1008 18:29:37.803111 4859 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-879f6c89f-9gpzj" Oct 08 18:29:37 crc kubenswrapper[4859]: I1008 18:29:37.855147 4859 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-xlfdd" Oct 08 18:29:37 crc kubenswrapper[4859]: I1008 18:29:37.907270 4859 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/460ad6e7-859d-4111-a555-b55c5b7228e0-config\") pod \"460ad6e7-859d-4111-a555-b55c5b7228e0\" (UID: \"460ad6e7-859d-4111-a555-b55c5b7228e0\") " Oct 08 18:29:37 crc kubenswrapper[4859]: I1008 18:29:37.907347 4859 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/ca64d874-5d86-4c49-8953-631a75d59788-serving-cert\") pod \"ca64d874-5d86-4c49-8953-631a75d59788\" (UID: \"ca64d874-5d86-4c49-8953-631a75d59788\") " Oct 08 18:29:37 crc kubenswrapper[4859]: I1008 18:29:37.907380 4859 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-gbpww\" (UniqueName: \"kubernetes.io/projected/460ad6e7-859d-4111-a555-b55c5b7228e0-kube-api-access-gbpww\") pod \"460ad6e7-859d-4111-a555-b55c5b7228e0\" (UID: \"460ad6e7-859d-4111-a555-b55c5b7228e0\") " Oct 08 18:29:37 crc kubenswrapper[4859]: I1008 18:29:37.907407 4859 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-9hnbb\" (UniqueName: \"kubernetes.io/projected/ca64d874-5d86-4c49-8953-631a75d59788-kube-api-access-9hnbb\") pod \"ca64d874-5d86-4c49-8953-631a75d59788\" (UID: \"ca64d874-5d86-4c49-8953-631a75d59788\") " Oct 08 18:29:37 crc kubenswrapper[4859]: I1008 18:29:37.907474 4859 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/460ad6e7-859d-4111-a555-b55c5b7228e0-proxy-ca-bundles\") pod \"460ad6e7-859d-4111-a555-b55c5b7228e0\" (UID: \"460ad6e7-859d-4111-a555-b55c5b7228e0\") " Oct 08 18:29:37 crc kubenswrapper[4859]: I1008 18:29:37.907527 4859 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/ca64d874-5d86-4c49-8953-631a75d59788-client-ca\") pod \"ca64d874-5d86-4c49-8953-631a75d59788\" (UID: \"ca64d874-5d86-4c49-8953-631a75d59788\") " Oct 08 18:29:37 crc kubenswrapper[4859]: I1008 18:29:37.907583 4859 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/ca64d874-5d86-4c49-8953-631a75d59788-config\") pod \"ca64d874-5d86-4c49-8953-631a75d59788\" (UID: \"ca64d874-5d86-4c49-8953-631a75d59788\") " Oct 08 18:29:37 crc kubenswrapper[4859]: I1008 18:29:37.907603 4859 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/460ad6e7-859d-4111-a555-b55c5b7228e0-serving-cert\") pod \"460ad6e7-859d-4111-a555-b55c5b7228e0\" (UID: \"460ad6e7-859d-4111-a555-b55c5b7228e0\") " Oct 08 18:29:37 crc kubenswrapper[4859]: I1008 18:29:37.907653 4859 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/460ad6e7-859d-4111-a555-b55c5b7228e0-client-ca\") pod \"460ad6e7-859d-4111-a555-b55c5b7228e0\" (UID: \"460ad6e7-859d-4111-a555-b55c5b7228e0\") " Oct 08 18:29:37 crc kubenswrapper[4859]: I1008 18:29:37.908456 4859 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/460ad6e7-859d-4111-a555-b55c5b7228e0-client-ca" (OuterVolumeSpecName: "client-ca") pod "460ad6e7-859d-4111-a555-b55c5b7228e0" (UID: "460ad6e7-859d-4111-a555-b55c5b7228e0"). InnerVolumeSpecName "client-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 08 18:29:37 crc kubenswrapper[4859]: I1008 18:29:37.908445 4859 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/460ad6e7-859d-4111-a555-b55c5b7228e0-proxy-ca-bundles" (OuterVolumeSpecName: "proxy-ca-bundles") pod "460ad6e7-859d-4111-a555-b55c5b7228e0" (UID: "460ad6e7-859d-4111-a555-b55c5b7228e0"). InnerVolumeSpecName "proxy-ca-bundles". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 08 18:29:37 crc kubenswrapper[4859]: I1008 18:29:37.908798 4859 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/ca64d874-5d86-4c49-8953-631a75d59788-client-ca" (OuterVolumeSpecName: "client-ca") pod "ca64d874-5d86-4c49-8953-631a75d59788" (UID: "ca64d874-5d86-4c49-8953-631a75d59788"). InnerVolumeSpecName "client-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 08 18:29:37 crc kubenswrapper[4859]: I1008 18:29:37.909165 4859 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/460ad6e7-859d-4111-a555-b55c5b7228e0-config" (OuterVolumeSpecName: "config") pod "460ad6e7-859d-4111-a555-b55c5b7228e0" (UID: "460ad6e7-859d-4111-a555-b55c5b7228e0"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 08 18:29:37 crc kubenswrapper[4859]: I1008 18:29:37.909516 4859 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/ca64d874-5d86-4c49-8953-631a75d59788-config" (OuterVolumeSpecName: "config") pod "ca64d874-5d86-4c49-8953-631a75d59788" (UID: "ca64d874-5d86-4c49-8953-631a75d59788"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 08 18:29:37 crc kubenswrapper[4859]: I1008 18:29:37.916845 4859 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/460ad6e7-859d-4111-a555-b55c5b7228e0-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "460ad6e7-859d-4111-a555-b55c5b7228e0" (UID: "460ad6e7-859d-4111-a555-b55c5b7228e0"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 08 18:29:37 crc kubenswrapper[4859]: I1008 18:29:37.917015 4859 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/ca64d874-5d86-4c49-8953-631a75d59788-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "ca64d874-5d86-4c49-8953-631a75d59788" (UID: "ca64d874-5d86-4c49-8953-631a75d59788"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 08 18:29:37 crc kubenswrapper[4859]: I1008 18:29:37.917582 4859 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/460ad6e7-859d-4111-a555-b55c5b7228e0-kube-api-access-gbpww" (OuterVolumeSpecName: "kube-api-access-gbpww") pod "460ad6e7-859d-4111-a555-b55c5b7228e0" (UID: "460ad6e7-859d-4111-a555-b55c5b7228e0"). InnerVolumeSpecName "kube-api-access-gbpww". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 08 18:29:37 crc kubenswrapper[4859]: I1008 18:29:37.920130 4859 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/ca64d874-5d86-4c49-8953-631a75d59788-kube-api-access-9hnbb" (OuterVolumeSpecName: "kube-api-access-9hnbb") pod "ca64d874-5d86-4c49-8953-631a75d59788" (UID: "ca64d874-5d86-4c49-8953-631a75d59788"). InnerVolumeSpecName "kube-api-access-9hnbb". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 08 18:29:38 crc kubenswrapper[4859]: I1008 18:29:38.009448 4859 reconciler_common.go:293] "Volume detached for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/460ad6e7-859d-4111-a555-b55c5b7228e0-proxy-ca-bundles\") on node \"crc\" DevicePath \"\"" Oct 08 18:29:38 crc kubenswrapper[4859]: I1008 18:29:38.009487 4859 reconciler_common.go:293] "Volume detached for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/ca64d874-5d86-4c49-8953-631a75d59788-client-ca\") on node \"crc\" DevicePath \"\"" Oct 08 18:29:38 crc kubenswrapper[4859]: I1008 18:29:38.009497 4859 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/ca64d874-5d86-4c49-8953-631a75d59788-config\") on node \"crc\" DevicePath \"\"" Oct 08 18:29:38 crc kubenswrapper[4859]: I1008 18:29:38.009506 4859 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/460ad6e7-859d-4111-a555-b55c5b7228e0-serving-cert\") on node \"crc\" DevicePath \"\"" Oct 08 18:29:38 crc kubenswrapper[4859]: I1008 18:29:38.009515 4859 reconciler_common.go:293] "Volume detached for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/460ad6e7-859d-4111-a555-b55c5b7228e0-client-ca\") on node \"crc\" DevicePath \"\"" Oct 08 18:29:38 crc kubenswrapper[4859]: I1008 18:29:38.009525 4859 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/460ad6e7-859d-4111-a555-b55c5b7228e0-config\") on node \"crc\" DevicePath \"\"" Oct 08 18:29:38 crc kubenswrapper[4859]: I1008 18:29:38.009533 4859 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/ca64d874-5d86-4c49-8953-631a75d59788-serving-cert\") on node \"crc\" DevicePath \"\"" Oct 08 18:29:38 crc kubenswrapper[4859]: I1008 18:29:38.009545 4859 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-9hnbb\" (UniqueName: \"kubernetes.io/projected/ca64d874-5d86-4c49-8953-631a75d59788-kube-api-access-9hnbb\") on node \"crc\" DevicePath \"\"" Oct 08 18:29:38 crc kubenswrapper[4859]: I1008 18:29:38.009556 4859 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-gbpww\" (UniqueName: \"kubernetes.io/projected/460ad6e7-859d-4111-a555-b55c5b7228e0-kube-api-access-gbpww\") on node \"crc\" DevicePath \"\"" Oct 08 18:29:38 crc kubenswrapper[4859]: I1008 18:29:38.094079 4859 generic.go:334] "Generic (PLEG): container finished" podID="460ad6e7-859d-4111-a555-b55c5b7228e0" containerID="46ee8cb842f339e6b4ba065cab54a4e76c9bac684c499161345db073e276bd89" exitCode=0 Oct 08 18:29:38 crc kubenswrapper[4859]: I1008 18:29:38.094130 4859 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-879f6c89f-9gpzj" event={"ID":"460ad6e7-859d-4111-a555-b55c5b7228e0","Type":"ContainerDied","Data":"46ee8cb842f339e6b4ba065cab54a4e76c9bac684c499161345db073e276bd89"} Oct 08 18:29:38 crc kubenswrapper[4859]: I1008 18:29:38.094185 4859 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-879f6c89f-9gpzj" event={"ID":"460ad6e7-859d-4111-a555-b55c5b7228e0","Type":"ContainerDied","Data":"1e7c718f9f31e46de3067304f0212127211b5576b88d7881e74ede8e5e3004a7"} Oct 08 18:29:38 crc kubenswrapper[4859]: I1008 18:29:38.094203 4859 scope.go:117] "RemoveContainer" containerID="46ee8cb842f339e6b4ba065cab54a4e76c9bac684c499161345db073e276bd89" Oct 08 18:29:38 crc kubenswrapper[4859]: I1008 18:29:38.094517 4859 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-879f6c89f-9gpzj" Oct 08 18:29:38 crc kubenswrapper[4859]: I1008 18:29:38.096980 4859 generic.go:334] "Generic (PLEG): container finished" podID="ca64d874-5d86-4c49-8953-631a75d59788" containerID="44d292882e24bdc475f139a3b68beb09c235ab83c773cff48466e5d53d3b8a19" exitCode=0 Oct 08 18:29:38 crc kubenswrapper[4859]: I1008 18:29:38.097010 4859 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-xlfdd" event={"ID":"ca64d874-5d86-4c49-8953-631a75d59788","Type":"ContainerDied","Data":"44d292882e24bdc475f139a3b68beb09c235ab83c773cff48466e5d53d3b8a19"} Oct 08 18:29:38 crc kubenswrapper[4859]: I1008 18:29:38.097030 4859 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-xlfdd" event={"ID":"ca64d874-5d86-4c49-8953-631a75d59788","Type":"ContainerDied","Data":"ba63a47285e8dd8deca0f8b02ae365771ce71a87162bf7b4eeedeeead2dd4b5b"} Oct 08 18:29:38 crc kubenswrapper[4859]: I1008 18:29:38.097035 4859 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-xlfdd" Oct 08 18:29:38 crc kubenswrapper[4859]: I1008 18:29:38.118837 4859 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-controller-manager/controller-manager-879f6c89f-9gpzj"] Oct 08 18:29:38 crc kubenswrapper[4859]: I1008 18:29:38.120355 4859 scope.go:117] "RemoveContainer" containerID="46ee8cb842f339e6b4ba065cab54a4e76c9bac684c499161345db073e276bd89" Oct 08 18:29:38 crc kubenswrapper[4859]: E1008 18:29:38.122839 4859 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"46ee8cb842f339e6b4ba065cab54a4e76c9bac684c499161345db073e276bd89\": container with ID starting with 46ee8cb842f339e6b4ba065cab54a4e76c9bac684c499161345db073e276bd89 not found: ID does not exist" containerID="46ee8cb842f339e6b4ba065cab54a4e76c9bac684c499161345db073e276bd89" Oct 08 18:29:38 crc kubenswrapper[4859]: I1008 18:29:38.122868 4859 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"46ee8cb842f339e6b4ba065cab54a4e76c9bac684c499161345db073e276bd89"} err="failed to get container status \"46ee8cb842f339e6b4ba065cab54a4e76c9bac684c499161345db073e276bd89\": rpc error: code = NotFound desc = could not find container \"46ee8cb842f339e6b4ba065cab54a4e76c9bac684c499161345db073e276bd89\": container with ID starting with 46ee8cb842f339e6b4ba065cab54a4e76c9bac684c499161345db073e276bd89 not found: ID does not exist" Oct 08 18:29:38 crc kubenswrapper[4859]: I1008 18:29:38.122901 4859 scope.go:117] "RemoveContainer" containerID="44d292882e24bdc475f139a3b68beb09c235ab83c773cff48466e5d53d3b8a19" Oct 08 18:29:38 crc kubenswrapper[4859]: I1008 18:29:38.125908 4859 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-controller-manager/controller-manager-879f6c89f-9gpzj"] Oct 08 18:29:38 crc kubenswrapper[4859]: I1008 18:29:38.138060 4859 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-6576b87f9c-xlfdd"] Oct 08 18:29:38 crc kubenswrapper[4859]: I1008 18:29:38.141346 4859 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-6576b87f9c-xlfdd"] Oct 08 18:29:38 crc kubenswrapper[4859]: I1008 18:29:38.144869 4859 scope.go:117] "RemoveContainer" containerID="44d292882e24bdc475f139a3b68beb09c235ab83c773cff48466e5d53d3b8a19" Oct 08 18:29:38 crc kubenswrapper[4859]: E1008 18:29:38.145476 4859 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"44d292882e24bdc475f139a3b68beb09c235ab83c773cff48466e5d53d3b8a19\": container with ID starting with 44d292882e24bdc475f139a3b68beb09c235ab83c773cff48466e5d53d3b8a19 not found: ID does not exist" containerID="44d292882e24bdc475f139a3b68beb09c235ab83c773cff48466e5d53d3b8a19" Oct 08 18:29:38 crc kubenswrapper[4859]: I1008 18:29:38.145514 4859 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"44d292882e24bdc475f139a3b68beb09c235ab83c773cff48466e5d53d3b8a19"} err="failed to get container status \"44d292882e24bdc475f139a3b68beb09c235ab83c773cff48466e5d53d3b8a19\": rpc error: code = NotFound desc = could not find container \"44d292882e24bdc475f139a3b68beb09c235ab83c773cff48466e5d53d3b8a19\": container with ID starting with 44d292882e24bdc475f139a3b68beb09c235ab83c773cff48466e5d53d3b8a19 not found: ID does not exist" Oct 08 18:29:38 crc kubenswrapper[4859]: I1008 18:29:38.280192 4859 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="metallb-system/metallb-operator-controller-manager-6789c5bc7c-k7fl4" Oct 08 18:29:38 crc kubenswrapper[4859]: I1008 18:29:38.477346 4859 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="460ad6e7-859d-4111-a555-b55c5b7228e0" path="/var/lib/kubelet/pods/460ad6e7-859d-4111-a555-b55c5b7228e0/volumes" Oct 08 18:29:38 crc kubenswrapper[4859]: I1008 18:29:38.479537 4859 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="ca64d874-5d86-4c49-8953-631a75d59788" path="/var/lib/kubelet/pods/ca64d874-5d86-4c49-8953-631a75d59788/volumes" Oct 08 18:29:38 crc kubenswrapper[4859]: I1008 18:29:38.932606 4859 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-route-controller-manager/route-controller-manager-5c55d4668-vqww5"] Oct 08 18:29:38 crc kubenswrapper[4859]: E1008 18:29:38.933065 4859 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ca64d874-5d86-4c49-8953-631a75d59788" containerName="route-controller-manager" Oct 08 18:29:38 crc kubenswrapper[4859]: I1008 18:29:38.933164 4859 state_mem.go:107] "Deleted CPUSet assignment" podUID="ca64d874-5d86-4c49-8953-631a75d59788" containerName="route-controller-manager" Oct 08 18:29:38 crc kubenswrapper[4859]: E1008 18:29:38.933197 4859 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="460ad6e7-859d-4111-a555-b55c5b7228e0" containerName="controller-manager" Oct 08 18:29:38 crc kubenswrapper[4859]: I1008 18:29:38.933210 4859 state_mem.go:107] "Deleted CPUSet assignment" podUID="460ad6e7-859d-4111-a555-b55c5b7228e0" containerName="controller-manager" Oct 08 18:29:38 crc kubenswrapper[4859]: I1008 18:29:38.933361 4859 memory_manager.go:354] "RemoveStaleState removing state" podUID="460ad6e7-859d-4111-a555-b55c5b7228e0" containerName="controller-manager" Oct 08 18:29:38 crc kubenswrapper[4859]: I1008 18:29:38.933385 4859 memory_manager.go:354] "RemoveStaleState removing state" podUID="ca64d874-5d86-4c49-8953-631a75d59788" containerName="route-controller-manager" Oct 08 18:29:38 crc kubenswrapper[4859]: I1008 18:29:38.934045 4859 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-5c55d4668-vqww5" Oct 08 18:29:38 crc kubenswrapper[4859]: I1008 18:29:38.937057 4859 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"kube-root-ca.crt" Oct 08 18:29:38 crc kubenswrapper[4859]: I1008 18:29:38.937636 4859 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-route-controller-manager"/"route-controller-manager-sa-dockercfg-h2zr2" Oct 08 18:29:38 crc kubenswrapper[4859]: I1008 18:29:38.937956 4859 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"config" Oct 08 18:29:38 crc kubenswrapper[4859]: I1008 18:29:38.938118 4859 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["metallb-system/frr-k8s-mfcd8"] Oct 08 18:29:38 crc kubenswrapper[4859]: I1008 18:29:38.938323 4859 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-route-controller-manager"/"serving-cert" Oct 08 18:29:38 crc kubenswrapper[4859]: I1008 18:29:38.938593 4859 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"openshift-service-ca.crt" Oct 08 18:29:38 crc kubenswrapper[4859]: I1008 18:29:38.938878 4859 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"client-ca" Oct 08 18:29:38 crc kubenswrapper[4859]: I1008 18:29:38.941110 4859 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/frr-k8s-mfcd8" Oct 08 18:29:38 crc kubenswrapper[4859]: I1008 18:29:38.947808 4859 reflector.go:368] Caches populated for *v1.ConfigMap from object-"metallb-system"/"frr-startup" Oct 08 18:29:38 crc kubenswrapper[4859]: I1008 18:29:38.948440 4859 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"frr-k8s-certs-secret" Oct 08 18:29:38 crc kubenswrapper[4859]: I1008 18:29:38.948659 4859 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"frr-k8s-daemon-dockercfg-ww9h2" Oct 08 18:29:38 crc kubenswrapper[4859]: I1008 18:29:38.957422 4859 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-5c55d4668-vqww5"] Oct 08 18:29:38 crc kubenswrapper[4859]: I1008 18:29:38.970795 4859 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["metallb-system/frr-k8s-webhook-server-64bf5d555-77xwq"] Oct 08 18:29:38 crc kubenswrapper[4859]: I1008 18:29:38.971828 4859 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/frr-k8s-webhook-server-64bf5d555-77xwq" Oct 08 18:29:38 crc kubenswrapper[4859]: I1008 18:29:38.974247 4859 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"frr-k8s-webhook-server-cert" Oct 08 18:29:38 crc kubenswrapper[4859]: I1008 18:29:38.984310 4859 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["metallb-system/frr-k8s-webhook-server-64bf5d555-77xwq"] Oct 08 18:29:39 crc kubenswrapper[4859]: I1008 18:29:39.021024 4859 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"frr-conf\" (UniqueName: \"kubernetes.io/empty-dir/69cc95cc-1a58-40c2-9c84-4808c217c2e5-frr-conf\") pod \"frr-k8s-mfcd8\" (UID: \"69cc95cc-1a58-40c2-9c84-4808c217c2e5\") " pod="metallb-system/frr-k8s-mfcd8" Oct 08 18:29:39 crc kubenswrapper[4859]: I1008 18:29:39.021117 4859 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-xx9s6\" (UniqueName: \"kubernetes.io/projected/69cc95cc-1a58-40c2-9c84-4808c217c2e5-kube-api-access-xx9s6\") pod \"frr-k8s-mfcd8\" (UID: \"69cc95cc-1a58-40c2-9c84-4808c217c2e5\") " pod="metallb-system/frr-k8s-mfcd8" Oct 08 18:29:39 crc kubenswrapper[4859]: I1008 18:29:39.021351 4859 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/4d682ed0-e5d8-48ad-8e52-cd29b2f65bd2-cert\") pod \"frr-k8s-webhook-server-64bf5d555-77xwq\" (UID: \"4d682ed0-e5d8-48ad-8e52-cd29b2f65bd2\") " pod="metallb-system/frr-k8s-webhook-server-64bf5d555-77xwq" Oct 08 18:29:39 crc kubenswrapper[4859]: I1008 18:29:39.021373 4859 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/c31d0cee-7935-4c85-8461-3ddfd06a8c1d-client-ca\") pod \"route-controller-manager-5c55d4668-vqww5\" (UID: \"c31d0cee-7935-4c85-8461-3ddfd06a8c1d\") " pod="openshift-route-controller-manager/route-controller-manager-5c55d4668-vqww5" Oct 08 18:29:39 crc kubenswrapper[4859]: I1008 18:29:39.021393 4859 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/c31d0cee-7935-4c85-8461-3ddfd06a8c1d-serving-cert\") pod \"route-controller-manager-5c55d4668-vqww5\" (UID: \"c31d0cee-7935-4c85-8461-3ddfd06a8c1d\") " pod="openshift-route-controller-manager/route-controller-manager-5c55d4668-vqww5" Oct 08 18:29:39 crc kubenswrapper[4859]: I1008 18:29:39.021418 4859 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-nxvb9\" (UniqueName: \"kubernetes.io/projected/c31d0cee-7935-4c85-8461-3ddfd06a8c1d-kube-api-access-nxvb9\") pod \"route-controller-manager-5c55d4668-vqww5\" (UID: \"c31d0cee-7935-4c85-8461-3ddfd06a8c1d\") " pod="openshift-route-controller-manager/route-controller-manager-5c55d4668-vqww5" Oct 08 18:29:39 crc kubenswrapper[4859]: I1008 18:29:39.021445 4859 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"frr-sockets\" (UniqueName: \"kubernetes.io/empty-dir/69cc95cc-1a58-40c2-9c84-4808c217c2e5-frr-sockets\") pod \"frr-k8s-mfcd8\" (UID: \"69cc95cc-1a58-40c2-9c84-4808c217c2e5\") " pod="metallb-system/frr-k8s-mfcd8" Oct 08 18:29:39 crc kubenswrapper[4859]: I1008 18:29:39.021465 4859 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics\" (UniqueName: \"kubernetes.io/empty-dir/69cc95cc-1a58-40c2-9c84-4808c217c2e5-metrics\") pod \"frr-k8s-mfcd8\" (UID: \"69cc95cc-1a58-40c2-9c84-4808c217c2e5\") " pod="metallb-system/frr-k8s-mfcd8" Oct 08 18:29:39 crc kubenswrapper[4859]: I1008 18:29:39.021486 4859 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/69cc95cc-1a58-40c2-9c84-4808c217c2e5-metrics-certs\") pod \"frr-k8s-mfcd8\" (UID: \"69cc95cc-1a58-40c2-9c84-4808c217c2e5\") " pod="metallb-system/frr-k8s-mfcd8" Oct 08 18:29:39 crc kubenswrapper[4859]: I1008 18:29:39.021521 4859 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"reloader\" (UniqueName: \"kubernetes.io/empty-dir/69cc95cc-1a58-40c2-9c84-4808c217c2e5-reloader\") pod \"frr-k8s-mfcd8\" (UID: \"69cc95cc-1a58-40c2-9c84-4808c217c2e5\") " pod="metallb-system/frr-k8s-mfcd8" Oct 08 18:29:39 crc kubenswrapper[4859]: I1008 18:29:39.021552 4859 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-frl82\" (UniqueName: \"kubernetes.io/projected/4d682ed0-e5d8-48ad-8e52-cd29b2f65bd2-kube-api-access-frl82\") pod \"frr-k8s-webhook-server-64bf5d555-77xwq\" (UID: \"4d682ed0-e5d8-48ad-8e52-cd29b2f65bd2\") " pod="metallb-system/frr-k8s-webhook-server-64bf5d555-77xwq" Oct 08 18:29:39 crc kubenswrapper[4859]: I1008 18:29:39.021573 4859 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"frr-startup\" (UniqueName: \"kubernetes.io/configmap/69cc95cc-1a58-40c2-9c84-4808c217c2e5-frr-startup\") pod \"frr-k8s-mfcd8\" (UID: \"69cc95cc-1a58-40c2-9c84-4808c217c2e5\") " pod="metallb-system/frr-k8s-mfcd8" Oct 08 18:29:39 crc kubenswrapper[4859]: I1008 18:29:39.021596 4859 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/c31d0cee-7935-4c85-8461-3ddfd06a8c1d-config\") pod \"route-controller-manager-5c55d4668-vqww5\" (UID: \"c31d0cee-7935-4c85-8461-3ddfd06a8c1d\") " pod="openshift-route-controller-manager/route-controller-manager-5c55d4668-vqww5" Oct 08 18:29:39 crc kubenswrapper[4859]: I1008 18:29:39.034588 4859 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-controller-manager/controller-manager-5b87458c9c-4krwv"] Oct 08 18:29:39 crc kubenswrapper[4859]: I1008 18:29:39.035567 4859 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-5b87458c9c-4krwv" Oct 08 18:29:39 crc kubenswrapper[4859]: I1008 18:29:39.038213 4859 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"openshift-service-ca.crt" Oct 08 18:29:39 crc kubenswrapper[4859]: I1008 18:29:39.038648 4859 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager"/"openshift-controller-manager-sa-dockercfg-msq4c" Oct 08 18:29:39 crc kubenswrapper[4859]: I1008 18:29:39.039379 4859 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"config" Oct 08 18:29:39 crc kubenswrapper[4859]: I1008 18:29:39.040186 4859 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"client-ca" Oct 08 18:29:39 crc kubenswrapper[4859]: I1008 18:29:39.040378 4859 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager"/"serving-cert" Oct 08 18:29:39 crc kubenswrapper[4859]: I1008 18:29:39.040569 4859 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"kube-root-ca.crt" Oct 08 18:29:39 crc kubenswrapper[4859]: I1008 18:29:39.050920 4859 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"openshift-global-ca" Oct 08 18:29:39 crc kubenswrapper[4859]: I1008 18:29:39.055716 4859 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-controller-manager/controller-manager-5b87458c9c-4krwv"] Oct 08 18:29:39 crc kubenswrapper[4859]: I1008 18:29:39.123530 4859 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/c31d0cee-7935-4c85-8461-3ddfd06a8c1d-serving-cert\") pod \"route-controller-manager-5c55d4668-vqww5\" (UID: \"c31d0cee-7935-4c85-8461-3ddfd06a8c1d\") " pod="openshift-route-controller-manager/route-controller-manager-5c55d4668-vqww5" Oct 08 18:29:39 crc kubenswrapper[4859]: I1008 18:29:39.123601 4859 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/0d0c4bdf-fe24-4589-9fed-d6d81cfc25b8-client-ca\") pod \"controller-manager-5b87458c9c-4krwv\" (UID: \"0d0c4bdf-fe24-4589-9fed-d6d81cfc25b8\") " pod="openshift-controller-manager/controller-manager-5b87458c9c-4krwv" Oct 08 18:29:39 crc kubenswrapper[4859]: I1008 18:29:39.123633 4859 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-nxvb9\" (UniqueName: \"kubernetes.io/projected/c31d0cee-7935-4c85-8461-3ddfd06a8c1d-kube-api-access-nxvb9\") pod \"route-controller-manager-5c55d4668-vqww5\" (UID: \"c31d0cee-7935-4c85-8461-3ddfd06a8c1d\") " pod="openshift-route-controller-manager/route-controller-manager-5c55d4668-vqww5" Oct 08 18:29:39 crc kubenswrapper[4859]: I1008 18:29:39.123663 4859 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"frr-sockets\" (UniqueName: \"kubernetes.io/empty-dir/69cc95cc-1a58-40c2-9c84-4808c217c2e5-frr-sockets\") pod \"frr-k8s-mfcd8\" (UID: \"69cc95cc-1a58-40c2-9c84-4808c217c2e5\") " pod="metallb-system/frr-k8s-mfcd8" Oct 08 18:29:39 crc kubenswrapper[4859]: I1008 18:29:39.123716 4859 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics\" (UniqueName: \"kubernetes.io/empty-dir/69cc95cc-1a58-40c2-9c84-4808c217c2e5-metrics\") pod \"frr-k8s-mfcd8\" (UID: \"69cc95cc-1a58-40c2-9c84-4808c217c2e5\") " pod="metallb-system/frr-k8s-mfcd8" Oct 08 18:29:39 crc kubenswrapper[4859]: I1008 18:29:39.123751 4859 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/0d0c4bdf-fe24-4589-9fed-d6d81cfc25b8-serving-cert\") pod \"controller-manager-5b87458c9c-4krwv\" (UID: \"0d0c4bdf-fe24-4589-9fed-d6d81cfc25b8\") " pod="openshift-controller-manager/controller-manager-5b87458c9c-4krwv" Oct 08 18:29:39 crc kubenswrapper[4859]: I1008 18:29:39.124198 4859 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/69cc95cc-1a58-40c2-9c84-4808c217c2e5-metrics-certs\") pod \"frr-k8s-mfcd8\" (UID: \"69cc95cc-1a58-40c2-9c84-4808c217c2e5\") " pod="metallb-system/frr-k8s-mfcd8" Oct 08 18:29:39 crc kubenswrapper[4859]: I1008 18:29:39.124367 4859 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-ts9rq\" (UniqueName: \"kubernetes.io/projected/0d0c4bdf-fe24-4589-9fed-d6d81cfc25b8-kube-api-access-ts9rq\") pod \"controller-manager-5b87458c9c-4krwv\" (UID: \"0d0c4bdf-fe24-4589-9fed-d6d81cfc25b8\") " pod="openshift-controller-manager/controller-manager-5b87458c9c-4krwv" Oct 08 18:29:39 crc kubenswrapper[4859]: I1008 18:29:39.124602 4859 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"frr-sockets\" (UniqueName: \"kubernetes.io/empty-dir/69cc95cc-1a58-40c2-9c84-4808c217c2e5-frr-sockets\") pod \"frr-k8s-mfcd8\" (UID: \"69cc95cc-1a58-40c2-9c84-4808c217c2e5\") " pod="metallb-system/frr-k8s-mfcd8" Oct 08 18:29:39 crc kubenswrapper[4859]: I1008 18:29:39.124909 4859 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics\" (UniqueName: \"kubernetes.io/empty-dir/69cc95cc-1a58-40c2-9c84-4808c217c2e5-metrics\") pod \"frr-k8s-mfcd8\" (UID: \"69cc95cc-1a58-40c2-9c84-4808c217c2e5\") " pod="metallb-system/frr-k8s-mfcd8" Oct 08 18:29:39 crc kubenswrapper[4859]: I1008 18:29:39.125392 4859 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"reloader\" (UniqueName: \"kubernetes.io/empty-dir/69cc95cc-1a58-40c2-9c84-4808c217c2e5-reloader\") pod \"frr-k8s-mfcd8\" (UID: \"69cc95cc-1a58-40c2-9c84-4808c217c2e5\") " pod="metallb-system/frr-k8s-mfcd8" Oct 08 18:29:39 crc kubenswrapper[4859]: I1008 18:29:39.125473 4859 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-frl82\" (UniqueName: \"kubernetes.io/projected/4d682ed0-e5d8-48ad-8e52-cd29b2f65bd2-kube-api-access-frl82\") pod \"frr-k8s-webhook-server-64bf5d555-77xwq\" (UID: \"4d682ed0-e5d8-48ad-8e52-cd29b2f65bd2\") " pod="metallb-system/frr-k8s-webhook-server-64bf5d555-77xwq" Oct 08 18:29:39 crc kubenswrapper[4859]: I1008 18:29:39.125744 4859 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"frr-startup\" (UniqueName: \"kubernetes.io/configmap/69cc95cc-1a58-40c2-9c84-4808c217c2e5-frr-startup\") pod \"frr-k8s-mfcd8\" (UID: \"69cc95cc-1a58-40c2-9c84-4808c217c2e5\") " pod="metallb-system/frr-k8s-mfcd8" Oct 08 18:29:39 crc kubenswrapper[4859]: I1008 18:29:39.125770 4859 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/c31d0cee-7935-4c85-8461-3ddfd06a8c1d-config\") pod \"route-controller-manager-5c55d4668-vqww5\" (UID: \"c31d0cee-7935-4c85-8461-3ddfd06a8c1d\") " pod="openshift-route-controller-manager/route-controller-manager-5c55d4668-vqww5" Oct 08 18:29:39 crc kubenswrapper[4859]: I1008 18:29:39.125793 4859 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/0d0c4bdf-fe24-4589-9fed-d6d81cfc25b8-config\") pod \"controller-manager-5b87458c9c-4krwv\" (UID: \"0d0c4bdf-fe24-4589-9fed-d6d81cfc25b8\") " pod="openshift-controller-manager/controller-manager-5b87458c9c-4krwv" Oct 08 18:29:39 crc kubenswrapper[4859]: I1008 18:29:39.125832 4859 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"frr-conf\" (UniqueName: \"kubernetes.io/empty-dir/69cc95cc-1a58-40c2-9c84-4808c217c2e5-frr-conf\") pod \"frr-k8s-mfcd8\" (UID: \"69cc95cc-1a58-40c2-9c84-4808c217c2e5\") " pod="metallb-system/frr-k8s-mfcd8" Oct 08 18:29:39 crc kubenswrapper[4859]: I1008 18:29:39.125871 4859 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-xx9s6\" (UniqueName: \"kubernetes.io/projected/69cc95cc-1a58-40c2-9c84-4808c217c2e5-kube-api-access-xx9s6\") pod \"frr-k8s-mfcd8\" (UID: \"69cc95cc-1a58-40c2-9c84-4808c217c2e5\") " pod="metallb-system/frr-k8s-mfcd8" Oct 08 18:29:39 crc kubenswrapper[4859]: I1008 18:29:39.125917 4859 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/0d0c4bdf-fe24-4589-9fed-d6d81cfc25b8-proxy-ca-bundles\") pod \"controller-manager-5b87458c9c-4krwv\" (UID: \"0d0c4bdf-fe24-4589-9fed-d6d81cfc25b8\") " pod="openshift-controller-manager/controller-manager-5b87458c9c-4krwv" Oct 08 18:29:39 crc kubenswrapper[4859]: I1008 18:29:39.125953 4859 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/4d682ed0-e5d8-48ad-8e52-cd29b2f65bd2-cert\") pod \"frr-k8s-webhook-server-64bf5d555-77xwq\" (UID: \"4d682ed0-e5d8-48ad-8e52-cd29b2f65bd2\") " pod="metallb-system/frr-k8s-webhook-server-64bf5d555-77xwq" Oct 08 18:29:39 crc kubenswrapper[4859]: I1008 18:29:39.125981 4859 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/c31d0cee-7935-4c85-8461-3ddfd06a8c1d-client-ca\") pod \"route-controller-manager-5c55d4668-vqww5\" (UID: \"c31d0cee-7935-4c85-8461-3ddfd06a8c1d\") " pod="openshift-route-controller-manager/route-controller-manager-5c55d4668-vqww5" Oct 08 18:29:39 crc kubenswrapper[4859]: I1008 18:29:39.126158 4859 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"reloader\" (UniqueName: \"kubernetes.io/empty-dir/69cc95cc-1a58-40c2-9c84-4808c217c2e5-reloader\") pod \"frr-k8s-mfcd8\" (UID: \"69cc95cc-1a58-40c2-9c84-4808c217c2e5\") " pod="metallb-system/frr-k8s-mfcd8" Oct 08 18:29:39 crc kubenswrapper[4859]: I1008 18:29:39.126876 4859 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"frr-conf\" (UniqueName: \"kubernetes.io/empty-dir/69cc95cc-1a58-40c2-9c84-4808c217c2e5-frr-conf\") pod \"frr-k8s-mfcd8\" (UID: \"69cc95cc-1a58-40c2-9c84-4808c217c2e5\") " pod="metallb-system/frr-k8s-mfcd8" Oct 08 18:29:39 crc kubenswrapper[4859]: I1008 18:29:39.126879 4859 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/c31d0cee-7935-4c85-8461-3ddfd06a8c1d-client-ca\") pod \"route-controller-manager-5c55d4668-vqww5\" (UID: \"c31d0cee-7935-4c85-8461-3ddfd06a8c1d\") " pod="openshift-route-controller-manager/route-controller-manager-5c55d4668-vqww5" Oct 08 18:29:39 crc kubenswrapper[4859]: I1008 18:29:39.127006 4859 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/c31d0cee-7935-4c85-8461-3ddfd06a8c1d-config\") pod \"route-controller-manager-5c55d4668-vqww5\" (UID: \"c31d0cee-7935-4c85-8461-3ddfd06a8c1d\") " pod="openshift-route-controller-manager/route-controller-manager-5c55d4668-vqww5" Oct 08 18:29:39 crc kubenswrapper[4859]: E1008 18:29:39.126966 4859 secret.go:188] Couldn't get secret metallb-system/frr-k8s-webhook-server-cert: secret "frr-k8s-webhook-server-cert" not found Oct 08 18:29:39 crc kubenswrapper[4859]: E1008 18:29:39.127184 4859 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/4d682ed0-e5d8-48ad-8e52-cd29b2f65bd2-cert podName:4d682ed0-e5d8-48ad-8e52-cd29b2f65bd2 nodeName:}" failed. No retries permitted until 2025-10-08 18:29:39.627165083 +0000 UTC m=+749.874004462 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "cert" (UniqueName: "kubernetes.io/secret/4d682ed0-e5d8-48ad-8e52-cd29b2f65bd2-cert") pod "frr-k8s-webhook-server-64bf5d555-77xwq" (UID: "4d682ed0-e5d8-48ad-8e52-cd29b2f65bd2") : secret "frr-k8s-webhook-server-cert" not found Oct 08 18:29:39 crc kubenswrapper[4859]: I1008 18:29:39.127886 4859 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"frr-startup\" (UniqueName: \"kubernetes.io/configmap/69cc95cc-1a58-40c2-9c84-4808c217c2e5-frr-startup\") pod \"frr-k8s-mfcd8\" (UID: \"69cc95cc-1a58-40c2-9c84-4808c217c2e5\") " pod="metallb-system/frr-k8s-mfcd8" Oct 08 18:29:39 crc kubenswrapper[4859]: I1008 18:29:39.135848 4859 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/c31d0cee-7935-4c85-8461-3ddfd06a8c1d-serving-cert\") pod \"route-controller-manager-5c55d4668-vqww5\" (UID: \"c31d0cee-7935-4c85-8461-3ddfd06a8c1d\") " pod="openshift-route-controller-manager/route-controller-manager-5c55d4668-vqww5" Oct 08 18:29:39 crc kubenswrapper[4859]: I1008 18:29:39.137450 4859 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/69cc95cc-1a58-40c2-9c84-4808c217c2e5-metrics-certs\") pod \"frr-k8s-mfcd8\" (UID: \"69cc95cc-1a58-40c2-9c84-4808c217c2e5\") " pod="metallb-system/frr-k8s-mfcd8" Oct 08 18:29:39 crc kubenswrapper[4859]: I1008 18:29:39.167224 4859 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["metallb-system/speaker-47524"] Oct 08 18:29:39 crc kubenswrapper[4859]: I1008 18:29:39.168402 4859 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/speaker-47524" Oct 08 18:29:39 crc kubenswrapper[4859]: I1008 18:29:39.175477 4859 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"speaker-certs-secret" Oct 08 18:29:39 crc kubenswrapper[4859]: I1008 18:29:39.175590 4859 reflector.go:368] Caches populated for *v1.ConfigMap from object-"metallb-system"/"metallb-excludel2" Oct 08 18:29:39 crc kubenswrapper[4859]: I1008 18:29:39.175713 4859 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"metallb-memberlist" Oct 08 18:29:39 crc kubenswrapper[4859]: I1008 18:29:39.176187 4859 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"speaker-dockercfg-wk8tw" Oct 08 18:29:39 crc kubenswrapper[4859]: I1008 18:29:39.192866 4859 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-xx9s6\" (UniqueName: \"kubernetes.io/projected/69cc95cc-1a58-40c2-9c84-4808c217c2e5-kube-api-access-xx9s6\") pod \"frr-k8s-mfcd8\" (UID: \"69cc95cc-1a58-40c2-9c84-4808c217c2e5\") " pod="metallb-system/frr-k8s-mfcd8" Oct 08 18:29:39 crc kubenswrapper[4859]: I1008 18:29:39.193577 4859 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-nxvb9\" (UniqueName: \"kubernetes.io/projected/c31d0cee-7935-4c85-8461-3ddfd06a8c1d-kube-api-access-nxvb9\") pod \"route-controller-manager-5c55d4668-vqww5\" (UID: \"c31d0cee-7935-4c85-8461-3ddfd06a8c1d\") " pod="openshift-route-controller-manager/route-controller-manager-5c55d4668-vqww5" Oct 08 18:29:39 crc kubenswrapper[4859]: I1008 18:29:39.201489 4859 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-frl82\" (UniqueName: \"kubernetes.io/projected/4d682ed0-e5d8-48ad-8e52-cd29b2f65bd2-kube-api-access-frl82\") pod \"frr-k8s-webhook-server-64bf5d555-77xwq\" (UID: \"4d682ed0-e5d8-48ad-8e52-cd29b2f65bd2\") " pod="metallb-system/frr-k8s-webhook-server-64bf5d555-77xwq" Oct 08 18:29:39 crc kubenswrapper[4859]: I1008 18:29:39.207857 4859 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["metallb-system/controller-68d546b9d8-cr5dk"] Oct 08 18:29:39 crc kubenswrapper[4859]: I1008 18:29:39.209015 4859 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/controller-68d546b9d8-cr5dk" Oct 08 18:29:39 crc kubenswrapper[4859]: I1008 18:29:39.214006 4859 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"controller-certs-secret" Oct 08 18:29:39 crc kubenswrapper[4859]: I1008 18:29:39.225984 4859 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["metallb-system/controller-68d546b9d8-cr5dk"] Oct 08 18:29:39 crc kubenswrapper[4859]: I1008 18:29:39.226816 4859 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"memberlist\" (UniqueName: \"kubernetes.io/secret/7557fe67-9517-441e-944c-9d5286f5735c-memberlist\") pod \"speaker-47524\" (UID: \"7557fe67-9517-441e-944c-9d5286f5735c\") " pod="metallb-system/speaker-47524" Oct 08 18:29:39 crc kubenswrapper[4859]: I1008 18:29:39.226891 4859 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/0d0c4bdf-fe24-4589-9fed-d6d81cfc25b8-serving-cert\") pod \"controller-manager-5b87458c9c-4krwv\" (UID: \"0d0c4bdf-fe24-4589-9fed-d6d81cfc25b8\") " pod="openshift-controller-manager/controller-manager-5b87458c9c-4krwv" Oct 08 18:29:39 crc kubenswrapper[4859]: I1008 18:29:39.226930 4859 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/7557fe67-9517-441e-944c-9d5286f5735c-metrics-certs\") pod \"speaker-47524\" (UID: \"7557fe67-9517-441e-944c-9d5286f5735c\") " pod="metallb-system/speaker-47524" Oct 08 18:29:39 crc kubenswrapper[4859]: I1008 18:29:39.226956 4859 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-ts9rq\" (UniqueName: \"kubernetes.io/projected/0d0c4bdf-fe24-4589-9fed-d6d81cfc25b8-kube-api-access-ts9rq\") pod \"controller-manager-5b87458c9c-4krwv\" (UID: \"0d0c4bdf-fe24-4589-9fed-d6d81cfc25b8\") " pod="openshift-controller-manager/controller-manager-5b87458c9c-4krwv" Oct 08 18:29:39 crc kubenswrapper[4859]: I1008 18:29:39.226978 4859 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metallb-excludel2\" (UniqueName: \"kubernetes.io/configmap/7557fe67-9517-441e-944c-9d5286f5735c-metallb-excludel2\") pod \"speaker-47524\" (UID: \"7557fe67-9517-441e-944c-9d5286f5735c\") " pod="metallb-system/speaker-47524" Oct 08 18:29:39 crc kubenswrapper[4859]: I1008 18:29:39.227001 4859 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-bbphq\" (UniqueName: \"kubernetes.io/projected/7557fe67-9517-441e-944c-9d5286f5735c-kube-api-access-bbphq\") pod \"speaker-47524\" (UID: \"7557fe67-9517-441e-944c-9d5286f5735c\") " pod="metallb-system/speaker-47524" Oct 08 18:29:39 crc kubenswrapper[4859]: I1008 18:29:39.227029 4859 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/0d0c4bdf-fe24-4589-9fed-d6d81cfc25b8-config\") pod \"controller-manager-5b87458c9c-4krwv\" (UID: \"0d0c4bdf-fe24-4589-9fed-d6d81cfc25b8\") " pod="openshift-controller-manager/controller-manager-5b87458c9c-4krwv" Oct 08 18:29:39 crc kubenswrapper[4859]: I1008 18:29:39.227081 4859 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/0d0c4bdf-fe24-4589-9fed-d6d81cfc25b8-proxy-ca-bundles\") pod \"controller-manager-5b87458c9c-4krwv\" (UID: \"0d0c4bdf-fe24-4589-9fed-d6d81cfc25b8\") " pod="openshift-controller-manager/controller-manager-5b87458c9c-4krwv" Oct 08 18:29:39 crc kubenswrapper[4859]: I1008 18:29:39.227137 4859 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/0d0c4bdf-fe24-4589-9fed-d6d81cfc25b8-client-ca\") pod \"controller-manager-5b87458c9c-4krwv\" (UID: \"0d0c4bdf-fe24-4589-9fed-d6d81cfc25b8\") " pod="openshift-controller-manager/controller-manager-5b87458c9c-4krwv" Oct 08 18:29:39 crc kubenswrapper[4859]: I1008 18:29:39.228327 4859 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/0d0c4bdf-fe24-4589-9fed-d6d81cfc25b8-client-ca\") pod \"controller-manager-5b87458c9c-4krwv\" (UID: \"0d0c4bdf-fe24-4589-9fed-d6d81cfc25b8\") " pod="openshift-controller-manager/controller-manager-5b87458c9c-4krwv" Oct 08 18:29:39 crc kubenswrapper[4859]: I1008 18:29:39.232653 4859 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/0d0c4bdf-fe24-4589-9fed-d6d81cfc25b8-serving-cert\") pod \"controller-manager-5b87458c9c-4krwv\" (UID: \"0d0c4bdf-fe24-4589-9fed-d6d81cfc25b8\") " pod="openshift-controller-manager/controller-manager-5b87458c9c-4krwv" Oct 08 18:29:39 crc kubenswrapper[4859]: I1008 18:29:39.232743 4859 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/0d0c4bdf-fe24-4589-9fed-d6d81cfc25b8-config\") pod \"controller-manager-5b87458c9c-4krwv\" (UID: \"0d0c4bdf-fe24-4589-9fed-d6d81cfc25b8\") " pod="openshift-controller-manager/controller-manager-5b87458c9c-4krwv" Oct 08 18:29:39 crc kubenswrapper[4859]: I1008 18:29:39.233796 4859 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/0d0c4bdf-fe24-4589-9fed-d6d81cfc25b8-proxy-ca-bundles\") pod \"controller-manager-5b87458c9c-4krwv\" (UID: \"0d0c4bdf-fe24-4589-9fed-d6d81cfc25b8\") " pod="openshift-controller-manager/controller-manager-5b87458c9c-4krwv" Oct 08 18:29:39 crc kubenswrapper[4859]: I1008 18:29:39.263356 4859 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-5c55d4668-vqww5" Oct 08 18:29:39 crc kubenswrapper[4859]: I1008 18:29:39.269829 4859 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/frr-k8s-mfcd8" Oct 08 18:29:39 crc kubenswrapper[4859]: I1008 18:29:39.272052 4859 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-ts9rq\" (UniqueName: \"kubernetes.io/projected/0d0c4bdf-fe24-4589-9fed-d6d81cfc25b8-kube-api-access-ts9rq\") pod \"controller-manager-5b87458c9c-4krwv\" (UID: \"0d0c4bdf-fe24-4589-9fed-d6d81cfc25b8\") " pod="openshift-controller-manager/controller-manager-5b87458c9c-4krwv" Oct 08 18:29:39 crc kubenswrapper[4859]: I1008 18:29:39.328598 4859 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"memberlist\" (UniqueName: \"kubernetes.io/secret/7557fe67-9517-441e-944c-9d5286f5735c-memberlist\") pod \"speaker-47524\" (UID: \"7557fe67-9517-441e-944c-9d5286f5735c\") " pod="metallb-system/speaker-47524" Oct 08 18:29:39 crc kubenswrapper[4859]: I1008 18:29:39.328656 4859 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/7f1dbe0e-c03a-4ec2-ba3c-7179d2c67e41-cert\") pod \"controller-68d546b9d8-cr5dk\" (UID: \"7f1dbe0e-c03a-4ec2-ba3c-7179d2c67e41\") " pod="metallb-system/controller-68d546b9d8-cr5dk" Oct 08 18:29:39 crc kubenswrapper[4859]: I1008 18:29:39.328703 4859 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-qppp2\" (UniqueName: \"kubernetes.io/projected/7f1dbe0e-c03a-4ec2-ba3c-7179d2c67e41-kube-api-access-qppp2\") pod \"controller-68d546b9d8-cr5dk\" (UID: \"7f1dbe0e-c03a-4ec2-ba3c-7179d2c67e41\") " pod="metallb-system/controller-68d546b9d8-cr5dk" Oct 08 18:29:39 crc kubenswrapper[4859]: I1008 18:29:39.328731 4859 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/7f1dbe0e-c03a-4ec2-ba3c-7179d2c67e41-metrics-certs\") pod \"controller-68d546b9d8-cr5dk\" (UID: \"7f1dbe0e-c03a-4ec2-ba3c-7179d2c67e41\") " pod="metallb-system/controller-68d546b9d8-cr5dk" Oct 08 18:29:39 crc kubenswrapper[4859]: I1008 18:29:39.328758 4859 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/7557fe67-9517-441e-944c-9d5286f5735c-metrics-certs\") pod \"speaker-47524\" (UID: \"7557fe67-9517-441e-944c-9d5286f5735c\") " pod="metallb-system/speaker-47524" Oct 08 18:29:39 crc kubenswrapper[4859]: I1008 18:29:39.328776 4859 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metallb-excludel2\" (UniqueName: \"kubernetes.io/configmap/7557fe67-9517-441e-944c-9d5286f5735c-metallb-excludel2\") pod \"speaker-47524\" (UID: \"7557fe67-9517-441e-944c-9d5286f5735c\") " pod="metallb-system/speaker-47524" Oct 08 18:29:39 crc kubenswrapper[4859]: I1008 18:29:39.328794 4859 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-bbphq\" (UniqueName: \"kubernetes.io/projected/7557fe67-9517-441e-944c-9d5286f5735c-kube-api-access-bbphq\") pod \"speaker-47524\" (UID: \"7557fe67-9517-441e-944c-9d5286f5735c\") " pod="metallb-system/speaker-47524" Oct 08 18:29:39 crc kubenswrapper[4859]: E1008 18:29:39.329179 4859 secret.go:188] Couldn't get secret metallb-system/metallb-memberlist: secret "metallb-memberlist" not found Oct 08 18:29:39 crc kubenswrapper[4859]: E1008 18:29:39.329229 4859 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/7557fe67-9517-441e-944c-9d5286f5735c-memberlist podName:7557fe67-9517-441e-944c-9d5286f5735c nodeName:}" failed. No retries permitted until 2025-10-08 18:29:39.829214604 +0000 UTC m=+750.076053983 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "memberlist" (UniqueName: "kubernetes.io/secret/7557fe67-9517-441e-944c-9d5286f5735c-memberlist") pod "speaker-47524" (UID: "7557fe67-9517-441e-944c-9d5286f5735c") : secret "metallb-memberlist" not found Oct 08 18:29:39 crc kubenswrapper[4859]: E1008 18:29:39.329372 4859 secret.go:188] Couldn't get secret metallb-system/speaker-certs-secret: secret "speaker-certs-secret" not found Oct 08 18:29:39 crc kubenswrapper[4859]: E1008 18:29:39.329401 4859 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/7557fe67-9517-441e-944c-9d5286f5735c-metrics-certs podName:7557fe67-9517-441e-944c-9d5286f5735c nodeName:}" failed. No retries permitted until 2025-10-08 18:29:39.829393099 +0000 UTC m=+750.076232478 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/7557fe67-9517-441e-944c-9d5286f5735c-metrics-certs") pod "speaker-47524" (UID: "7557fe67-9517-441e-944c-9d5286f5735c") : secret "speaker-certs-secret" not found Oct 08 18:29:39 crc kubenswrapper[4859]: I1008 18:29:39.330339 4859 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metallb-excludel2\" (UniqueName: \"kubernetes.io/configmap/7557fe67-9517-441e-944c-9d5286f5735c-metallb-excludel2\") pod \"speaker-47524\" (UID: \"7557fe67-9517-441e-944c-9d5286f5735c\") " pod="metallb-system/speaker-47524" Oct 08 18:29:39 crc kubenswrapper[4859]: I1008 18:29:39.361936 4859 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-5b87458c9c-4krwv" Oct 08 18:29:39 crc kubenswrapper[4859]: I1008 18:29:39.370024 4859 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-bbphq\" (UniqueName: \"kubernetes.io/projected/7557fe67-9517-441e-944c-9d5286f5735c-kube-api-access-bbphq\") pod \"speaker-47524\" (UID: \"7557fe67-9517-441e-944c-9d5286f5735c\") " pod="metallb-system/speaker-47524" Oct 08 18:29:39 crc kubenswrapper[4859]: I1008 18:29:39.429828 4859 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/7f1dbe0e-c03a-4ec2-ba3c-7179d2c67e41-cert\") pod \"controller-68d546b9d8-cr5dk\" (UID: \"7f1dbe0e-c03a-4ec2-ba3c-7179d2c67e41\") " pod="metallb-system/controller-68d546b9d8-cr5dk" Oct 08 18:29:39 crc kubenswrapper[4859]: I1008 18:29:39.429875 4859 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-qppp2\" (UniqueName: \"kubernetes.io/projected/7f1dbe0e-c03a-4ec2-ba3c-7179d2c67e41-kube-api-access-qppp2\") pod \"controller-68d546b9d8-cr5dk\" (UID: \"7f1dbe0e-c03a-4ec2-ba3c-7179d2c67e41\") " pod="metallb-system/controller-68d546b9d8-cr5dk" Oct 08 18:29:39 crc kubenswrapper[4859]: I1008 18:29:39.429905 4859 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/7f1dbe0e-c03a-4ec2-ba3c-7179d2c67e41-metrics-certs\") pod \"controller-68d546b9d8-cr5dk\" (UID: \"7f1dbe0e-c03a-4ec2-ba3c-7179d2c67e41\") " pod="metallb-system/controller-68d546b9d8-cr5dk" Oct 08 18:29:39 crc kubenswrapper[4859]: E1008 18:29:39.430099 4859 secret.go:188] Couldn't get secret metallb-system/controller-certs-secret: secret "controller-certs-secret" not found Oct 08 18:29:39 crc kubenswrapper[4859]: E1008 18:29:39.430176 4859 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/7f1dbe0e-c03a-4ec2-ba3c-7179d2c67e41-metrics-certs podName:7f1dbe0e-c03a-4ec2-ba3c-7179d2c67e41 nodeName:}" failed. No retries permitted until 2025-10-08 18:29:39.930155152 +0000 UTC m=+750.176994531 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/7f1dbe0e-c03a-4ec2-ba3c-7179d2c67e41-metrics-certs") pod "controller-68d546b9d8-cr5dk" (UID: "7f1dbe0e-c03a-4ec2-ba3c-7179d2c67e41") : secret "controller-certs-secret" not found Oct 08 18:29:39 crc kubenswrapper[4859]: I1008 18:29:39.434870 4859 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"metallb-webhook-cert" Oct 08 18:29:39 crc kubenswrapper[4859]: I1008 18:29:39.445171 4859 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert\" (UniqueName: \"kubernetes.io/secret/7f1dbe0e-c03a-4ec2-ba3c-7179d2c67e41-cert\") pod \"controller-68d546b9d8-cr5dk\" (UID: \"7f1dbe0e-c03a-4ec2-ba3c-7179d2c67e41\") " pod="metallb-system/controller-68d546b9d8-cr5dk" Oct 08 18:29:39 crc kubenswrapper[4859]: I1008 18:29:39.457110 4859 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-qppp2\" (UniqueName: \"kubernetes.io/projected/7f1dbe0e-c03a-4ec2-ba3c-7179d2c67e41-kube-api-access-qppp2\") pod \"controller-68d546b9d8-cr5dk\" (UID: \"7f1dbe0e-c03a-4ec2-ba3c-7179d2c67e41\") " pod="metallb-system/controller-68d546b9d8-cr5dk" Oct 08 18:29:39 crc kubenswrapper[4859]: I1008 18:29:39.611806 4859 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-5c55d4668-vqww5"] Oct 08 18:29:39 crc kubenswrapper[4859]: W1008 18:29:39.627911 4859 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podc31d0cee_7935_4c85_8461_3ddfd06a8c1d.slice/crio-84beb0fce3dc5dbc41af9f68245e4f457248563acb9c879cd2d6bd15d5545a2a WatchSource:0}: Error finding container 84beb0fce3dc5dbc41af9f68245e4f457248563acb9c879cd2d6bd15d5545a2a: Status 404 returned error can't find the container with id 84beb0fce3dc5dbc41af9f68245e4f457248563acb9c879cd2d6bd15d5545a2a Oct 08 18:29:39 crc kubenswrapper[4859]: I1008 18:29:39.633086 4859 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/4d682ed0-e5d8-48ad-8e52-cd29b2f65bd2-cert\") pod \"frr-k8s-webhook-server-64bf5d555-77xwq\" (UID: \"4d682ed0-e5d8-48ad-8e52-cd29b2f65bd2\") " pod="metallb-system/frr-k8s-webhook-server-64bf5d555-77xwq" Oct 08 18:29:39 crc kubenswrapper[4859]: I1008 18:29:39.643982 4859 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert\" (UniqueName: \"kubernetes.io/secret/4d682ed0-e5d8-48ad-8e52-cd29b2f65bd2-cert\") pod \"frr-k8s-webhook-server-64bf5d555-77xwq\" (UID: \"4d682ed0-e5d8-48ad-8e52-cd29b2f65bd2\") " pod="metallb-system/frr-k8s-webhook-server-64bf5d555-77xwq" Oct 08 18:29:39 crc kubenswrapper[4859]: I1008 18:29:39.657972 4859 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-controller-manager/controller-manager-5b87458c9c-4krwv"] Oct 08 18:29:39 crc kubenswrapper[4859]: I1008 18:29:39.835772 4859 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"memberlist\" (UniqueName: \"kubernetes.io/secret/7557fe67-9517-441e-944c-9d5286f5735c-memberlist\") pod \"speaker-47524\" (UID: \"7557fe67-9517-441e-944c-9d5286f5735c\") " pod="metallb-system/speaker-47524" Oct 08 18:29:39 crc kubenswrapper[4859]: I1008 18:29:39.836174 4859 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/7557fe67-9517-441e-944c-9d5286f5735c-metrics-certs\") pod \"speaker-47524\" (UID: \"7557fe67-9517-441e-944c-9d5286f5735c\") " pod="metallb-system/speaker-47524" Oct 08 18:29:39 crc kubenswrapper[4859]: E1008 18:29:39.837384 4859 secret.go:188] Couldn't get secret metallb-system/metallb-memberlist: secret "metallb-memberlist" not found Oct 08 18:29:39 crc kubenswrapper[4859]: E1008 18:29:39.837499 4859 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/7557fe67-9517-441e-944c-9d5286f5735c-memberlist podName:7557fe67-9517-441e-944c-9d5286f5735c nodeName:}" failed. No retries permitted until 2025-10-08 18:29:40.837475388 +0000 UTC m=+751.084314767 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "memberlist" (UniqueName: "kubernetes.io/secret/7557fe67-9517-441e-944c-9d5286f5735c-memberlist") pod "speaker-47524" (UID: "7557fe67-9517-441e-944c-9d5286f5735c") : secret "metallb-memberlist" not found Oct 08 18:29:39 crc kubenswrapper[4859]: I1008 18:29:39.840349 4859 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/7557fe67-9517-441e-944c-9d5286f5735c-metrics-certs\") pod \"speaker-47524\" (UID: \"7557fe67-9517-441e-944c-9d5286f5735c\") " pod="metallb-system/speaker-47524" Oct 08 18:29:39 crc kubenswrapper[4859]: I1008 18:29:39.891945 4859 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/frr-k8s-webhook-server-64bf5d555-77xwq" Oct 08 18:29:39 crc kubenswrapper[4859]: I1008 18:29:39.937792 4859 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/7f1dbe0e-c03a-4ec2-ba3c-7179d2c67e41-metrics-certs\") pod \"controller-68d546b9d8-cr5dk\" (UID: \"7f1dbe0e-c03a-4ec2-ba3c-7179d2c67e41\") " pod="metallb-system/controller-68d546b9d8-cr5dk" Oct 08 18:29:39 crc kubenswrapper[4859]: I1008 18:29:39.942245 4859 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/7f1dbe0e-c03a-4ec2-ba3c-7179d2c67e41-metrics-certs\") pod \"controller-68d546b9d8-cr5dk\" (UID: \"7f1dbe0e-c03a-4ec2-ba3c-7179d2c67e41\") " pod="metallb-system/controller-68d546b9d8-cr5dk" Oct 08 18:29:40 crc kubenswrapper[4859]: I1008 18:29:40.177929 4859 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/controller-68d546b9d8-cr5dk" Oct 08 18:29:40 crc kubenswrapper[4859]: I1008 18:29:40.189752 4859 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["metallb-system/frr-k8s-webhook-server-64bf5d555-77xwq"] Oct 08 18:29:40 crc kubenswrapper[4859]: W1008 18:29:40.195485 4859 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod4d682ed0_e5d8_48ad_8e52_cd29b2f65bd2.slice/crio-96171b543d7dcac1a40afc3a06500c239e637e3e4f5b3c107d1cb8fa0238b979 WatchSource:0}: Error finding container 96171b543d7dcac1a40afc3a06500c239e637e3e4f5b3c107d1cb8fa0238b979: Status 404 returned error can't find the container with id 96171b543d7dcac1a40afc3a06500c239e637e3e4f5b3c107d1cb8fa0238b979 Oct 08 18:29:40 crc kubenswrapper[4859]: I1008 18:29:40.197534 4859 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-5b87458c9c-4krwv" event={"ID":"0d0c4bdf-fe24-4589-9fed-d6d81cfc25b8","Type":"ContainerStarted","Data":"5d9762fa8aec5cbb2529ea189fbd2753ccdd362ad8d2f939031ca8cc6a8ca52e"} Oct 08 18:29:40 crc kubenswrapper[4859]: I1008 18:29:40.197589 4859 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-5b87458c9c-4krwv" event={"ID":"0d0c4bdf-fe24-4589-9fed-d6d81cfc25b8","Type":"ContainerStarted","Data":"91adc3d6c8511b92aaa9ad829653db5b494b8a10af830ae2158312538dc74d44"} Oct 08 18:29:40 crc kubenswrapper[4859]: I1008 18:29:40.198067 4859 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-controller-manager/controller-manager-5b87458c9c-4krwv" Oct 08 18:29:40 crc kubenswrapper[4859]: I1008 18:29:40.199488 4859 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-mfcd8" event={"ID":"69cc95cc-1a58-40c2-9c84-4808c217c2e5","Type":"ContainerStarted","Data":"dcf40357854726409e06b6e6e225b4e3cc494a94627da3ec5821596f33cfe85e"} Oct 08 18:29:40 crc kubenswrapper[4859]: I1008 18:29:40.203767 4859 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-controller-manager/controller-manager-5b87458c9c-4krwv" Oct 08 18:29:40 crc kubenswrapper[4859]: I1008 18:29:40.203963 4859 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-5c55d4668-vqww5" event={"ID":"c31d0cee-7935-4c85-8461-3ddfd06a8c1d","Type":"ContainerStarted","Data":"897e30c5bd369af24075edc07f8f6247b794fa23c8388c5afa2fdb279c42d3c6"} Oct 08 18:29:40 crc kubenswrapper[4859]: I1008 18:29:40.203999 4859 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-5c55d4668-vqww5" event={"ID":"c31d0cee-7935-4c85-8461-3ddfd06a8c1d","Type":"ContainerStarted","Data":"84beb0fce3dc5dbc41af9f68245e4f457248563acb9c879cd2d6bd15d5545a2a"} Oct 08 18:29:40 crc kubenswrapper[4859]: I1008 18:29:40.204269 4859 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-route-controller-manager/route-controller-manager-5c55d4668-vqww5" Oct 08 18:29:40 crc kubenswrapper[4859]: I1008 18:29:40.266636 4859 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-route-controller-manager/route-controller-manager-5c55d4668-vqww5" podStartSLOduration=2.266613714 podStartE2EDuration="2.266613714s" podCreationTimestamp="2025-10-08 18:29:38 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-08 18:29:40.262439752 +0000 UTC m=+750.509279141" watchObservedRunningTime="2025-10-08 18:29:40.266613714 +0000 UTC m=+750.513453093" Oct 08 18:29:40 crc kubenswrapper[4859]: I1008 18:29:40.268427 4859 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-controller-manager/controller-manager-5b87458c9c-4krwv" podStartSLOduration=3.268409287 podStartE2EDuration="3.268409287s" podCreationTimestamp="2025-10-08 18:29:37 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-08 18:29:40.239648364 +0000 UTC m=+750.486487743" watchObservedRunningTime="2025-10-08 18:29:40.268409287 +0000 UTC m=+750.515248676" Oct 08 18:29:40 crc kubenswrapper[4859]: I1008 18:29:40.333593 4859 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-route-controller-manager/route-controller-manager-5c55d4668-vqww5" Oct 08 18:29:40 crc kubenswrapper[4859]: I1008 18:29:40.701354 4859 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["metallb-system/controller-68d546b9d8-cr5dk"] Oct 08 18:29:40 crc kubenswrapper[4859]: W1008 18:29:40.713238 4859 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod7f1dbe0e_c03a_4ec2_ba3c_7179d2c67e41.slice/crio-edff858684c08b954208039275350d441b94915d7ab48fdad5f8a1497387c1df WatchSource:0}: Error finding container edff858684c08b954208039275350d441b94915d7ab48fdad5f8a1497387c1df: Status 404 returned error can't find the container with id edff858684c08b954208039275350d441b94915d7ab48fdad5f8a1497387c1df Oct 08 18:29:40 crc kubenswrapper[4859]: I1008 18:29:40.857388 4859 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"memberlist\" (UniqueName: \"kubernetes.io/secret/7557fe67-9517-441e-944c-9d5286f5735c-memberlist\") pod \"speaker-47524\" (UID: \"7557fe67-9517-441e-944c-9d5286f5735c\") " pod="metallb-system/speaker-47524" Oct 08 18:29:40 crc kubenswrapper[4859]: E1008 18:29:40.857627 4859 secret.go:188] Couldn't get secret metallb-system/metallb-memberlist: secret "metallb-memberlist" not found Oct 08 18:29:40 crc kubenswrapper[4859]: E1008 18:29:40.857742 4859 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/7557fe67-9517-441e-944c-9d5286f5735c-memberlist podName:7557fe67-9517-441e-944c-9d5286f5735c nodeName:}" failed. No retries permitted until 2025-10-08 18:29:42.857721696 +0000 UTC m=+753.104561065 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "memberlist" (UniqueName: "kubernetes.io/secret/7557fe67-9517-441e-944c-9d5286f5735c-memberlist") pod "speaker-47524" (UID: "7557fe67-9517-441e-944c-9d5286f5735c") : secret "metallb-memberlist" not found Oct 08 18:29:41 crc kubenswrapper[4859]: I1008 18:29:41.211572 4859 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-webhook-server-64bf5d555-77xwq" event={"ID":"4d682ed0-e5d8-48ad-8e52-cd29b2f65bd2","Type":"ContainerStarted","Data":"96171b543d7dcac1a40afc3a06500c239e637e3e4f5b3c107d1cb8fa0238b979"} Oct 08 18:29:41 crc kubenswrapper[4859]: I1008 18:29:41.213743 4859 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/controller-68d546b9d8-cr5dk" event={"ID":"7f1dbe0e-c03a-4ec2-ba3c-7179d2c67e41","Type":"ContainerStarted","Data":"c7e29a1659d75d4d4b074cbdb529ea8721a4d49e9e341da975fb36a4017eb70b"} Oct 08 18:29:41 crc kubenswrapper[4859]: I1008 18:29:41.213911 4859 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/controller-68d546b9d8-cr5dk" event={"ID":"7f1dbe0e-c03a-4ec2-ba3c-7179d2c67e41","Type":"ContainerStarted","Data":"2d79f5ad684e1145acd055adf271f578f39a2896ef53773b57b2c9b8fb707339"} Oct 08 18:29:41 crc kubenswrapper[4859]: I1008 18:29:41.214011 4859 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/controller-68d546b9d8-cr5dk" event={"ID":"7f1dbe0e-c03a-4ec2-ba3c-7179d2c67e41","Type":"ContainerStarted","Data":"edff858684c08b954208039275350d441b94915d7ab48fdad5f8a1497387c1df"} Oct 08 18:29:41 crc kubenswrapper[4859]: I1008 18:29:41.214259 4859 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="metallb-system/controller-68d546b9d8-cr5dk" Oct 08 18:29:41 crc kubenswrapper[4859]: I1008 18:29:41.253221 4859 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="metallb-system/controller-68d546b9d8-cr5dk" podStartSLOduration=2.253198475 podStartE2EDuration="2.253198475s" podCreationTimestamp="2025-10-08 18:29:39 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-08 18:29:41.247871409 +0000 UTC m=+751.494710808" watchObservedRunningTime="2025-10-08 18:29:41.253198475 +0000 UTC m=+751.500037854" Oct 08 18:29:42 crc kubenswrapper[4859]: I1008 18:29:42.888601 4859 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"memberlist\" (UniqueName: \"kubernetes.io/secret/7557fe67-9517-441e-944c-9d5286f5735c-memberlist\") pod \"speaker-47524\" (UID: \"7557fe67-9517-441e-944c-9d5286f5735c\") " pod="metallb-system/speaker-47524" Oct 08 18:29:42 crc kubenswrapper[4859]: I1008 18:29:42.896370 4859 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"memberlist\" (UniqueName: \"kubernetes.io/secret/7557fe67-9517-441e-944c-9d5286f5735c-memberlist\") pod \"speaker-47524\" (UID: \"7557fe67-9517-441e-944c-9d5286f5735c\") " pod="metallb-system/speaker-47524" Oct 08 18:29:43 crc kubenswrapper[4859]: I1008 18:29:43.150996 4859 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/speaker-47524" Oct 08 18:29:43 crc kubenswrapper[4859]: W1008 18:29:43.174336 4859 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod7557fe67_9517_441e_944c_9d5286f5735c.slice/crio-247f7ed6fc7ff1e9ad0d2c332f89f7455b29edfe194579181facd5dc20d9aad7 WatchSource:0}: Error finding container 247f7ed6fc7ff1e9ad0d2c332f89f7455b29edfe194579181facd5dc20d9aad7: Status 404 returned error can't find the container with id 247f7ed6fc7ff1e9ad0d2c332f89f7455b29edfe194579181facd5dc20d9aad7 Oct 08 18:29:43 crc kubenswrapper[4859]: I1008 18:29:43.250403 4859 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/speaker-47524" event={"ID":"7557fe67-9517-441e-944c-9d5286f5735c","Type":"ContainerStarted","Data":"247f7ed6fc7ff1e9ad0d2c332f89f7455b29edfe194579181facd5dc20d9aad7"} Oct 08 18:29:44 crc kubenswrapper[4859]: I1008 18:29:44.263411 4859 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/speaker-47524" event={"ID":"7557fe67-9517-441e-944c-9d5286f5735c","Type":"ContainerStarted","Data":"eaa2404ff42f611d9de786327532580218326a95c36b34c1d791a4e79181f157"} Oct 08 18:29:44 crc kubenswrapper[4859]: I1008 18:29:44.263831 4859 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="metallb-system/speaker-47524" Oct 08 18:29:44 crc kubenswrapper[4859]: I1008 18:29:44.263843 4859 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/speaker-47524" event={"ID":"7557fe67-9517-441e-944c-9d5286f5735c","Type":"ContainerStarted","Data":"58974c21ba0ce7362294c888da3a8facfd7f02075c578a4f02849d3b7c63c7eb"} Oct 08 18:29:44 crc kubenswrapper[4859]: I1008 18:29:44.296023 4859 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="metallb-system/speaker-47524" podStartSLOduration=5.295988464 podStartE2EDuration="5.295988464s" podCreationTimestamp="2025-10-08 18:29:39 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-08 18:29:44.282672924 +0000 UTC m=+754.529512303" watchObservedRunningTime="2025-10-08 18:29:44.295988464 +0000 UTC m=+754.542827843" Oct 08 18:29:45 crc kubenswrapper[4859]: I1008 18:29:45.083520 4859 dynamic_cafile_content.go:123] "Loaded a new CA Bundle and Verifier" name="client-ca-bundle::/etc/kubernetes/kubelet-ca.crt" Oct 08 18:29:47 crc kubenswrapper[4859]: I1008 18:29:47.708013 4859 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-rw68k"] Oct 08 18:29:47 crc kubenswrapper[4859]: I1008 18:29:47.709347 4859 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-rw68k" Oct 08 18:29:47 crc kubenswrapper[4859]: I1008 18:29:47.725187 4859 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-rw68k"] Oct 08 18:29:47 crc kubenswrapper[4859]: I1008 18:29:47.774678 4859 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/453a83ad-e06c-4485-bba7-524b7128462f-utilities\") pod \"community-operators-rw68k\" (UID: \"453a83ad-e06c-4485-bba7-524b7128462f\") " pod="openshift-marketplace/community-operators-rw68k" Oct 08 18:29:47 crc kubenswrapper[4859]: I1008 18:29:47.774786 4859 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-w8tlg\" (UniqueName: \"kubernetes.io/projected/453a83ad-e06c-4485-bba7-524b7128462f-kube-api-access-w8tlg\") pod \"community-operators-rw68k\" (UID: \"453a83ad-e06c-4485-bba7-524b7128462f\") " pod="openshift-marketplace/community-operators-rw68k" Oct 08 18:29:47 crc kubenswrapper[4859]: I1008 18:29:47.774877 4859 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/453a83ad-e06c-4485-bba7-524b7128462f-catalog-content\") pod \"community-operators-rw68k\" (UID: \"453a83ad-e06c-4485-bba7-524b7128462f\") " pod="openshift-marketplace/community-operators-rw68k" Oct 08 18:29:47 crc kubenswrapper[4859]: I1008 18:29:47.876143 4859 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/453a83ad-e06c-4485-bba7-524b7128462f-catalog-content\") pod \"community-operators-rw68k\" (UID: \"453a83ad-e06c-4485-bba7-524b7128462f\") " pod="openshift-marketplace/community-operators-rw68k" Oct 08 18:29:47 crc kubenswrapper[4859]: I1008 18:29:47.876231 4859 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/453a83ad-e06c-4485-bba7-524b7128462f-utilities\") pod \"community-operators-rw68k\" (UID: \"453a83ad-e06c-4485-bba7-524b7128462f\") " pod="openshift-marketplace/community-operators-rw68k" Oct 08 18:29:47 crc kubenswrapper[4859]: I1008 18:29:47.876280 4859 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-w8tlg\" (UniqueName: \"kubernetes.io/projected/453a83ad-e06c-4485-bba7-524b7128462f-kube-api-access-w8tlg\") pod \"community-operators-rw68k\" (UID: \"453a83ad-e06c-4485-bba7-524b7128462f\") " pod="openshift-marketplace/community-operators-rw68k" Oct 08 18:29:47 crc kubenswrapper[4859]: I1008 18:29:47.876903 4859 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/453a83ad-e06c-4485-bba7-524b7128462f-utilities\") pod \"community-operators-rw68k\" (UID: \"453a83ad-e06c-4485-bba7-524b7128462f\") " pod="openshift-marketplace/community-operators-rw68k" Oct 08 18:29:47 crc kubenswrapper[4859]: I1008 18:29:47.877207 4859 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/453a83ad-e06c-4485-bba7-524b7128462f-catalog-content\") pod \"community-operators-rw68k\" (UID: \"453a83ad-e06c-4485-bba7-524b7128462f\") " pod="openshift-marketplace/community-operators-rw68k" Oct 08 18:29:47 crc kubenswrapper[4859]: I1008 18:29:47.899056 4859 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-w8tlg\" (UniqueName: \"kubernetes.io/projected/453a83ad-e06c-4485-bba7-524b7128462f-kube-api-access-w8tlg\") pod \"community-operators-rw68k\" (UID: \"453a83ad-e06c-4485-bba7-524b7128462f\") " pod="openshift-marketplace/community-operators-rw68k" Oct 08 18:29:48 crc kubenswrapper[4859]: I1008 18:29:48.027136 4859 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-rw68k" Oct 08 18:29:48 crc kubenswrapper[4859]: I1008 18:29:48.311222 4859 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-webhook-server-64bf5d555-77xwq" event={"ID":"4d682ed0-e5d8-48ad-8e52-cd29b2f65bd2","Type":"ContainerStarted","Data":"da8b2ded8463474b112cd660deb15817b526839d3444cabfe365811362c56ef8"} Oct 08 18:29:48 crc kubenswrapper[4859]: I1008 18:29:48.311410 4859 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="metallb-system/frr-k8s-webhook-server-64bf5d555-77xwq" Oct 08 18:29:48 crc kubenswrapper[4859]: I1008 18:29:48.316958 4859 generic.go:334] "Generic (PLEG): container finished" podID="69cc95cc-1a58-40c2-9c84-4808c217c2e5" containerID="c8d2a19dc472e6366b5206dd39e37f0a61405c2dc931779c7cd46599e571f4cd" exitCode=0 Oct 08 18:29:48 crc kubenswrapper[4859]: I1008 18:29:48.317000 4859 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-mfcd8" event={"ID":"69cc95cc-1a58-40c2-9c84-4808c217c2e5","Type":"ContainerDied","Data":"c8d2a19dc472e6366b5206dd39e37f0a61405c2dc931779c7cd46599e571f4cd"} Oct 08 18:29:48 crc kubenswrapper[4859]: I1008 18:29:48.336298 4859 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="metallb-system/frr-k8s-webhook-server-64bf5d555-77xwq" podStartSLOduration=3.080933167 podStartE2EDuration="10.336273343s" podCreationTimestamp="2025-10-08 18:29:38 +0000 UTC" firstStartedPulling="2025-10-08 18:29:40.204446193 +0000 UTC m=+750.451285582" lastFinishedPulling="2025-10-08 18:29:47.459786379 +0000 UTC m=+757.706625758" observedRunningTime="2025-10-08 18:29:48.330420502 +0000 UTC m=+758.577259881" watchObservedRunningTime="2025-10-08 18:29:48.336273343 +0000 UTC m=+758.583112722" Oct 08 18:29:48 crc kubenswrapper[4859]: I1008 18:29:48.520393 4859 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-rw68k"] Oct 08 18:29:48 crc kubenswrapper[4859]: W1008 18:29:48.525860 4859 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod453a83ad_e06c_4485_bba7_524b7128462f.slice/crio-37bdb42b252deeede466ca0ad9ac6d183f362902426a666760327cc69b1726ab WatchSource:0}: Error finding container 37bdb42b252deeede466ca0ad9ac6d183f362902426a666760327cc69b1726ab: Status 404 returned error can't find the container with id 37bdb42b252deeede466ca0ad9ac6d183f362902426a666760327cc69b1726ab Oct 08 18:29:49 crc kubenswrapper[4859]: I1008 18:29:49.326500 4859 generic.go:334] "Generic (PLEG): container finished" podID="69cc95cc-1a58-40c2-9c84-4808c217c2e5" containerID="22a775d34adb4cf9ea887a2911e206f6bfba9354cf459f3495c9a30e1a2ffa63" exitCode=0 Oct 08 18:29:49 crc kubenswrapper[4859]: I1008 18:29:49.326555 4859 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-mfcd8" event={"ID":"69cc95cc-1a58-40c2-9c84-4808c217c2e5","Type":"ContainerDied","Data":"22a775d34adb4cf9ea887a2911e206f6bfba9354cf459f3495c9a30e1a2ffa63"} Oct 08 18:29:49 crc kubenswrapper[4859]: I1008 18:29:49.328492 4859 generic.go:334] "Generic (PLEG): container finished" podID="453a83ad-e06c-4485-bba7-524b7128462f" containerID="81a60605b657849fe152794e839a2430aeff961f7243f05721bc11729b9978fe" exitCode=0 Oct 08 18:29:49 crc kubenswrapper[4859]: I1008 18:29:49.328621 4859 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-rw68k" event={"ID":"453a83ad-e06c-4485-bba7-524b7128462f","Type":"ContainerDied","Data":"81a60605b657849fe152794e839a2430aeff961f7243f05721bc11729b9978fe"} Oct 08 18:29:49 crc kubenswrapper[4859]: I1008 18:29:49.328664 4859 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-rw68k" event={"ID":"453a83ad-e06c-4485-bba7-524b7128462f","Type":"ContainerStarted","Data":"37bdb42b252deeede466ca0ad9ac6d183f362902426a666760327cc69b1726ab"} Oct 08 18:29:50 crc kubenswrapper[4859]: I1008 18:29:50.182573 4859 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="metallb-system/controller-68d546b9d8-cr5dk" Oct 08 18:29:50 crc kubenswrapper[4859]: I1008 18:29:50.340965 4859 generic.go:334] "Generic (PLEG): container finished" podID="69cc95cc-1a58-40c2-9c84-4808c217c2e5" containerID="985de5d81925a1d16a700f3177163c0b2e461fb0ba196dd06b0edf30d5cea466" exitCode=0 Oct 08 18:29:50 crc kubenswrapper[4859]: I1008 18:29:50.341034 4859 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-mfcd8" event={"ID":"69cc95cc-1a58-40c2-9c84-4808c217c2e5","Type":"ContainerDied","Data":"985de5d81925a1d16a700f3177163c0b2e461fb0ba196dd06b0edf30d5cea466"} Oct 08 18:29:50 crc kubenswrapper[4859]: I1008 18:29:50.343516 4859 generic.go:334] "Generic (PLEG): container finished" podID="453a83ad-e06c-4485-bba7-524b7128462f" containerID="00337353f6a1de8a59cfbd400f17ad2da3b4dadf1240a970d8c7175b49a46e8c" exitCode=0 Oct 08 18:29:50 crc kubenswrapper[4859]: I1008 18:29:50.343549 4859 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-rw68k" event={"ID":"453a83ad-e06c-4485-bba7-524b7128462f","Type":"ContainerDied","Data":"00337353f6a1de8a59cfbd400f17ad2da3b4dadf1240a970d8c7175b49a46e8c"} Oct 08 18:29:50 crc kubenswrapper[4859]: E1008 18:29:50.404671 4859 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod453a83ad_e06c_4485_bba7_524b7128462f.slice/crio-conmon-00337353f6a1de8a59cfbd400f17ad2da3b4dadf1240a970d8c7175b49a46e8c.scope\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod453a83ad_e06c_4485_bba7_524b7128462f.slice/crio-00337353f6a1de8a59cfbd400f17ad2da3b4dadf1240a970d8c7175b49a46e8c.scope\": RecentStats: unable to find data in memory cache]" Oct 08 18:29:51 crc kubenswrapper[4859]: I1008 18:29:51.359979 4859 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-mfcd8" event={"ID":"69cc95cc-1a58-40c2-9c84-4808c217c2e5","Type":"ContainerStarted","Data":"1a888d73bc9c14095ab8809d95d6621e305e647574246fd8ea6d7ffa9b81016d"} Oct 08 18:29:51 crc kubenswrapper[4859]: I1008 18:29:51.360407 4859 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-mfcd8" event={"ID":"69cc95cc-1a58-40c2-9c84-4808c217c2e5","Type":"ContainerStarted","Data":"b63d6cc614d446d7694bb875c5d72f7fe35600f1965cc6959356f172620b643b"} Oct 08 18:29:51 crc kubenswrapper[4859]: I1008 18:29:51.360425 4859 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-mfcd8" event={"ID":"69cc95cc-1a58-40c2-9c84-4808c217c2e5","Type":"ContainerStarted","Data":"59b5d1098247acbf52f2eff84fa08d9948fcc13085dab24d8365cc89f75cc5ca"} Oct 08 18:29:51 crc kubenswrapper[4859]: I1008 18:29:51.360438 4859 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-mfcd8" event={"ID":"69cc95cc-1a58-40c2-9c84-4808c217c2e5","Type":"ContainerStarted","Data":"ef3a3fcb00e591c16a62871dd3b0012337b615f92da33a4d16821bec2a6ec2b5"} Oct 08 18:29:51 crc kubenswrapper[4859]: I1008 18:29:51.360450 4859 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-mfcd8" event={"ID":"69cc95cc-1a58-40c2-9c84-4808c217c2e5","Type":"ContainerStarted","Data":"15bd416b9dd03856eaa9016d86526064167726af6b1dd39a515e9eeb2c9a2e5d"} Oct 08 18:29:51 crc kubenswrapper[4859]: I1008 18:29:51.363861 4859 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-rw68k" event={"ID":"453a83ad-e06c-4485-bba7-524b7128462f","Type":"ContainerStarted","Data":"9fb031553cfab809d6b97efdae195d8a0f6b611557913e13da708cfdf26c41bc"} Oct 08 18:29:51 crc kubenswrapper[4859]: I1008 18:29:51.382151 4859 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-rw68k" podStartSLOduration=2.95330432 podStartE2EDuration="4.382132362s" podCreationTimestamp="2025-10-08 18:29:47 +0000 UTC" firstStartedPulling="2025-10-08 18:29:49.329747987 +0000 UTC m=+759.576587386" lastFinishedPulling="2025-10-08 18:29:50.758576039 +0000 UTC m=+761.005415428" observedRunningTime="2025-10-08 18:29:51.381393211 +0000 UTC m=+761.628232610" watchObservedRunningTime="2025-10-08 18:29:51.382132362 +0000 UTC m=+761.628971741" Oct 08 18:29:52 crc kubenswrapper[4859]: I1008 18:29:52.378155 4859 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-mfcd8" event={"ID":"69cc95cc-1a58-40c2-9c84-4808c217c2e5","Type":"ContainerStarted","Data":"2f3e3f3bb488bcc0a0246751417382f222520e2f085da0c91437f33798b98395"} Oct 08 18:29:52 crc kubenswrapper[4859]: I1008 18:29:52.378337 4859 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="metallb-system/frr-k8s-mfcd8" Oct 08 18:29:52 crc kubenswrapper[4859]: I1008 18:29:52.407679 4859 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="metallb-system/frr-k8s-mfcd8" podStartSLOduration=6.423770278 podStartE2EDuration="14.407658694s" podCreationTimestamp="2025-10-08 18:29:38 +0000 UTC" firstStartedPulling="2025-10-08 18:29:39.443892425 +0000 UTC m=+749.690731804" lastFinishedPulling="2025-10-08 18:29:47.427780831 +0000 UTC m=+757.674620220" observedRunningTime="2025-10-08 18:29:52.40443849 +0000 UTC m=+762.651277869" watchObservedRunningTime="2025-10-08 18:29:52.407658694 +0000 UTC m=+762.654498073" Oct 08 18:29:53 crc kubenswrapper[4859]: I1008 18:29:53.158576 4859 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="metallb-system/speaker-47524" Oct 08 18:29:54 crc kubenswrapper[4859]: I1008 18:29:54.270861 4859 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="metallb-system/frr-k8s-mfcd8" Oct 08 18:29:54 crc kubenswrapper[4859]: I1008 18:29:54.308233 4859 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="metallb-system/frr-k8s-mfcd8" Oct 08 18:29:55 crc kubenswrapper[4859]: I1008 18:29:55.381110 4859 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-mqxtr"] Oct 08 18:29:55 crc kubenswrapper[4859]: I1008 18:29:55.382530 4859 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-mqxtr" Oct 08 18:29:55 crc kubenswrapper[4859]: I1008 18:29:55.408931 4859 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-mqxtr"] Oct 08 18:29:55 crc kubenswrapper[4859]: I1008 18:29:55.494245 4859 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/95364dee-6272-4f36-b265-139af1823b2f-catalog-content\") pod \"redhat-marketplace-mqxtr\" (UID: \"95364dee-6272-4f36-b265-139af1823b2f\") " pod="openshift-marketplace/redhat-marketplace-mqxtr" Oct 08 18:29:55 crc kubenswrapper[4859]: I1008 18:29:55.494311 4859 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-qskx8\" (UniqueName: \"kubernetes.io/projected/95364dee-6272-4f36-b265-139af1823b2f-kube-api-access-qskx8\") pod \"redhat-marketplace-mqxtr\" (UID: \"95364dee-6272-4f36-b265-139af1823b2f\") " pod="openshift-marketplace/redhat-marketplace-mqxtr" Oct 08 18:29:55 crc kubenswrapper[4859]: I1008 18:29:55.494340 4859 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/95364dee-6272-4f36-b265-139af1823b2f-utilities\") pod \"redhat-marketplace-mqxtr\" (UID: \"95364dee-6272-4f36-b265-139af1823b2f\") " pod="openshift-marketplace/redhat-marketplace-mqxtr" Oct 08 18:29:55 crc kubenswrapper[4859]: I1008 18:29:55.596146 4859 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/95364dee-6272-4f36-b265-139af1823b2f-catalog-content\") pod \"redhat-marketplace-mqxtr\" (UID: \"95364dee-6272-4f36-b265-139af1823b2f\") " pod="openshift-marketplace/redhat-marketplace-mqxtr" Oct 08 18:29:55 crc kubenswrapper[4859]: I1008 18:29:55.596235 4859 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-qskx8\" (UniqueName: \"kubernetes.io/projected/95364dee-6272-4f36-b265-139af1823b2f-kube-api-access-qskx8\") pod \"redhat-marketplace-mqxtr\" (UID: \"95364dee-6272-4f36-b265-139af1823b2f\") " pod="openshift-marketplace/redhat-marketplace-mqxtr" Oct 08 18:29:55 crc kubenswrapper[4859]: I1008 18:29:55.596283 4859 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/95364dee-6272-4f36-b265-139af1823b2f-utilities\") pod \"redhat-marketplace-mqxtr\" (UID: \"95364dee-6272-4f36-b265-139af1823b2f\") " pod="openshift-marketplace/redhat-marketplace-mqxtr" Oct 08 18:29:55 crc kubenswrapper[4859]: I1008 18:29:55.596978 4859 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/95364dee-6272-4f36-b265-139af1823b2f-catalog-content\") pod \"redhat-marketplace-mqxtr\" (UID: \"95364dee-6272-4f36-b265-139af1823b2f\") " pod="openshift-marketplace/redhat-marketplace-mqxtr" Oct 08 18:29:55 crc kubenswrapper[4859]: I1008 18:29:55.597178 4859 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/95364dee-6272-4f36-b265-139af1823b2f-utilities\") pod \"redhat-marketplace-mqxtr\" (UID: \"95364dee-6272-4f36-b265-139af1823b2f\") " pod="openshift-marketplace/redhat-marketplace-mqxtr" Oct 08 18:29:55 crc kubenswrapper[4859]: I1008 18:29:55.630268 4859 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-qskx8\" (UniqueName: \"kubernetes.io/projected/95364dee-6272-4f36-b265-139af1823b2f-kube-api-access-qskx8\") pod \"redhat-marketplace-mqxtr\" (UID: \"95364dee-6272-4f36-b265-139af1823b2f\") " pod="openshift-marketplace/redhat-marketplace-mqxtr" Oct 08 18:29:55 crc kubenswrapper[4859]: I1008 18:29:55.710117 4859 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-mqxtr" Oct 08 18:29:56 crc kubenswrapper[4859]: I1008 18:29:56.153883 4859 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-mqxtr"] Oct 08 18:29:56 crc kubenswrapper[4859]: I1008 18:29:56.411142 4859 generic.go:334] "Generic (PLEG): container finished" podID="95364dee-6272-4f36-b265-139af1823b2f" containerID="4aed0306378fe02319bdb45b1ea1d11659b097d1af1456c3f0d11256e9423c5a" exitCode=0 Oct 08 18:29:56 crc kubenswrapper[4859]: I1008 18:29:56.411277 4859 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-mqxtr" event={"ID":"95364dee-6272-4f36-b265-139af1823b2f","Type":"ContainerDied","Data":"4aed0306378fe02319bdb45b1ea1d11659b097d1af1456c3f0d11256e9423c5a"} Oct 08 18:29:56 crc kubenswrapper[4859]: I1008 18:29:56.411664 4859 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-mqxtr" event={"ID":"95364dee-6272-4f36-b265-139af1823b2f","Type":"ContainerStarted","Data":"b909a71c2a25df1ccfa5dfe3e4462cc1d8f0554231e4b84e2aef2296cedae777"} Oct 08 18:29:58 crc kubenswrapper[4859]: I1008 18:29:58.028283 4859 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-rw68k" Oct 08 18:29:58 crc kubenswrapper[4859]: I1008 18:29:58.029012 4859 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-rw68k" Oct 08 18:29:58 crc kubenswrapper[4859]: I1008 18:29:58.086176 4859 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-rw68k" Oct 08 18:29:58 crc kubenswrapper[4859]: I1008 18:29:58.433616 4859 generic.go:334] "Generic (PLEG): container finished" podID="95364dee-6272-4f36-b265-139af1823b2f" containerID="79c64bd519aa17ec98f7eedfee51a09a08351a3c1948f72ee80c20b1dbeb3cc9" exitCode=0 Oct 08 18:29:58 crc kubenswrapper[4859]: I1008 18:29:58.433729 4859 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-mqxtr" event={"ID":"95364dee-6272-4f36-b265-139af1823b2f","Type":"ContainerDied","Data":"79c64bd519aa17ec98f7eedfee51a09a08351a3c1948f72ee80c20b1dbeb3cc9"} Oct 08 18:29:58 crc kubenswrapper[4859]: I1008 18:29:58.501924 4859 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-rw68k" Oct 08 18:29:59 crc kubenswrapper[4859]: I1008 18:29:59.375467 4859 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/openstack-operator-index-z8f5z"] Oct 08 18:29:59 crc kubenswrapper[4859]: I1008 18:29:59.376890 4859 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-operator-index-z8f5z" Oct 08 18:29:59 crc kubenswrapper[4859]: I1008 18:29:59.379769 4859 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"openstack-operator-index-dockercfg-4k8wq" Oct 08 18:29:59 crc kubenswrapper[4859]: I1008 18:29:59.380054 4859 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack-operators"/"kube-root-ca.crt" Oct 08 18:29:59 crc kubenswrapper[4859]: I1008 18:29:59.382076 4859 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack-operators"/"openshift-service-ca.crt" Oct 08 18:29:59 crc kubenswrapper[4859]: I1008 18:29:59.383806 4859 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/openstack-operator-index-z8f5z"] Oct 08 18:29:59 crc kubenswrapper[4859]: I1008 18:29:59.459343 4859 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-mqxtr" event={"ID":"95364dee-6272-4f36-b265-139af1823b2f","Type":"ContainerStarted","Data":"d7799dc202759262536424e4806e4acb338441b91eb0d76626da83855dd30d75"} Oct 08 18:29:59 crc kubenswrapper[4859]: I1008 18:29:59.487458 4859 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-mqxtr" podStartSLOduration=1.935226434 podStartE2EDuration="4.487430266s" podCreationTimestamp="2025-10-08 18:29:55 +0000 UTC" firstStartedPulling="2025-10-08 18:29:56.414322009 +0000 UTC m=+766.661161388" lastFinishedPulling="2025-10-08 18:29:58.966525841 +0000 UTC m=+769.213365220" observedRunningTime="2025-10-08 18:29:59.485323674 +0000 UTC m=+769.732163073" watchObservedRunningTime="2025-10-08 18:29:59.487430266 +0000 UTC m=+769.734269645" Oct 08 18:29:59 crc kubenswrapper[4859]: I1008 18:29:59.552045 4859 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-4f2k4\" (UniqueName: \"kubernetes.io/projected/6c1463a8-c973-42db-8671-f274a513230a-kube-api-access-4f2k4\") pod \"openstack-operator-index-z8f5z\" (UID: \"6c1463a8-c973-42db-8671-f274a513230a\") " pod="openstack-operators/openstack-operator-index-z8f5z" Oct 08 18:29:59 crc kubenswrapper[4859]: I1008 18:29:59.654353 4859 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-4f2k4\" (UniqueName: \"kubernetes.io/projected/6c1463a8-c973-42db-8671-f274a513230a-kube-api-access-4f2k4\") pod \"openstack-operator-index-z8f5z\" (UID: \"6c1463a8-c973-42db-8671-f274a513230a\") " pod="openstack-operators/openstack-operator-index-z8f5z" Oct 08 18:29:59 crc kubenswrapper[4859]: I1008 18:29:59.680662 4859 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-4f2k4\" (UniqueName: \"kubernetes.io/projected/6c1463a8-c973-42db-8671-f274a513230a-kube-api-access-4f2k4\") pod \"openstack-operator-index-z8f5z\" (UID: \"6c1463a8-c973-42db-8671-f274a513230a\") " pod="openstack-operators/openstack-operator-index-z8f5z" Oct 08 18:29:59 crc kubenswrapper[4859]: I1008 18:29:59.712647 4859 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-operator-index-z8f5z" Oct 08 18:29:59 crc kubenswrapper[4859]: I1008 18:29:59.901064 4859 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="metallb-system/frr-k8s-webhook-server-64bf5d555-77xwq" Oct 08 18:30:00 crc kubenswrapper[4859]: I1008 18:30:00.144481 4859 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29332470-sc2qz"] Oct 08 18:30:00 crc kubenswrapper[4859]: I1008 18:30:00.145387 4859 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29332470-sc2qz" Oct 08 18:30:00 crc kubenswrapper[4859]: I1008 18:30:00.149427 4859 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"collect-profiles-dockercfg-kzf4t" Oct 08 18:30:00 crc kubenswrapper[4859]: I1008 18:30:00.149973 4859 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"collect-profiles-config" Oct 08 18:30:00 crc kubenswrapper[4859]: I1008 18:30:00.156125 4859 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29332470-sc2qz"] Oct 08 18:30:00 crc kubenswrapper[4859]: I1008 18:30:00.195050 4859 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/openstack-operator-index-z8f5z"] Oct 08 18:30:00 crc kubenswrapper[4859]: I1008 18:30:00.262057 4859 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/af0b7fda-8561-4159-b38f-882401a5fc29-config-volume\") pod \"collect-profiles-29332470-sc2qz\" (UID: \"af0b7fda-8561-4159-b38f-882401a5fc29\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29332470-sc2qz" Oct 08 18:30:00 crc kubenswrapper[4859]: I1008 18:30:00.262505 4859 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/af0b7fda-8561-4159-b38f-882401a5fc29-secret-volume\") pod \"collect-profiles-29332470-sc2qz\" (UID: \"af0b7fda-8561-4159-b38f-882401a5fc29\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29332470-sc2qz" Oct 08 18:30:00 crc kubenswrapper[4859]: I1008 18:30:00.262617 4859 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-zmft9\" (UniqueName: \"kubernetes.io/projected/af0b7fda-8561-4159-b38f-882401a5fc29-kube-api-access-zmft9\") pod \"collect-profiles-29332470-sc2qz\" (UID: \"af0b7fda-8561-4159-b38f-882401a5fc29\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29332470-sc2qz" Oct 08 18:30:00 crc kubenswrapper[4859]: I1008 18:30:00.363860 4859 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/af0b7fda-8561-4159-b38f-882401a5fc29-config-volume\") pod \"collect-profiles-29332470-sc2qz\" (UID: \"af0b7fda-8561-4159-b38f-882401a5fc29\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29332470-sc2qz" Oct 08 18:30:00 crc kubenswrapper[4859]: I1008 18:30:00.363969 4859 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/af0b7fda-8561-4159-b38f-882401a5fc29-secret-volume\") pod \"collect-profiles-29332470-sc2qz\" (UID: \"af0b7fda-8561-4159-b38f-882401a5fc29\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29332470-sc2qz" Oct 08 18:30:00 crc kubenswrapper[4859]: I1008 18:30:00.364027 4859 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-zmft9\" (UniqueName: \"kubernetes.io/projected/af0b7fda-8561-4159-b38f-882401a5fc29-kube-api-access-zmft9\") pod \"collect-profiles-29332470-sc2qz\" (UID: \"af0b7fda-8561-4159-b38f-882401a5fc29\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29332470-sc2qz" Oct 08 18:30:00 crc kubenswrapper[4859]: I1008 18:30:00.365776 4859 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/af0b7fda-8561-4159-b38f-882401a5fc29-config-volume\") pod \"collect-profiles-29332470-sc2qz\" (UID: \"af0b7fda-8561-4159-b38f-882401a5fc29\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29332470-sc2qz" Oct 08 18:30:00 crc kubenswrapper[4859]: I1008 18:30:00.372238 4859 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/af0b7fda-8561-4159-b38f-882401a5fc29-secret-volume\") pod \"collect-profiles-29332470-sc2qz\" (UID: \"af0b7fda-8561-4159-b38f-882401a5fc29\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29332470-sc2qz" Oct 08 18:30:00 crc kubenswrapper[4859]: I1008 18:30:00.387789 4859 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-zmft9\" (UniqueName: \"kubernetes.io/projected/af0b7fda-8561-4159-b38f-882401a5fc29-kube-api-access-zmft9\") pod \"collect-profiles-29332470-sc2qz\" (UID: \"af0b7fda-8561-4159-b38f-882401a5fc29\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29332470-sc2qz" Oct 08 18:30:00 crc kubenswrapper[4859]: I1008 18:30:00.463569 4859 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29332470-sc2qz" Oct 08 18:30:00 crc kubenswrapper[4859]: I1008 18:30:00.465593 4859 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-operator-index-z8f5z" event={"ID":"6c1463a8-c973-42db-8671-f274a513230a","Type":"ContainerStarted","Data":"a9b00a607bf154c9f0238158cd3b10565a8067a4bfe9d4f4c614664eca198087"} Oct 08 18:30:00 crc kubenswrapper[4859]: I1008 18:30:00.863066 4859 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29332470-sc2qz"] Oct 08 18:30:00 crc kubenswrapper[4859]: W1008 18:30:00.869111 4859 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podaf0b7fda_8561_4159_b38f_882401a5fc29.slice/crio-f9252d7418340ed9b9d80466ecff579decc5478d550ca74ecf90489eacbecada WatchSource:0}: Error finding container f9252d7418340ed9b9d80466ecff579decc5478d550ca74ecf90489eacbecada: Status 404 returned error can't find the container with id f9252d7418340ed9b9d80466ecff579decc5478d550ca74ecf90489eacbecada Oct 08 18:30:01 crc kubenswrapper[4859]: I1008 18:30:01.482621 4859 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29332470-sc2qz" event={"ID":"af0b7fda-8561-4159-b38f-882401a5fc29","Type":"ContainerStarted","Data":"562b43d35c2214b9dffd19a414e87c80d6a071857c70d6baefe20e372475003c"} Oct 08 18:30:01 crc kubenswrapper[4859]: I1008 18:30:01.483072 4859 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29332470-sc2qz" event={"ID":"af0b7fda-8561-4159-b38f-882401a5fc29","Type":"ContainerStarted","Data":"f9252d7418340ed9b9d80466ecff579decc5478d550ca74ecf90489eacbecada"} Oct 08 18:30:01 crc kubenswrapper[4859]: I1008 18:30:01.501844 4859 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-operator-lifecycle-manager/collect-profiles-29332470-sc2qz" podStartSLOduration=1.501824858 podStartE2EDuration="1.501824858s" podCreationTimestamp="2025-10-08 18:30:00 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-08 18:30:01.499142179 +0000 UTC m=+771.745981578" watchObservedRunningTime="2025-10-08 18:30:01.501824858 +0000 UTC m=+771.748664237" Oct 08 18:30:02 crc kubenswrapper[4859]: I1008 18:30:02.492148 4859 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-operator-index-z8f5z" event={"ID":"6c1463a8-c973-42db-8671-f274a513230a","Type":"ContainerStarted","Data":"c27243f00bac70f486d35bbf79e1222dc50013f412ba904d4ce1c9a03b855e53"} Oct 08 18:30:02 crc kubenswrapper[4859]: I1008 18:30:02.494791 4859 generic.go:334] "Generic (PLEG): container finished" podID="af0b7fda-8561-4159-b38f-882401a5fc29" containerID="562b43d35c2214b9dffd19a414e87c80d6a071857c70d6baefe20e372475003c" exitCode=0 Oct 08 18:30:02 crc kubenswrapper[4859]: I1008 18:30:02.494863 4859 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29332470-sc2qz" event={"ID":"af0b7fda-8561-4159-b38f-882401a5fc29","Type":"ContainerDied","Data":"562b43d35c2214b9dffd19a414e87c80d6a071857c70d6baefe20e372475003c"} Oct 08 18:30:02 crc kubenswrapper[4859]: I1008 18:30:02.542983 4859 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/openstack-operator-index-z8f5z" podStartSLOduration=2.059705131 podStartE2EDuration="3.542960928s" podCreationTimestamp="2025-10-08 18:29:59 +0000 UTC" firstStartedPulling="2025-10-08 18:30:00.208314691 +0000 UTC m=+770.455154080" lastFinishedPulling="2025-10-08 18:30:01.691570468 +0000 UTC m=+771.938409877" observedRunningTime="2025-10-08 18:30:02.51847169 +0000 UTC m=+772.765311069" watchObservedRunningTime="2025-10-08 18:30:02.542960928 +0000 UTC m=+772.789800317" Oct 08 18:30:03 crc kubenswrapper[4859]: I1008 18:30:03.162966 4859 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-rw68k"] Oct 08 18:30:03 crc kubenswrapper[4859]: I1008 18:30:03.163440 4859 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/community-operators-rw68k" podUID="453a83ad-e06c-4485-bba7-524b7128462f" containerName="registry-server" containerID="cri-o://9fb031553cfab809d6b97efdae195d8a0f6b611557913e13da708cfdf26c41bc" gracePeriod=2 Oct 08 18:30:03 crc kubenswrapper[4859]: I1008 18:30:03.505553 4859 generic.go:334] "Generic (PLEG): container finished" podID="453a83ad-e06c-4485-bba7-524b7128462f" containerID="9fb031553cfab809d6b97efdae195d8a0f6b611557913e13da708cfdf26c41bc" exitCode=0 Oct 08 18:30:03 crc kubenswrapper[4859]: I1008 18:30:03.505633 4859 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-rw68k" event={"ID":"453a83ad-e06c-4485-bba7-524b7128462f","Type":"ContainerDied","Data":"9fb031553cfab809d6b97efdae195d8a0f6b611557913e13da708cfdf26c41bc"} Oct 08 18:30:03 crc kubenswrapper[4859]: I1008 18:30:03.865694 4859 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29332470-sc2qz" Oct 08 18:30:04 crc kubenswrapper[4859]: I1008 18:30:04.027880 4859 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/af0b7fda-8561-4159-b38f-882401a5fc29-secret-volume\") pod \"af0b7fda-8561-4159-b38f-882401a5fc29\" (UID: \"af0b7fda-8561-4159-b38f-882401a5fc29\") " Oct 08 18:30:04 crc kubenswrapper[4859]: I1008 18:30:04.027966 4859 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/af0b7fda-8561-4159-b38f-882401a5fc29-config-volume\") pod \"af0b7fda-8561-4159-b38f-882401a5fc29\" (UID: \"af0b7fda-8561-4159-b38f-882401a5fc29\") " Oct 08 18:30:04 crc kubenswrapper[4859]: I1008 18:30:04.028011 4859 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-zmft9\" (UniqueName: \"kubernetes.io/projected/af0b7fda-8561-4159-b38f-882401a5fc29-kube-api-access-zmft9\") pod \"af0b7fda-8561-4159-b38f-882401a5fc29\" (UID: \"af0b7fda-8561-4159-b38f-882401a5fc29\") " Oct 08 18:30:04 crc kubenswrapper[4859]: I1008 18:30:04.029268 4859 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/af0b7fda-8561-4159-b38f-882401a5fc29-config-volume" (OuterVolumeSpecName: "config-volume") pod "af0b7fda-8561-4159-b38f-882401a5fc29" (UID: "af0b7fda-8561-4159-b38f-882401a5fc29"). InnerVolumeSpecName "config-volume". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 08 18:30:04 crc kubenswrapper[4859]: I1008 18:30:04.034000 4859 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/af0b7fda-8561-4159-b38f-882401a5fc29-secret-volume" (OuterVolumeSpecName: "secret-volume") pod "af0b7fda-8561-4159-b38f-882401a5fc29" (UID: "af0b7fda-8561-4159-b38f-882401a5fc29"). InnerVolumeSpecName "secret-volume". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 08 18:30:04 crc kubenswrapper[4859]: I1008 18:30:04.034200 4859 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/af0b7fda-8561-4159-b38f-882401a5fc29-kube-api-access-zmft9" (OuterVolumeSpecName: "kube-api-access-zmft9") pod "af0b7fda-8561-4159-b38f-882401a5fc29" (UID: "af0b7fda-8561-4159-b38f-882401a5fc29"). InnerVolumeSpecName "kube-api-access-zmft9". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 08 18:30:04 crc kubenswrapper[4859]: I1008 18:30:04.129298 4859 reconciler_common.go:293] "Volume detached for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/af0b7fda-8561-4159-b38f-882401a5fc29-secret-volume\") on node \"crc\" DevicePath \"\"" Oct 08 18:30:04 crc kubenswrapper[4859]: I1008 18:30:04.129338 4859 reconciler_common.go:293] "Volume detached for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/af0b7fda-8561-4159-b38f-882401a5fc29-config-volume\") on node \"crc\" DevicePath \"\"" Oct 08 18:30:04 crc kubenswrapper[4859]: I1008 18:30:04.129351 4859 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-zmft9\" (UniqueName: \"kubernetes.io/projected/af0b7fda-8561-4159-b38f-882401a5fc29-kube-api-access-zmft9\") on node \"crc\" DevicePath \"\"" Oct 08 18:30:04 crc kubenswrapper[4859]: I1008 18:30:04.514654 4859 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29332470-sc2qz" event={"ID":"af0b7fda-8561-4159-b38f-882401a5fc29","Type":"ContainerDied","Data":"f9252d7418340ed9b9d80466ecff579decc5478d550ca74ecf90489eacbecada"} Oct 08 18:30:04 crc kubenswrapper[4859]: I1008 18:30:04.514725 4859 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="f9252d7418340ed9b9d80466ecff579decc5478d550ca74ecf90489eacbecada" Oct 08 18:30:04 crc kubenswrapper[4859]: I1008 18:30:04.514780 4859 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29332470-sc2qz" Oct 08 18:30:05 crc kubenswrapper[4859]: I1008 18:30:05.122395 4859 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-rw68k" Oct 08 18:30:05 crc kubenswrapper[4859]: I1008 18:30:05.243392 4859 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-w8tlg\" (UniqueName: \"kubernetes.io/projected/453a83ad-e06c-4485-bba7-524b7128462f-kube-api-access-w8tlg\") pod \"453a83ad-e06c-4485-bba7-524b7128462f\" (UID: \"453a83ad-e06c-4485-bba7-524b7128462f\") " Oct 08 18:30:05 crc kubenswrapper[4859]: I1008 18:30:05.243591 4859 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/453a83ad-e06c-4485-bba7-524b7128462f-catalog-content\") pod \"453a83ad-e06c-4485-bba7-524b7128462f\" (UID: \"453a83ad-e06c-4485-bba7-524b7128462f\") " Oct 08 18:30:05 crc kubenswrapper[4859]: I1008 18:30:05.243814 4859 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/453a83ad-e06c-4485-bba7-524b7128462f-utilities\") pod \"453a83ad-e06c-4485-bba7-524b7128462f\" (UID: \"453a83ad-e06c-4485-bba7-524b7128462f\") " Oct 08 18:30:05 crc kubenswrapper[4859]: I1008 18:30:05.245604 4859 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/453a83ad-e06c-4485-bba7-524b7128462f-utilities" (OuterVolumeSpecName: "utilities") pod "453a83ad-e06c-4485-bba7-524b7128462f" (UID: "453a83ad-e06c-4485-bba7-524b7128462f"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 08 18:30:05 crc kubenswrapper[4859]: I1008 18:30:05.249896 4859 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/453a83ad-e06c-4485-bba7-524b7128462f-kube-api-access-w8tlg" (OuterVolumeSpecName: "kube-api-access-w8tlg") pod "453a83ad-e06c-4485-bba7-524b7128462f" (UID: "453a83ad-e06c-4485-bba7-524b7128462f"). InnerVolumeSpecName "kube-api-access-w8tlg". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 08 18:30:05 crc kubenswrapper[4859]: I1008 18:30:05.291579 4859 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/453a83ad-e06c-4485-bba7-524b7128462f-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "453a83ad-e06c-4485-bba7-524b7128462f" (UID: "453a83ad-e06c-4485-bba7-524b7128462f"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 08 18:30:05 crc kubenswrapper[4859]: I1008 18:30:05.347144 4859 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/453a83ad-e06c-4485-bba7-524b7128462f-utilities\") on node \"crc\" DevicePath \"\"" Oct 08 18:30:05 crc kubenswrapper[4859]: I1008 18:30:05.347218 4859 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-w8tlg\" (UniqueName: \"kubernetes.io/projected/453a83ad-e06c-4485-bba7-524b7128462f-kube-api-access-w8tlg\") on node \"crc\" DevicePath \"\"" Oct 08 18:30:05 crc kubenswrapper[4859]: I1008 18:30:05.347251 4859 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/453a83ad-e06c-4485-bba7-524b7128462f-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 08 18:30:05 crc kubenswrapper[4859]: I1008 18:30:05.527007 4859 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-rw68k" event={"ID":"453a83ad-e06c-4485-bba7-524b7128462f","Type":"ContainerDied","Data":"37bdb42b252deeede466ca0ad9ac6d183f362902426a666760327cc69b1726ab"} Oct 08 18:30:05 crc kubenswrapper[4859]: I1008 18:30:05.527088 4859 scope.go:117] "RemoveContainer" containerID="9fb031553cfab809d6b97efdae195d8a0f6b611557913e13da708cfdf26c41bc" Oct 08 18:30:05 crc kubenswrapper[4859]: I1008 18:30:05.527136 4859 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-rw68k" Oct 08 18:30:05 crc kubenswrapper[4859]: I1008 18:30:05.549795 4859 scope.go:117] "RemoveContainer" containerID="00337353f6a1de8a59cfbd400f17ad2da3b4dadf1240a970d8c7175b49a46e8c" Oct 08 18:30:05 crc kubenswrapper[4859]: I1008 18:30:05.562512 4859 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-rw68k"] Oct 08 18:30:05 crc kubenswrapper[4859]: I1008 18:30:05.569238 4859 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/community-operators-rw68k"] Oct 08 18:30:05 crc kubenswrapper[4859]: I1008 18:30:05.586053 4859 scope.go:117] "RemoveContainer" containerID="81a60605b657849fe152794e839a2430aeff961f7243f05721bc11729b9978fe" Oct 08 18:30:05 crc kubenswrapper[4859]: I1008 18:30:05.711224 4859 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-mqxtr" Oct 08 18:30:05 crc kubenswrapper[4859]: I1008 18:30:05.711311 4859 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-mqxtr" Oct 08 18:30:05 crc kubenswrapper[4859]: I1008 18:30:05.781852 4859 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-mqxtr" Oct 08 18:30:06 crc kubenswrapper[4859]: I1008 18:30:06.477351 4859 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="453a83ad-e06c-4485-bba7-524b7128462f" path="/var/lib/kubelet/pods/453a83ad-e06c-4485-bba7-524b7128462f/volumes" Oct 08 18:30:06 crc kubenswrapper[4859]: I1008 18:30:06.606539 4859 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-mqxtr" Oct 08 18:30:09 crc kubenswrapper[4859]: I1008 18:30:09.275978 4859 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="metallb-system/frr-k8s-mfcd8" Oct 08 18:30:09 crc kubenswrapper[4859]: I1008 18:30:09.713197 4859 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack-operators/openstack-operator-index-z8f5z" Oct 08 18:30:09 crc kubenswrapper[4859]: I1008 18:30:09.713304 4859 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/openstack-operator-index-z8f5z" Oct 08 18:30:09 crc kubenswrapper[4859]: I1008 18:30:09.740193 4859 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack-operators/openstack-operator-index-z8f5z" Oct 08 18:30:10 crc kubenswrapper[4859]: I1008 18:30:10.558834 4859 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-mqxtr"] Oct 08 18:30:10 crc kubenswrapper[4859]: I1008 18:30:10.560168 4859 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-marketplace-mqxtr" podUID="95364dee-6272-4f36-b265-139af1823b2f" containerName="registry-server" containerID="cri-o://d7799dc202759262536424e4806e4acb338441b91eb0d76626da83855dd30d75" gracePeriod=2 Oct 08 18:30:10 crc kubenswrapper[4859]: I1008 18:30:10.602670 4859 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/openstack-operator-index-z8f5z" Oct 08 18:30:11 crc kubenswrapper[4859]: I1008 18:30:11.055171 4859 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-mqxtr" Oct 08 18:30:11 crc kubenswrapper[4859]: I1008 18:30:11.232938 4859 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-qskx8\" (UniqueName: \"kubernetes.io/projected/95364dee-6272-4f36-b265-139af1823b2f-kube-api-access-qskx8\") pod \"95364dee-6272-4f36-b265-139af1823b2f\" (UID: \"95364dee-6272-4f36-b265-139af1823b2f\") " Oct 08 18:30:11 crc kubenswrapper[4859]: I1008 18:30:11.233295 4859 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/95364dee-6272-4f36-b265-139af1823b2f-catalog-content\") pod \"95364dee-6272-4f36-b265-139af1823b2f\" (UID: \"95364dee-6272-4f36-b265-139af1823b2f\") " Oct 08 18:30:11 crc kubenswrapper[4859]: I1008 18:30:11.233486 4859 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/95364dee-6272-4f36-b265-139af1823b2f-utilities\") pod \"95364dee-6272-4f36-b265-139af1823b2f\" (UID: \"95364dee-6272-4f36-b265-139af1823b2f\") " Oct 08 18:30:11 crc kubenswrapper[4859]: I1008 18:30:11.235048 4859 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/95364dee-6272-4f36-b265-139af1823b2f-utilities" (OuterVolumeSpecName: "utilities") pod "95364dee-6272-4f36-b265-139af1823b2f" (UID: "95364dee-6272-4f36-b265-139af1823b2f"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 08 18:30:11 crc kubenswrapper[4859]: I1008 18:30:11.238720 4859 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/95364dee-6272-4f36-b265-139af1823b2f-kube-api-access-qskx8" (OuterVolumeSpecName: "kube-api-access-qskx8") pod "95364dee-6272-4f36-b265-139af1823b2f" (UID: "95364dee-6272-4f36-b265-139af1823b2f"). InnerVolumeSpecName "kube-api-access-qskx8". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 08 18:30:11 crc kubenswrapper[4859]: I1008 18:30:11.245200 4859 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/95364dee-6272-4f36-b265-139af1823b2f-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "95364dee-6272-4f36-b265-139af1823b2f" (UID: "95364dee-6272-4f36-b265-139af1823b2f"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 08 18:30:11 crc kubenswrapper[4859]: I1008 18:30:11.334618 4859 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/95364dee-6272-4f36-b265-139af1823b2f-utilities\") on node \"crc\" DevicePath \"\"" Oct 08 18:30:11 crc kubenswrapper[4859]: I1008 18:30:11.334667 4859 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-qskx8\" (UniqueName: \"kubernetes.io/projected/95364dee-6272-4f36-b265-139af1823b2f-kube-api-access-qskx8\") on node \"crc\" DevicePath \"\"" Oct 08 18:30:11 crc kubenswrapper[4859]: I1008 18:30:11.334707 4859 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/95364dee-6272-4f36-b265-139af1823b2f-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 08 18:30:11 crc kubenswrapper[4859]: I1008 18:30:11.576175 4859 generic.go:334] "Generic (PLEG): container finished" podID="95364dee-6272-4f36-b265-139af1823b2f" containerID="d7799dc202759262536424e4806e4acb338441b91eb0d76626da83855dd30d75" exitCode=0 Oct 08 18:30:11 crc kubenswrapper[4859]: I1008 18:30:11.576298 4859 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-mqxtr" Oct 08 18:30:11 crc kubenswrapper[4859]: I1008 18:30:11.576292 4859 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-mqxtr" event={"ID":"95364dee-6272-4f36-b265-139af1823b2f","Type":"ContainerDied","Data":"d7799dc202759262536424e4806e4acb338441b91eb0d76626da83855dd30d75"} Oct 08 18:30:11 crc kubenswrapper[4859]: I1008 18:30:11.576362 4859 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-mqxtr" event={"ID":"95364dee-6272-4f36-b265-139af1823b2f","Type":"ContainerDied","Data":"b909a71c2a25df1ccfa5dfe3e4462cc1d8f0554231e4b84e2aef2296cedae777"} Oct 08 18:30:11 crc kubenswrapper[4859]: I1008 18:30:11.576394 4859 scope.go:117] "RemoveContainer" containerID="d7799dc202759262536424e4806e4acb338441b91eb0d76626da83855dd30d75" Oct 08 18:30:11 crc kubenswrapper[4859]: I1008 18:30:11.601581 4859 scope.go:117] "RemoveContainer" containerID="79c64bd519aa17ec98f7eedfee51a09a08351a3c1948f72ee80c20b1dbeb3cc9" Oct 08 18:30:11 crc kubenswrapper[4859]: I1008 18:30:11.608207 4859 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-mqxtr"] Oct 08 18:30:11 crc kubenswrapper[4859]: I1008 18:30:11.612347 4859 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-marketplace-mqxtr"] Oct 08 18:30:11 crc kubenswrapper[4859]: I1008 18:30:11.632616 4859 scope.go:117] "RemoveContainer" containerID="4aed0306378fe02319bdb45b1ea1d11659b097d1af1456c3f0d11256e9423c5a" Oct 08 18:30:11 crc kubenswrapper[4859]: I1008 18:30:11.658905 4859 scope.go:117] "RemoveContainer" containerID="d7799dc202759262536424e4806e4acb338441b91eb0d76626da83855dd30d75" Oct 08 18:30:11 crc kubenswrapper[4859]: E1008 18:30:11.659448 4859 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"d7799dc202759262536424e4806e4acb338441b91eb0d76626da83855dd30d75\": container with ID starting with d7799dc202759262536424e4806e4acb338441b91eb0d76626da83855dd30d75 not found: ID does not exist" containerID="d7799dc202759262536424e4806e4acb338441b91eb0d76626da83855dd30d75" Oct 08 18:30:11 crc kubenswrapper[4859]: I1008 18:30:11.659488 4859 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"d7799dc202759262536424e4806e4acb338441b91eb0d76626da83855dd30d75"} err="failed to get container status \"d7799dc202759262536424e4806e4acb338441b91eb0d76626da83855dd30d75\": rpc error: code = NotFound desc = could not find container \"d7799dc202759262536424e4806e4acb338441b91eb0d76626da83855dd30d75\": container with ID starting with d7799dc202759262536424e4806e4acb338441b91eb0d76626da83855dd30d75 not found: ID does not exist" Oct 08 18:30:11 crc kubenswrapper[4859]: I1008 18:30:11.659517 4859 scope.go:117] "RemoveContainer" containerID="79c64bd519aa17ec98f7eedfee51a09a08351a3c1948f72ee80c20b1dbeb3cc9" Oct 08 18:30:11 crc kubenswrapper[4859]: E1008 18:30:11.660001 4859 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"79c64bd519aa17ec98f7eedfee51a09a08351a3c1948f72ee80c20b1dbeb3cc9\": container with ID starting with 79c64bd519aa17ec98f7eedfee51a09a08351a3c1948f72ee80c20b1dbeb3cc9 not found: ID does not exist" containerID="79c64bd519aa17ec98f7eedfee51a09a08351a3c1948f72ee80c20b1dbeb3cc9" Oct 08 18:30:11 crc kubenswrapper[4859]: I1008 18:30:11.660071 4859 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"79c64bd519aa17ec98f7eedfee51a09a08351a3c1948f72ee80c20b1dbeb3cc9"} err="failed to get container status \"79c64bd519aa17ec98f7eedfee51a09a08351a3c1948f72ee80c20b1dbeb3cc9\": rpc error: code = NotFound desc = could not find container \"79c64bd519aa17ec98f7eedfee51a09a08351a3c1948f72ee80c20b1dbeb3cc9\": container with ID starting with 79c64bd519aa17ec98f7eedfee51a09a08351a3c1948f72ee80c20b1dbeb3cc9 not found: ID does not exist" Oct 08 18:30:11 crc kubenswrapper[4859]: I1008 18:30:11.660115 4859 scope.go:117] "RemoveContainer" containerID="4aed0306378fe02319bdb45b1ea1d11659b097d1af1456c3f0d11256e9423c5a" Oct 08 18:30:11 crc kubenswrapper[4859]: E1008 18:30:11.660483 4859 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"4aed0306378fe02319bdb45b1ea1d11659b097d1af1456c3f0d11256e9423c5a\": container with ID starting with 4aed0306378fe02319bdb45b1ea1d11659b097d1af1456c3f0d11256e9423c5a not found: ID does not exist" containerID="4aed0306378fe02319bdb45b1ea1d11659b097d1af1456c3f0d11256e9423c5a" Oct 08 18:30:11 crc kubenswrapper[4859]: I1008 18:30:11.660514 4859 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"4aed0306378fe02319bdb45b1ea1d11659b097d1af1456c3f0d11256e9423c5a"} err="failed to get container status \"4aed0306378fe02319bdb45b1ea1d11659b097d1af1456c3f0d11256e9423c5a\": rpc error: code = NotFound desc = could not find container \"4aed0306378fe02319bdb45b1ea1d11659b097d1af1456c3f0d11256e9423c5a\": container with ID starting with 4aed0306378fe02319bdb45b1ea1d11659b097d1af1456c3f0d11256e9423c5a not found: ID does not exist" Oct 08 18:30:12 crc kubenswrapper[4859]: I1008 18:30:12.481920 4859 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="95364dee-6272-4f36-b265-139af1823b2f" path="/var/lib/kubelet/pods/95364dee-6272-4f36-b265-139af1823b2f/volumes" Oct 08 18:30:14 crc kubenswrapper[4859]: I1008 18:30:14.614946 4859 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/1bfea55cf6f540853d0c0d919aed51b5fd6e2d105cc44a68cb64399f8bxp4tb"] Oct 08 18:30:14 crc kubenswrapper[4859]: E1008 18:30:14.615308 4859 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="453a83ad-e06c-4485-bba7-524b7128462f" containerName="registry-server" Oct 08 18:30:14 crc kubenswrapper[4859]: I1008 18:30:14.615329 4859 state_mem.go:107] "Deleted CPUSet assignment" podUID="453a83ad-e06c-4485-bba7-524b7128462f" containerName="registry-server" Oct 08 18:30:14 crc kubenswrapper[4859]: E1008 18:30:14.615348 4859 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="453a83ad-e06c-4485-bba7-524b7128462f" containerName="extract-content" Oct 08 18:30:14 crc kubenswrapper[4859]: I1008 18:30:14.615359 4859 state_mem.go:107] "Deleted CPUSet assignment" podUID="453a83ad-e06c-4485-bba7-524b7128462f" containerName="extract-content" Oct 08 18:30:14 crc kubenswrapper[4859]: E1008 18:30:14.615375 4859 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="95364dee-6272-4f36-b265-139af1823b2f" containerName="extract-utilities" Oct 08 18:30:14 crc kubenswrapper[4859]: I1008 18:30:14.615384 4859 state_mem.go:107] "Deleted CPUSet assignment" podUID="95364dee-6272-4f36-b265-139af1823b2f" containerName="extract-utilities" Oct 08 18:30:14 crc kubenswrapper[4859]: E1008 18:30:14.615397 4859 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="95364dee-6272-4f36-b265-139af1823b2f" containerName="extract-content" Oct 08 18:30:14 crc kubenswrapper[4859]: I1008 18:30:14.615405 4859 state_mem.go:107] "Deleted CPUSet assignment" podUID="95364dee-6272-4f36-b265-139af1823b2f" containerName="extract-content" Oct 08 18:30:14 crc kubenswrapper[4859]: E1008 18:30:14.615417 4859 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="af0b7fda-8561-4159-b38f-882401a5fc29" containerName="collect-profiles" Oct 08 18:30:14 crc kubenswrapper[4859]: I1008 18:30:14.615426 4859 state_mem.go:107] "Deleted CPUSet assignment" podUID="af0b7fda-8561-4159-b38f-882401a5fc29" containerName="collect-profiles" Oct 08 18:30:14 crc kubenswrapper[4859]: E1008 18:30:14.615439 4859 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="453a83ad-e06c-4485-bba7-524b7128462f" containerName="extract-utilities" Oct 08 18:30:14 crc kubenswrapper[4859]: I1008 18:30:14.615447 4859 state_mem.go:107] "Deleted CPUSet assignment" podUID="453a83ad-e06c-4485-bba7-524b7128462f" containerName="extract-utilities" Oct 08 18:30:14 crc kubenswrapper[4859]: E1008 18:30:14.615461 4859 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="95364dee-6272-4f36-b265-139af1823b2f" containerName="registry-server" Oct 08 18:30:14 crc kubenswrapper[4859]: I1008 18:30:14.615470 4859 state_mem.go:107] "Deleted CPUSet assignment" podUID="95364dee-6272-4f36-b265-139af1823b2f" containerName="registry-server" Oct 08 18:30:14 crc kubenswrapper[4859]: I1008 18:30:14.615644 4859 memory_manager.go:354] "RemoveStaleState removing state" podUID="af0b7fda-8561-4159-b38f-882401a5fc29" containerName="collect-profiles" Oct 08 18:30:14 crc kubenswrapper[4859]: I1008 18:30:14.615673 4859 memory_manager.go:354] "RemoveStaleState removing state" podUID="453a83ad-e06c-4485-bba7-524b7128462f" containerName="registry-server" Oct 08 18:30:14 crc kubenswrapper[4859]: I1008 18:30:14.615730 4859 memory_manager.go:354] "RemoveStaleState removing state" podUID="95364dee-6272-4f36-b265-139af1823b2f" containerName="registry-server" Oct 08 18:30:14 crc kubenswrapper[4859]: I1008 18:30:14.616854 4859 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/1bfea55cf6f540853d0c0d919aed51b5fd6e2d105cc44a68cb64399f8bxp4tb" Oct 08 18:30:14 crc kubenswrapper[4859]: I1008 18:30:14.623049 4859 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"default-dockercfg-6q4rf" Oct 08 18:30:14 crc kubenswrapper[4859]: I1008 18:30:14.628459 4859 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/1bfea55cf6f540853d0c0d919aed51b5fd6e2d105cc44a68cb64399f8bxp4tb"] Oct 08 18:30:14 crc kubenswrapper[4859]: I1008 18:30:14.790136 4859 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-clw5d\" (UniqueName: \"kubernetes.io/projected/70914ba1-396a-4131-835f-817debc89338-kube-api-access-clw5d\") pod \"1bfea55cf6f540853d0c0d919aed51b5fd6e2d105cc44a68cb64399f8bxp4tb\" (UID: \"70914ba1-396a-4131-835f-817debc89338\") " pod="openstack-operators/1bfea55cf6f540853d0c0d919aed51b5fd6e2d105cc44a68cb64399f8bxp4tb" Oct 08 18:30:14 crc kubenswrapper[4859]: I1008 18:30:14.790265 4859 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/70914ba1-396a-4131-835f-817debc89338-bundle\") pod \"1bfea55cf6f540853d0c0d919aed51b5fd6e2d105cc44a68cb64399f8bxp4tb\" (UID: \"70914ba1-396a-4131-835f-817debc89338\") " pod="openstack-operators/1bfea55cf6f540853d0c0d919aed51b5fd6e2d105cc44a68cb64399f8bxp4tb" Oct 08 18:30:14 crc kubenswrapper[4859]: I1008 18:30:14.790328 4859 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/70914ba1-396a-4131-835f-817debc89338-util\") pod \"1bfea55cf6f540853d0c0d919aed51b5fd6e2d105cc44a68cb64399f8bxp4tb\" (UID: \"70914ba1-396a-4131-835f-817debc89338\") " pod="openstack-operators/1bfea55cf6f540853d0c0d919aed51b5fd6e2d105cc44a68cb64399f8bxp4tb" Oct 08 18:30:14 crc kubenswrapper[4859]: I1008 18:30:14.891260 4859 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/70914ba1-396a-4131-835f-817debc89338-bundle\") pod \"1bfea55cf6f540853d0c0d919aed51b5fd6e2d105cc44a68cb64399f8bxp4tb\" (UID: \"70914ba1-396a-4131-835f-817debc89338\") " pod="openstack-operators/1bfea55cf6f540853d0c0d919aed51b5fd6e2d105cc44a68cb64399f8bxp4tb" Oct 08 18:30:14 crc kubenswrapper[4859]: I1008 18:30:14.891325 4859 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/70914ba1-396a-4131-835f-817debc89338-util\") pod \"1bfea55cf6f540853d0c0d919aed51b5fd6e2d105cc44a68cb64399f8bxp4tb\" (UID: \"70914ba1-396a-4131-835f-817debc89338\") " pod="openstack-operators/1bfea55cf6f540853d0c0d919aed51b5fd6e2d105cc44a68cb64399f8bxp4tb" Oct 08 18:30:14 crc kubenswrapper[4859]: I1008 18:30:14.891414 4859 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-clw5d\" (UniqueName: \"kubernetes.io/projected/70914ba1-396a-4131-835f-817debc89338-kube-api-access-clw5d\") pod \"1bfea55cf6f540853d0c0d919aed51b5fd6e2d105cc44a68cb64399f8bxp4tb\" (UID: \"70914ba1-396a-4131-835f-817debc89338\") " pod="openstack-operators/1bfea55cf6f540853d0c0d919aed51b5fd6e2d105cc44a68cb64399f8bxp4tb" Oct 08 18:30:14 crc kubenswrapper[4859]: I1008 18:30:14.891812 4859 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/70914ba1-396a-4131-835f-817debc89338-bundle\") pod \"1bfea55cf6f540853d0c0d919aed51b5fd6e2d105cc44a68cb64399f8bxp4tb\" (UID: \"70914ba1-396a-4131-835f-817debc89338\") " pod="openstack-operators/1bfea55cf6f540853d0c0d919aed51b5fd6e2d105cc44a68cb64399f8bxp4tb" Oct 08 18:30:14 crc kubenswrapper[4859]: I1008 18:30:14.891881 4859 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/70914ba1-396a-4131-835f-817debc89338-util\") pod \"1bfea55cf6f540853d0c0d919aed51b5fd6e2d105cc44a68cb64399f8bxp4tb\" (UID: \"70914ba1-396a-4131-835f-817debc89338\") " pod="openstack-operators/1bfea55cf6f540853d0c0d919aed51b5fd6e2d105cc44a68cb64399f8bxp4tb" Oct 08 18:30:14 crc kubenswrapper[4859]: I1008 18:30:14.923822 4859 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-clw5d\" (UniqueName: \"kubernetes.io/projected/70914ba1-396a-4131-835f-817debc89338-kube-api-access-clw5d\") pod \"1bfea55cf6f540853d0c0d919aed51b5fd6e2d105cc44a68cb64399f8bxp4tb\" (UID: \"70914ba1-396a-4131-835f-817debc89338\") " pod="openstack-operators/1bfea55cf6f540853d0c0d919aed51b5fd6e2d105cc44a68cb64399f8bxp4tb" Oct 08 18:30:14 crc kubenswrapper[4859]: I1008 18:30:14.933225 4859 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/1bfea55cf6f540853d0c0d919aed51b5fd6e2d105cc44a68cb64399f8bxp4tb" Oct 08 18:30:15 crc kubenswrapper[4859]: I1008 18:30:15.373309 4859 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/1bfea55cf6f540853d0c0d919aed51b5fd6e2d105cc44a68cb64399f8bxp4tb"] Oct 08 18:30:15 crc kubenswrapper[4859]: W1008 18:30:15.384276 4859 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod70914ba1_396a_4131_835f_817debc89338.slice/crio-dc426bf08ee5983ab8f115947715c1cb8996d3a85d37ef327d364f6bb3c302d8 WatchSource:0}: Error finding container dc426bf08ee5983ab8f115947715c1cb8996d3a85d37ef327d364f6bb3c302d8: Status 404 returned error can't find the container with id dc426bf08ee5983ab8f115947715c1cb8996d3a85d37ef327d364f6bb3c302d8 Oct 08 18:30:15 crc kubenswrapper[4859]: I1008 18:30:15.616360 4859 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/1bfea55cf6f540853d0c0d919aed51b5fd6e2d105cc44a68cb64399f8bxp4tb" event={"ID":"70914ba1-396a-4131-835f-817debc89338","Type":"ContainerStarted","Data":"74af3338d841bf8a6c46ff4e22f63a1a885b9ccf077444bb1bbd7ae6dadf676a"} Oct 08 18:30:15 crc kubenswrapper[4859]: I1008 18:30:15.616907 4859 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/1bfea55cf6f540853d0c0d919aed51b5fd6e2d105cc44a68cb64399f8bxp4tb" event={"ID":"70914ba1-396a-4131-835f-817debc89338","Type":"ContainerStarted","Data":"dc426bf08ee5983ab8f115947715c1cb8996d3a85d37ef327d364f6bb3c302d8"} Oct 08 18:30:16 crc kubenswrapper[4859]: I1008 18:30:16.622619 4859 generic.go:334] "Generic (PLEG): container finished" podID="70914ba1-396a-4131-835f-817debc89338" containerID="74af3338d841bf8a6c46ff4e22f63a1a885b9ccf077444bb1bbd7ae6dadf676a" exitCode=0 Oct 08 18:30:16 crc kubenswrapper[4859]: I1008 18:30:16.622671 4859 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/1bfea55cf6f540853d0c0d919aed51b5fd6e2d105cc44a68cb64399f8bxp4tb" event={"ID":"70914ba1-396a-4131-835f-817debc89338","Type":"ContainerDied","Data":"74af3338d841bf8a6c46ff4e22f63a1a885b9ccf077444bb1bbd7ae6dadf676a"} Oct 08 18:30:17 crc kubenswrapper[4859]: I1008 18:30:17.631564 4859 generic.go:334] "Generic (PLEG): container finished" podID="70914ba1-396a-4131-835f-817debc89338" containerID="15ea1f9a721c0ca037458da7db29f28ef2117af5b988c0d744729438952ca634" exitCode=0 Oct 08 18:30:17 crc kubenswrapper[4859]: I1008 18:30:17.631620 4859 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/1bfea55cf6f540853d0c0d919aed51b5fd6e2d105cc44a68cb64399f8bxp4tb" event={"ID":"70914ba1-396a-4131-835f-817debc89338","Type":"ContainerDied","Data":"15ea1f9a721c0ca037458da7db29f28ef2117af5b988c0d744729438952ca634"} Oct 08 18:30:17 crc kubenswrapper[4859]: I1008 18:30:17.924682 4859 patch_prober.go:28] interesting pod/machine-config-daemon-82s52 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 08 18:30:17 crc kubenswrapper[4859]: I1008 18:30:17.925163 4859 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-82s52" podUID="b23a6a6c-9d92-4e7b-840e-55cfda873a2d" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 08 18:30:18 crc kubenswrapper[4859]: I1008 18:30:18.646808 4859 generic.go:334] "Generic (PLEG): container finished" podID="70914ba1-396a-4131-835f-817debc89338" containerID="537e5e0ca46116d98d1620ef29555f36c9905a693e35a02cbf2f9a9bfe209066" exitCode=0 Oct 08 18:30:18 crc kubenswrapper[4859]: I1008 18:30:18.646923 4859 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/1bfea55cf6f540853d0c0d919aed51b5fd6e2d105cc44a68cb64399f8bxp4tb" event={"ID":"70914ba1-396a-4131-835f-817debc89338","Type":"ContainerDied","Data":"537e5e0ca46116d98d1620ef29555f36c9905a693e35a02cbf2f9a9bfe209066"} Oct 08 18:30:19 crc kubenswrapper[4859]: I1008 18:30:19.954034 4859 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack-operators/1bfea55cf6f540853d0c0d919aed51b5fd6e2d105cc44a68cb64399f8bxp4tb" Oct 08 18:30:19 crc kubenswrapper[4859]: I1008 18:30:19.960143 4859 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/70914ba1-396a-4131-835f-817debc89338-util\") pod \"70914ba1-396a-4131-835f-817debc89338\" (UID: \"70914ba1-396a-4131-835f-817debc89338\") " Oct 08 18:30:19 crc kubenswrapper[4859]: I1008 18:30:19.960209 4859 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/70914ba1-396a-4131-835f-817debc89338-bundle\") pod \"70914ba1-396a-4131-835f-817debc89338\" (UID: \"70914ba1-396a-4131-835f-817debc89338\") " Oct 08 18:30:19 crc kubenswrapper[4859]: I1008 18:30:19.960266 4859 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-clw5d\" (UniqueName: \"kubernetes.io/projected/70914ba1-396a-4131-835f-817debc89338-kube-api-access-clw5d\") pod \"70914ba1-396a-4131-835f-817debc89338\" (UID: \"70914ba1-396a-4131-835f-817debc89338\") " Oct 08 18:30:19 crc kubenswrapper[4859]: I1008 18:30:19.961156 4859 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/70914ba1-396a-4131-835f-817debc89338-bundle" (OuterVolumeSpecName: "bundle") pod "70914ba1-396a-4131-835f-817debc89338" (UID: "70914ba1-396a-4131-835f-817debc89338"). InnerVolumeSpecName "bundle". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 08 18:30:19 crc kubenswrapper[4859]: I1008 18:30:19.967209 4859 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/70914ba1-396a-4131-835f-817debc89338-kube-api-access-clw5d" (OuterVolumeSpecName: "kube-api-access-clw5d") pod "70914ba1-396a-4131-835f-817debc89338" (UID: "70914ba1-396a-4131-835f-817debc89338"). InnerVolumeSpecName "kube-api-access-clw5d". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 08 18:30:19 crc kubenswrapper[4859]: I1008 18:30:19.980278 4859 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/70914ba1-396a-4131-835f-817debc89338-util" (OuterVolumeSpecName: "util") pod "70914ba1-396a-4131-835f-817debc89338" (UID: "70914ba1-396a-4131-835f-817debc89338"). InnerVolumeSpecName "util". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 08 18:30:20 crc kubenswrapper[4859]: I1008 18:30:20.061629 4859 reconciler_common.go:293] "Volume detached for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/70914ba1-396a-4131-835f-817debc89338-util\") on node \"crc\" DevicePath \"\"" Oct 08 18:30:20 crc kubenswrapper[4859]: I1008 18:30:20.061677 4859 reconciler_common.go:293] "Volume detached for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/70914ba1-396a-4131-835f-817debc89338-bundle\") on node \"crc\" DevicePath \"\"" Oct 08 18:30:20 crc kubenswrapper[4859]: I1008 18:30:20.061699 4859 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-clw5d\" (UniqueName: \"kubernetes.io/projected/70914ba1-396a-4131-835f-817debc89338-kube-api-access-clw5d\") on node \"crc\" DevicePath \"\"" Oct 08 18:30:20 crc kubenswrapper[4859]: I1008 18:30:20.661504 4859 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/1bfea55cf6f540853d0c0d919aed51b5fd6e2d105cc44a68cb64399f8bxp4tb" event={"ID":"70914ba1-396a-4131-835f-817debc89338","Type":"ContainerDied","Data":"dc426bf08ee5983ab8f115947715c1cb8996d3a85d37ef327d364f6bb3c302d8"} Oct 08 18:30:20 crc kubenswrapper[4859]: I1008 18:30:20.661921 4859 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="dc426bf08ee5983ab8f115947715c1cb8996d3a85d37ef327d364f6bb3c302d8" Oct 08 18:30:20 crc kubenswrapper[4859]: I1008 18:30:20.661559 4859 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack-operators/1bfea55cf6f540853d0c0d919aed51b5fd6e2d105cc44a68cb64399f8bxp4tb" Oct 08 18:30:26 crc kubenswrapper[4859]: I1008 18:30:26.053411 4859 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/openstack-operator-controller-operator-bd6bc67fb-jkvgb"] Oct 08 18:30:26 crc kubenswrapper[4859]: E1008 18:30:26.054250 4859 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="70914ba1-396a-4131-835f-817debc89338" containerName="pull" Oct 08 18:30:26 crc kubenswrapper[4859]: I1008 18:30:26.054267 4859 state_mem.go:107] "Deleted CPUSet assignment" podUID="70914ba1-396a-4131-835f-817debc89338" containerName="pull" Oct 08 18:30:26 crc kubenswrapper[4859]: E1008 18:30:26.054281 4859 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="70914ba1-396a-4131-835f-817debc89338" containerName="util" Oct 08 18:30:26 crc kubenswrapper[4859]: I1008 18:30:26.054287 4859 state_mem.go:107] "Deleted CPUSet assignment" podUID="70914ba1-396a-4131-835f-817debc89338" containerName="util" Oct 08 18:30:26 crc kubenswrapper[4859]: E1008 18:30:26.054308 4859 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="70914ba1-396a-4131-835f-817debc89338" containerName="extract" Oct 08 18:30:26 crc kubenswrapper[4859]: I1008 18:30:26.054315 4859 state_mem.go:107] "Deleted CPUSet assignment" podUID="70914ba1-396a-4131-835f-817debc89338" containerName="extract" Oct 08 18:30:26 crc kubenswrapper[4859]: I1008 18:30:26.054446 4859 memory_manager.go:354] "RemoveStaleState removing state" podUID="70914ba1-396a-4131-835f-817debc89338" containerName="extract" Oct 08 18:30:26 crc kubenswrapper[4859]: I1008 18:30:26.055281 4859 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-operator-controller-operator-bd6bc67fb-jkvgb" Oct 08 18:30:26 crc kubenswrapper[4859]: I1008 18:30:26.059103 4859 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"openstack-operator-controller-operator-dockercfg-zzlzg" Oct 08 18:30:26 crc kubenswrapper[4859]: I1008 18:30:26.082722 4859 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/openstack-operator-controller-operator-bd6bc67fb-jkvgb"] Oct 08 18:30:26 crc kubenswrapper[4859]: I1008 18:30:26.143470 4859 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-hvpc8\" (UniqueName: \"kubernetes.io/projected/f60d2d10-2f0f-464f-9d1e-e63a5ef48972-kube-api-access-hvpc8\") pod \"openstack-operator-controller-operator-bd6bc67fb-jkvgb\" (UID: \"f60d2d10-2f0f-464f-9d1e-e63a5ef48972\") " pod="openstack-operators/openstack-operator-controller-operator-bd6bc67fb-jkvgb" Oct 08 18:30:26 crc kubenswrapper[4859]: I1008 18:30:26.245050 4859 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-hvpc8\" (UniqueName: \"kubernetes.io/projected/f60d2d10-2f0f-464f-9d1e-e63a5ef48972-kube-api-access-hvpc8\") pod \"openstack-operator-controller-operator-bd6bc67fb-jkvgb\" (UID: \"f60d2d10-2f0f-464f-9d1e-e63a5ef48972\") " pod="openstack-operators/openstack-operator-controller-operator-bd6bc67fb-jkvgb" Oct 08 18:30:26 crc kubenswrapper[4859]: I1008 18:30:26.268472 4859 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-hvpc8\" (UniqueName: \"kubernetes.io/projected/f60d2d10-2f0f-464f-9d1e-e63a5ef48972-kube-api-access-hvpc8\") pod \"openstack-operator-controller-operator-bd6bc67fb-jkvgb\" (UID: \"f60d2d10-2f0f-464f-9d1e-e63a5ef48972\") " pod="openstack-operators/openstack-operator-controller-operator-bd6bc67fb-jkvgb" Oct 08 18:30:26 crc kubenswrapper[4859]: I1008 18:30:26.374861 4859 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-operator-controller-operator-bd6bc67fb-jkvgb" Oct 08 18:30:26 crc kubenswrapper[4859]: I1008 18:30:26.602484 4859 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/openstack-operator-controller-operator-bd6bc67fb-jkvgb"] Oct 08 18:30:26 crc kubenswrapper[4859]: I1008 18:30:26.706731 4859 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-operator-controller-operator-bd6bc67fb-jkvgb" event={"ID":"f60d2d10-2f0f-464f-9d1e-e63a5ef48972","Type":"ContainerStarted","Data":"2fc6f41c817a8445e7c75f0b53dd4579b76cf375f37063996aabc75e6ed8207a"} Oct 08 18:30:31 crc kubenswrapper[4859]: I1008 18:30:31.743751 4859 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-operator-controller-operator-bd6bc67fb-jkvgb" event={"ID":"f60d2d10-2f0f-464f-9d1e-e63a5ef48972","Type":"ContainerStarted","Data":"83be3697b793b861bdd6f53dce5d5732bf7c737ace132b0668474894602fb819"} Oct 08 18:30:33 crc kubenswrapper[4859]: I1008 18:30:33.758706 4859 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-operator-controller-operator-bd6bc67fb-jkvgb" event={"ID":"f60d2d10-2f0f-464f-9d1e-e63a5ef48972","Type":"ContainerStarted","Data":"8e781f076f9b047f3bfd2a8e45e0c3065552af72b0d256fb3f9e9374cafefe64"} Oct 08 18:30:33 crc kubenswrapper[4859]: I1008 18:30:33.759625 4859 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/openstack-operator-controller-operator-bd6bc67fb-jkvgb" Oct 08 18:30:33 crc kubenswrapper[4859]: I1008 18:30:33.793561 4859 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/openstack-operator-controller-operator-bd6bc67fb-jkvgb" podStartSLOduration=0.817789311 podStartE2EDuration="7.793541225s" podCreationTimestamp="2025-10-08 18:30:26 +0000 UTC" firstStartedPulling="2025-10-08 18:30:26.612146889 +0000 UTC m=+796.858986268" lastFinishedPulling="2025-10-08 18:30:33.587898803 +0000 UTC m=+803.834738182" observedRunningTime="2025-10-08 18:30:33.791480477 +0000 UTC m=+804.038319876" watchObservedRunningTime="2025-10-08 18:30:33.793541225 +0000 UTC m=+804.040380604" Oct 08 18:30:36 crc kubenswrapper[4859]: I1008 18:30:36.377958 4859 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/openstack-operator-controller-operator-bd6bc67fb-jkvgb" Oct 08 18:30:41 crc kubenswrapper[4859]: I1008 18:30:41.621845 4859 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-jws9w"] Oct 08 18:30:41 crc kubenswrapper[4859]: I1008 18:30:41.624006 4859 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-jws9w" Oct 08 18:30:41 crc kubenswrapper[4859]: I1008 18:30:41.632399 4859 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-jws9w"] Oct 08 18:30:41 crc kubenswrapper[4859]: I1008 18:30:41.654524 4859 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/51932a59-6c7a-4c64-9acf-59984e192d91-utilities\") pod \"redhat-operators-jws9w\" (UID: \"51932a59-6c7a-4c64-9acf-59984e192d91\") " pod="openshift-marketplace/redhat-operators-jws9w" Oct 08 18:30:41 crc kubenswrapper[4859]: I1008 18:30:41.654730 4859 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/51932a59-6c7a-4c64-9acf-59984e192d91-catalog-content\") pod \"redhat-operators-jws9w\" (UID: \"51932a59-6c7a-4c64-9acf-59984e192d91\") " pod="openshift-marketplace/redhat-operators-jws9w" Oct 08 18:30:41 crc kubenswrapper[4859]: I1008 18:30:41.654804 4859 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-m7pzh\" (UniqueName: \"kubernetes.io/projected/51932a59-6c7a-4c64-9acf-59984e192d91-kube-api-access-m7pzh\") pod \"redhat-operators-jws9w\" (UID: \"51932a59-6c7a-4c64-9acf-59984e192d91\") " pod="openshift-marketplace/redhat-operators-jws9w" Oct 08 18:30:41 crc kubenswrapper[4859]: I1008 18:30:41.755848 4859 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-m7pzh\" (UniqueName: \"kubernetes.io/projected/51932a59-6c7a-4c64-9acf-59984e192d91-kube-api-access-m7pzh\") pod \"redhat-operators-jws9w\" (UID: \"51932a59-6c7a-4c64-9acf-59984e192d91\") " pod="openshift-marketplace/redhat-operators-jws9w" Oct 08 18:30:41 crc kubenswrapper[4859]: I1008 18:30:41.755931 4859 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/51932a59-6c7a-4c64-9acf-59984e192d91-utilities\") pod \"redhat-operators-jws9w\" (UID: \"51932a59-6c7a-4c64-9acf-59984e192d91\") " pod="openshift-marketplace/redhat-operators-jws9w" Oct 08 18:30:41 crc kubenswrapper[4859]: I1008 18:30:41.755985 4859 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/51932a59-6c7a-4c64-9acf-59984e192d91-catalog-content\") pod \"redhat-operators-jws9w\" (UID: \"51932a59-6c7a-4c64-9acf-59984e192d91\") " pod="openshift-marketplace/redhat-operators-jws9w" Oct 08 18:30:41 crc kubenswrapper[4859]: I1008 18:30:41.756452 4859 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/51932a59-6c7a-4c64-9acf-59984e192d91-catalog-content\") pod \"redhat-operators-jws9w\" (UID: \"51932a59-6c7a-4c64-9acf-59984e192d91\") " pod="openshift-marketplace/redhat-operators-jws9w" Oct 08 18:30:41 crc kubenswrapper[4859]: I1008 18:30:41.756859 4859 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/51932a59-6c7a-4c64-9acf-59984e192d91-utilities\") pod \"redhat-operators-jws9w\" (UID: \"51932a59-6c7a-4c64-9acf-59984e192d91\") " pod="openshift-marketplace/redhat-operators-jws9w" Oct 08 18:30:41 crc kubenswrapper[4859]: I1008 18:30:41.782804 4859 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-m7pzh\" (UniqueName: \"kubernetes.io/projected/51932a59-6c7a-4c64-9acf-59984e192d91-kube-api-access-m7pzh\") pod \"redhat-operators-jws9w\" (UID: \"51932a59-6c7a-4c64-9acf-59984e192d91\") " pod="openshift-marketplace/redhat-operators-jws9w" Oct 08 18:30:41 crc kubenswrapper[4859]: I1008 18:30:41.941357 4859 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-jws9w" Oct 08 18:30:42 crc kubenswrapper[4859]: I1008 18:30:42.293287 4859 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-jws9w"] Oct 08 18:30:42 crc kubenswrapper[4859]: W1008 18:30:42.301235 4859 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod51932a59_6c7a_4c64_9acf_59984e192d91.slice/crio-0c93203584618cbe5643074d3bf05d482d24b5076b251592048ec50a965ed19e WatchSource:0}: Error finding container 0c93203584618cbe5643074d3bf05d482d24b5076b251592048ec50a965ed19e: Status 404 returned error can't find the container with id 0c93203584618cbe5643074d3bf05d482d24b5076b251592048ec50a965ed19e Oct 08 18:30:42 crc kubenswrapper[4859]: I1008 18:30:42.822554 4859 generic.go:334] "Generic (PLEG): container finished" podID="51932a59-6c7a-4c64-9acf-59984e192d91" containerID="1223abb72da61b77cb9987f243aabc4723c02474786911c7a3414b9a47285daa" exitCode=0 Oct 08 18:30:42 crc kubenswrapper[4859]: I1008 18:30:42.822658 4859 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-jws9w" event={"ID":"51932a59-6c7a-4c64-9acf-59984e192d91","Type":"ContainerDied","Data":"1223abb72da61b77cb9987f243aabc4723c02474786911c7a3414b9a47285daa"} Oct 08 18:30:42 crc kubenswrapper[4859]: I1008 18:30:42.822940 4859 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-jws9w" event={"ID":"51932a59-6c7a-4c64-9acf-59984e192d91","Type":"ContainerStarted","Data":"0c93203584618cbe5643074d3bf05d482d24b5076b251592048ec50a965ed19e"} Oct 08 18:30:43 crc kubenswrapper[4859]: I1008 18:30:43.838267 4859 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-jws9w" event={"ID":"51932a59-6c7a-4c64-9acf-59984e192d91","Type":"ContainerStarted","Data":"2f03ba0f6776e9c43ffc9f973cc293af86a987d1343b0e1ce3daf57cba93123c"} Oct 08 18:30:44 crc kubenswrapper[4859]: I1008 18:30:44.846187 4859 generic.go:334] "Generic (PLEG): container finished" podID="51932a59-6c7a-4c64-9acf-59984e192d91" containerID="2f03ba0f6776e9c43ffc9f973cc293af86a987d1343b0e1ce3daf57cba93123c" exitCode=0 Oct 08 18:30:44 crc kubenswrapper[4859]: I1008 18:30:44.846555 4859 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-jws9w" event={"ID":"51932a59-6c7a-4c64-9acf-59984e192d91","Type":"ContainerDied","Data":"2f03ba0f6776e9c43ffc9f973cc293af86a987d1343b0e1ce3daf57cba93123c"} Oct 08 18:30:45 crc kubenswrapper[4859]: I1008 18:30:45.858619 4859 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-jws9w" event={"ID":"51932a59-6c7a-4c64-9acf-59984e192d91","Type":"ContainerStarted","Data":"51de042cfa3748c000a73b19681f8fba61c0263e8fa8c6232162000607ec37bb"} Oct 08 18:30:45 crc kubenswrapper[4859]: I1008 18:30:45.878522 4859 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-jws9w" podStartSLOduration=2.344999261 podStartE2EDuration="4.878499026s" podCreationTimestamp="2025-10-08 18:30:41 +0000 UTC" firstStartedPulling="2025-10-08 18:30:42.824000305 +0000 UTC m=+813.070839684" lastFinishedPulling="2025-10-08 18:30:45.35750005 +0000 UTC m=+815.604339449" observedRunningTime="2025-10-08 18:30:45.874984467 +0000 UTC m=+816.121823866" watchObservedRunningTime="2025-10-08 18:30:45.878499026 +0000 UTC m=+816.125338405" Oct 08 18:30:47 crc kubenswrapper[4859]: I1008 18:30:47.924939 4859 patch_prober.go:28] interesting pod/machine-config-daemon-82s52 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 08 18:30:47 crc kubenswrapper[4859]: I1008 18:30:47.925255 4859 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-82s52" podUID="b23a6a6c-9d92-4e7b-840e-55cfda873a2d" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 08 18:30:51 crc kubenswrapper[4859]: I1008 18:30:51.941967 4859 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-jws9w" Oct 08 18:30:51 crc kubenswrapper[4859]: I1008 18:30:51.942604 4859 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-jws9w" Oct 08 18:30:51 crc kubenswrapper[4859]: I1008 18:30:51.999303 4859 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-jws9w" Oct 08 18:30:52 crc kubenswrapper[4859]: I1008 18:30:52.934709 4859 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/barbican-operator-controller-manager-658bdf4b74-8bfst"] Oct 08 18:30:52 crc kubenswrapper[4859]: I1008 18:30:52.935647 4859 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/barbican-operator-controller-manager-658bdf4b74-8bfst" Oct 08 18:30:52 crc kubenswrapper[4859]: I1008 18:30:52.938215 4859 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"barbican-operator-controller-manager-dockercfg-f69qd" Oct 08 18:30:52 crc kubenswrapper[4859]: I1008 18:30:52.943616 4859 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/cinder-operator-controller-manager-7b7fb68549-lbw64"] Oct 08 18:30:52 crc kubenswrapper[4859]: I1008 18:30:52.944812 4859 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/cinder-operator-controller-manager-7b7fb68549-lbw64" Oct 08 18:30:52 crc kubenswrapper[4859]: I1008 18:30:52.950990 4859 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"cinder-operator-controller-manager-dockercfg-z75cp" Oct 08 18:30:52 crc kubenswrapper[4859]: I1008 18:30:52.961419 4859 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/barbican-operator-controller-manager-658bdf4b74-8bfst"] Oct 08 18:30:52 crc kubenswrapper[4859]: I1008 18:30:52.983742 4859 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/designate-operator-controller-manager-85d5d9dd78-nhqkh"] Oct 08 18:30:52 crc kubenswrapper[4859]: I1008 18:30:52.985527 4859 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/designate-operator-controller-manager-85d5d9dd78-nhqkh" Oct 08 18:30:52 crc kubenswrapper[4859]: I1008 18:30:52.995176 4859 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"designate-operator-controller-manager-dockercfg-zb6jc" Oct 08 18:30:52 crc kubenswrapper[4859]: I1008 18:30:52.995206 4859 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-jws9w" Oct 08 18:30:53 crc kubenswrapper[4859]: I1008 18:30:53.011576 4859 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/cinder-operator-controller-manager-7b7fb68549-lbw64"] Oct 08 18:30:53 crc kubenswrapper[4859]: I1008 18:30:53.021086 4859 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-4gs5v\" (UniqueName: \"kubernetes.io/projected/d097cd11-b23b-4d38-80d7-5d85c257cef4-kube-api-access-4gs5v\") pod \"barbican-operator-controller-manager-658bdf4b74-8bfst\" (UID: \"d097cd11-b23b-4d38-80d7-5d85c257cef4\") " pod="openstack-operators/barbican-operator-controller-manager-658bdf4b74-8bfst" Oct 08 18:30:53 crc kubenswrapper[4859]: I1008 18:30:53.038595 4859 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/designate-operator-controller-manager-85d5d9dd78-nhqkh"] Oct 08 18:30:53 crc kubenswrapper[4859]: I1008 18:30:53.077258 4859 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/glance-operator-controller-manager-84b9b84486-d9gmv"] Oct 08 18:30:53 crc kubenswrapper[4859]: I1008 18:30:53.078351 4859 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/glance-operator-controller-manager-84b9b84486-d9gmv" Oct 08 18:30:53 crc kubenswrapper[4859]: I1008 18:30:53.080317 4859 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"glance-operator-controller-manager-dockercfg-9r2lr" Oct 08 18:30:53 crc kubenswrapper[4859]: I1008 18:30:53.083436 4859 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/heat-operator-controller-manager-858f76bbdd-b5bfz"] Oct 08 18:30:53 crc kubenswrapper[4859]: I1008 18:30:53.084527 4859 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/heat-operator-controller-manager-858f76bbdd-b5bfz" Oct 08 18:30:53 crc kubenswrapper[4859]: I1008 18:30:53.088902 4859 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"heat-operator-controller-manager-dockercfg-mgrvx" Oct 08 18:30:53 crc kubenswrapper[4859]: I1008 18:30:53.089611 4859 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/horizon-operator-controller-manager-7ffbcb7588-86z4n"] Oct 08 18:30:53 crc kubenswrapper[4859]: I1008 18:30:53.090372 4859 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/horizon-operator-controller-manager-7ffbcb7588-86z4n" Oct 08 18:30:53 crc kubenswrapper[4859]: I1008 18:30:53.101066 4859 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"horizon-operator-controller-manager-dockercfg-shbst" Oct 08 18:30:53 crc kubenswrapper[4859]: I1008 18:30:53.122459 4859 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-hfvxh\" (UniqueName: \"kubernetes.io/projected/d9a50bbe-f28f-46d8-9266-fd54e44388a3-kube-api-access-hfvxh\") pod \"designate-operator-controller-manager-85d5d9dd78-nhqkh\" (UID: \"d9a50bbe-f28f-46d8-9266-fd54e44388a3\") " pod="openstack-operators/designate-operator-controller-manager-85d5d9dd78-nhqkh" Oct 08 18:30:53 crc kubenswrapper[4859]: I1008 18:30:53.122498 4859 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-zdtfv\" (UniqueName: \"kubernetes.io/projected/e2b104df-4109-45ab-8a5a-3569da53caa0-kube-api-access-zdtfv\") pod \"cinder-operator-controller-manager-7b7fb68549-lbw64\" (UID: \"e2b104df-4109-45ab-8a5a-3569da53caa0\") " pod="openstack-operators/cinder-operator-controller-manager-7b7fb68549-lbw64" Oct 08 18:30:53 crc kubenswrapper[4859]: I1008 18:30:53.122578 4859 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-4gs5v\" (UniqueName: \"kubernetes.io/projected/d097cd11-b23b-4d38-80d7-5d85c257cef4-kube-api-access-4gs5v\") pod \"barbican-operator-controller-manager-658bdf4b74-8bfst\" (UID: \"d097cd11-b23b-4d38-80d7-5d85c257cef4\") " pod="openstack-operators/barbican-operator-controller-manager-658bdf4b74-8bfst" Oct 08 18:30:53 crc kubenswrapper[4859]: I1008 18:30:53.129494 4859 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/infra-operator-controller-manager-656bcbd775-ctjz9"] Oct 08 18:30:53 crc kubenswrapper[4859]: I1008 18:30:53.130530 4859 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/infra-operator-controller-manager-656bcbd775-ctjz9" Oct 08 18:30:53 crc kubenswrapper[4859]: I1008 18:30:53.137004 4859 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"infra-operator-controller-manager-dockercfg-4f958" Oct 08 18:30:53 crc kubenswrapper[4859]: I1008 18:30:53.137148 4859 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"infra-operator-webhook-server-cert" Oct 08 18:30:53 crc kubenswrapper[4859]: I1008 18:30:53.137168 4859 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/ironic-operator-controller-manager-9c5c78d49-z9952"] Oct 08 18:30:53 crc kubenswrapper[4859]: I1008 18:30:53.138263 4859 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/ironic-operator-controller-manager-9c5c78d49-z9952" Oct 08 18:30:53 crc kubenswrapper[4859]: I1008 18:30:53.144064 4859 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"ironic-operator-controller-manager-dockercfg-gt5d5" Oct 08 18:30:53 crc kubenswrapper[4859]: I1008 18:30:53.149785 4859 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/heat-operator-controller-manager-858f76bbdd-b5bfz"] Oct 08 18:30:53 crc kubenswrapper[4859]: I1008 18:30:53.167219 4859 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-4gs5v\" (UniqueName: \"kubernetes.io/projected/d097cd11-b23b-4d38-80d7-5d85c257cef4-kube-api-access-4gs5v\") pod \"barbican-operator-controller-manager-658bdf4b74-8bfst\" (UID: \"d097cd11-b23b-4d38-80d7-5d85c257cef4\") " pod="openstack-operators/barbican-operator-controller-manager-658bdf4b74-8bfst" Oct 08 18:30:53 crc kubenswrapper[4859]: I1008 18:30:53.171734 4859 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/glance-operator-controller-manager-84b9b84486-d9gmv"] Oct 08 18:30:53 crc kubenswrapper[4859]: I1008 18:30:53.189755 4859 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/infra-operator-controller-manager-656bcbd775-ctjz9"] Oct 08 18:30:53 crc kubenswrapper[4859]: I1008 18:30:53.199737 4859 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/horizon-operator-controller-manager-7ffbcb7588-86z4n"] Oct 08 18:30:53 crc kubenswrapper[4859]: I1008 18:30:53.202493 4859 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/keystone-operator-controller-manager-55b6b7c7b8-rsbgr"] Oct 08 18:30:53 crc kubenswrapper[4859]: I1008 18:30:53.212995 4859 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/ironic-operator-controller-manager-9c5c78d49-z9952"] Oct 08 18:30:53 crc kubenswrapper[4859]: I1008 18:30:53.213057 4859 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-jws9w"] Oct 08 18:30:53 crc kubenswrapper[4859]: I1008 18:30:53.213174 4859 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/keystone-operator-controller-manager-55b6b7c7b8-rsbgr" Oct 08 18:30:53 crc kubenswrapper[4859]: I1008 18:30:53.220411 4859 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"keystone-operator-controller-manager-dockercfg-nsjq6" Oct 08 18:30:53 crc kubenswrapper[4859]: I1008 18:30:53.224189 4859 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-k5drv\" (UniqueName: \"kubernetes.io/projected/2b67febf-c4c8-4e48-9350-29dd496439b4-kube-api-access-k5drv\") pod \"ironic-operator-controller-manager-9c5c78d49-z9952\" (UID: \"2b67febf-c4c8-4e48-9350-29dd496439b4\") " pod="openstack-operators/ironic-operator-controller-manager-9c5c78d49-z9952" Oct 08 18:30:53 crc kubenswrapper[4859]: I1008 18:30:53.224233 4859 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-f7g7k\" (UniqueName: \"kubernetes.io/projected/ca71782e-3569-4ac2-a178-239f0ae200d0-kube-api-access-f7g7k\") pod \"horizon-operator-controller-manager-7ffbcb7588-86z4n\" (UID: \"ca71782e-3569-4ac2-a178-239f0ae200d0\") " pod="openstack-operators/horizon-operator-controller-manager-7ffbcb7588-86z4n" Oct 08 18:30:53 crc kubenswrapper[4859]: I1008 18:30:53.224258 4859 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-5sgbf\" (UniqueName: \"kubernetes.io/projected/0b068d16-0072-41f8-ad01-4194fd428014-kube-api-access-5sgbf\") pod \"glance-operator-controller-manager-84b9b84486-d9gmv\" (UID: \"0b068d16-0072-41f8-ad01-4194fd428014\") " pod="openstack-operators/glance-operator-controller-manager-84b9b84486-d9gmv" Oct 08 18:30:53 crc kubenswrapper[4859]: I1008 18:30:53.224280 4859 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-wlqqz\" (UniqueName: \"kubernetes.io/projected/1a1cfd66-310a-4077-8020-df7e6e069c53-kube-api-access-wlqqz\") pod \"infra-operator-controller-manager-656bcbd775-ctjz9\" (UID: \"1a1cfd66-310a-4077-8020-df7e6e069c53\") " pod="openstack-operators/infra-operator-controller-manager-656bcbd775-ctjz9" Oct 08 18:30:53 crc kubenswrapper[4859]: I1008 18:30:53.224301 4859 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-qslr5\" (UniqueName: \"kubernetes.io/projected/19fa1d1d-0731-4c21-9faa-b1665ec4af36-kube-api-access-qslr5\") pod \"heat-operator-controller-manager-858f76bbdd-b5bfz\" (UID: \"19fa1d1d-0731-4c21-9faa-b1665ec4af36\") " pod="openstack-operators/heat-operator-controller-manager-858f76bbdd-b5bfz" Oct 08 18:30:53 crc kubenswrapper[4859]: I1008 18:30:53.224327 4859 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-hfvxh\" (UniqueName: \"kubernetes.io/projected/d9a50bbe-f28f-46d8-9266-fd54e44388a3-kube-api-access-hfvxh\") pod \"designate-operator-controller-manager-85d5d9dd78-nhqkh\" (UID: \"d9a50bbe-f28f-46d8-9266-fd54e44388a3\") " pod="openstack-operators/designate-operator-controller-manager-85d5d9dd78-nhqkh" Oct 08 18:30:53 crc kubenswrapper[4859]: I1008 18:30:53.224346 4859 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-zdtfv\" (UniqueName: \"kubernetes.io/projected/e2b104df-4109-45ab-8a5a-3569da53caa0-kube-api-access-zdtfv\") pod \"cinder-operator-controller-manager-7b7fb68549-lbw64\" (UID: \"e2b104df-4109-45ab-8a5a-3569da53caa0\") " pod="openstack-operators/cinder-operator-controller-manager-7b7fb68549-lbw64" Oct 08 18:30:53 crc kubenswrapper[4859]: I1008 18:30:53.224402 4859 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/1a1cfd66-310a-4077-8020-df7e6e069c53-cert\") pod \"infra-operator-controller-manager-656bcbd775-ctjz9\" (UID: \"1a1cfd66-310a-4077-8020-df7e6e069c53\") " pod="openstack-operators/infra-operator-controller-manager-656bcbd775-ctjz9" Oct 08 18:30:53 crc kubenswrapper[4859]: I1008 18:30:53.239095 4859 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/keystone-operator-controller-manager-55b6b7c7b8-rsbgr"] Oct 08 18:30:53 crc kubenswrapper[4859]: I1008 18:30:53.246004 4859 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/manila-operator-controller-manager-5f67fbc655-xg74g"] Oct 08 18:30:53 crc kubenswrapper[4859]: I1008 18:30:53.247215 4859 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/manila-operator-controller-manager-5f67fbc655-xg74g" Oct 08 18:30:53 crc kubenswrapper[4859]: I1008 18:30:53.260592 4859 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"manila-operator-controller-manager-dockercfg-lctp8" Oct 08 18:30:53 crc kubenswrapper[4859]: I1008 18:30:53.268347 4859 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/barbican-operator-controller-manager-658bdf4b74-8bfst" Oct 08 18:30:53 crc kubenswrapper[4859]: I1008 18:30:53.299899 4859 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/mariadb-operator-controller-manager-f9fb45f8f-ptpqs"] Oct 08 18:30:53 crc kubenswrapper[4859]: I1008 18:30:53.301324 4859 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/mariadb-operator-controller-manager-f9fb45f8f-ptpqs" Oct 08 18:30:53 crc kubenswrapper[4859]: I1008 18:30:53.308273 4859 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"mariadb-operator-controller-manager-dockercfg-x5s7x" Oct 08 18:30:53 crc kubenswrapper[4859]: I1008 18:30:53.309810 4859 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-hfvxh\" (UniqueName: \"kubernetes.io/projected/d9a50bbe-f28f-46d8-9266-fd54e44388a3-kube-api-access-hfvxh\") pod \"designate-operator-controller-manager-85d5d9dd78-nhqkh\" (UID: \"d9a50bbe-f28f-46d8-9266-fd54e44388a3\") " pod="openstack-operators/designate-operator-controller-manager-85d5d9dd78-nhqkh" Oct 08 18:30:53 crc kubenswrapper[4859]: I1008 18:30:53.313232 4859 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/neutron-operator-controller-manager-79d585cb66-zn65q"] Oct 08 18:30:53 crc kubenswrapper[4859]: I1008 18:30:53.318413 4859 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/neutron-operator-controller-manager-79d585cb66-zn65q" Oct 08 18:30:53 crc kubenswrapper[4859]: I1008 18:30:53.321874 4859 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/manila-operator-controller-manager-5f67fbc655-xg74g"] Oct 08 18:30:53 crc kubenswrapper[4859]: I1008 18:30:53.327158 4859 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-vqs92\" (UniqueName: \"kubernetes.io/projected/0228033e-c835-4618-9603-2c67cf9ce57d-kube-api-access-vqs92\") pod \"keystone-operator-controller-manager-55b6b7c7b8-rsbgr\" (UID: \"0228033e-c835-4618-9603-2c67cf9ce57d\") " pod="openstack-operators/keystone-operator-controller-manager-55b6b7c7b8-rsbgr" Oct 08 18:30:53 crc kubenswrapper[4859]: I1008 18:30:53.327242 4859 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-k5drv\" (UniqueName: \"kubernetes.io/projected/2b67febf-c4c8-4e48-9350-29dd496439b4-kube-api-access-k5drv\") pod \"ironic-operator-controller-manager-9c5c78d49-z9952\" (UID: \"2b67febf-c4c8-4e48-9350-29dd496439b4\") " pod="openstack-operators/ironic-operator-controller-manager-9c5c78d49-z9952" Oct 08 18:30:53 crc kubenswrapper[4859]: I1008 18:30:53.327276 4859 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-f7g7k\" (UniqueName: \"kubernetes.io/projected/ca71782e-3569-4ac2-a178-239f0ae200d0-kube-api-access-f7g7k\") pod \"horizon-operator-controller-manager-7ffbcb7588-86z4n\" (UID: \"ca71782e-3569-4ac2-a178-239f0ae200d0\") " pod="openstack-operators/horizon-operator-controller-manager-7ffbcb7588-86z4n" Oct 08 18:30:53 crc kubenswrapper[4859]: I1008 18:30:53.327298 4859 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-5sgbf\" (UniqueName: \"kubernetes.io/projected/0b068d16-0072-41f8-ad01-4194fd428014-kube-api-access-5sgbf\") pod \"glance-operator-controller-manager-84b9b84486-d9gmv\" (UID: \"0b068d16-0072-41f8-ad01-4194fd428014\") " pod="openstack-operators/glance-operator-controller-manager-84b9b84486-d9gmv" Oct 08 18:30:53 crc kubenswrapper[4859]: I1008 18:30:53.327324 4859 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-wlqqz\" (UniqueName: \"kubernetes.io/projected/1a1cfd66-310a-4077-8020-df7e6e069c53-kube-api-access-wlqqz\") pod \"infra-operator-controller-manager-656bcbd775-ctjz9\" (UID: \"1a1cfd66-310a-4077-8020-df7e6e069c53\") " pod="openstack-operators/infra-operator-controller-manager-656bcbd775-ctjz9" Oct 08 18:30:53 crc kubenswrapper[4859]: I1008 18:30:53.327345 4859 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-qslr5\" (UniqueName: \"kubernetes.io/projected/19fa1d1d-0731-4c21-9faa-b1665ec4af36-kube-api-access-qslr5\") pod \"heat-operator-controller-manager-858f76bbdd-b5bfz\" (UID: \"19fa1d1d-0731-4c21-9faa-b1665ec4af36\") " pod="openstack-operators/heat-operator-controller-manager-858f76bbdd-b5bfz" Oct 08 18:30:53 crc kubenswrapper[4859]: I1008 18:30:53.327395 4859 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/1a1cfd66-310a-4077-8020-df7e6e069c53-cert\") pod \"infra-operator-controller-manager-656bcbd775-ctjz9\" (UID: \"1a1cfd66-310a-4077-8020-df7e6e069c53\") " pod="openstack-operators/infra-operator-controller-manager-656bcbd775-ctjz9" Oct 08 18:30:53 crc kubenswrapper[4859]: E1008 18:30:53.327537 4859 secret.go:188] Couldn't get secret openstack-operators/infra-operator-webhook-server-cert: secret "infra-operator-webhook-server-cert" not found Oct 08 18:30:53 crc kubenswrapper[4859]: E1008 18:30:53.327602 4859 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/1a1cfd66-310a-4077-8020-df7e6e069c53-cert podName:1a1cfd66-310a-4077-8020-df7e6e069c53 nodeName:}" failed. No retries permitted until 2025-10-08 18:30:53.827581893 +0000 UTC m=+824.074421272 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "cert" (UniqueName: "kubernetes.io/secret/1a1cfd66-310a-4077-8020-df7e6e069c53-cert") pod "infra-operator-controller-manager-656bcbd775-ctjz9" (UID: "1a1cfd66-310a-4077-8020-df7e6e069c53") : secret "infra-operator-webhook-server-cert" not found Oct 08 18:30:53 crc kubenswrapper[4859]: I1008 18:30:53.328483 4859 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/designate-operator-controller-manager-85d5d9dd78-nhqkh" Oct 08 18:30:53 crc kubenswrapper[4859]: I1008 18:30:53.333123 4859 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/nova-operator-controller-manager-5df598886f-p99nw"] Oct 08 18:30:53 crc kubenswrapper[4859]: I1008 18:30:53.334412 4859 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/nova-operator-controller-manager-5df598886f-p99nw" Oct 08 18:30:53 crc kubenswrapper[4859]: I1008 18:30:53.339101 4859 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/neutron-operator-controller-manager-79d585cb66-zn65q"] Oct 08 18:30:53 crc kubenswrapper[4859]: I1008 18:30:53.345438 4859 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/mariadb-operator-controller-manager-f9fb45f8f-ptpqs"] Oct 08 18:30:53 crc kubenswrapper[4859]: I1008 18:30:53.349639 4859 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"neutron-operator-controller-manager-dockercfg-hhxsw" Oct 08 18:30:53 crc kubenswrapper[4859]: I1008 18:30:53.350044 4859 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"nova-operator-controller-manager-dockercfg-2nzzj" Oct 08 18:30:53 crc kubenswrapper[4859]: I1008 18:30:53.355574 4859 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-zdtfv\" (UniqueName: \"kubernetes.io/projected/e2b104df-4109-45ab-8a5a-3569da53caa0-kube-api-access-zdtfv\") pod \"cinder-operator-controller-manager-7b7fb68549-lbw64\" (UID: \"e2b104df-4109-45ab-8a5a-3569da53caa0\") " pod="openstack-operators/cinder-operator-controller-manager-7b7fb68549-lbw64" Oct 08 18:30:53 crc kubenswrapper[4859]: I1008 18:30:53.361616 4859 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/octavia-operator-controller-manager-69fdcfc5f5-lsl8l"] Oct 08 18:30:53 crc kubenswrapper[4859]: I1008 18:30:53.362838 4859 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/octavia-operator-controller-manager-69fdcfc5f5-lsl8l" Oct 08 18:30:53 crc kubenswrapper[4859]: I1008 18:30:53.372604 4859 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/nova-operator-controller-manager-5df598886f-p99nw"] Oct 08 18:30:53 crc kubenswrapper[4859]: I1008 18:30:53.373190 4859 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-5sgbf\" (UniqueName: \"kubernetes.io/projected/0b068d16-0072-41f8-ad01-4194fd428014-kube-api-access-5sgbf\") pod \"glance-operator-controller-manager-84b9b84486-d9gmv\" (UID: \"0b068d16-0072-41f8-ad01-4194fd428014\") " pod="openstack-operators/glance-operator-controller-manager-84b9b84486-d9gmv" Oct 08 18:30:53 crc kubenswrapper[4859]: I1008 18:30:53.376285 4859 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"octavia-operator-controller-manager-dockercfg-6jrk9" Oct 08 18:30:53 crc kubenswrapper[4859]: I1008 18:30:53.377947 4859 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/octavia-operator-controller-manager-69fdcfc5f5-lsl8l"] Oct 08 18:30:53 crc kubenswrapper[4859]: I1008 18:30:53.379699 4859 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-k5drv\" (UniqueName: \"kubernetes.io/projected/2b67febf-c4c8-4e48-9350-29dd496439b4-kube-api-access-k5drv\") pod \"ironic-operator-controller-manager-9c5c78d49-z9952\" (UID: \"2b67febf-c4c8-4e48-9350-29dd496439b4\") " pod="openstack-operators/ironic-operator-controller-manager-9c5c78d49-z9952" Oct 08 18:30:53 crc kubenswrapper[4859]: I1008 18:30:53.383357 4859 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/ovn-operator-controller-manager-79db49b9fb-hhgrl"] Oct 08 18:30:53 crc kubenswrapper[4859]: I1008 18:30:53.385714 4859 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/ovn-operator-controller-manager-79db49b9fb-hhgrl" Oct 08 18:30:53 crc kubenswrapper[4859]: I1008 18:30:53.392506 4859 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/openstack-baremetal-operator-controller-manager-677c5f5bff5pg7p"] Oct 08 18:30:53 crc kubenswrapper[4859]: I1008 18:30:53.393593 4859 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-baremetal-operator-controller-manager-677c5f5bff5pg7p" Oct 08 18:30:53 crc kubenswrapper[4859]: I1008 18:30:53.394233 4859 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"ovn-operator-controller-manager-dockercfg-2xwxw" Oct 08 18:30:53 crc kubenswrapper[4859]: I1008 18:30:53.396561 4859 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-wlqqz\" (UniqueName: \"kubernetes.io/projected/1a1cfd66-310a-4077-8020-df7e6e069c53-kube-api-access-wlqqz\") pod \"infra-operator-controller-manager-656bcbd775-ctjz9\" (UID: \"1a1cfd66-310a-4077-8020-df7e6e069c53\") " pod="openstack-operators/infra-operator-controller-manager-656bcbd775-ctjz9" Oct 08 18:30:53 crc kubenswrapper[4859]: I1008 18:30:53.409079 4859 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/ovn-operator-controller-manager-79db49b9fb-hhgrl"] Oct 08 18:30:53 crc kubenswrapper[4859]: I1008 18:30:53.411941 4859 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/glance-operator-controller-manager-84b9b84486-d9gmv" Oct 08 18:30:53 crc kubenswrapper[4859]: I1008 18:30:53.415656 4859 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-qslr5\" (UniqueName: \"kubernetes.io/projected/19fa1d1d-0731-4c21-9faa-b1665ec4af36-kube-api-access-qslr5\") pod \"heat-operator-controller-manager-858f76bbdd-b5bfz\" (UID: \"19fa1d1d-0731-4c21-9faa-b1665ec4af36\") " pod="openstack-operators/heat-operator-controller-manager-858f76bbdd-b5bfz" Oct 08 18:30:53 crc kubenswrapper[4859]: I1008 18:30:53.416226 4859 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"openstack-baremetal-operator-webhook-server-cert" Oct 08 18:30:53 crc kubenswrapper[4859]: I1008 18:30:53.420256 4859 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-f7g7k\" (UniqueName: \"kubernetes.io/projected/ca71782e-3569-4ac2-a178-239f0ae200d0-kube-api-access-f7g7k\") pod \"horizon-operator-controller-manager-7ffbcb7588-86z4n\" (UID: \"ca71782e-3569-4ac2-a178-239f0ae200d0\") " pod="openstack-operators/horizon-operator-controller-manager-7ffbcb7588-86z4n" Oct 08 18:30:53 crc kubenswrapper[4859]: I1008 18:30:53.440228 4859 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"openstack-baremetal-operator-controller-manager-dockercfg-p5p5g" Oct 08 18:30:53 crc kubenswrapper[4859]: I1008 18:30:53.440729 4859 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-fnbtf\" (UniqueName: \"kubernetes.io/projected/e2889a4a-89ec-47de-99df-7e749c3e5514-kube-api-access-fnbtf\") pod \"nova-operator-controller-manager-5df598886f-p99nw\" (UID: \"e2889a4a-89ec-47de-99df-7e749c3e5514\") " pod="openstack-operators/nova-operator-controller-manager-5df598886f-p99nw" Oct 08 18:30:53 crc kubenswrapper[4859]: I1008 18:30:53.440784 4859 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-9qb6z\" (UniqueName: \"kubernetes.io/projected/b6a2cffa-8421-43c4-bad8-5d4d341a40cc-kube-api-access-9qb6z\") pod \"mariadb-operator-controller-manager-f9fb45f8f-ptpqs\" (UID: \"b6a2cffa-8421-43c4-bad8-5d4d341a40cc\") " pod="openstack-operators/mariadb-operator-controller-manager-f9fb45f8f-ptpqs" Oct 08 18:30:53 crc kubenswrapper[4859]: I1008 18:30:53.439579 4859 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/heat-operator-controller-manager-858f76bbdd-b5bfz" Oct 08 18:30:53 crc kubenswrapper[4859]: I1008 18:30:53.440929 4859 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-jqbg7\" (UniqueName: \"kubernetes.io/projected/fbce123a-545f-4c4c-86ca-2d68f81aa3f3-kube-api-access-jqbg7\") pod \"manila-operator-controller-manager-5f67fbc655-xg74g\" (UID: \"fbce123a-545f-4c4c-86ca-2d68f81aa3f3\") " pod="openstack-operators/manila-operator-controller-manager-5f67fbc655-xg74g" Oct 08 18:30:53 crc kubenswrapper[4859]: I1008 18:30:53.441029 4859 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-vqs92\" (UniqueName: \"kubernetes.io/projected/0228033e-c835-4618-9603-2c67cf9ce57d-kube-api-access-vqs92\") pod \"keystone-operator-controller-manager-55b6b7c7b8-rsbgr\" (UID: \"0228033e-c835-4618-9603-2c67cf9ce57d\") " pod="openstack-operators/keystone-operator-controller-manager-55b6b7c7b8-rsbgr" Oct 08 18:30:53 crc kubenswrapper[4859]: I1008 18:30:53.441059 4859 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-zj6wn\" (UniqueName: \"kubernetes.io/projected/e69529d1-1a2a-433c-a16b-3c6de62a0587-kube-api-access-zj6wn\") pod \"neutron-operator-controller-manager-79d585cb66-zn65q\" (UID: \"e69529d1-1a2a-433c-a16b-3c6de62a0587\") " pod="openstack-operators/neutron-operator-controller-manager-79d585cb66-zn65q" Oct 08 18:30:53 crc kubenswrapper[4859]: I1008 18:30:53.441255 4859 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/horizon-operator-controller-manager-7ffbcb7588-86z4n" Oct 08 18:30:53 crc kubenswrapper[4859]: I1008 18:30:53.468726 4859 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/ironic-operator-controller-manager-9c5c78d49-z9952" Oct 08 18:30:53 crc kubenswrapper[4859]: I1008 18:30:53.479797 4859 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/placement-operator-controller-manager-68b6c87b68-xgjpk"] Oct 08 18:30:53 crc kubenswrapper[4859]: I1008 18:30:53.493102 4859 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/placement-operator-controller-manager-68b6c87b68-xgjpk" Oct 08 18:30:53 crc kubenswrapper[4859]: I1008 18:30:53.516478 4859 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-vqs92\" (UniqueName: \"kubernetes.io/projected/0228033e-c835-4618-9603-2c67cf9ce57d-kube-api-access-vqs92\") pod \"keystone-operator-controller-manager-55b6b7c7b8-rsbgr\" (UID: \"0228033e-c835-4618-9603-2c67cf9ce57d\") " pod="openstack-operators/keystone-operator-controller-manager-55b6b7c7b8-rsbgr" Oct 08 18:30:53 crc kubenswrapper[4859]: I1008 18:30:53.529416 4859 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"placement-operator-controller-manager-dockercfg-dnbsc" Oct 08 18:30:53 crc kubenswrapper[4859]: I1008 18:30:53.538665 4859 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/openstack-baremetal-operator-controller-manager-677c5f5bff5pg7p"] Oct 08 18:30:53 crc kubenswrapper[4859]: I1008 18:30:53.546209 4859 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/keystone-operator-controller-manager-55b6b7c7b8-rsbgr" Oct 08 18:30:53 crc kubenswrapper[4859]: I1008 18:30:53.563471 4859 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-n6kv2\" (UniqueName: \"kubernetes.io/projected/550bb7bb-6f9c-4b20-9d92-2c85773e05a2-kube-api-access-n6kv2\") pod \"octavia-operator-controller-manager-69fdcfc5f5-lsl8l\" (UID: \"550bb7bb-6f9c-4b20-9d92-2c85773e05a2\") " pod="openstack-operators/octavia-operator-controller-manager-69fdcfc5f5-lsl8l" Oct 08 18:30:53 crc kubenswrapper[4859]: I1008 18:30:53.563566 4859 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-4w2pq\" (UniqueName: \"kubernetes.io/projected/dd924000-5690-490e-a4db-e29ed7f821ec-kube-api-access-4w2pq\") pod \"openstack-baremetal-operator-controller-manager-677c5f5bff5pg7p\" (UID: \"dd924000-5690-490e-a4db-e29ed7f821ec\") " pod="openstack-operators/openstack-baremetal-operator-controller-manager-677c5f5bff5pg7p" Oct 08 18:30:53 crc kubenswrapper[4859]: I1008 18:30:53.563619 4859 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/dd924000-5690-490e-a4db-e29ed7f821ec-cert\") pod \"openstack-baremetal-operator-controller-manager-677c5f5bff5pg7p\" (UID: \"dd924000-5690-490e-a4db-e29ed7f821ec\") " pod="openstack-operators/openstack-baremetal-operator-controller-manager-677c5f5bff5pg7p" Oct 08 18:30:53 crc kubenswrapper[4859]: I1008 18:30:53.563643 4859 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-jqbg7\" (UniqueName: \"kubernetes.io/projected/fbce123a-545f-4c4c-86ca-2d68f81aa3f3-kube-api-access-jqbg7\") pod \"manila-operator-controller-manager-5f67fbc655-xg74g\" (UID: \"fbce123a-545f-4c4c-86ca-2d68f81aa3f3\") " pod="openstack-operators/manila-operator-controller-manager-5f67fbc655-xg74g" Oct 08 18:30:53 crc kubenswrapper[4859]: I1008 18:30:53.574821 4859 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/placement-operator-controller-manager-68b6c87b68-xgjpk"] Oct 08 18:30:53 crc kubenswrapper[4859]: I1008 18:30:53.580465 4859 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-zj6wn\" (UniqueName: \"kubernetes.io/projected/e69529d1-1a2a-433c-a16b-3c6de62a0587-kube-api-access-zj6wn\") pod \"neutron-operator-controller-manager-79d585cb66-zn65q\" (UID: \"e69529d1-1a2a-433c-a16b-3c6de62a0587\") " pod="openstack-operators/neutron-operator-controller-manager-79d585cb66-zn65q" Oct 08 18:30:53 crc kubenswrapper[4859]: I1008 18:30:53.580577 4859 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-59mk9\" (UniqueName: \"kubernetes.io/projected/e2675d43-51ec-4a8c-aa2d-5aedc14f649a-kube-api-access-59mk9\") pod \"ovn-operator-controller-manager-79db49b9fb-hhgrl\" (UID: \"e2675d43-51ec-4a8c-aa2d-5aedc14f649a\") " pod="openstack-operators/ovn-operator-controller-manager-79db49b9fb-hhgrl" Oct 08 18:30:53 crc kubenswrapper[4859]: I1008 18:30:53.580612 4859 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-fnbtf\" (UniqueName: \"kubernetes.io/projected/e2889a4a-89ec-47de-99df-7e749c3e5514-kube-api-access-fnbtf\") pod \"nova-operator-controller-manager-5df598886f-p99nw\" (UID: \"e2889a4a-89ec-47de-99df-7e749c3e5514\") " pod="openstack-operators/nova-operator-controller-manager-5df598886f-p99nw" Oct 08 18:30:53 crc kubenswrapper[4859]: I1008 18:30:53.580665 4859 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-9qb6z\" (UniqueName: \"kubernetes.io/projected/b6a2cffa-8421-43c4-bad8-5d4d341a40cc-kube-api-access-9qb6z\") pod \"mariadb-operator-controller-manager-f9fb45f8f-ptpqs\" (UID: \"b6a2cffa-8421-43c4-bad8-5d4d341a40cc\") " pod="openstack-operators/mariadb-operator-controller-manager-f9fb45f8f-ptpqs" Oct 08 18:30:53 crc kubenswrapper[4859]: I1008 18:30:53.601769 4859 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/cinder-operator-controller-manager-7b7fb68549-lbw64" Oct 08 18:30:53 crc kubenswrapper[4859]: I1008 18:30:53.606931 4859 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/swift-operator-controller-manager-db6d7f97b-zjkwm"] Oct 08 18:30:53 crc kubenswrapper[4859]: I1008 18:30:53.608629 4859 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/swift-operator-controller-manager-db6d7f97b-zjkwm" Oct 08 18:30:53 crc kubenswrapper[4859]: I1008 18:30:53.627945 4859 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-jqbg7\" (UniqueName: \"kubernetes.io/projected/fbce123a-545f-4c4c-86ca-2d68f81aa3f3-kube-api-access-jqbg7\") pod \"manila-operator-controller-manager-5f67fbc655-xg74g\" (UID: \"fbce123a-545f-4c4c-86ca-2d68f81aa3f3\") " pod="openstack-operators/manila-operator-controller-manager-5f67fbc655-xg74g" Oct 08 18:30:53 crc kubenswrapper[4859]: I1008 18:30:53.630255 4859 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"swift-operator-controller-manager-dockercfg-kd5k9" Oct 08 18:30:53 crc kubenswrapper[4859]: I1008 18:30:53.630719 4859 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-9qb6z\" (UniqueName: \"kubernetes.io/projected/b6a2cffa-8421-43c4-bad8-5d4d341a40cc-kube-api-access-9qb6z\") pod \"mariadb-operator-controller-manager-f9fb45f8f-ptpqs\" (UID: \"b6a2cffa-8421-43c4-bad8-5d4d341a40cc\") " pod="openstack-operators/mariadb-operator-controller-manager-f9fb45f8f-ptpqs" Oct 08 18:30:53 crc kubenswrapper[4859]: I1008 18:30:53.642406 4859 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-fnbtf\" (UniqueName: \"kubernetes.io/projected/e2889a4a-89ec-47de-99df-7e749c3e5514-kube-api-access-fnbtf\") pod \"nova-operator-controller-manager-5df598886f-p99nw\" (UID: \"e2889a4a-89ec-47de-99df-7e749c3e5514\") " pod="openstack-operators/nova-operator-controller-manager-5df598886f-p99nw" Oct 08 18:30:53 crc kubenswrapper[4859]: I1008 18:30:53.643395 4859 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-zj6wn\" (UniqueName: \"kubernetes.io/projected/e69529d1-1a2a-433c-a16b-3c6de62a0587-kube-api-access-zj6wn\") pod \"neutron-operator-controller-manager-79d585cb66-zn65q\" (UID: \"e69529d1-1a2a-433c-a16b-3c6de62a0587\") " pod="openstack-operators/neutron-operator-controller-manager-79d585cb66-zn65q" Oct 08 18:30:53 crc kubenswrapper[4859]: I1008 18:30:53.671662 4859 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/swift-operator-controller-manager-db6d7f97b-zjkwm"] Oct 08 18:30:53 crc kubenswrapper[4859]: I1008 18:30:53.686663 4859 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/dd924000-5690-490e-a4db-e29ed7f821ec-cert\") pod \"openstack-baremetal-operator-controller-manager-677c5f5bff5pg7p\" (UID: \"dd924000-5690-490e-a4db-e29ed7f821ec\") " pod="openstack-operators/openstack-baremetal-operator-controller-manager-677c5f5bff5pg7p" Oct 08 18:30:53 crc kubenswrapper[4859]: I1008 18:30:53.686755 4859 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-59mk9\" (UniqueName: \"kubernetes.io/projected/e2675d43-51ec-4a8c-aa2d-5aedc14f649a-kube-api-access-59mk9\") pod \"ovn-operator-controller-manager-79db49b9fb-hhgrl\" (UID: \"e2675d43-51ec-4a8c-aa2d-5aedc14f649a\") " pod="openstack-operators/ovn-operator-controller-manager-79db49b9fb-hhgrl" Oct 08 18:30:53 crc kubenswrapper[4859]: I1008 18:30:53.686793 4859 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-cjng4\" (UniqueName: \"kubernetes.io/projected/f56e7a0c-7b3a-4dc8-9367-cfeef5c5cef9-kube-api-access-cjng4\") pod \"placement-operator-controller-manager-68b6c87b68-xgjpk\" (UID: \"f56e7a0c-7b3a-4dc8-9367-cfeef5c5cef9\") " pod="openstack-operators/placement-operator-controller-manager-68b6c87b68-xgjpk" Oct 08 18:30:53 crc kubenswrapper[4859]: I1008 18:30:53.686818 4859 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-n6kv2\" (UniqueName: \"kubernetes.io/projected/550bb7bb-6f9c-4b20-9d92-2c85773e05a2-kube-api-access-n6kv2\") pod \"octavia-operator-controller-manager-69fdcfc5f5-lsl8l\" (UID: \"550bb7bb-6f9c-4b20-9d92-2c85773e05a2\") " pod="openstack-operators/octavia-operator-controller-manager-69fdcfc5f5-lsl8l" Oct 08 18:30:53 crc kubenswrapper[4859]: I1008 18:30:53.686846 4859 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-4w2pq\" (UniqueName: \"kubernetes.io/projected/dd924000-5690-490e-a4db-e29ed7f821ec-kube-api-access-4w2pq\") pod \"openstack-baremetal-operator-controller-manager-677c5f5bff5pg7p\" (UID: \"dd924000-5690-490e-a4db-e29ed7f821ec\") " pod="openstack-operators/openstack-baremetal-operator-controller-manager-677c5f5bff5pg7p" Oct 08 18:30:53 crc kubenswrapper[4859]: I1008 18:30:53.686867 4859 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-n4bpm\" (UniqueName: \"kubernetes.io/projected/44980d79-924e-436f-950d-8fcdb4c42af1-kube-api-access-n4bpm\") pod \"swift-operator-controller-manager-db6d7f97b-zjkwm\" (UID: \"44980d79-924e-436f-950d-8fcdb4c42af1\") " pod="openstack-operators/swift-operator-controller-manager-db6d7f97b-zjkwm" Oct 08 18:30:53 crc kubenswrapper[4859]: E1008 18:30:53.687011 4859 secret.go:188] Couldn't get secret openstack-operators/openstack-baremetal-operator-webhook-server-cert: secret "openstack-baremetal-operator-webhook-server-cert" not found Oct 08 18:30:53 crc kubenswrapper[4859]: E1008 18:30:53.687048 4859 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/dd924000-5690-490e-a4db-e29ed7f821ec-cert podName:dd924000-5690-490e-a4db-e29ed7f821ec nodeName:}" failed. No retries permitted until 2025-10-08 18:30:54.187036649 +0000 UTC m=+824.433876028 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "cert" (UniqueName: "kubernetes.io/secret/dd924000-5690-490e-a4db-e29ed7f821ec-cert") pod "openstack-baremetal-operator-controller-manager-677c5f5bff5pg7p" (UID: "dd924000-5690-490e-a4db-e29ed7f821ec") : secret "openstack-baremetal-operator-webhook-server-cert" not found Oct 08 18:30:53 crc kubenswrapper[4859]: I1008 18:30:53.687567 4859 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/telemetry-operator-controller-manager-76796d4c6b-kjbfp"] Oct 08 18:30:53 crc kubenswrapper[4859]: I1008 18:30:53.688599 4859 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/telemetry-operator-controller-manager-76796d4c6b-kjbfp" Oct 08 18:30:53 crc kubenswrapper[4859]: I1008 18:30:53.708527 4859 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/telemetry-operator-controller-manager-76796d4c6b-kjbfp"] Oct 08 18:30:53 crc kubenswrapper[4859]: I1008 18:30:53.710999 4859 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"telemetry-operator-controller-manager-dockercfg-4dlv7" Oct 08 18:30:53 crc kubenswrapper[4859]: I1008 18:30:53.734101 4859 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-n6kv2\" (UniqueName: \"kubernetes.io/projected/550bb7bb-6f9c-4b20-9d92-2c85773e05a2-kube-api-access-n6kv2\") pod \"octavia-operator-controller-manager-69fdcfc5f5-lsl8l\" (UID: \"550bb7bb-6f9c-4b20-9d92-2c85773e05a2\") " pod="openstack-operators/octavia-operator-controller-manager-69fdcfc5f5-lsl8l" Oct 08 18:30:53 crc kubenswrapper[4859]: I1008 18:30:53.743016 4859 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/test-operator-controller-manager-56c698c775-d2g8b"] Oct 08 18:30:53 crc kubenswrapper[4859]: I1008 18:30:53.744343 4859 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/test-operator-controller-manager-56c698c775-d2g8b" Oct 08 18:30:53 crc kubenswrapper[4859]: I1008 18:30:53.749921 4859 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"test-operator-controller-manager-dockercfg-wplnk" Oct 08 18:30:53 crc kubenswrapper[4859]: I1008 18:30:53.757607 4859 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-4w2pq\" (UniqueName: \"kubernetes.io/projected/dd924000-5690-490e-a4db-e29ed7f821ec-kube-api-access-4w2pq\") pod \"openstack-baremetal-operator-controller-manager-677c5f5bff5pg7p\" (UID: \"dd924000-5690-490e-a4db-e29ed7f821ec\") " pod="openstack-operators/openstack-baremetal-operator-controller-manager-677c5f5bff5pg7p" Oct 08 18:30:53 crc kubenswrapper[4859]: I1008 18:30:53.758877 4859 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-59mk9\" (UniqueName: \"kubernetes.io/projected/e2675d43-51ec-4a8c-aa2d-5aedc14f649a-kube-api-access-59mk9\") pod \"ovn-operator-controller-manager-79db49b9fb-hhgrl\" (UID: \"e2675d43-51ec-4a8c-aa2d-5aedc14f649a\") " pod="openstack-operators/ovn-operator-controller-manager-79db49b9fb-hhgrl" Oct 08 18:30:53 crc kubenswrapper[4859]: I1008 18:30:53.767301 4859 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/test-operator-controller-manager-56c698c775-d2g8b"] Oct 08 18:30:53 crc kubenswrapper[4859]: I1008 18:30:53.767868 4859 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/mariadb-operator-controller-manager-f9fb45f8f-ptpqs" Oct 08 18:30:53 crc kubenswrapper[4859]: I1008 18:30:53.771308 4859 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/neutron-operator-controller-manager-79d585cb66-zn65q" Oct 08 18:30:53 crc kubenswrapper[4859]: I1008 18:30:53.793624 4859 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-n4bpm\" (UniqueName: \"kubernetes.io/projected/44980d79-924e-436f-950d-8fcdb4c42af1-kube-api-access-n4bpm\") pod \"swift-operator-controller-manager-db6d7f97b-zjkwm\" (UID: \"44980d79-924e-436f-950d-8fcdb4c42af1\") " pod="openstack-operators/swift-operator-controller-manager-db6d7f97b-zjkwm" Oct 08 18:30:53 crc kubenswrapper[4859]: I1008 18:30:53.793667 4859 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-2rtrv\" (UniqueName: \"kubernetes.io/projected/ffa2914a-717c-4763-8b1c-4a1ff18858f5-kube-api-access-2rtrv\") pod \"telemetry-operator-controller-manager-76796d4c6b-kjbfp\" (UID: \"ffa2914a-717c-4763-8b1c-4a1ff18858f5\") " pod="openstack-operators/telemetry-operator-controller-manager-76796d4c6b-kjbfp" Oct 08 18:30:53 crc kubenswrapper[4859]: I1008 18:30:53.793873 4859 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-tfwsj\" (UniqueName: \"kubernetes.io/projected/2d6cf85d-c5cc-4466-a471-bb17aa8fca9d-kube-api-access-tfwsj\") pod \"test-operator-controller-manager-56c698c775-d2g8b\" (UID: \"2d6cf85d-c5cc-4466-a471-bb17aa8fca9d\") " pod="openstack-operators/test-operator-controller-manager-56c698c775-d2g8b" Oct 08 18:30:53 crc kubenswrapper[4859]: I1008 18:30:53.793901 4859 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cjng4\" (UniqueName: \"kubernetes.io/projected/f56e7a0c-7b3a-4dc8-9367-cfeef5c5cef9-kube-api-access-cjng4\") pod \"placement-operator-controller-manager-68b6c87b68-xgjpk\" (UID: \"f56e7a0c-7b3a-4dc8-9367-cfeef5c5cef9\") " pod="openstack-operators/placement-operator-controller-manager-68b6c87b68-xgjpk" Oct 08 18:30:53 crc kubenswrapper[4859]: I1008 18:30:53.800790 4859 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/watcher-operator-controller-manager-7794bc6bd-2v65r"] Oct 08 18:30:53 crc kubenswrapper[4859]: I1008 18:30:53.802034 4859 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/watcher-operator-controller-manager-7794bc6bd-2v65r"] Oct 08 18:30:53 crc kubenswrapper[4859]: I1008 18:30:53.802143 4859 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/watcher-operator-controller-manager-7794bc6bd-2v65r" Oct 08 18:30:53 crc kubenswrapper[4859]: I1008 18:30:53.818825 4859 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"watcher-operator-controller-manager-dockercfg-sjlmb" Oct 08 18:30:53 crc kubenswrapper[4859]: I1008 18:30:53.819001 4859 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/nova-operator-controller-manager-5df598886f-p99nw" Oct 08 18:30:53 crc kubenswrapper[4859]: I1008 18:30:53.834117 4859 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/openstack-operator-controller-manager-598c4c5b5-6l55k"] Oct 08 18:30:53 crc kubenswrapper[4859]: I1008 18:30:53.835304 4859 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-operator-controller-manager-598c4c5b5-6l55k" Oct 08 18:30:53 crc kubenswrapper[4859]: I1008 18:30:53.835990 4859 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/octavia-operator-controller-manager-69fdcfc5f5-lsl8l" Oct 08 18:30:53 crc kubenswrapper[4859]: I1008 18:30:53.844582 4859 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-cjng4\" (UniqueName: \"kubernetes.io/projected/f56e7a0c-7b3a-4dc8-9367-cfeef5c5cef9-kube-api-access-cjng4\") pod \"placement-operator-controller-manager-68b6c87b68-xgjpk\" (UID: \"f56e7a0c-7b3a-4dc8-9367-cfeef5c5cef9\") " pod="openstack-operators/placement-operator-controller-manager-68b6c87b68-xgjpk" Oct 08 18:30:53 crc kubenswrapper[4859]: I1008 18:30:53.849444 4859 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"openstack-operator-controller-manager-dockercfg-5bq9p" Oct 08 18:30:53 crc kubenswrapper[4859]: I1008 18:30:53.851188 4859 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"webhook-server-cert" Oct 08 18:30:53 crc kubenswrapper[4859]: I1008 18:30:53.857985 4859 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/openstack-operator-controller-manager-598c4c5b5-6l55k"] Oct 08 18:30:53 crc kubenswrapper[4859]: I1008 18:30:53.862341 4859 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/ovn-operator-controller-manager-79db49b9fb-hhgrl" Oct 08 18:30:53 crc kubenswrapper[4859]: I1008 18:30:53.864442 4859 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-n4bpm\" (UniqueName: \"kubernetes.io/projected/44980d79-924e-436f-950d-8fcdb4c42af1-kube-api-access-n4bpm\") pod \"swift-operator-controller-manager-db6d7f97b-zjkwm\" (UID: \"44980d79-924e-436f-950d-8fcdb4c42af1\") " pod="openstack-operators/swift-operator-controller-manager-db6d7f97b-zjkwm" Oct 08 18:30:53 crc kubenswrapper[4859]: I1008 18:30:53.882880 4859 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/manila-operator-controller-manager-5f67fbc655-xg74g" Oct 08 18:30:53 crc kubenswrapper[4859]: I1008 18:30:53.895936 4859 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-tfwsj\" (UniqueName: \"kubernetes.io/projected/2d6cf85d-c5cc-4466-a471-bb17aa8fca9d-kube-api-access-tfwsj\") pod \"test-operator-controller-manager-56c698c775-d2g8b\" (UID: \"2d6cf85d-c5cc-4466-a471-bb17aa8fca9d\") " pod="openstack-operators/test-operator-controller-manager-56c698c775-d2g8b" Oct 08 18:30:53 crc kubenswrapper[4859]: I1008 18:30:53.895986 4859 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/182ef4ca-109f-46bd-a935-091d05275271-cert\") pod \"openstack-operator-controller-manager-598c4c5b5-6l55k\" (UID: \"182ef4ca-109f-46bd-a935-091d05275271\") " pod="openstack-operators/openstack-operator-controller-manager-598c4c5b5-6l55k" Oct 08 18:30:53 crc kubenswrapper[4859]: I1008 18:30:53.896020 4859 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-wzlgx\" (UniqueName: \"kubernetes.io/projected/182ef4ca-109f-46bd-a935-091d05275271-kube-api-access-wzlgx\") pod \"openstack-operator-controller-manager-598c4c5b5-6l55k\" (UID: \"182ef4ca-109f-46bd-a935-091d05275271\") " pod="openstack-operators/openstack-operator-controller-manager-598c4c5b5-6l55k" Oct 08 18:30:53 crc kubenswrapper[4859]: I1008 18:30:53.896046 4859 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-2rtrv\" (UniqueName: \"kubernetes.io/projected/ffa2914a-717c-4763-8b1c-4a1ff18858f5-kube-api-access-2rtrv\") pod \"telemetry-operator-controller-manager-76796d4c6b-kjbfp\" (UID: \"ffa2914a-717c-4763-8b1c-4a1ff18858f5\") " pod="openstack-operators/telemetry-operator-controller-manager-76796d4c6b-kjbfp" Oct 08 18:30:53 crc kubenswrapper[4859]: I1008 18:30:53.896086 4859 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/1a1cfd66-310a-4077-8020-df7e6e069c53-cert\") pod \"infra-operator-controller-manager-656bcbd775-ctjz9\" (UID: \"1a1cfd66-310a-4077-8020-df7e6e069c53\") " pod="openstack-operators/infra-operator-controller-manager-656bcbd775-ctjz9" Oct 08 18:30:53 crc kubenswrapper[4859]: I1008 18:30:53.896111 4859 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-94589\" (UniqueName: \"kubernetes.io/projected/cf5b6741-336e-435c-8209-82b8e19f4896-kube-api-access-94589\") pod \"watcher-operator-controller-manager-7794bc6bd-2v65r\" (UID: \"cf5b6741-336e-435c-8209-82b8e19f4896\") " pod="openstack-operators/watcher-operator-controller-manager-7794bc6bd-2v65r" Oct 08 18:30:53 crc kubenswrapper[4859]: E1008 18:30:53.896389 4859 secret.go:188] Couldn't get secret openstack-operators/infra-operator-webhook-server-cert: secret "infra-operator-webhook-server-cert" not found Oct 08 18:30:53 crc kubenswrapper[4859]: E1008 18:30:53.896440 4859 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/1a1cfd66-310a-4077-8020-df7e6e069c53-cert podName:1a1cfd66-310a-4077-8020-df7e6e069c53 nodeName:}" failed. No retries permitted until 2025-10-08 18:30:54.896424657 +0000 UTC m=+825.143264036 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "cert" (UniqueName: "kubernetes.io/secret/1a1cfd66-310a-4077-8020-df7e6e069c53-cert") pod "infra-operator-controller-manager-656bcbd775-ctjz9" (UID: "1a1cfd66-310a-4077-8020-df7e6e069c53") : secret "infra-operator-webhook-server-cert" not found Oct 08 18:30:53 crc kubenswrapper[4859]: I1008 18:30:53.920705 4859 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/placement-operator-controller-manager-68b6c87b68-xgjpk" Oct 08 18:30:53 crc kubenswrapper[4859]: I1008 18:30:53.942273 4859 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-tfwsj\" (UniqueName: \"kubernetes.io/projected/2d6cf85d-c5cc-4466-a471-bb17aa8fca9d-kube-api-access-tfwsj\") pod \"test-operator-controller-manager-56c698c775-d2g8b\" (UID: \"2d6cf85d-c5cc-4466-a471-bb17aa8fca9d\") " pod="openstack-operators/test-operator-controller-manager-56c698c775-d2g8b" Oct 08 18:30:53 crc kubenswrapper[4859]: I1008 18:30:53.948067 4859 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-2rtrv\" (UniqueName: \"kubernetes.io/projected/ffa2914a-717c-4763-8b1c-4a1ff18858f5-kube-api-access-2rtrv\") pod \"telemetry-operator-controller-manager-76796d4c6b-kjbfp\" (UID: \"ffa2914a-717c-4763-8b1c-4a1ff18858f5\") " pod="openstack-operators/telemetry-operator-controller-manager-76796d4c6b-kjbfp" Oct 08 18:30:53 crc kubenswrapper[4859]: I1008 18:30:53.953278 4859 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/rabbitmq-cluster-operator-manager-5f97d8c699-dxrl8"] Oct 08 18:30:53 crc kubenswrapper[4859]: I1008 18:30:53.954245 4859 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/rabbitmq-cluster-operator-manager-5f97d8c699-dxrl8" Oct 08 18:30:53 crc kubenswrapper[4859]: I1008 18:30:53.962221 4859 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"rabbitmq-cluster-operator-controller-manager-dockercfg-p9kkr" Oct 08 18:30:54 crc kubenswrapper[4859]: I1008 18:30:53.999347 4859 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/rabbitmq-cluster-operator-manager-5f97d8c699-dxrl8"] Oct 08 18:30:54 crc kubenswrapper[4859]: I1008 18:30:54.000247 4859 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-94589\" (UniqueName: \"kubernetes.io/projected/cf5b6741-336e-435c-8209-82b8e19f4896-kube-api-access-94589\") pod \"watcher-operator-controller-manager-7794bc6bd-2v65r\" (UID: \"cf5b6741-336e-435c-8209-82b8e19f4896\") " pod="openstack-operators/watcher-operator-controller-manager-7794bc6bd-2v65r" Oct 08 18:30:54 crc kubenswrapper[4859]: I1008 18:30:54.000318 4859 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/182ef4ca-109f-46bd-a935-091d05275271-cert\") pod \"openstack-operator-controller-manager-598c4c5b5-6l55k\" (UID: \"182ef4ca-109f-46bd-a935-091d05275271\") " pod="openstack-operators/openstack-operator-controller-manager-598c4c5b5-6l55k" Oct 08 18:30:54 crc kubenswrapper[4859]: I1008 18:30:54.000350 4859 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-wzlgx\" (UniqueName: \"kubernetes.io/projected/182ef4ca-109f-46bd-a935-091d05275271-kube-api-access-wzlgx\") pod \"openstack-operator-controller-manager-598c4c5b5-6l55k\" (UID: \"182ef4ca-109f-46bd-a935-091d05275271\") " pod="openstack-operators/openstack-operator-controller-manager-598c4c5b5-6l55k" Oct 08 18:30:54 crc kubenswrapper[4859]: E1008 18:30:54.000867 4859 secret.go:188] Couldn't get secret openstack-operators/webhook-server-cert: secret "webhook-server-cert" not found Oct 08 18:30:54 crc kubenswrapper[4859]: E1008 18:30:54.000947 4859 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/182ef4ca-109f-46bd-a935-091d05275271-cert podName:182ef4ca-109f-46bd-a935-091d05275271 nodeName:}" failed. No retries permitted until 2025-10-08 18:30:54.50091662 +0000 UTC m=+824.747756209 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "cert" (UniqueName: "kubernetes.io/secret/182ef4ca-109f-46bd-a935-091d05275271-cert") pod "openstack-operator-controller-manager-598c4c5b5-6l55k" (UID: "182ef4ca-109f-46bd-a935-091d05275271") : secret "webhook-server-cert" not found Oct 08 18:30:54 crc kubenswrapper[4859]: I1008 18:30:54.043666 4859 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/swift-operator-controller-manager-db6d7f97b-zjkwm" Oct 08 18:30:54 crc kubenswrapper[4859]: I1008 18:30:54.044906 4859 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-94589\" (UniqueName: \"kubernetes.io/projected/cf5b6741-336e-435c-8209-82b8e19f4896-kube-api-access-94589\") pod \"watcher-operator-controller-manager-7794bc6bd-2v65r\" (UID: \"cf5b6741-336e-435c-8209-82b8e19f4896\") " pod="openstack-operators/watcher-operator-controller-manager-7794bc6bd-2v65r" Oct 08 18:30:54 crc kubenswrapper[4859]: I1008 18:30:54.057706 4859 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-wzlgx\" (UniqueName: \"kubernetes.io/projected/182ef4ca-109f-46bd-a935-091d05275271-kube-api-access-wzlgx\") pod \"openstack-operator-controller-manager-598c4c5b5-6l55k\" (UID: \"182ef4ca-109f-46bd-a935-091d05275271\") " pod="openstack-operators/openstack-operator-controller-manager-598c4c5b5-6l55k" Oct 08 18:30:54 crc kubenswrapper[4859]: I1008 18:30:54.064937 4859 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/telemetry-operator-controller-manager-76796d4c6b-kjbfp" Oct 08 18:30:54 crc kubenswrapper[4859]: I1008 18:30:54.093168 4859 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/test-operator-controller-manager-56c698c775-d2g8b" Oct 08 18:30:54 crc kubenswrapper[4859]: I1008 18:30:54.107052 4859 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-k2nrr\" (UniqueName: \"kubernetes.io/projected/e1dbe8d7-f993-42b1-9a53-7d0c342780c1-kube-api-access-k2nrr\") pod \"rabbitmq-cluster-operator-manager-5f97d8c699-dxrl8\" (UID: \"e1dbe8d7-f993-42b1-9a53-7d0c342780c1\") " pod="openstack-operators/rabbitmq-cluster-operator-manager-5f97d8c699-dxrl8" Oct 08 18:30:54 crc kubenswrapper[4859]: I1008 18:30:54.156992 4859 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/watcher-operator-controller-manager-7794bc6bd-2v65r" Oct 08 18:30:54 crc kubenswrapper[4859]: I1008 18:30:54.214901 4859 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/dd924000-5690-490e-a4db-e29ed7f821ec-cert\") pod \"openstack-baremetal-operator-controller-manager-677c5f5bff5pg7p\" (UID: \"dd924000-5690-490e-a4db-e29ed7f821ec\") " pod="openstack-operators/openstack-baremetal-operator-controller-manager-677c5f5bff5pg7p" Oct 08 18:30:54 crc kubenswrapper[4859]: I1008 18:30:54.215029 4859 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-k2nrr\" (UniqueName: \"kubernetes.io/projected/e1dbe8d7-f993-42b1-9a53-7d0c342780c1-kube-api-access-k2nrr\") pod \"rabbitmq-cluster-operator-manager-5f97d8c699-dxrl8\" (UID: \"e1dbe8d7-f993-42b1-9a53-7d0c342780c1\") " pod="openstack-operators/rabbitmq-cluster-operator-manager-5f97d8c699-dxrl8" Oct 08 18:30:54 crc kubenswrapper[4859]: I1008 18:30:54.230565 4859 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert\" (UniqueName: \"kubernetes.io/secret/dd924000-5690-490e-a4db-e29ed7f821ec-cert\") pod \"openstack-baremetal-operator-controller-manager-677c5f5bff5pg7p\" (UID: \"dd924000-5690-490e-a4db-e29ed7f821ec\") " pod="openstack-operators/openstack-baremetal-operator-controller-manager-677c5f5bff5pg7p" Oct 08 18:30:54 crc kubenswrapper[4859]: I1008 18:30:54.238406 4859 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-k2nrr\" (UniqueName: \"kubernetes.io/projected/e1dbe8d7-f993-42b1-9a53-7d0c342780c1-kube-api-access-k2nrr\") pod \"rabbitmq-cluster-operator-manager-5f97d8c699-dxrl8\" (UID: \"e1dbe8d7-f993-42b1-9a53-7d0c342780c1\") " pod="openstack-operators/rabbitmq-cluster-operator-manager-5f97d8c699-dxrl8" Oct 08 18:30:54 crc kubenswrapper[4859]: I1008 18:30:54.282090 4859 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/barbican-operator-controller-manager-658bdf4b74-8bfst"] Oct 08 18:30:54 crc kubenswrapper[4859]: I1008 18:30:54.308881 4859 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/rabbitmq-cluster-operator-manager-5f97d8c699-dxrl8" Oct 08 18:30:54 crc kubenswrapper[4859]: I1008 18:30:54.406968 4859 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/designate-operator-controller-manager-85d5d9dd78-nhqkh"] Oct 08 18:30:54 crc kubenswrapper[4859]: I1008 18:30:54.486839 4859 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-baremetal-operator-controller-manager-677c5f5bff5pg7p" Oct 08 18:30:54 crc kubenswrapper[4859]: I1008 18:30:54.524607 4859 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/182ef4ca-109f-46bd-a935-091d05275271-cert\") pod \"openstack-operator-controller-manager-598c4c5b5-6l55k\" (UID: \"182ef4ca-109f-46bd-a935-091d05275271\") " pod="openstack-operators/openstack-operator-controller-manager-598c4c5b5-6l55k" Oct 08 18:30:54 crc kubenswrapper[4859]: I1008 18:30:54.534126 4859 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert\" (UniqueName: \"kubernetes.io/secret/182ef4ca-109f-46bd-a935-091d05275271-cert\") pod \"openstack-operator-controller-manager-598c4c5b5-6l55k\" (UID: \"182ef4ca-109f-46bd-a935-091d05275271\") " pod="openstack-operators/openstack-operator-controller-manager-598c4c5b5-6l55k" Oct 08 18:30:54 crc kubenswrapper[4859]: I1008 18:30:54.563134 4859 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/heat-operator-controller-manager-858f76bbdd-b5bfz"] Oct 08 18:30:54 crc kubenswrapper[4859]: W1008 18:30:54.638269 4859 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod19fa1d1d_0731_4c21_9faa_b1665ec4af36.slice/crio-ee848f8bdd4be59b7d53f2a6b022efdce8e05e613f276e0b4fce4189814663ab WatchSource:0}: Error finding container ee848f8bdd4be59b7d53f2a6b022efdce8e05e613f276e0b4fce4189814663ab: Status 404 returned error can't find the container with id ee848f8bdd4be59b7d53f2a6b022efdce8e05e613f276e0b4fce4189814663ab Oct 08 18:30:54 crc kubenswrapper[4859]: I1008 18:30:54.644814 4859 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/horizon-operator-controller-manager-7ffbcb7588-86z4n"] Oct 08 18:30:54 crc kubenswrapper[4859]: I1008 18:30:54.781253 4859 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/glance-operator-controller-manager-84b9b84486-d9gmv"] Oct 08 18:30:54 crc kubenswrapper[4859]: I1008 18:30:54.782877 4859 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-operator-controller-manager-598c4c5b5-6l55k" Oct 08 18:30:54 crc kubenswrapper[4859]: I1008 18:30:54.789836 4859 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/ironic-operator-controller-manager-9c5c78d49-z9952"] Oct 08 18:30:54 crc kubenswrapper[4859]: I1008 18:30:54.931952 4859 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/ironic-operator-controller-manager-9c5c78d49-z9952" event={"ID":"2b67febf-c4c8-4e48-9350-29dd496439b4","Type":"ContainerStarted","Data":"5a32cdaa829e1ba4a7143fade3236dd21caf258d4abad3f8aa50794fa4bb02af"} Oct 08 18:30:54 crc kubenswrapper[4859]: I1008 18:30:54.932397 4859 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/1a1cfd66-310a-4077-8020-df7e6e069c53-cert\") pod \"infra-operator-controller-manager-656bcbd775-ctjz9\" (UID: \"1a1cfd66-310a-4077-8020-df7e6e069c53\") " pod="openstack-operators/infra-operator-controller-manager-656bcbd775-ctjz9" Oct 08 18:30:54 crc kubenswrapper[4859]: I1008 18:30:54.933412 4859 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/barbican-operator-controller-manager-658bdf4b74-8bfst" event={"ID":"d097cd11-b23b-4d38-80d7-5d85c257cef4","Type":"ContainerStarted","Data":"4a5c2fdfc40283e2714008db9fc66bbbd6c96220b80c6ace2447a49c66f32cd5"} Oct 08 18:30:54 crc kubenswrapper[4859]: I1008 18:30:54.934606 4859 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/glance-operator-controller-manager-84b9b84486-d9gmv" event={"ID":"0b068d16-0072-41f8-ad01-4194fd428014","Type":"ContainerStarted","Data":"14adbf85ce9b105adc95350be5794a43b3781b935109881850d4cef253eaf1ff"} Oct 08 18:30:54 crc kubenswrapper[4859]: I1008 18:30:54.936805 4859 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/horizon-operator-controller-manager-7ffbcb7588-86z4n" event={"ID":"ca71782e-3569-4ac2-a178-239f0ae200d0","Type":"ContainerStarted","Data":"c115e3eb6e262780e634c3c3da8efbcb69c45cc11f16a88d7a10f9b611293d80"} Oct 08 18:30:54 crc kubenswrapper[4859]: I1008 18:30:54.937884 4859 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/designate-operator-controller-manager-85d5d9dd78-nhqkh" event={"ID":"d9a50bbe-f28f-46d8-9266-fd54e44388a3","Type":"ContainerStarted","Data":"9dd72d4ff9b2a849776824124b6abb1a90329e65e7f09cb9fdc7ff028462d841"} Oct 08 18:30:54 crc kubenswrapper[4859]: I1008 18:30:54.939593 4859 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/heat-operator-controller-manager-858f76bbdd-b5bfz" event={"ID":"19fa1d1d-0731-4c21-9faa-b1665ec4af36","Type":"ContainerStarted","Data":"ee848f8bdd4be59b7d53f2a6b022efdce8e05e613f276e0b4fce4189814663ab"} Oct 08 18:30:54 crc kubenswrapper[4859]: I1008 18:30:54.939893 4859 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-operators-jws9w" podUID="51932a59-6c7a-4c64-9acf-59984e192d91" containerName="registry-server" containerID="cri-o://51de042cfa3748c000a73b19681f8fba61c0263e8fa8c6232162000607ec37bb" gracePeriod=2 Oct 08 18:30:54 crc kubenswrapper[4859]: I1008 18:30:54.943063 4859 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert\" (UniqueName: \"kubernetes.io/secret/1a1cfd66-310a-4077-8020-df7e6e069c53-cert\") pod \"infra-operator-controller-manager-656bcbd775-ctjz9\" (UID: \"1a1cfd66-310a-4077-8020-df7e6e069c53\") " pod="openstack-operators/infra-operator-controller-manager-656bcbd775-ctjz9" Oct 08 18:30:54 crc kubenswrapper[4859]: I1008 18:30:54.957224 4859 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/infra-operator-controller-manager-656bcbd775-ctjz9" Oct 08 18:30:55 crc kubenswrapper[4859]: I1008 18:30:55.239518 4859 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/keystone-operator-controller-manager-55b6b7c7b8-rsbgr"] Oct 08 18:30:55 crc kubenswrapper[4859]: I1008 18:30:55.252130 4859 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/mariadb-operator-controller-manager-f9fb45f8f-ptpqs"] Oct 08 18:30:55 crc kubenswrapper[4859]: I1008 18:30:55.260669 4859 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/neutron-operator-controller-manager-79d585cb66-zn65q"] Oct 08 18:30:55 crc kubenswrapper[4859]: I1008 18:30:55.280482 4859 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/nova-operator-controller-manager-5df598886f-p99nw"] Oct 08 18:30:55 crc kubenswrapper[4859]: I1008 18:30:55.294320 4859 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/octavia-operator-controller-manager-69fdcfc5f5-lsl8l"] Oct 08 18:30:55 crc kubenswrapper[4859]: I1008 18:30:55.302605 4859 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/ovn-operator-controller-manager-79db49b9fb-hhgrl"] Oct 08 18:30:55 crc kubenswrapper[4859]: I1008 18:30:55.307529 4859 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/manila-operator-controller-manager-5f67fbc655-xg74g"] Oct 08 18:30:55 crc kubenswrapper[4859]: I1008 18:30:55.317608 4859 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/swift-operator-controller-manager-db6d7f97b-zjkwm"] Oct 08 18:30:55 crc kubenswrapper[4859]: I1008 18:30:55.317657 4859 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/cinder-operator-controller-manager-7b7fb68549-lbw64"] Oct 08 18:30:55 crc kubenswrapper[4859]: I1008 18:30:55.387916 4859 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-b8rjh"] Oct 08 18:30:55 crc kubenswrapper[4859]: I1008 18:30:55.389491 4859 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-b8rjh" Oct 08 18:30:55 crc kubenswrapper[4859]: I1008 18:30:55.407833 4859 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-b8rjh"] Oct 08 18:30:55 crc kubenswrapper[4859]: I1008 18:30:55.540845 4859 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/aefd0d7b-0c10-443e-9b21-6c47cb711eb7-utilities\") pod \"certified-operators-b8rjh\" (UID: \"aefd0d7b-0c10-443e-9b21-6c47cb711eb7\") " pod="openshift-marketplace/certified-operators-b8rjh" Oct 08 18:30:55 crc kubenswrapper[4859]: I1008 18:30:55.541367 4859 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-rhbnl\" (UniqueName: \"kubernetes.io/projected/aefd0d7b-0c10-443e-9b21-6c47cb711eb7-kube-api-access-rhbnl\") pod \"certified-operators-b8rjh\" (UID: \"aefd0d7b-0c10-443e-9b21-6c47cb711eb7\") " pod="openshift-marketplace/certified-operators-b8rjh" Oct 08 18:30:55 crc kubenswrapper[4859]: I1008 18:30:55.541550 4859 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/aefd0d7b-0c10-443e-9b21-6c47cb711eb7-catalog-content\") pod \"certified-operators-b8rjh\" (UID: \"aefd0d7b-0c10-443e-9b21-6c47cb711eb7\") " pod="openshift-marketplace/certified-operators-b8rjh" Oct 08 18:30:55 crc kubenswrapper[4859]: E1008 18:30:55.602152 4859 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:manager,Image:quay.io/openstack-k8s-operators/placement-operator@sha256:d33c1f507e1f5b9a4bf226ad98917e92101ac66b36e19d35cbe04ae7014f6bff,Command:[/manager],Args:[--health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080 --leader-elect],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LEASE_DURATION,Value:30,ValueFrom:nil,},EnvVar{Name:RENEW_DEADLINE,Value:20,ValueFrom:nil,},EnvVar{Name:RETRY_PERIOD,Value:5,ValueFrom:nil,},EnvVar{Name:ENABLE_WEBHOOKS,Value:false,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{536870912 0} {} BinarySI},},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{268435456 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-cjng4,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/healthz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:15,TimeoutSeconds:1,PeriodSeconds:20,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/readyz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:1,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod placement-operator-controller-manager-68b6c87b68-xgjpk_openstack-operators(f56e7a0c-7b3a-4dc8-9367-cfeef5c5cef9): ErrImagePull: pull QPS exceeded" logger="UnhandledError" Oct 08 18:30:55 crc kubenswrapper[4859]: I1008 18:30:55.610877 4859 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/telemetry-operator-controller-manager-76796d4c6b-kjbfp"] Oct 08 18:30:55 crc kubenswrapper[4859]: E1008 18:30:55.612444 4859 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:manager,Image:quay.io/openstack-k8s-operators/telemetry-operator@sha256:9d26476523320d70d6d457b91663e8c233ed320d77032a7c57a89ce1aedd3931,Command:[/manager],Args:[--health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080 --leader-elect],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LEASE_DURATION,Value:30,ValueFrom:nil,},EnvVar{Name:RENEW_DEADLINE,Value:20,ValueFrom:nil,},EnvVar{Name:RETRY_PERIOD,Value:5,ValueFrom:nil,},EnvVar{Name:ENABLE_WEBHOOKS,Value:false,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{536870912 0} {} BinarySI},},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{268435456 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-2rtrv,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/healthz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:15,TimeoutSeconds:1,PeriodSeconds:20,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/readyz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:1,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod telemetry-operator-controller-manager-76796d4c6b-kjbfp_openstack-operators(ffa2914a-717c-4763-8b1c-4a1ff18858f5): ErrImagePull: pull QPS exceeded" logger="UnhandledError" Oct 08 18:30:55 crc kubenswrapper[4859]: I1008 18:30:55.617598 4859 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/rabbitmq-cluster-operator-manager-5f97d8c699-dxrl8"] Oct 08 18:30:55 crc kubenswrapper[4859]: I1008 18:30:55.623964 4859 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/placement-operator-controller-manager-68b6c87b68-xgjpk"] Oct 08 18:30:55 crc kubenswrapper[4859]: E1008 18:30:55.627504 4859 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:manager,Image:quay.io/openstack-k8s-operators/test-operator@sha256:efa8fb78cffb573d299ffcc7bab1099affd2dbbab222152092b313074306e0a9,Command:[/manager],Args:[--health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080 --leader-elect],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LEASE_DURATION,Value:30,ValueFrom:nil,},EnvVar{Name:RENEW_DEADLINE,Value:20,ValueFrom:nil,},EnvVar{Name:RETRY_PERIOD,Value:5,ValueFrom:nil,},EnvVar{Name:ENABLE_WEBHOOKS,Value:false,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{536870912 0} {} BinarySI},},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{268435456 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-tfwsj,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/healthz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:15,TimeoutSeconds:1,PeriodSeconds:20,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/readyz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:1,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod test-operator-controller-manager-56c698c775-d2g8b_openstack-operators(2d6cf85d-c5cc-4466-a471-bb17aa8fca9d): ErrImagePull: pull QPS exceeded" logger="UnhandledError" Oct 08 18:30:55 crc kubenswrapper[4859]: I1008 18:30:55.643088 4859 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/aefd0d7b-0c10-443e-9b21-6c47cb711eb7-catalog-content\") pod \"certified-operators-b8rjh\" (UID: \"aefd0d7b-0c10-443e-9b21-6c47cb711eb7\") " pod="openshift-marketplace/certified-operators-b8rjh" Oct 08 18:30:55 crc kubenswrapper[4859]: I1008 18:30:55.643195 4859 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/aefd0d7b-0c10-443e-9b21-6c47cb711eb7-utilities\") pod \"certified-operators-b8rjh\" (UID: \"aefd0d7b-0c10-443e-9b21-6c47cb711eb7\") " pod="openshift-marketplace/certified-operators-b8rjh" Oct 08 18:30:55 crc kubenswrapper[4859]: I1008 18:30:55.643255 4859 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rhbnl\" (UniqueName: \"kubernetes.io/projected/aefd0d7b-0c10-443e-9b21-6c47cb711eb7-kube-api-access-rhbnl\") pod \"certified-operators-b8rjh\" (UID: \"aefd0d7b-0c10-443e-9b21-6c47cb711eb7\") " pod="openshift-marketplace/certified-operators-b8rjh" Oct 08 18:30:55 crc kubenswrapper[4859]: I1008 18:30:55.644258 4859 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/aefd0d7b-0c10-443e-9b21-6c47cb711eb7-catalog-content\") pod \"certified-operators-b8rjh\" (UID: \"aefd0d7b-0c10-443e-9b21-6c47cb711eb7\") " pod="openshift-marketplace/certified-operators-b8rjh" Oct 08 18:30:55 crc kubenswrapper[4859]: I1008 18:30:55.644588 4859 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/aefd0d7b-0c10-443e-9b21-6c47cb711eb7-utilities\") pod \"certified-operators-b8rjh\" (UID: \"aefd0d7b-0c10-443e-9b21-6c47cb711eb7\") " pod="openshift-marketplace/certified-operators-b8rjh" Oct 08 18:30:55 crc kubenswrapper[4859]: I1008 18:30:55.646882 4859 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/test-operator-controller-manager-56c698c775-d2g8b"] Oct 08 18:30:55 crc kubenswrapper[4859]: E1008 18:30:55.651130 4859 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:manager,Image:quay.io/openstack-k8s-operators/watcher-operator@sha256:e4ae07e859166fc5e2cb4f8e0e2c3358b9d2e2d6721a3864d2e0c651d36698ca,Command:[/manager],Args:[--health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080 --leader-elect],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LEASE_DURATION,Value:30,ValueFrom:nil,},EnvVar{Name:RENEW_DEADLINE,Value:20,ValueFrom:nil,},EnvVar{Name:RETRY_PERIOD,Value:5,ValueFrom:nil,},EnvVar{Name:ENABLE_WEBHOOKS,Value:false,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{536870912 0} {} BinarySI},},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{268435456 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-94589,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/healthz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:15,TimeoutSeconds:1,PeriodSeconds:20,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/readyz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:1,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod watcher-operator-controller-manager-7794bc6bd-2v65r_openstack-operators(cf5b6741-336e-435c-8209-82b8e19f4896): ErrImagePull: pull QPS exceeded" logger="UnhandledError" Oct 08 18:30:55 crc kubenswrapper[4859]: I1008 18:30:55.658279 4859 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/watcher-operator-controller-manager-7794bc6bd-2v65r"] Oct 08 18:30:55 crc kubenswrapper[4859]: I1008 18:30:55.665209 4859 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rhbnl\" (UniqueName: \"kubernetes.io/projected/aefd0d7b-0c10-443e-9b21-6c47cb711eb7-kube-api-access-rhbnl\") pod \"certified-operators-b8rjh\" (UID: \"aefd0d7b-0c10-443e-9b21-6c47cb711eb7\") " pod="openshift-marketplace/certified-operators-b8rjh" Oct 08 18:30:55 crc kubenswrapper[4859]: I1008 18:30:55.669898 4859 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/openstack-operator-controller-manager-598c4c5b5-6l55k"] Oct 08 18:30:55 crc kubenswrapper[4859]: I1008 18:30:55.680133 4859 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/openstack-baremetal-operator-controller-manager-677c5f5bff5pg7p"] Oct 08 18:30:55 crc kubenswrapper[4859]: I1008 18:30:55.698903 4859 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/infra-operator-controller-manager-656bcbd775-ctjz9"] Oct 08 18:30:55 crc kubenswrapper[4859]: I1008 18:30:55.715534 4859 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-b8rjh" Oct 08 18:30:55 crc kubenswrapper[4859]: E1008 18:30:55.797956 4859 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:manager,Image:quay.io/openstack-k8s-operators/openstack-baremetal-operator@sha256:a17fc270857869fd1efe5020b2a1cb8c2abbd838f08de88f3a6a59e8754ec351,Command:[/manager],Args:[--health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080 --leader-elect],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LEASE_DURATION,Value:30,ValueFrom:nil,},EnvVar{Name:RENEW_DEADLINE,Value:20,ValueFrom:nil,},EnvVar{Name:RETRY_PERIOD,Value:5,ValueFrom:nil,},EnvVar{Name:ENABLE_WEBHOOKS,Value:true,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_AGENT_IMAGE_URL_DEFAULT,Value:quay.io/openstack-k8s-operators/openstack-baremetal-operator-agent@sha256:03b4f3db4b373515f7e4095984b97197c05a14f87b2a0a525eb5d7be1d7bda66,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_ANSIBLEEE_IMAGE_URL_DEFAULT,Value:quay.io/openstack-k8s-operators/openstack-ansibleee-runner@sha256:6722a752fb7cbffbae811f6ad6567120fbd4ebbe8c38a83ec2df02850a3276bd,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_AODH_API_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-aodh-api@sha256:2115452234aedb505ed4efc6cd9b9a4ce3b9809aa7d0128d8fbeeee84dad1a69,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_AODH_EVALUATOR_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-aodh-evaluator@sha256:50597a8eaa6c4383f357574dcab8358b698729797b4156d932985a08ab86b7cd,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_AODH_LISTENER_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-aodh-listener@sha256:cb4997d62c7b2534233a676cb92e19cf85dda07e2fb9fa642c28aab30489f69a,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_AODH_NOTIFIER_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-aodh-notifier@sha256:1ccbf3f6cf24c9ee91bed71467491e22b8cb4b95bce90250f4174fae936b0fa1,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_APACHE_IMAGE_URL_DEFAULT,Value:registry.redhat.io/ubi9/httpd-24@sha256:e91d58021b54c46883595ff66be65882de54abdb3be2ca53c4162b20d18b5f48,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_BARBICAN_API_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-barbican-api@sha256:cbe345acb37e57986ecf6685d28c72d0e639bdb493a18e9d3ba947d6c3a16384,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_BARBICAN_KEYSTONE_LISTENER_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-barbican-keystone-listener@sha256:e7dcc3bf23d5e0393ac173e3c43d4ae85f4613a4fd16b3c147dc32ae491d49bf,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_BARBICAN_WORKER_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-barbican-worker@sha256:2a1a8b582c6e4cc31081bd8b0887acf45e31c1d14596c4e361d27d08fef0debf,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_CEILOMETER_CENTRAL_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-ceilometer-central@sha256:86daeb9c834bfcedb533086dff59a6b5b6e832b94ce2a9116337f8736bb80032,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_CEILOMETER_COMPUTE_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-ceilometer-compute@sha256:5d4fdf424fad33a3650163e9e7423f92e97de3305508c2b7c6435822e0313189,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_CEILOMETER_IPMI_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-ceilometer-ipmi@sha256:6d28de018f6e1672e775a75735e3bc16b63da41acd8fb5196ee0b06856c07133,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_CEILOMETER_MYSQLD_EXPORTER_IMAGE_URL_DEFAULT,Value:quay.io/prometheus/mysqld-exporter@sha256:7211a617ec657701ca819aa0ba28e1d5750f5bf2c1391b755cc4a48cc360b0fa,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_CEILOMETER_NOTIFICATION_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-ceilometer-notification@sha256:c5fc9b72fc593bcf3b569c7ed24a256448eb1afab1504e668a3822e978be1306,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_CEILOMETER_SGCORE_IMAGE_URL_DEFAULT,Value:quay.io/openstack-k8s-operators/sg-core@sha256:09b5017c95d7697e66b9c64846bc48ef5826a009cba89b956ec54561e5f4a2d1,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_CINDER_API_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-cinder-api@sha256:85c75d60e1bd2f8a9ea0a2bb21a8df64c0a6f7b504cc1a05a355981d4b90e92f,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_CINDER_BACKUP_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-cinder-backup@sha256:88b99249f15470f359fb554f7f3a56974b743f4655e3f0c982c0260f75a67697,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_CINDER_SCHEDULER_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-cinder-scheduler@sha256:e861d66785047d39eb68d9bac23e3f57ac84d9bd95593502d9b3b913b99fd1a4,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_CINDER_VOLUME_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-cinder-volume@sha256:b95f09bf3d259f9eacf3b63931977483f5c3c332f49b95ee8a69d8e3fb71d082,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_DESIGNATE_API_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-designate-api@sha256:6fc7801c0d18d41b9f11484b1cdb342de9cebd93072ec2205dbe40945715184f,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_DESIGNATE_BACKENDBIND9_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-designate-backend-bind9@sha256:d4d824b80cbed683543d9e8c7045ac97e080774f45a5067ccbca26404e067821,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_DESIGNATE_CENTRAL_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-designate-central@sha256:182ec75938d8d3fb7d8f916373368add24062fec90489aa57776a81d0b36ea20,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_DESIGNATE_MDNS_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-designate-mdns@sha256:9507ba5ab74cbae902e2dc07f89c7b3b5b76d8079e444365fe0eee6000fd7aaa,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_DESIGNATE_PRODUCER_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-designate-producer@sha256:17db080dcc4099f8a20aa0f238b6bca5c104672ae46743adeab9d1637725ecaa,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_DESIGNATE_UNBOUND_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-unbound@sha256:fd55cf3d73bfdc518419c9ba0b0cbef275140ae2d3bd0342a7310f81d57c2d78,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_DESIGNATE_WORKER_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-designate-worker@sha256:d164a9bd383f50df69fc22e7422f4650cd5076c90ed19278fc0f04e54345a63d,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_EDPM_FRR_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-frr@sha256:6beffe7d0bd75f9d1f495aeb7ab2334a2414af2c581d4833363df8441ed01018,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_EDPM_ISCSID_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-iscsid@sha256:261e76f60c6bc6b172dc3608504552c63e83358a4fa3c0952a671544d83aa83f,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_EDPM_KEPLER_IMAGE_URL_DEFAULT,Value:quay.io/sustainable_computing_io/kepler@sha256:581b65b646301e0fcb07582150ba63438f1353a85bf9acf1eb2acb4ce71c58bd,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_EDPM_LOGROTATE_CROND_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-cron@sha256:2308c7b6c3d0aabbadfc9a06d84d67d2243f27fe8eed740ee96b1ce910203f62,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_EDPM_MULTIPATHD_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-multipathd@sha256:02d33f59749441cd5751c319e9d7cff97ab1004844c0e992650d340c6e8fbf43,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_EDPM_NEUTRON_DHCP_AGENT_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-neutron-dhcp-agent@sha256:9cf0ca292340f1f978603955ef682effbf24316d6e2376b1c89906d84c3f06d0,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_EDPM_NEUTRON_METADATA_AGENT_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-neutron-metadata-agent-ovn@sha256:c3e651f35b930bcf1a3084be8910c2f3f34d22a976c5379cf518a68d9994bfa7,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_EDPM_NEUTRON_OVN_AGENT_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-neutron-ovn-agent@sha256:58f678016d7f6c8fe579abe886fd138ef853642faa6766ca60639feac12d82ac,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_EDPM_NEUTRON_SRIOV_AGENT_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-neutron-sriov-agent@sha256:46f92909153aaf03a585374b77d103c536509747e3270558d9a533295c46a7c5,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_EDPM_NODE_EXPORTER_IMAGE_URL_DEFAULT,Value:quay.io/prometheus/node-exporter@sha256:39c642b2b337e38c18e80266fb14383754178202f40103646337722a594d984c,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_EDPM_OVN_BGP_AGENT_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-ovn-bgp-agent@sha256:7fe367f51638c5c302fd3f8e66a31b09cb3b11519a7f72ef142b6c6fe8b91694,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_EDPM_PODMAN_EXPORTER_IMAGE_URL_DEFAULT,Value:quay.io/navidys/prometheus-podman-exporter@sha256:d339ba049bbd1adccb795962bf163f5b22fd84dea865d88b9eb525e46247d6bd,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_GLANCE_API_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-glance-api@sha256:9ebf424d4107275a2e3f21f7a18ef257ff2f97c1298109ac7c802a5a4f4794f2,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_HEAT_API_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-heat-api@sha256:4fcbe0d9a3c845708ecc32102ad4abbcbd947d87e5cf91f186de75b5d84ec681,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_HEAT_CFNAPI_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-heat-api-cfn@sha256:58a4e9a4dea86635c93ce37a2bb3c60ece62b3d656f6ee6a8845347cbb3e90fd,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_HEAT_ENGINE_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-heat-engine@sha256:6f2b843bc9f4ceb1ee873972d69e6bae6e1dbd378b486995bc3697d8bcff6339,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_HORIZON_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-horizon@sha256:03b4bb79b71d5ca7792d19c4c0ee08a5e5a407ad844c087305c42dd909ee7490,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_INFRA_MEMCACHED_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-memcached@sha256:773daada6402d9cad089cdc809d6c0335456d057ac1a25441ab5d82add2f70f4,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_INFRA_REDIS_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-redis@sha256:7323406a63fb3fdbb3eea4da0f7e8ed89c94c9bd0ad5ecd6c18fa4a4c2c550c4,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_IRONIC_API_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-ironic-api@sha256:7ae82068011e2d2e5ddc88c943fd32ff4a11902793e7a1df729811b2e27122a0,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_IRONIC_CONDUCTOR_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-ironic-conductor@sha256:0c762c15d9d98d39cc9dc3d1f9a70f9188fef58d4e2f3b0c69c896cab8da5e48,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_IRONIC_INSPECTOR_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-ironic-inspector@sha256:febf65561eeef5b36b70d0d65ee83f6451e43ec97bfab4d826e14215da6ff19b,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_IRONIC_NEUTRON_AGENT_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-ironic-neutron-agent@sha256:b8aadfc3d547c5ef1e27fcb573d4760cf8c2f2271eefe1793c35a0d46b640837,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_IRONIC_PXE_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-ironic-pxe@sha256:ecc91fd5079ee6d0c6ae1b11e97da790e33864d0e1930e574f959da2bddfa59a,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_IRONIC_PYTHON_AGENT_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/ironic-python-agent@sha256:2e981e93f99c929a3f04e5e41c8f645d44d390a9aeee3c5193cce7ec2edcbf3a,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_KEYSTONE_API_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-keystone@sha256:1e5714637b6e1a24c2858fe6d9bbb3f00bc61d69ad74a657b1c23682bf4cb2b7,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_KSM_IMAGE_URL_DEFAULT,Value:registry.k8s.io/kube-state-metrics/kube-state-metrics@sha256:db384bf43222b066c378e77027a675d4cd9911107adba46c2922b3a55e10d6fb,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_MANILA_API_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-manila-api@sha256:35b8dcf27dc3b67f3840fa0e693ff312f74f7e22c634dff206a5c4d0133c716c,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_MANILA_SCHEDULER_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-manila-scheduler@sha256:e109e4863e05e803dbfe04917756fd52231c560c65353170a2000be6cc2bb53d,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_MANILA_SHARE_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-manila-share@sha256:6df0bebd9318ce11624413249e7e9781311638f276f8877668d3b382fe90e62f,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_MARIADB_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-mariadb@sha256:56b75d97f4a48c8cf58b3a7c18c43618efb308bf0188124f6301142e61299b0c,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_NET_UTILS_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-netutils@sha256:a51ed62767206067aa501142dbf01f20b3d65325d30faf1b4d6424d5b17dfba5,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_NEUTRON_API_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-neutron-server@sha256:c4e71b2158fd939dad8b8e705273493051d3023273d23b279f2699dce6db33df,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_NOVA_API_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-nova-api@sha256:592e3cd32d3cc97a69093ad905b449aa374ffbb1b2644b738bb6c1434476d1f6,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_NOVA_COMPUTE_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-nova-compute@sha256:5f179b847f2dc32d9110b8f2be9fe65f1aeada1e18105dffdaf052981215d844,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_NOVA_CONDUCTOR_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-nova-conductor@sha256:9596452e283febbe08204d0ef0fd1992af3395d0969f7ac76663ed7c8be5b4d4,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_NOVA_NOVNC_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-nova-novncproxy@sha256:d61005a10bef1b37762a8a41e6755c1169241e36cc5f92886bca6f4f6b9c381a,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_NOVA_SCHEDULER_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-nova-scheduler@sha256:e6a4335bcbeed3cd3e73ac879f754e314761e4a417a67539ca88e96a79346328,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_OCTAVIA_API_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-octavia-api@sha256:97d88fc53421b699fc91983313d7beec4a0f177089e95bdf5ba15c3f521db9a9,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_OCTAVIA_HEALTHMANAGER_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-octavia-health-manager@sha256:5365e5c9c3ad2ede1b6945255b2cc6b009d642c39babdf25e0655282cfa646fe,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_OCTAVIA_HOUSEKEEPING_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-octavia-housekeeping@sha256:5b55795d774e0ea160ff8a7fd491ed41cf2d93c7d821694abb3a879eaffcefeb,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_OCTAVIA_RSYSLOG_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-rsyslog@sha256:26e955c46a6063eafcfeb79430bf3d9268dbe95687c00e63a624b3ec5a846f5a,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_OCTAVIA_WORKER_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-octavia-worker@sha256:58939baa18ab09e2b24996c5f3665ae52274b781f661ea06a67c991e9a832d5a,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_OPENSTACK_CLIENT_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-openstackclient@sha256:b8bff6857fec93c3c1521f1a8c23de21bcb86fc0f960972e81f6c3f95d4185be,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_OPENSTACK_MUST_GATHER_DEFAULT,Value:quay.io/openstack-k8s-operators/openstack-must-gather@sha256:943eee724277e252795909137538a553ef5284c8103ad01b9be7b0138c66d14d,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_OPENSTACK_NETWORK_EXPORTER_IMAGE_URL_DEFAULT,Value:quay.io/openstack-k8s-operators/openstack-network-exporter@sha256:ecd56e6733c475f2d441344fd98f288c3eac0261ba113695fec7520a954ccbc7,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_OS_CONTAINER_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/edpm-hardened-uefi@sha256:d97b08fd421065c8c33a523973822ac468500cbe853069aa9214393fbda7a908,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_OVN_CONTROLLER_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-ovn-controller@sha256:d76f7d6620930cc2e9ac070492bbeb525f83ce5ff4947463e3784bf1ce04a857,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_OVN_CONTROLLER_OVS_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-ovn-base@sha256:289dea3beea1cd4405895fc42e44372b35e4a941e31c59e102c333471a3ca9b7,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_OVN_NB_DBCLUSTER_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-ovn-nb-db-server@sha256:9b19894fa67a81bf8ba4159b55b49f38877c670aeb97e2021c341cef2a9294e4,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_OVN_NORTHD_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-ovn-northd@sha256:ea164961ad30453ad0301c6b73364e1f1024f689634c88dd98265f9c7048e31d,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_OVN_SB_DBCLUSTER_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-ovn-sb-db-server@sha256:6f9f2ea45f0271f6da8eb05a5f74cf5ce6769479346f5c2f407ee6f31a9c7ff3,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_PLACEMENT_API_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-placement-api@sha256:59448516174fc3bab679b9a8dd62cb9a9d16b5734aadbeb98e960e3b7c79bd22,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_RABBITMQ_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-rabbitmq@sha256:adcdeb8ecd601fb03c3b0901d5b5111af2ca48f7dd443e22224db6daaf08f5d0,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_SWIFT_ACCOUNT_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-swift-account@sha256:2bf32d9b95899d7637dfe19d07cf1ecc9a06593984faff57a3c0dce060012edb,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_SWIFT_CONTAINER_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-swift-container@sha256:7a452cd18b64d522e8a1e25bdcea543e9fe5f5b76e1c5e044c2b5334e06a326b,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_SWIFT_OBJECT_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-swift-object@sha256:6a46aa13aa359b8e782a22d67db42db02bbf2bb7e35df4b684ac1daeda38cde3,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_SWIFT_PROXY_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-swift-proxy-server@sha256:f6824854bea6b2acbb00c34639799b4744818d4adbdd40e37dc5088f9ae18d58,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_TEST_TEMPEST_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-tempest-all@sha256:a66d2fdc21f25c690f02e643d2666dbe7df43a64cd55086ec33d6755e6d809b9,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_WATCHER_API_IMAGE_URL_DEFAULT,Value:quay.io/podified-master-centos9/openstack-watcher-api@sha256:e0e84e3fb8f9f12e9280d3d28b415639abfad5b5e46787a61b42beb97f32c76e,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_WATCHER_APPLIER_IMAGE_URL_DEFAULT,Value:quay.io/podified-master-centos9/openstack-watcher-applier@sha256:bd5376b807eca8a409ea663906d9a990cd95a109b8f1f2c56c06f9201be1f6ca,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_WATCHER_DECISION_ENGINE_IMAGE_URL_DEFAULT,Value:quay.io/podified-master-centos9/openstack-watcher-decision-engine@sha256:366de9fadc79d3ea77de93ef1c81e4b8525764f0704e3f30d1a8afa2745c8166,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{536870912 0} {} BinarySI},},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{268435456 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:cert,ReadOnly:true,MountPath:/tmp/k8s-webhook-server/serving-certs,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-4w2pq,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/healthz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:15,TimeoutSeconds:1,PeriodSeconds:20,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/readyz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:1,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000660000,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod openstack-baremetal-operator-controller-manager-677c5f5bff5pg7p_openstack-operators(dd924000-5690-490e-a4db-e29ed7f821ec): ErrImagePull: pull QPS exceeded" logger="UnhandledError" Oct 08 18:30:55 crc kubenswrapper[4859]: I1008 18:30:55.975931 4859 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/neutron-operator-controller-manager-79d585cb66-zn65q" event={"ID":"e69529d1-1a2a-433c-a16b-3c6de62a0587","Type":"ContainerStarted","Data":"cf844bbe6511caaf20d525dff3afb6ba526b7d608cf1faf83f83c3d2549bf308"} Oct 08 18:30:55 crc kubenswrapper[4859]: I1008 18:30:55.986859 4859 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/rabbitmq-cluster-operator-manager-5f97d8c699-dxrl8" event={"ID":"e1dbe8d7-f993-42b1-9a53-7d0c342780c1","Type":"ContainerStarted","Data":"51c5dc2f1577b9ae3793cae0df0548f4c12357bc148e4c6c96d4226e08eabe3d"} Oct 08 18:30:55 crc kubenswrapper[4859]: I1008 18:30:55.998839 4859 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-baremetal-operator-controller-manager-677c5f5bff5pg7p" event={"ID":"dd924000-5690-490e-a4db-e29ed7f821ec","Type":"ContainerStarted","Data":"1bdf1a40bda87f25679fbc6af28c45d60764e2110ad4079c15c735d78343bae8"} Oct 08 18:30:56 crc kubenswrapper[4859]: I1008 18:30:56.001483 4859 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/mariadb-operator-controller-manager-f9fb45f8f-ptpqs" event={"ID":"b6a2cffa-8421-43c4-bad8-5d4d341a40cc","Type":"ContainerStarted","Data":"937b90dc57b858573e1fa2630f4e0451fc1038296d184b59ab72f381570aa4a8"} Oct 08 18:30:56 crc kubenswrapper[4859]: I1008 18:30:56.002510 4859 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/keystone-operator-controller-manager-55b6b7c7b8-rsbgr" event={"ID":"0228033e-c835-4618-9603-2c67cf9ce57d","Type":"ContainerStarted","Data":"0d825bc06c3d3154589e1b07e3bc19f8534874240c0fc0956933924429605934"} Oct 08 18:30:56 crc kubenswrapper[4859]: I1008 18:30:56.003214 4859 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/watcher-operator-controller-manager-7794bc6bd-2v65r" event={"ID":"cf5b6741-336e-435c-8209-82b8e19f4896","Type":"ContainerStarted","Data":"d3d2665bd4be4e98686da615053de87ec145ff4d1bdd8cd30d3005ddc23f01f3"} Oct 08 18:30:56 crc kubenswrapper[4859]: I1008 18:30:56.003849 4859 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/nova-operator-controller-manager-5df598886f-p99nw" event={"ID":"e2889a4a-89ec-47de-99df-7e749c3e5514","Type":"ContainerStarted","Data":"0daaac2ded0f6e6eaebbeb482c60ef742245a430bf22d10deff4f4df9f0d8a2d"} Oct 08 18:30:56 crc kubenswrapper[4859]: I1008 18:30:56.008834 4859 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/cinder-operator-controller-manager-7b7fb68549-lbw64" event={"ID":"e2b104df-4109-45ab-8a5a-3569da53caa0","Type":"ContainerStarted","Data":"cbfde52696f3fc39bd5ac88e5aacb416d119d7c331bb60c3cdb9ce104eeda704"} Oct 08 18:30:56 crc kubenswrapper[4859]: I1008 18:30:56.029995 4859 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/manila-operator-controller-manager-5f67fbc655-xg74g" event={"ID":"fbce123a-545f-4c4c-86ca-2d68f81aa3f3","Type":"ContainerStarted","Data":"6f718b0df1eede1800a30b996e5594ec5761bc2a715ffc6c16eeb41154c577b3"} Oct 08 18:30:56 crc kubenswrapper[4859]: I1008 18:30:56.049930 4859 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-operator-controller-manager-598c4c5b5-6l55k" event={"ID":"182ef4ca-109f-46bd-a935-091d05275271","Type":"ContainerStarted","Data":"20b6735ecfca50d22c510ad976f94309d43f78e5bcf50e67a1155c257fa96d4b"} Oct 08 18:30:56 crc kubenswrapper[4859]: I1008 18:30:56.068904 4859 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/infra-operator-controller-manager-656bcbd775-ctjz9" event={"ID":"1a1cfd66-310a-4077-8020-df7e6e069c53","Type":"ContainerStarted","Data":"61c2b4cd600b5638a4911f0e0202fc158afd49020bb9f9922f4136cd7dd9fb31"} Oct 08 18:30:56 crc kubenswrapper[4859]: I1008 18:30:56.085961 4859 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/telemetry-operator-controller-manager-76796d4c6b-kjbfp" event={"ID":"ffa2914a-717c-4763-8b1c-4a1ff18858f5","Type":"ContainerStarted","Data":"24c0c5d7e597fdffa0ec7b2e49ffe7197ccff5297cb6df6641f86cbaf6a86e56"} Oct 08 18:30:56 crc kubenswrapper[4859]: I1008 18:30:56.092199 4859 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/placement-operator-controller-manager-68b6c87b68-xgjpk" event={"ID":"f56e7a0c-7b3a-4dc8-9367-cfeef5c5cef9","Type":"ContainerStarted","Data":"d7abfa028088305ee534b2836ca43420cfd97a37b83133f401e2684c94dfcb51"} Oct 08 18:30:56 crc kubenswrapper[4859]: I1008 18:30:56.097580 4859 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/ovn-operator-controller-manager-79db49b9fb-hhgrl" event={"ID":"e2675d43-51ec-4a8c-aa2d-5aedc14f649a","Type":"ContainerStarted","Data":"96d2e80bd8243b02aaeb80ab0133fe9b38893c88750e461736c5acf186ea3e43"} Oct 08 18:30:56 crc kubenswrapper[4859]: I1008 18:30:56.106066 4859 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/octavia-operator-controller-manager-69fdcfc5f5-lsl8l" event={"ID":"550bb7bb-6f9c-4b20-9d92-2c85773e05a2","Type":"ContainerStarted","Data":"09da9f9c63c58343a23a13ee89762b5552c7842cfcc8e3f59ce3a8a07ccc3801"} Oct 08 18:30:56 crc kubenswrapper[4859]: I1008 18:30:56.118374 4859 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/swift-operator-controller-manager-db6d7f97b-zjkwm" event={"ID":"44980d79-924e-436f-950d-8fcdb4c42af1","Type":"ContainerStarted","Data":"bb10012e7b11cc5786898000e95ca6a2e99839569216c241a7c949701e4b8a38"} Oct 08 18:30:56 crc kubenswrapper[4859]: I1008 18:30:56.136745 4859 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/test-operator-controller-manager-56c698c775-d2g8b" event={"ID":"2d6cf85d-c5cc-4466-a471-bb17aa8fca9d","Type":"ContainerStarted","Data":"4630c3dcdd1af12070cfac2ab527aa205d402461b71e264e2007a6093466ab65"} Oct 08 18:30:56 crc kubenswrapper[4859]: I1008 18:30:56.430474 4859 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-b8rjh"] Oct 08 18:30:56 crc kubenswrapper[4859]: W1008 18:30:56.500556 4859 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podaefd0d7b_0c10_443e_9b21_6c47cb711eb7.slice/crio-5f8885dec4565219c199bd035c2888192ab214ea38bec07e68b6d2664903dd26 WatchSource:0}: Error finding container 5f8885dec4565219c199bd035c2888192ab214ea38bec07e68b6d2664903dd26: Status 404 returned error can't find the container with id 5f8885dec4565219c199bd035c2888192ab214ea38bec07e68b6d2664903dd26 Oct 08 18:30:57 crc kubenswrapper[4859]: I1008 18:30:57.148861 4859 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/placement-operator-controller-manager-68b6c87b68-xgjpk" event={"ID":"f56e7a0c-7b3a-4dc8-9367-cfeef5c5cef9","Type":"ContainerStarted","Data":"aa03aef680ed3796618f1d7db1f78df99e90a2c85a9583b2589a3f7a5e601023"} Oct 08 18:30:57 crc kubenswrapper[4859]: I1008 18:30:57.152488 4859 generic.go:334] "Generic (PLEG): container finished" podID="51932a59-6c7a-4c64-9acf-59984e192d91" containerID="51de042cfa3748c000a73b19681f8fba61c0263e8fa8c6232162000607ec37bb" exitCode=0 Oct 08 18:30:57 crc kubenswrapper[4859]: I1008 18:30:57.152555 4859 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-jws9w" event={"ID":"51932a59-6c7a-4c64-9acf-59984e192d91","Type":"ContainerDied","Data":"51de042cfa3748c000a73b19681f8fba61c0263e8fa8c6232162000607ec37bb"} Oct 08 18:30:57 crc kubenswrapper[4859]: I1008 18:30:57.153606 4859 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-b8rjh" event={"ID":"aefd0d7b-0c10-443e-9b21-6c47cb711eb7","Type":"ContainerStarted","Data":"5f8885dec4565219c199bd035c2888192ab214ea38bec07e68b6d2664903dd26"} Oct 08 18:30:57 crc kubenswrapper[4859]: I1008 18:30:57.155051 4859 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-baremetal-operator-controller-manager-677c5f5bff5pg7p" event={"ID":"dd924000-5690-490e-a4db-e29ed7f821ec","Type":"ContainerStarted","Data":"e9c2c1d36e5bbffa856fdc4636580b76ccda341bec5a7176949c6f8dc437f92e"} Oct 08 18:30:57 crc kubenswrapper[4859]: I1008 18:30:57.157056 4859 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/watcher-operator-controller-manager-7794bc6bd-2v65r" event={"ID":"cf5b6741-336e-435c-8209-82b8e19f4896","Type":"ContainerStarted","Data":"bc24f705719fb4d618afac3c808f76bfcecb8c7fd9cf37b05dbef48838730469"} Oct 08 18:30:57 crc kubenswrapper[4859]: I1008 18:30:57.158730 4859 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/test-operator-controller-manager-56c698c775-d2g8b" event={"ID":"2d6cf85d-c5cc-4466-a471-bb17aa8fca9d","Type":"ContainerStarted","Data":"7df123ee224dec70bb8f11e45c8a03c6cc724cdafd333935727eb4d6cf651628"} Oct 08 18:30:57 crc kubenswrapper[4859]: I1008 18:30:57.160249 4859 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/telemetry-operator-controller-manager-76796d4c6b-kjbfp" event={"ID":"ffa2914a-717c-4763-8b1c-4a1ff18858f5","Type":"ContainerStarted","Data":"5d6c46ac50d8498cc4f7bb67ce89ab8cbba2ed7bc019efacc0007b5cb8b527d7"} Oct 08 18:30:57 crc kubenswrapper[4859]: I1008 18:30:57.169008 4859 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-operator-controller-manager-598c4c5b5-6l55k" event={"ID":"182ef4ca-109f-46bd-a935-091d05275271","Type":"ContainerStarted","Data":"d842546ef25675db57839383d987d61a4392409dcbc6882b38d99474870c54d1"} Oct 08 18:30:57 crc kubenswrapper[4859]: E1008 18:30:57.240912 4859 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ErrImagePull: \"pull QPS exceeded\"" pod="openstack-operators/placement-operator-controller-manager-68b6c87b68-xgjpk" podUID="f56e7a0c-7b3a-4dc8-9367-cfeef5c5cef9" Oct 08 18:30:57 crc kubenswrapper[4859]: E1008 18:30:57.363513 4859 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ErrImagePull: \"pull QPS exceeded\"" pod="openstack-operators/test-operator-controller-manager-56c698c775-d2g8b" podUID="2d6cf85d-c5cc-4466-a471-bb17aa8fca9d" Oct 08 18:30:57 crc kubenswrapper[4859]: E1008 18:30:57.364423 4859 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ErrImagePull: \"pull QPS exceeded\"" pod="openstack-operators/openstack-baremetal-operator-controller-manager-677c5f5bff5pg7p" podUID="dd924000-5690-490e-a4db-e29ed7f821ec" Oct 08 18:30:57 crc kubenswrapper[4859]: E1008 18:30:57.364760 4859 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ErrImagePull: \"pull QPS exceeded\"" pod="openstack-operators/telemetry-operator-controller-manager-76796d4c6b-kjbfp" podUID="ffa2914a-717c-4763-8b1c-4a1ff18858f5" Oct 08 18:30:57 crc kubenswrapper[4859]: E1008 18:30:57.366532 4859 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ErrImagePull: \"pull QPS exceeded\"" pod="openstack-operators/watcher-operator-controller-manager-7794bc6bd-2v65r" podUID="cf5b6741-336e-435c-8209-82b8e19f4896" Oct 08 18:30:57 crc kubenswrapper[4859]: I1008 18:30:57.805214 4859 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-jws9w" Oct 08 18:30:57 crc kubenswrapper[4859]: I1008 18:30:57.993225 4859 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/51932a59-6c7a-4c64-9acf-59984e192d91-catalog-content\") pod \"51932a59-6c7a-4c64-9acf-59984e192d91\" (UID: \"51932a59-6c7a-4c64-9acf-59984e192d91\") " Oct 08 18:30:57 crc kubenswrapper[4859]: I1008 18:30:57.993581 4859 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/51932a59-6c7a-4c64-9acf-59984e192d91-utilities\") pod \"51932a59-6c7a-4c64-9acf-59984e192d91\" (UID: \"51932a59-6c7a-4c64-9acf-59984e192d91\") " Oct 08 18:30:57 crc kubenswrapper[4859]: I1008 18:30:57.993647 4859 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-m7pzh\" (UniqueName: \"kubernetes.io/projected/51932a59-6c7a-4c64-9acf-59984e192d91-kube-api-access-m7pzh\") pod \"51932a59-6c7a-4c64-9acf-59984e192d91\" (UID: \"51932a59-6c7a-4c64-9acf-59984e192d91\") " Oct 08 18:30:57 crc kubenswrapper[4859]: I1008 18:30:57.994566 4859 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/51932a59-6c7a-4c64-9acf-59984e192d91-utilities" (OuterVolumeSpecName: "utilities") pod "51932a59-6c7a-4c64-9acf-59984e192d91" (UID: "51932a59-6c7a-4c64-9acf-59984e192d91"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 08 18:30:58 crc kubenswrapper[4859]: I1008 18:30:58.028130 4859 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/51932a59-6c7a-4c64-9acf-59984e192d91-kube-api-access-m7pzh" (OuterVolumeSpecName: "kube-api-access-m7pzh") pod "51932a59-6c7a-4c64-9acf-59984e192d91" (UID: "51932a59-6c7a-4c64-9acf-59984e192d91"). InnerVolumeSpecName "kube-api-access-m7pzh". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 08 18:30:58 crc kubenswrapper[4859]: I1008 18:30:58.086229 4859 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/51932a59-6c7a-4c64-9acf-59984e192d91-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "51932a59-6c7a-4c64-9acf-59984e192d91" (UID: "51932a59-6c7a-4c64-9acf-59984e192d91"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 08 18:30:58 crc kubenswrapper[4859]: I1008 18:30:58.094997 4859 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/51932a59-6c7a-4c64-9acf-59984e192d91-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 08 18:30:58 crc kubenswrapper[4859]: I1008 18:30:58.095041 4859 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/51932a59-6c7a-4c64-9acf-59984e192d91-utilities\") on node \"crc\" DevicePath \"\"" Oct 08 18:30:58 crc kubenswrapper[4859]: I1008 18:30:58.095052 4859 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-m7pzh\" (UniqueName: \"kubernetes.io/projected/51932a59-6c7a-4c64-9acf-59984e192d91-kube-api-access-m7pzh\") on node \"crc\" DevicePath \"\"" Oct 08 18:30:58 crc kubenswrapper[4859]: I1008 18:30:58.182787 4859 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-operator-controller-manager-598c4c5b5-6l55k" event={"ID":"182ef4ca-109f-46bd-a935-091d05275271","Type":"ContainerStarted","Data":"7d0a6aac59bf8f6715d08ea9a18d420f7721e7c4e15f773765823fd3e97ace45"} Oct 08 18:30:58 crc kubenswrapper[4859]: I1008 18:30:58.183017 4859 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/openstack-operator-controller-manager-598c4c5b5-6l55k" Oct 08 18:30:58 crc kubenswrapper[4859]: I1008 18:30:58.190812 4859 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-jws9w" event={"ID":"51932a59-6c7a-4c64-9acf-59984e192d91","Type":"ContainerDied","Data":"0c93203584618cbe5643074d3bf05d482d24b5076b251592048ec50a965ed19e"} Oct 08 18:30:58 crc kubenswrapper[4859]: I1008 18:30:58.190869 4859 scope.go:117] "RemoveContainer" containerID="51de042cfa3748c000a73b19681f8fba61c0263e8fa8c6232162000607ec37bb" Oct 08 18:30:58 crc kubenswrapper[4859]: I1008 18:30:58.190977 4859 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-jws9w" Oct 08 18:30:58 crc kubenswrapper[4859]: I1008 18:30:58.205925 4859 generic.go:334] "Generic (PLEG): container finished" podID="aefd0d7b-0c10-443e-9b21-6c47cb711eb7" containerID="d933baaac4cf8d90574c5c0f78b0557752e9fdb5a092f5f3890fbe21c3263e1d" exitCode=0 Oct 08 18:30:58 crc kubenswrapper[4859]: I1008 18:30:58.206075 4859 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-b8rjh" event={"ID":"aefd0d7b-0c10-443e-9b21-6c47cb711eb7","Type":"ContainerDied","Data":"d933baaac4cf8d90574c5c0f78b0557752e9fdb5a092f5f3890fbe21c3263e1d"} Oct 08 18:30:58 crc kubenswrapper[4859]: E1008 18:30:58.207760 4859 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/watcher-operator@sha256:e4ae07e859166fc5e2cb4f8e0e2c3358b9d2e2d6721a3864d2e0c651d36698ca\\\"\"" pod="openstack-operators/watcher-operator-controller-manager-7794bc6bd-2v65r" podUID="cf5b6741-336e-435c-8209-82b8e19f4896" Oct 08 18:30:58 crc kubenswrapper[4859]: E1008 18:30:58.208176 4859 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/openstack-baremetal-operator@sha256:a17fc270857869fd1efe5020b2a1cb8c2abbd838f08de88f3a6a59e8754ec351\\\"\"" pod="openstack-operators/openstack-baremetal-operator-controller-manager-677c5f5bff5pg7p" podUID="dd924000-5690-490e-a4db-e29ed7f821ec" Oct 08 18:30:58 crc kubenswrapper[4859]: E1008 18:30:58.208183 4859 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/placement-operator@sha256:d33c1f507e1f5b9a4bf226ad98917e92101ac66b36e19d35cbe04ae7014f6bff\\\"\"" pod="openstack-operators/placement-operator-controller-manager-68b6c87b68-xgjpk" podUID="f56e7a0c-7b3a-4dc8-9367-cfeef5c5cef9" Oct 08 18:30:58 crc kubenswrapper[4859]: E1008 18:30:58.208609 4859 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/test-operator@sha256:efa8fb78cffb573d299ffcc7bab1099affd2dbbab222152092b313074306e0a9\\\"\"" pod="openstack-operators/test-operator-controller-manager-56c698c775-d2g8b" podUID="2d6cf85d-c5cc-4466-a471-bb17aa8fca9d" Oct 08 18:30:58 crc kubenswrapper[4859]: E1008 18:30:58.208905 4859 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/telemetry-operator@sha256:9d26476523320d70d6d457b91663e8c233ed320d77032a7c57a89ce1aedd3931\\\"\"" pod="openstack-operators/telemetry-operator-controller-manager-76796d4c6b-kjbfp" podUID="ffa2914a-717c-4763-8b1c-4a1ff18858f5" Oct 08 18:30:58 crc kubenswrapper[4859]: I1008 18:30:58.226016 4859 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/openstack-operator-controller-manager-598c4c5b5-6l55k" podStartSLOduration=5.225994103 podStartE2EDuration="5.225994103s" podCreationTimestamp="2025-10-08 18:30:53 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-08 18:30:58.22022407 +0000 UTC m=+828.467063469" watchObservedRunningTime="2025-10-08 18:30:58.225994103 +0000 UTC m=+828.472833482" Oct 08 18:30:58 crc kubenswrapper[4859]: I1008 18:30:58.270485 4859 scope.go:117] "RemoveContainer" containerID="2f03ba0f6776e9c43ffc9f973cc293af86a987d1343b0e1ce3daf57cba93123c" Oct 08 18:30:58 crc kubenswrapper[4859]: I1008 18:30:58.275641 4859 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-jws9w"] Oct 08 18:30:58 crc kubenswrapper[4859]: I1008 18:30:58.281310 4859 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-operators-jws9w"] Oct 08 18:30:58 crc kubenswrapper[4859]: I1008 18:30:58.370277 4859 scope.go:117] "RemoveContainer" containerID="1223abb72da61b77cb9987f243aabc4723c02474786911c7a3414b9a47285daa" Oct 08 18:30:58 crc kubenswrapper[4859]: I1008 18:30:58.486546 4859 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="51932a59-6c7a-4c64-9acf-59984e192d91" path="/var/lib/kubelet/pods/51932a59-6c7a-4c64-9acf-59984e192d91/volumes" Oct 08 18:31:04 crc kubenswrapper[4859]: I1008 18:31:04.789112 4859 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/openstack-operator-controller-manager-598c4c5b5-6l55k" Oct 08 18:31:08 crc kubenswrapper[4859]: I1008 18:31:08.301857 4859 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/glance-operator-controller-manager-84b9b84486-d9gmv" event={"ID":"0b068d16-0072-41f8-ad01-4194fd428014","Type":"ContainerStarted","Data":"6394bcff45b759173b04340f2b2063e923f781e24da83bd7315b77480732fe93"} Oct 08 18:31:08 crc kubenswrapper[4859]: I1008 18:31:08.313228 4859 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/manila-operator-controller-manager-5f67fbc655-xg74g" event={"ID":"fbce123a-545f-4c4c-86ca-2d68f81aa3f3","Type":"ContainerStarted","Data":"71f6674af821070b8fdbdf5d1828140294e061a289cc275b2409530fa8812b4b"} Oct 08 18:31:08 crc kubenswrapper[4859]: I1008 18:31:08.325147 4859 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/octavia-operator-controller-manager-69fdcfc5f5-lsl8l" event={"ID":"550bb7bb-6f9c-4b20-9d92-2c85773e05a2","Type":"ContainerStarted","Data":"9381eee0213c8a1cf1b8678056302ff2e58e877feb4c1c027bafaf43e7520cf4"} Oct 08 18:31:08 crc kubenswrapper[4859]: I1008 18:31:08.338060 4859 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/horizon-operator-controller-manager-7ffbcb7588-86z4n" event={"ID":"ca71782e-3569-4ac2-a178-239f0ae200d0","Type":"ContainerStarted","Data":"286dc862f97460e6769d0700c6206ceaaf9170b97f2e4311ffd5a6611cc879d6"} Oct 08 18:31:08 crc kubenswrapper[4859]: I1008 18:31:08.355127 4859 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/infra-operator-controller-manager-656bcbd775-ctjz9" event={"ID":"1a1cfd66-310a-4077-8020-df7e6e069c53","Type":"ContainerStarted","Data":"423d3500a54ab7c446f6bf6b982a556887196edfda5707ecd3a2f1349fe5e845"} Oct 08 18:31:08 crc kubenswrapper[4859]: I1008 18:31:08.366207 4859 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/heat-operator-controller-manager-858f76bbdd-b5bfz" event={"ID":"19fa1d1d-0731-4c21-9faa-b1665ec4af36","Type":"ContainerStarted","Data":"2e792e4b1ad5600059f28ee1e65359d330cdd00c7774407f0fceea19ad45ec56"} Oct 08 18:31:08 crc kubenswrapper[4859]: I1008 18:31:08.368411 4859 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/ironic-operator-controller-manager-9c5c78d49-z9952" event={"ID":"2b67febf-c4c8-4e48-9350-29dd496439b4","Type":"ContainerStarted","Data":"db4b2579fc6c3fcb92844edd809c216f60be8f30cfa986662c81055893cf0513"} Oct 08 18:31:08 crc kubenswrapper[4859]: I1008 18:31:08.378325 4859 generic.go:334] "Generic (PLEG): container finished" podID="aefd0d7b-0c10-443e-9b21-6c47cb711eb7" containerID="2570425de29f3db779ddcaded0192a9d6e247e3e1514413fba87a1bad871db4f" exitCode=0 Oct 08 18:31:08 crc kubenswrapper[4859]: I1008 18:31:08.378445 4859 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-b8rjh" event={"ID":"aefd0d7b-0c10-443e-9b21-6c47cb711eb7","Type":"ContainerDied","Data":"2570425de29f3db779ddcaded0192a9d6e247e3e1514413fba87a1bad871db4f"} Oct 08 18:31:08 crc kubenswrapper[4859]: I1008 18:31:08.396327 4859 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/ovn-operator-controller-manager-79db49b9fb-hhgrl" event={"ID":"e2675d43-51ec-4a8c-aa2d-5aedc14f649a","Type":"ContainerStarted","Data":"e0e47212f19ba99a7458eff81b20232430005db33b28fc54442a4bb3a53449af"} Oct 08 18:31:08 crc kubenswrapper[4859]: I1008 18:31:08.419439 4859 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/swift-operator-controller-manager-db6d7f97b-zjkwm" event={"ID":"44980d79-924e-436f-950d-8fcdb4c42af1","Type":"ContainerStarted","Data":"0e942e3ddad06f86acb0c24f09aa7a83185e8f956d80909040979ab91a9b0318"} Oct 08 18:31:08 crc kubenswrapper[4859]: I1008 18:31:08.437111 4859 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/mariadb-operator-controller-manager-f9fb45f8f-ptpqs" event={"ID":"b6a2cffa-8421-43c4-bad8-5d4d341a40cc","Type":"ContainerStarted","Data":"ed003c66a238d9ecca75d5dee5e416704f764f54318980b42bc75bf02aef56ea"} Oct 08 18:31:08 crc kubenswrapper[4859]: I1008 18:31:08.461551 4859 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/neutron-operator-controller-manager-79d585cb66-zn65q" event={"ID":"e69529d1-1a2a-433c-a16b-3c6de62a0587","Type":"ContainerStarted","Data":"12b622a58c0aa614976ae03c83cfd7567e8227da26b78637cd21650f459710eb"} Oct 08 18:31:08 crc kubenswrapper[4859]: I1008 18:31:08.490636 4859 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/designate-operator-controller-manager-85d5d9dd78-nhqkh" event={"ID":"d9a50bbe-f28f-46d8-9266-fd54e44388a3","Type":"ContainerStarted","Data":"d61df182cd28f30dc09f9e42d6a6d1d4ebfcadf86a98944b0600d3dc7ddd163e"} Oct 08 18:31:08 crc kubenswrapper[4859]: I1008 18:31:08.505507 4859 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/rabbitmq-cluster-operator-manager-5f97d8c699-dxrl8" event={"ID":"e1dbe8d7-f993-42b1-9a53-7d0c342780c1","Type":"ContainerStarted","Data":"4670d4cb573fd5d4b03ecc240f920627063542aeae7c0761ce7de6c4b8a4bfad"} Oct 08 18:31:08 crc kubenswrapper[4859]: I1008 18:31:08.515404 4859 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/nova-operator-controller-manager-5df598886f-p99nw" event={"ID":"e2889a4a-89ec-47de-99df-7e749c3e5514","Type":"ContainerStarted","Data":"dc213a2a43340d068dbd34c03aab6e350ddec6a591a94c5d0e02d9c111d9abca"} Oct 08 18:31:08 crc kubenswrapper[4859]: I1008 18:31:08.566129 4859 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/cinder-operator-controller-manager-7b7fb68549-lbw64" event={"ID":"e2b104df-4109-45ab-8a5a-3569da53caa0","Type":"ContainerStarted","Data":"34ebb8a3a2738f8c0399048efc1ba30ddba9b7d0d1d8efab8acef696bd0fbc1d"} Oct 08 18:31:08 crc kubenswrapper[4859]: I1008 18:31:08.591105 4859 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/rabbitmq-cluster-operator-manager-5f97d8c699-dxrl8" podStartSLOduration=4.863956465 podStartE2EDuration="15.591085118s" podCreationTimestamp="2025-10-08 18:30:53 +0000 UTC" firstStartedPulling="2025-10-08 18:30:55.597672608 +0000 UTC m=+825.844511997" lastFinishedPulling="2025-10-08 18:31:06.324801271 +0000 UTC m=+836.571640650" observedRunningTime="2025-10-08 18:31:08.581634192 +0000 UTC m=+838.828473571" watchObservedRunningTime="2025-10-08 18:31:08.591085118 +0000 UTC m=+838.837924497" Oct 08 18:31:08 crc kubenswrapper[4859]: I1008 18:31:08.591294 4859 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/barbican-operator-controller-manager-658bdf4b74-8bfst" event={"ID":"d097cd11-b23b-4d38-80d7-5d85c257cef4","Type":"ContainerStarted","Data":"bb74d978ddc34653386844cfe64ab401e6e41322589f16ef34bf0853c84ccfbd"} Oct 08 18:31:08 crc kubenswrapper[4859]: I1008 18:31:08.637933 4859 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/keystone-operator-controller-manager-55b6b7c7b8-rsbgr" event={"ID":"0228033e-c835-4618-9603-2c67cf9ce57d","Type":"ContainerStarted","Data":"890d0c2a902d1954ffd0962342e8561eeeed8152191abe3b6bd8555d528916b5"} Oct 08 18:31:09 crc kubenswrapper[4859]: I1008 18:31:09.648538 4859 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/barbican-operator-controller-manager-658bdf4b74-8bfst" event={"ID":"d097cd11-b23b-4d38-80d7-5d85c257cef4","Type":"ContainerStarted","Data":"b051811f6ce95cb2df5ad34f660033a7f579e660596a744f4c967f268d4ba455"} Oct 08 18:31:09 crc kubenswrapper[4859]: I1008 18:31:09.649017 4859 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/barbican-operator-controller-manager-658bdf4b74-8bfst" Oct 08 18:31:09 crc kubenswrapper[4859]: I1008 18:31:09.655166 4859 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/octavia-operator-controller-manager-69fdcfc5f5-lsl8l" event={"ID":"550bb7bb-6f9c-4b20-9d92-2c85773e05a2","Type":"ContainerStarted","Data":"82fd8ab861ea49b9619bdb8cd0d1281f8bde47cb83cdcf0c4c4fee095bae0600"} Oct 08 18:31:09 crc kubenswrapper[4859]: I1008 18:31:09.655976 4859 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/octavia-operator-controller-manager-69fdcfc5f5-lsl8l" Oct 08 18:31:09 crc kubenswrapper[4859]: I1008 18:31:09.657369 4859 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/mariadb-operator-controller-manager-f9fb45f8f-ptpqs" event={"ID":"b6a2cffa-8421-43c4-bad8-5d4d341a40cc","Type":"ContainerStarted","Data":"117a1537907c9ca5c75a590701e93f7be22bebec969f6ef25924db901fe6b8b8"} Oct 08 18:31:09 crc kubenswrapper[4859]: I1008 18:31:09.657627 4859 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/mariadb-operator-controller-manager-f9fb45f8f-ptpqs" Oct 08 18:31:09 crc kubenswrapper[4859]: I1008 18:31:09.660174 4859 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/cinder-operator-controller-manager-7b7fb68549-lbw64" event={"ID":"e2b104df-4109-45ab-8a5a-3569da53caa0","Type":"ContainerStarted","Data":"20daa8b99f0d981f41867c0147dac2f92dbaa52e5753f41bed528e0166ae22c2"} Oct 08 18:31:09 crc kubenswrapper[4859]: I1008 18:31:09.660786 4859 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/cinder-operator-controller-manager-7b7fb68549-lbw64" Oct 08 18:31:09 crc kubenswrapper[4859]: I1008 18:31:09.662677 4859 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/designate-operator-controller-manager-85d5d9dd78-nhqkh" event={"ID":"d9a50bbe-f28f-46d8-9266-fd54e44388a3","Type":"ContainerStarted","Data":"b552ef850e872dfda514514e99b474032a4fa6951cc729b8e3e4b76a22f7c90c"} Oct 08 18:31:09 crc kubenswrapper[4859]: I1008 18:31:09.663178 4859 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/designate-operator-controller-manager-85d5d9dd78-nhqkh" Oct 08 18:31:09 crc kubenswrapper[4859]: I1008 18:31:09.671899 4859 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/barbican-operator-controller-manager-658bdf4b74-8bfst" podStartSLOduration=5.736513295 podStartE2EDuration="17.671885552s" podCreationTimestamp="2025-10-08 18:30:52 +0000 UTC" firstStartedPulling="2025-10-08 18:30:54.332723297 +0000 UTC m=+824.579562676" lastFinishedPulling="2025-10-08 18:31:06.268095554 +0000 UTC m=+836.514934933" observedRunningTime="2025-10-08 18:31:09.670900395 +0000 UTC m=+839.917739794" watchObservedRunningTime="2025-10-08 18:31:09.671885552 +0000 UTC m=+839.918724931" Oct 08 18:31:09 crc kubenswrapper[4859]: I1008 18:31:09.675360 4859 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/infra-operator-controller-manager-656bcbd775-ctjz9" event={"ID":"1a1cfd66-310a-4077-8020-df7e6e069c53","Type":"ContainerStarted","Data":"64cc31946226e5e4e32c5f3a1a3d1065a2b1f388d5f58c8f871ecfd16163e7cc"} Oct 08 18:31:09 crc kubenswrapper[4859]: I1008 18:31:09.675806 4859 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/infra-operator-controller-manager-656bcbd775-ctjz9" Oct 08 18:31:09 crc kubenswrapper[4859]: I1008 18:31:09.685047 4859 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/heat-operator-controller-manager-858f76bbdd-b5bfz" event={"ID":"19fa1d1d-0731-4c21-9faa-b1665ec4af36","Type":"ContainerStarted","Data":"bb520335fc499f90095e635a0379da2734d1ab31feb9f93051a3a592a20e2496"} Oct 08 18:31:09 crc kubenswrapper[4859]: I1008 18:31:09.685163 4859 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/heat-operator-controller-manager-858f76bbdd-b5bfz" Oct 08 18:31:09 crc kubenswrapper[4859]: I1008 18:31:09.693514 4859 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/ironic-operator-controller-manager-9c5c78d49-z9952" event={"ID":"2b67febf-c4c8-4e48-9350-29dd496439b4","Type":"ContainerStarted","Data":"a1f52641601ba33195963e4c3c4042530e81c674a31d3aeea5152bd5420df6f9"} Oct 08 18:31:09 crc kubenswrapper[4859]: I1008 18:31:09.693830 4859 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/ironic-operator-controller-manager-9c5c78d49-z9952" Oct 08 18:31:09 crc kubenswrapper[4859]: I1008 18:31:09.703060 4859 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/glance-operator-controller-manager-84b9b84486-d9gmv" event={"ID":"0b068d16-0072-41f8-ad01-4194fd428014","Type":"ContainerStarted","Data":"e9a847e6074167570144e03c64318bfae51213107df387e2cf4009d2367daf8a"} Oct 08 18:31:09 crc kubenswrapper[4859]: I1008 18:31:09.703939 4859 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/glance-operator-controller-manager-84b9b84486-d9gmv" Oct 08 18:31:09 crc kubenswrapper[4859]: I1008 18:31:09.705907 4859 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/designate-operator-controller-manager-85d5d9dd78-nhqkh" podStartSLOduration=5.900277988 podStartE2EDuration="17.70589262s" podCreationTimestamp="2025-10-08 18:30:52 +0000 UTC" firstStartedPulling="2025-10-08 18:30:54.495396979 +0000 UTC m=+824.742236358" lastFinishedPulling="2025-10-08 18:31:06.301011611 +0000 UTC m=+836.547850990" observedRunningTime="2025-10-08 18:31:09.700235991 +0000 UTC m=+839.947075370" watchObservedRunningTime="2025-10-08 18:31:09.70589262 +0000 UTC m=+839.952732009" Oct 08 18:31:09 crc kubenswrapper[4859]: I1008 18:31:09.711598 4859 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/neutron-operator-controller-manager-79d585cb66-zn65q" event={"ID":"e69529d1-1a2a-433c-a16b-3c6de62a0587","Type":"ContainerStarted","Data":"874a4b55db097c6a6ba710fc26d76490829a62a800bf0fdca0fd33c6bf0e2fff"} Oct 08 18:31:09 crc kubenswrapper[4859]: I1008 18:31:09.712462 4859 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/neutron-operator-controller-manager-79d585cb66-zn65q" Oct 08 18:31:09 crc kubenswrapper[4859]: I1008 18:31:09.723780 4859 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/cinder-operator-controller-manager-7b7fb68549-lbw64" podStartSLOduration=6.805667611 podStartE2EDuration="17.723751933s" podCreationTimestamp="2025-10-08 18:30:52 +0000 UTC" firstStartedPulling="2025-10-08 18:30:55.34392289 +0000 UTC m=+825.590762259" lastFinishedPulling="2025-10-08 18:31:06.262007192 +0000 UTC m=+836.508846581" observedRunningTime="2025-10-08 18:31:09.720610265 +0000 UTC m=+839.967449644" watchObservedRunningTime="2025-10-08 18:31:09.723751933 +0000 UTC m=+839.970591302" Oct 08 18:31:09 crc kubenswrapper[4859]: I1008 18:31:09.724487 4859 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/ovn-operator-controller-manager-79db49b9fb-hhgrl" event={"ID":"e2675d43-51ec-4a8c-aa2d-5aedc14f649a","Type":"ContainerStarted","Data":"b9c2f1f6e80d6b8b11f104a3924651a063c4ffc07fe9ae23aee0907402a7774e"} Oct 08 18:31:09 crc kubenswrapper[4859]: I1008 18:31:09.724678 4859 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/ovn-operator-controller-manager-79db49b9fb-hhgrl" Oct 08 18:31:09 crc kubenswrapper[4859]: I1008 18:31:09.732550 4859 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/horizon-operator-controller-manager-7ffbcb7588-86z4n" event={"ID":"ca71782e-3569-4ac2-a178-239f0ae200d0","Type":"ContainerStarted","Data":"c01886ae9b0119a881d88eb50e70c2c9009dad2e995dab5adc33d342d075e731"} Oct 08 18:31:09 crc kubenswrapper[4859]: I1008 18:31:09.732808 4859 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/horizon-operator-controller-manager-7ffbcb7588-86z4n" Oct 08 18:31:09 crc kubenswrapper[4859]: I1008 18:31:09.735175 4859 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/keystone-operator-controller-manager-55b6b7c7b8-rsbgr" event={"ID":"0228033e-c835-4618-9603-2c67cf9ce57d","Type":"ContainerStarted","Data":"c2acde374d3a44013ffec32c0923d809e0947ad0d1e64b66e8e2f0ea7bc0415a"} Oct 08 18:31:09 crc kubenswrapper[4859]: I1008 18:31:09.735636 4859 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/keystone-operator-controller-manager-55b6b7c7b8-rsbgr" Oct 08 18:31:09 crc kubenswrapper[4859]: I1008 18:31:09.744052 4859 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-b8rjh" event={"ID":"aefd0d7b-0c10-443e-9b21-6c47cb711eb7","Type":"ContainerStarted","Data":"97c3d33e64a8245acdfb332d454850d2a9d7d87cafb6bbb784f824a373d3ca04"} Oct 08 18:31:09 crc kubenswrapper[4859]: I1008 18:31:09.746318 4859 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/nova-operator-controller-manager-5df598886f-p99nw" event={"ID":"e2889a4a-89ec-47de-99df-7e749c3e5514","Type":"ContainerStarted","Data":"ce726b0b60addf0464a29f4461d98982383335fa5c4ccf5aee24916babf0ca69"} Oct 08 18:31:09 crc kubenswrapper[4859]: I1008 18:31:09.746483 4859 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/nova-operator-controller-manager-5df598886f-p99nw" Oct 08 18:31:09 crc kubenswrapper[4859]: I1008 18:31:09.749592 4859 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/octavia-operator-controller-manager-69fdcfc5f5-lsl8l" podStartSLOduration=5.785935276 podStartE2EDuration="16.749578521s" podCreationTimestamp="2025-10-08 18:30:53 +0000 UTC" firstStartedPulling="2025-10-08 18:30:55.31833448 +0000 UTC m=+825.565173859" lastFinishedPulling="2025-10-08 18:31:06.281977715 +0000 UTC m=+836.528817104" observedRunningTime="2025-10-08 18:31:09.748318435 +0000 UTC m=+839.995157814" watchObservedRunningTime="2025-10-08 18:31:09.749578521 +0000 UTC m=+839.996417900" Oct 08 18:31:09 crc kubenswrapper[4859]: I1008 18:31:09.758349 4859 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/manila-operator-controller-manager-5f67fbc655-xg74g" event={"ID":"fbce123a-545f-4c4c-86ca-2d68f81aa3f3","Type":"ContainerStarted","Data":"3b32789ffe45dce78ce960d827ad6fde10d7328074570b4be5c1e51a9f893543"} Oct 08 18:31:09 crc kubenswrapper[4859]: I1008 18:31:09.759083 4859 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/manila-operator-controller-manager-5f67fbc655-xg74g" Oct 08 18:31:09 crc kubenswrapper[4859]: I1008 18:31:09.762619 4859 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/swift-operator-controller-manager-db6d7f97b-zjkwm" event={"ID":"44980d79-924e-436f-950d-8fcdb4c42af1","Type":"ContainerStarted","Data":"f53817748fca127a1a510a33f2bbc8961dacbc2499a38d4df4550ed28dbd8c2f"} Oct 08 18:31:09 crc kubenswrapper[4859]: I1008 18:31:09.763795 4859 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/swift-operator-controller-manager-db6d7f97b-zjkwm" Oct 08 18:31:09 crc kubenswrapper[4859]: I1008 18:31:09.773486 4859 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/mariadb-operator-controller-manager-f9fb45f8f-ptpqs" podStartSLOduration=5.738672905 podStartE2EDuration="16.773468164s" podCreationTimestamp="2025-10-08 18:30:53 +0000 UTC" firstStartedPulling="2025-10-08 18:30:55.257837376 +0000 UTC m=+825.504676755" lastFinishedPulling="2025-10-08 18:31:06.292632635 +0000 UTC m=+836.539472014" observedRunningTime="2025-10-08 18:31:09.769148572 +0000 UTC m=+840.015987951" watchObservedRunningTime="2025-10-08 18:31:09.773468164 +0000 UTC m=+840.020307543" Oct 08 18:31:09 crc kubenswrapper[4859]: I1008 18:31:09.816528 4859 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/ironic-operator-controller-manager-9c5c78d49-z9952" podStartSLOduration=5.312972873 podStartE2EDuration="16.816506596s" podCreationTimestamp="2025-10-08 18:30:53 +0000 UTC" firstStartedPulling="2025-10-08 18:30:54.799739542 +0000 UTC m=+825.046578921" lastFinishedPulling="2025-10-08 18:31:06.303273265 +0000 UTC m=+836.550112644" observedRunningTime="2025-10-08 18:31:09.796138862 +0000 UTC m=+840.042978241" watchObservedRunningTime="2025-10-08 18:31:09.816506596 +0000 UTC m=+840.063345985" Oct 08 18:31:09 crc kubenswrapper[4859]: I1008 18:31:09.820557 4859 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/glance-operator-controller-manager-84b9b84486-d9gmv" podStartSLOduration=5.331981288 podStartE2EDuration="16.82054757s" podCreationTimestamp="2025-10-08 18:30:53 +0000 UTC" firstStartedPulling="2025-10-08 18:30:54.798953329 +0000 UTC m=+825.045792708" lastFinishedPulling="2025-10-08 18:31:06.287519601 +0000 UTC m=+836.534358990" observedRunningTime="2025-10-08 18:31:09.81594212 +0000 UTC m=+840.062781519" watchObservedRunningTime="2025-10-08 18:31:09.82054757 +0000 UTC m=+840.067386949" Oct 08 18:31:09 crc kubenswrapper[4859]: I1008 18:31:09.833816 4859 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/ovn-operator-controller-manager-79db49b9fb-hhgrl" podStartSLOduration=5.886132098 podStartE2EDuration="16.833801543s" podCreationTimestamp="2025-10-08 18:30:53 +0000 UTC" firstStartedPulling="2025-10-08 18:30:55.321354195 +0000 UTC m=+825.568193574" lastFinishedPulling="2025-10-08 18:31:06.26902364 +0000 UTC m=+836.515863019" observedRunningTime="2025-10-08 18:31:09.83261519 +0000 UTC m=+840.079454579" watchObservedRunningTime="2025-10-08 18:31:09.833801543 +0000 UTC m=+840.080640922" Oct 08 18:31:09 crc kubenswrapper[4859]: I1008 18:31:09.858243 4859 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/neutron-operator-controller-manager-79d585cb66-zn65q" podStartSLOduration=5.850780322 podStartE2EDuration="16.858220321s" podCreationTimestamp="2025-10-08 18:30:53 +0000 UTC" firstStartedPulling="2025-10-08 18:30:55.295669531 +0000 UTC m=+825.542508910" lastFinishedPulling="2025-10-08 18:31:06.30310953 +0000 UTC m=+836.549948909" observedRunningTime="2025-10-08 18:31:09.854599539 +0000 UTC m=+840.101438928" watchObservedRunningTime="2025-10-08 18:31:09.858220321 +0000 UTC m=+840.105059700" Oct 08 18:31:09 crc kubenswrapper[4859]: I1008 18:31:09.880157 4859 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/manila-operator-controller-manager-5f67fbc655-xg74g" podStartSLOduration=5.901183932 podStartE2EDuration="16.880137388s" podCreationTimestamp="2025-10-08 18:30:53 +0000 UTC" firstStartedPulling="2025-10-08 18:30:55.324064091 +0000 UTC m=+825.570903470" lastFinishedPulling="2025-10-08 18:31:06.303017547 +0000 UTC m=+836.549856926" observedRunningTime="2025-10-08 18:31:09.877730051 +0000 UTC m=+840.124569430" watchObservedRunningTime="2025-10-08 18:31:09.880137388 +0000 UTC m=+840.126976767" Oct 08 18:31:09 crc kubenswrapper[4859]: I1008 18:31:09.899547 4859 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/nova-operator-controller-manager-5df598886f-p99nw" podStartSLOduration=5.849575479 podStartE2EDuration="16.899529275s" podCreationTimestamp="2025-10-08 18:30:53 +0000 UTC" firstStartedPulling="2025-10-08 18:30:55.286490183 +0000 UTC m=+825.533329562" lastFinishedPulling="2025-10-08 18:31:06.336443979 +0000 UTC m=+836.583283358" observedRunningTime="2025-10-08 18:31:09.896776537 +0000 UTC m=+840.143615936" watchObservedRunningTime="2025-10-08 18:31:09.899529275 +0000 UTC m=+840.146368654" Oct 08 18:31:09 crc kubenswrapper[4859]: I1008 18:31:09.916908 4859 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/horizon-operator-controller-manager-7ffbcb7588-86z4n" podStartSLOduration=5.31034138 podStartE2EDuration="16.916883284s" podCreationTimestamp="2025-10-08 18:30:53 +0000 UTC" firstStartedPulling="2025-10-08 18:30:54.662419334 +0000 UTC m=+824.909258713" lastFinishedPulling="2025-10-08 18:31:06.268961238 +0000 UTC m=+836.515800617" observedRunningTime="2025-10-08 18:31:09.916596625 +0000 UTC m=+840.163436004" watchObservedRunningTime="2025-10-08 18:31:09.916883284 +0000 UTC m=+840.163722663" Oct 08 18:31:09 crc kubenswrapper[4859]: I1008 18:31:09.933148 4859 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/keystone-operator-controller-manager-55b6b7c7b8-rsbgr" podStartSLOduration=5.887919539 podStartE2EDuration="16.933124921s" podCreationTimestamp="2025-10-08 18:30:53 +0000 UTC" firstStartedPulling="2025-10-08 18:30:55.257490426 +0000 UTC m=+825.504329805" lastFinishedPulling="2025-10-08 18:31:06.302695808 +0000 UTC m=+836.549535187" observedRunningTime="2025-10-08 18:31:09.928637175 +0000 UTC m=+840.175476564" watchObservedRunningTime="2025-10-08 18:31:09.933124921 +0000 UTC m=+840.179964290" Oct 08 18:31:09 crc kubenswrapper[4859]: I1008 18:31:09.944802 4859 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/heat-operator-controller-manager-858f76bbdd-b5bfz" podStartSLOduration=5.345223282 podStartE2EDuration="16.944776679s" podCreationTimestamp="2025-10-08 18:30:53 +0000 UTC" firstStartedPulling="2025-10-08 18:30:54.662347602 +0000 UTC m=+824.909186981" lastFinishedPulling="2025-10-08 18:31:06.261900999 +0000 UTC m=+836.508740378" observedRunningTime="2025-10-08 18:31:09.943905355 +0000 UTC m=+840.190744744" watchObservedRunningTime="2025-10-08 18:31:09.944776679 +0000 UTC m=+840.191616058" Oct 08 18:31:09 crc kubenswrapper[4859]: I1008 18:31:09.982494 4859 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/infra-operator-controller-manager-656bcbd775-ctjz9" podStartSLOduration=6.496928984 podStartE2EDuration="16.982465981s" podCreationTimestamp="2025-10-08 18:30:53 +0000 UTC" firstStartedPulling="2025-10-08 18:30:55.7965324 +0000 UTC m=+826.043371779" lastFinishedPulling="2025-10-08 18:31:06.282069387 +0000 UTC m=+836.528908776" observedRunningTime="2025-10-08 18:31:09.975363761 +0000 UTC m=+840.222203140" watchObservedRunningTime="2025-10-08 18:31:09.982465981 +0000 UTC m=+840.229305360" Oct 08 18:31:10 crc kubenswrapper[4859]: I1008 18:31:10.003635 4859 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-b8rjh" podStartSLOduration=4.334820587 podStartE2EDuration="15.003611817s" podCreationTimestamp="2025-10-08 18:30:55 +0000 UTC" firstStartedPulling="2025-10-08 18:30:58.208221192 +0000 UTC m=+828.455060571" lastFinishedPulling="2025-10-08 18:31:08.877012422 +0000 UTC m=+839.123851801" observedRunningTime="2025-10-08 18:31:09.996932118 +0000 UTC m=+840.243771497" watchObservedRunningTime="2025-10-08 18:31:10.003611817 +0000 UTC m=+840.250451196" Oct 08 18:31:10 crc kubenswrapper[4859]: I1008 18:31:10.013592 4859 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/swift-operator-controller-manager-db6d7f97b-zjkwm" podStartSLOduration=6.065303325 podStartE2EDuration="17.013568097s" podCreationTimestamp="2025-10-08 18:30:53 +0000 UTC" firstStartedPulling="2025-10-08 18:30:55.320753318 +0000 UTC m=+825.567592697" lastFinishedPulling="2025-10-08 18:31:06.26901807 +0000 UTC m=+836.515857469" observedRunningTime="2025-10-08 18:31:10.012542128 +0000 UTC m=+840.259381517" watchObservedRunningTime="2025-10-08 18:31:10.013568097 +0000 UTC m=+840.260407486" Oct 08 18:31:11 crc kubenswrapper[4859]: I1008 18:31:11.785854 4859 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/placement-operator-controller-manager-68b6c87b68-xgjpk" event={"ID":"f56e7a0c-7b3a-4dc8-9367-cfeef5c5cef9","Type":"ContainerStarted","Data":"5ccc7d12e82cdfdcef48be367b4e5090cc5240bbd66b51ecabaf7a26e7f12f02"} Oct 08 18:31:11 crc kubenswrapper[4859]: I1008 18:31:11.808617 4859 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/placement-operator-controller-manager-68b6c87b68-xgjpk" podStartSLOduration=3.437665119 podStartE2EDuration="18.808594289s" podCreationTimestamp="2025-10-08 18:30:53 +0000 UTC" firstStartedPulling="2025-10-08 18:30:55.60199816 +0000 UTC m=+825.848837539" lastFinishedPulling="2025-10-08 18:31:10.97292733 +0000 UTC m=+841.219766709" observedRunningTime="2025-10-08 18:31:11.802173968 +0000 UTC m=+842.049013537" watchObservedRunningTime="2025-10-08 18:31:11.808594289 +0000 UTC m=+842.055433678" Oct 08 18:31:13 crc kubenswrapper[4859]: I1008 18:31:13.273463 4859 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/barbican-operator-controller-manager-658bdf4b74-8bfst" Oct 08 18:31:13 crc kubenswrapper[4859]: I1008 18:31:13.334164 4859 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/designate-operator-controller-manager-85d5d9dd78-nhqkh" Oct 08 18:31:13 crc kubenswrapper[4859]: I1008 18:31:13.424279 4859 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/glance-operator-controller-manager-84b9b84486-d9gmv" Oct 08 18:31:13 crc kubenswrapper[4859]: I1008 18:31:13.443830 4859 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/horizon-operator-controller-manager-7ffbcb7588-86z4n" Oct 08 18:31:13 crc kubenswrapper[4859]: I1008 18:31:13.453569 4859 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/heat-operator-controller-manager-858f76bbdd-b5bfz" Oct 08 18:31:13 crc kubenswrapper[4859]: I1008 18:31:13.478355 4859 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/ironic-operator-controller-manager-9c5c78d49-z9952" Oct 08 18:31:13 crc kubenswrapper[4859]: I1008 18:31:13.549192 4859 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/keystone-operator-controller-manager-55b6b7c7b8-rsbgr" Oct 08 18:31:13 crc kubenswrapper[4859]: I1008 18:31:13.604900 4859 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/cinder-operator-controller-manager-7b7fb68549-lbw64" Oct 08 18:31:13 crc kubenswrapper[4859]: I1008 18:31:13.772470 4859 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/mariadb-operator-controller-manager-f9fb45f8f-ptpqs" Oct 08 18:31:13 crc kubenswrapper[4859]: I1008 18:31:13.774714 4859 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/neutron-operator-controller-manager-79d585cb66-zn65q" Oct 08 18:31:13 crc kubenswrapper[4859]: I1008 18:31:13.821929 4859 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/nova-operator-controller-manager-5df598886f-p99nw" Oct 08 18:31:13 crc kubenswrapper[4859]: I1008 18:31:13.842983 4859 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/octavia-operator-controller-manager-69fdcfc5f5-lsl8l" Oct 08 18:31:13 crc kubenswrapper[4859]: I1008 18:31:13.866302 4859 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/ovn-operator-controller-manager-79db49b9fb-hhgrl" Oct 08 18:31:13 crc kubenswrapper[4859]: I1008 18:31:13.887011 4859 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/manila-operator-controller-manager-5f67fbc655-xg74g" Oct 08 18:31:13 crc kubenswrapper[4859]: I1008 18:31:13.926865 4859 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/placement-operator-controller-manager-68b6c87b68-xgjpk" Oct 08 18:31:14 crc kubenswrapper[4859]: I1008 18:31:14.048878 4859 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/swift-operator-controller-manager-db6d7f97b-zjkwm" Oct 08 18:31:14 crc kubenswrapper[4859]: I1008 18:31:14.964060 4859 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/infra-operator-controller-manager-656bcbd775-ctjz9" Oct 08 18:31:15 crc kubenswrapper[4859]: I1008 18:31:15.716407 4859 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-b8rjh" Oct 08 18:31:15 crc kubenswrapper[4859]: I1008 18:31:15.716485 4859 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-b8rjh" Oct 08 18:31:15 crc kubenswrapper[4859]: I1008 18:31:15.765680 4859 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-b8rjh" Oct 08 18:31:15 crc kubenswrapper[4859]: I1008 18:31:15.857529 4859 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-b8rjh" Oct 08 18:31:16 crc kubenswrapper[4859]: I1008 18:31:16.002396 4859 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-b8rjh"] Oct 08 18:31:17 crc kubenswrapper[4859]: I1008 18:31:17.831372 4859 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/certified-operators-b8rjh" podUID="aefd0d7b-0c10-443e-9b21-6c47cb711eb7" containerName="registry-server" containerID="cri-o://97c3d33e64a8245acdfb332d454850d2a9d7d87cafb6bbb784f824a373d3ca04" gracePeriod=2 Oct 08 18:31:17 crc kubenswrapper[4859]: I1008 18:31:17.924718 4859 patch_prober.go:28] interesting pod/machine-config-daemon-82s52 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 08 18:31:17 crc kubenswrapper[4859]: I1008 18:31:17.924789 4859 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-82s52" podUID="b23a6a6c-9d92-4e7b-840e-55cfda873a2d" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 08 18:31:17 crc kubenswrapper[4859]: I1008 18:31:17.924847 4859 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-82s52" Oct 08 18:31:17 crc kubenswrapper[4859]: I1008 18:31:17.925584 4859 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"6ca72342b251d747b10d72d072b3e566097d846a52f53943f7da38fe2a34e100"} pod="openshift-machine-config-operator/machine-config-daemon-82s52" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Oct 08 18:31:17 crc kubenswrapper[4859]: I1008 18:31:17.925657 4859 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-82s52" podUID="b23a6a6c-9d92-4e7b-840e-55cfda873a2d" containerName="machine-config-daemon" containerID="cri-o://6ca72342b251d747b10d72d072b3e566097d846a52f53943f7da38fe2a34e100" gracePeriod=600 Oct 08 18:31:18 crc kubenswrapper[4859]: I1008 18:31:18.843740 4859 generic.go:334] "Generic (PLEG): container finished" podID="b23a6a6c-9d92-4e7b-840e-55cfda873a2d" containerID="6ca72342b251d747b10d72d072b3e566097d846a52f53943f7da38fe2a34e100" exitCode=0 Oct 08 18:31:18 crc kubenswrapper[4859]: I1008 18:31:18.843810 4859 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-82s52" event={"ID":"b23a6a6c-9d92-4e7b-840e-55cfda873a2d","Type":"ContainerDied","Data":"6ca72342b251d747b10d72d072b3e566097d846a52f53943f7da38fe2a34e100"} Oct 08 18:31:18 crc kubenswrapper[4859]: I1008 18:31:18.844118 4859 scope.go:117] "RemoveContainer" containerID="487a3a1048ef98faa7383425fe83462232329f063579bf1e501fc8c04727fcea" Oct 08 18:31:19 crc kubenswrapper[4859]: I1008 18:31:19.859123 4859 generic.go:334] "Generic (PLEG): container finished" podID="aefd0d7b-0c10-443e-9b21-6c47cb711eb7" containerID="97c3d33e64a8245acdfb332d454850d2a9d7d87cafb6bbb784f824a373d3ca04" exitCode=0 Oct 08 18:31:19 crc kubenswrapper[4859]: I1008 18:31:19.859199 4859 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-b8rjh" event={"ID":"aefd0d7b-0c10-443e-9b21-6c47cb711eb7","Type":"ContainerDied","Data":"97c3d33e64a8245acdfb332d454850d2a9d7d87cafb6bbb784f824a373d3ca04"} Oct 08 18:31:20 crc kubenswrapper[4859]: I1008 18:31:20.885169 4859 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-b8rjh" event={"ID":"aefd0d7b-0c10-443e-9b21-6c47cb711eb7","Type":"ContainerDied","Data":"5f8885dec4565219c199bd035c2888192ab214ea38bec07e68b6d2664903dd26"} Oct 08 18:31:20 crc kubenswrapper[4859]: I1008 18:31:20.885918 4859 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="5f8885dec4565219c199bd035c2888192ab214ea38bec07e68b6d2664903dd26" Oct 08 18:31:20 crc kubenswrapper[4859]: I1008 18:31:20.889334 4859 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-82s52" event={"ID":"b23a6a6c-9d92-4e7b-840e-55cfda873a2d","Type":"ContainerStarted","Data":"091916d0315d4cd4d10d5ce64ba3d175b9444c9b90ffd7170b4b5fd2fb8bb62d"} Oct 08 18:31:20 crc kubenswrapper[4859]: I1008 18:31:20.891697 4859 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-b8rjh" Oct 08 18:31:21 crc kubenswrapper[4859]: I1008 18:31:21.073214 4859 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/aefd0d7b-0c10-443e-9b21-6c47cb711eb7-utilities\") pod \"aefd0d7b-0c10-443e-9b21-6c47cb711eb7\" (UID: \"aefd0d7b-0c10-443e-9b21-6c47cb711eb7\") " Oct 08 18:31:21 crc kubenswrapper[4859]: I1008 18:31:21.073446 4859 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/aefd0d7b-0c10-443e-9b21-6c47cb711eb7-catalog-content\") pod \"aefd0d7b-0c10-443e-9b21-6c47cb711eb7\" (UID: \"aefd0d7b-0c10-443e-9b21-6c47cb711eb7\") " Oct 08 18:31:21 crc kubenswrapper[4859]: I1008 18:31:21.073487 4859 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-rhbnl\" (UniqueName: \"kubernetes.io/projected/aefd0d7b-0c10-443e-9b21-6c47cb711eb7-kube-api-access-rhbnl\") pod \"aefd0d7b-0c10-443e-9b21-6c47cb711eb7\" (UID: \"aefd0d7b-0c10-443e-9b21-6c47cb711eb7\") " Oct 08 18:31:21 crc kubenswrapper[4859]: I1008 18:31:21.074932 4859 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/aefd0d7b-0c10-443e-9b21-6c47cb711eb7-utilities" (OuterVolumeSpecName: "utilities") pod "aefd0d7b-0c10-443e-9b21-6c47cb711eb7" (UID: "aefd0d7b-0c10-443e-9b21-6c47cb711eb7"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 08 18:31:21 crc kubenswrapper[4859]: I1008 18:31:21.082003 4859 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/aefd0d7b-0c10-443e-9b21-6c47cb711eb7-kube-api-access-rhbnl" (OuterVolumeSpecName: "kube-api-access-rhbnl") pod "aefd0d7b-0c10-443e-9b21-6c47cb711eb7" (UID: "aefd0d7b-0c10-443e-9b21-6c47cb711eb7"). InnerVolumeSpecName "kube-api-access-rhbnl". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 08 18:31:21 crc kubenswrapper[4859]: I1008 18:31:21.127575 4859 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/aefd0d7b-0c10-443e-9b21-6c47cb711eb7-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "aefd0d7b-0c10-443e-9b21-6c47cb711eb7" (UID: "aefd0d7b-0c10-443e-9b21-6c47cb711eb7"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 08 18:31:21 crc kubenswrapper[4859]: I1008 18:31:21.175242 4859 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/aefd0d7b-0c10-443e-9b21-6c47cb711eb7-utilities\") on node \"crc\" DevicePath \"\"" Oct 08 18:31:21 crc kubenswrapper[4859]: I1008 18:31:21.175295 4859 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/aefd0d7b-0c10-443e-9b21-6c47cb711eb7-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 08 18:31:21 crc kubenswrapper[4859]: I1008 18:31:21.175316 4859 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-rhbnl\" (UniqueName: \"kubernetes.io/projected/aefd0d7b-0c10-443e-9b21-6c47cb711eb7-kube-api-access-rhbnl\") on node \"crc\" DevicePath \"\"" Oct 08 18:31:21 crc kubenswrapper[4859]: I1008 18:31:21.897378 4859 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-b8rjh" Oct 08 18:31:21 crc kubenswrapper[4859]: I1008 18:31:21.933346 4859 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-b8rjh"] Oct 08 18:31:21 crc kubenswrapper[4859]: I1008 18:31:21.939661 4859 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/certified-operators-b8rjh"] Oct 08 18:31:22 crc kubenswrapper[4859]: E1008 18:31:22.011466 4859 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podaefd0d7b_0c10_443e_9b21_6c47cb711eb7.slice/crio-5f8885dec4565219c199bd035c2888192ab214ea38bec07e68b6d2664903dd26\": RecentStats: unable to find data in memory cache]" Oct 08 18:31:22 crc kubenswrapper[4859]: I1008 18:31:22.479325 4859 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="aefd0d7b-0c10-443e-9b21-6c47cb711eb7" path="/var/lib/kubelet/pods/aefd0d7b-0c10-443e-9b21-6c47cb711eb7/volumes" Oct 08 18:31:22 crc kubenswrapper[4859]: I1008 18:31:22.906411 4859 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/test-operator-controller-manager-56c698c775-d2g8b" event={"ID":"2d6cf85d-c5cc-4466-a471-bb17aa8fca9d","Type":"ContainerStarted","Data":"99a4190da7f37ee722b6f11b93fd4bda60a8e8d31d0a8ef97ddcdb9b08d8909a"} Oct 08 18:31:22 crc kubenswrapper[4859]: I1008 18:31:22.906907 4859 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/test-operator-controller-manager-56c698c775-d2g8b" Oct 08 18:31:22 crc kubenswrapper[4859]: I1008 18:31:22.908944 4859 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/telemetry-operator-controller-manager-76796d4c6b-kjbfp" event={"ID":"ffa2914a-717c-4763-8b1c-4a1ff18858f5","Type":"ContainerStarted","Data":"8222e588411dcf1212d3908c9fbeaaaa28ff2a4d4f84814b295ed61b7c6f06bc"} Oct 08 18:31:22 crc kubenswrapper[4859]: I1008 18:31:22.909152 4859 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/telemetry-operator-controller-manager-76796d4c6b-kjbfp" Oct 08 18:31:22 crc kubenswrapper[4859]: I1008 18:31:22.910987 4859 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-baremetal-operator-controller-manager-677c5f5bff5pg7p" event={"ID":"dd924000-5690-490e-a4db-e29ed7f821ec","Type":"ContainerStarted","Data":"7186cace6305d406adf55724e12b5f6410ce533163d9c63afd9d680b087684b7"} Oct 08 18:31:22 crc kubenswrapper[4859]: I1008 18:31:22.911130 4859 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/openstack-baremetal-operator-controller-manager-677c5f5bff5pg7p" Oct 08 18:31:22 crc kubenswrapper[4859]: I1008 18:31:22.912745 4859 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/watcher-operator-controller-manager-7794bc6bd-2v65r" event={"ID":"cf5b6741-336e-435c-8209-82b8e19f4896","Type":"ContainerStarted","Data":"1abe568b8888d1b557d9873abc5a92eac6896fb5790d2bed17d6877bda5a4a9a"} Oct 08 18:31:22 crc kubenswrapper[4859]: I1008 18:31:22.912894 4859 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/watcher-operator-controller-manager-7794bc6bd-2v65r" Oct 08 18:31:22 crc kubenswrapper[4859]: I1008 18:31:22.928494 4859 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/test-operator-controller-manager-56c698c775-d2g8b" podStartSLOduration=3.396683395 podStartE2EDuration="29.928469935s" podCreationTimestamp="2025-10-08 18:30:53 +0000 UTC" firstStartedPulling="2025-10-08 18:30:55.627331424 +0000 UTC m=+825.874170803" lastFinishedPulling="2025-10-08 18:31:22.159117964 +0000 UTC m=+852.405957343" observedRunningTime="2025-10-08 18:31:22.924491698 +0000 UTC m=+853.171331077" watchObservedRunningTime="2025-10-08 18:31:22.928469935 +0000 UTC m=+853.175309314" Oct 08 18:31:22 crc kubenswrapper[4859]: I1008 18:31:22.958603 4859 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/openstack-baremetal-operator-controller-manager-677c5f5bff5pg7p" podStartSLOduration=3.600179395 podStartE2EDuration="29.958575237s" podCreationTimestamp="2025-10-08 18:30:53 +0000 UTC" firstStartedPulling="2025-10-08 18:30:55.797060575 +0000 UTC m=+826.043899954" lastFinishedPulling="2025-10-08 18:31:22.155456417 +0000 UTC m=+852.402295796" observedRunningTime="2025-10-08 18:31:22.955095195 +0000 UTC m=+853.201934594" watchObservedRunningTime="2025-10-08 18:31:22.958575237 +0000 UTC m=+853.205414616" Oct 08 18:31:22 crc kubenswrapper[4859]: I1008 18:31:22.973290 4859 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/telemetry-operator-controller-manager-76796d4c6b-kjbfp" podStartSLOduration=3.407985765 podStartE2EDuration="29.973266128s" podCreationTimestamp="2025-10-08 18:30:53 +0000 UTC" firstStartedPulling="2025-10-08 18:30:55.610478419 +0000 UTC m=+825.857317798" lastFinishedPulling="2025-10-08 18:31:22.175758782 +0000 UTC m=+852.422598161" observedRunningTime="2025-10-08 18:31:22.970637281 +0000 UTC m=+853.217476660" watchObservedRunningTime="2025-10-08 18:31:22.973266128 +0000 UTC m=+853.220105507" Oct 08 18:31:22 crc kubenswrapper[4859]: I1008 18:31:22.989743 4859 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/watcher-operator-controller-manager-7794bc6bd-2v65r" podStartSLOduration=3.484435839 podStartE2EDuration="29.98972337s" podCreationTimestamp="2025-10-08 18:30:53 +0000 UTC" firstStartedPulling="2025-10-08 18:30:55.65098873 +0000 UTC m=+825.897828109" lastFinishedPulling="2025-10-08 18:31:22.156276261 +0000 UTC m=+852.403115640" observedRunningTime="2025-10-08 18:31:22.985394683 +0000 UTC m=+853.232234072" watchObservedRunningTime="2025-10-08 18:31:22.98972337 +0000 UTC m=+853.236562749" Oct 08 18:31:23 crc kubenswrapper[4859]: I1008 18:31:23.925103 4859 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/placement-operator-controller-manager-68b6c87b68-xgjpk" Oct 08 18:31:34 crc kubenswrapper[4859]: I1008 18:31:34.069970 4859 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/telemetry-operator-controller-manager-76796d4c6b-kjbfp" Oct 08 18:31:34 crc kubenswrapper[4859]: I1008 18:31:34.098363 4859 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/test-operator-controller-manager-56c698c775-d2g8b" Oct 08 18:31:34 crc kubenswrapper[4859]: I1008 18:31:34.161136 4859 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/watcher-operator-controller-manager-7794bc6bd-2v65r" Oct 08 18:31:34 crc kubenswrapper[4859]: I1008 18:31:34.492927 4859 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/openstack-baremetal-operator-controller-manager-677c5f5bff5pg7p" Oct 08 18:31:50 crc kubenswrapper[4859]: I1008 18:31:50.318933 4859 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-7bfcb9d745-ntml8"] Oct 08 18:31:50 crc kubenswrapper[4859]: E1008 18:31:50.319929 4859 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="51932a59-6c7a-4c64-9acf-59984e192d91" containerName="extract-utilities" Oct 08 18:31:50 crc kubenswrapper[4859]: I1008 18:31:50.319950 4859 state_mem.go:107] "Deleted CPUSet assignment" podUID="51932a59-6c7a-4c64-9acf-59984e192d91" containerName="extract-utilities" Oct 08 18:31:50 crc kubenswrapper[4859]: E1008 18:31:50.320000 4859 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="aefd0d7b-0c10-443e-9b21-6c47cb711eb7" containerName="registry-server" Oct 08 18:31:50 crc kubenswrapper[4859]: I1008 18:31:50.320009 4859 state_mem.go:107] "Deleted CPUSet assignment" podUID="aefd0d7b-0c10-443e-9b21-6c47cb711eb7" containerName="registry-server" Oct 08 18:31:50 crc kubenswrapper[4859]: E1008 18:31:50.320031 4859 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="aefd0d7b-0c10-443e-9b21-6c47cb711eb7" containerName="extract-content" Oct 08 18:31:50 crc kubenswrapper[4859]: I1008 18:31:50.320040 4859 state_mem.go:107] "Deleted CPUSet assignment" podUID="aefd0d7b-0c10-443e-9b21-6c47cb711eb7" containerName="extract-content" Oct 08 18:31:50 crc kubenswrapper[4859]: E1008 18:31:50.320057 4859 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="51932a59-6c7a-4c64-9acf-59984e192d91" containerName="extract-content" Oct 08 18:31:50 crc kubenswrapper[4859]: I1008 18:31:50.320066 4859 state_mem.go:107] "Deleted CPUSet assignment" podUID="51932a59-6c7a-4c64-9acf-59984e192d91" containerName="extract-content" Oct 08 18:31:50 crc kubenswrapper[4859]: E1008 18:31:50.320082 4859 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="aefd0d7b-0c10-443e-9b21-6c47cb711eb7" containerName="extract-utilities" Oct 08 18:31:50 crc kubenswrapper[4859]: I1008 18:31:50.320090 4859 state_mem.go:107] "Deleted CPUSet assignment" podUID="aefd0d7b-0c10-443e-9b21-6c47cb711eb7" containerName="extract-utilities" Oct 08 18:31:50 crc kubenswrapper[4859]: E1008 18:31:50.320101 4859 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="51932a59-6c7a-4c64-9acf-59984e192d91" containerName="registry-server" Oct 08 18:31:50 crc kubenswrapper[4859]: I1008 18:31:50.320108 4859 state_mem.go:107] "Deleted CPUSet assignment" podUID="51932a59-6c7a-4c64-9acf-59984e192d91" containerName="registry-server" Oct 08 18:31:50 crc kubenswrapper[4859]: I1008 18:31:50.320286 4859 memory_manager.go:354] "RemoveStaleState removing state" podUID="aefd0d7b-0c10-443e-9b21-6c47cb711eb7" containerName="registry-server" Oct 08 18:31:50 crc kubenswrapper[4859]: I1008 18:31:50.320303 4859 memory_manager.go:354] "RemoveStaleState removing state" podUID="51932a59-6c7a-4c64-9acf-59984e192d91" containerName="registry-server" Oct 08 18:31:50 crc kubenswrapper[4859]: I1008 18:31:50.321224 4859 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-7bfcb9d745-ntml8" Oct 08 18:31:50 crc kubenswrapper[4859]: I1008 18:31:50.325351 4859 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"kube-root-ca.crt" Oct 08 18:31:50 crc kubenswrapper[4859]: I1008 18:31:50.325556 4859 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openshift-service-ca.crt" Oct 08 18:31:50 crc kubenswrapper[4859]: I1008 18:31:50.325884 4859 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dnsmasq-dns-dockercfg-r4k9l" Oct 08 18:31:50 crc kubenswrapper[4859]: I1008 18:31:50.330964 4859 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-7bfcb9d745-ntml8"] Oct 08 18:31:50 crc kubenswrapper[4859]: I1008 18:31:50.332393 4859 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"dns" Oct 08 18:31:50 crc kubenswrapper[4859]: I1008 18:31:50.352863 4859 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-plgwt\" (UniqueName: \"kubernetes.io/projected/e66f484b-21c2-45ff-9783-f73530d8ba73-kube-api-access-plgwt\") pod \"dnsmasq-dns-7bfcb9d745-ntml8\" (UID: \"e66f484b-21c2-45ff-9783-f73530d8ba73\") " pod="openstack/dnsmasq-dns-7bfcb9d745-ntml8" Oct 08 18:31:50 crc kubenswrapper[4859]: I1008 18:31:50.352980 4859 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/e66f484b-21c2-45ff-9783-f73530d8ba73-config\") pod \"dnsmasq-dns-7bfcb9d745-ntml8\" (UID: \"e66f484b-21c2-45ff-9783-f73530d8ba73\") " pod="openstack/dnsmasq-dns-7bfcb9d745-ntml8" Oct 08 18:31:50 crc kubenswrapper[4859]: I1008 18:31:50.404064 4859 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-758b79db4c-ncfwx"] Oct 08 18:31:50 crc kubenswrapper[4859]: I1008 18:31:50.405763 4859 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-758b79db4c-ncfwx" Oct 08 18:31:50 crc kubenswrapper[4859]: I1008 18:31:50.412645 4859 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"dns-svc" Oct 08 18:31:50 crc kubenswrapper[4859]: I1008 18:31:50.421629 4859 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-758b79db4c-ncfwx"] Oct 08 18:31:50 crc kubenswrapper[4859]: I1008 18:31:50.457439 4859 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/1b55c542-e6b1-4fd1-9459-f5f6555f8e21-dns-svc\") pod \"dnsmasq-dns-758b79db4c-ncfwx\" (UID: \"1b55c542-e6b1-4fd1-9459-f5f6555f8e21\") " pod="openstack/dnsmasq-dns-758b79db4c-ncfwx" Oct 08 18:31:50 crc kubenswrapper[4859]: I1008 18:31:50.457517 4859 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/e66f484b-21c2-45ff-9783-f73530d8ba73-config\") pod \"dnsmasq-dns-7bfcb9d745-ntml8\" (UID: \"e66f484b-21c2-45ff-9783-f73530d8ba73\") " pod="openstack/dnsmasq-dns-7bfcb9d745-ntml8" Oct 08 18:31:50 crc kubenswrapper[4859]: I1008 18:31:50.457637 4859 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-hgj4x\" (UniqueName: \"kubernetes.io/projected/1b55c542-e6b1-4fd1-9459-f5f6555f8e21-kube-api-access-hgj4x\") pod \"dnsmasq-dns-758b79db4c-ncfwx\" (UID: \"1b55c542-e6b1-4fd1-9459-f5f6555f8e21\") " pod="openstack/dnsmasq-dns-758b79db4c-ncfwx" Oct 08 18:31:50 crc kubenswrapper[4859]: I1008 18:31:50.457664 4859 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-plgwt\" (UniqueName: \"kubernetes.io/projected/e66f484b-21c2-45ff-9783-f73530d8ba73-kube-api-access-plgwt\") pod \"dnsmasq-dns-7bfcb9d745-ntml8\" (UID: \"e66f484b-21c2-45ff-9783-f73530d8ba73\") " pod="openstack/dnsmasq-dns-7bfcb9d745-ntml8" Oct 08 18:31:50 crc kubenswrapper[4859]: I1008 18:31:50.457783 4859 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/1b55c542-e6b1-4fd1-9459-f5f6555f8e21-config\") pod \"dnsmasq-dns-758b79db4c-ncfwx\" (UID: \"1b55c542-e6b1-4fd1-9459-f5f6555f8e21\") " pod="openstack/dnsmasq-dns-758b79db4c-ncfwx" Oct 08 18:31:50 crc kubenswrapper[4859]: I1008 18:31:50.458997 4859 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/e66f484b-21c2-45ff-9783-f73530d8ba73-config\") pod \"dnsmasq-dns-7bfcb9d745-ntml8\" (UID: \"e66f484b-21c2-45ff-9783-f73530d8ba73\") " pod="openstack/dnsmasq-dns-7bfcb9d745-ntml8" Oct 08 18:31:50 crc kubenswrapper[4859]: I1008 18:31:50.480736 4859 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-plgwt\" (UniqueName: \"kubernetes.io/projected/e66f484b-21c2-45ff-9783-f73530d8ba73-kube-api-access-plgwt\") pod \"dnsmasq-dns-7bfcb9d745-ntml8\" (UID: \"e66f484b-21c2-45ff-9783-f73530d8ba73\") " pod="openstack/dnsmasq-dns-7bfcb9d745-ntml8" Oct 08 18:31:50 crc kubenswrapper[4859]: I1008 18:31:50.559226 4859 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-hgj4x\" (UniqueName: \"kubernetes.io/projected/1b55c542-e6b1-4fd1-9459-f5f6555f8e21-kube-api-access-hgj4x\") pod \"dnsmasq-dns-758b79db4c-ncfwx\" (UID: \"1b55c542-e6b1-4fd1-9459-f5f6555f8e21\") " pod="openstack/dnsmasq-dns-758b79db4c-ncfwx" Oct 08 18:31:50 crc kubenswrapper[4859]: I1008 18:31:50.559316 4859 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/1b55c542-e6b1-4fd1-9459-f5f6555f8e21-config\") pod \"dnsmasq-dns-758b79db4c-ncfwx\" (UID: \"1b55c542-e6b1-4fd1-9459-f5f6555f8e21\") " pod="openstack/dnsmasq-dns-758b79db4c-ncfwx" Oct 08 18:31:50 crc kubenswrapper[4859]: I1008 18:31:50.559387 4859 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/1b55c542-e6b1-4fd1-9459-f5f6555f8e21-dns-svc\") pod \"dnsmasq-dns-758b79db4c-ncfwx\" (UID: \"1b55c542-e6b1-4fd1-9459-f5f6555f8e21\") " pod="openstack/dnsmasq-dns-758b79db4c-ncfwx" Oct 08 18:31:50 crc kubenswrapper[4859]: I1008 18:31:50.560244 4859 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/1b55c542-e6b1-4fd1-9459-f5f6555f8e21-dns-svc\") pod \"dnsmasq-dns-758b79db4c-ncfwx\" (UID: \"1b55c542-e6b1-4fd1-9459-f5f6555f8e21\") " pod="openstack/dnsmasq-dns-758b79db4c-ncfwx" Oct 08 18:31:50 crc kubenswrapper[4859]: I1008 18:31:50.560370 4859 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/1b55c542-e6b1-4fd1-9459-f5f6555f8e21-config\") pod \"dnsmasq-dns-758b79db4c-ncfwx\" (UID: \"1b55c542-e6b1-4fd1-9459-f5f6555f8e21\") " pod="openstack/dnsmasq-dns-758b79db4c-ncfwx" Oct 08 18:31:50 crc kubenswrapper[4859]: I1008 18:31:50.587128 4859 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-hgj4x\" (UniqueName: \"kubernetes.io/projected/1b55c542-e6b1-4fd1-9459-f5f6555f8e21-kube-api-access-hgj4x\") pod \"dnsmasq-dns-758b79db4c-ncfwx\" (UID: \"1b55c542-e6b1-4fd1-9459-f5f6555f8e21\") " pod="openstack/dnsmasq-dns-758b79db4c-ncfwx" Oct 08 18:31:50 crc kubenswrapper[4859]: I1008 18:31:50.641988 4859 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-7bfcb9d745-ntml8" Oct 08 18:31:50 crc kubenswrapper[4859]: I1008 18:31:50.731996 4859 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-758b79db4c-ncfwx" Oct 08 18:31:51 crc kubenswrapper[4859]: I1008 18:31:51.088090 4859 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-7bfcb9d745-ntml8"] Oct 08 18:31:51 crc kubenswrapper[4859]: W1008 18:31:51.094102 4859 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pode66f484b_21c2_45ff_9783_f73530d8ba73.slice/crio-32ac335953cb8ee65305bcaaadbeaf862c442a1a5c5ad7da82a336877b0a933f WatchSource:0}: Error finding container 32ac335953cb8ee65305bcaaadbeaf862c442a1a5c5ad7da82a336877b0a933f: Status 404 returned error can't find the container with id 32ac335953cb8ee65305bcaaadbeaf862c442a1a5c5ad7da82a336877b0a933f Oct 08 18:31:51 crc kubenswrapper[4859]: I1008 18:31:51.157088 4859 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-7bfcb9d745-ntml8" event={"ID":"e66f484b-21c2-45ff-9783-f73530d8ba73","Type":"ContainerStarted","Data":"32ac335953cb8ee65305bcaaadbeaf862c442a1a5c5ad7da82a336877b0a933f"} Oct 08 18:31:51 crc kubenswrapper[4859]: I1008 18:31:51.270529 4859 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-758b79db4c-ncfwx"] Oct 08 18:31:51 crc kubenswrapper[4859]: W1008 18:31:51.279284 4859 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod1b55c542_e6b1_4fd1_9459_f5f6555f8e21.slice/crio-2a163aafcd54edb49185c5e4b31adcbe370d9f9b9e038d4c4477f5eb92aea65f WatchSource:0}: Error finding container 2a163aafcd54edb49185c5e4b31adcbe370d9f9b9e038d4c4477f5eb92aea65f: Status 404 returned error can't find the container with id 2a163aafcd54edb49185c5e4b31adcbe370d9f9b9e038d4c4477f5eb92aea65f Oct 08 18:31:52 crc kubenswrapper[4859]: I1008 18:31:52.167191 4859 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-758b79db4c-ncfwx" event={"ID":"1b55c542-e6b1-4fd1-9459-f5f6555f8e21","Type":"ContainerStarted","Data":"2a163aafcd54edb49185c5e4b31adcbe370d9f9b9e038d4c4477f5eb92aea65f"} Oct 08 18:31:53 crc kubenswrapper[4859]: I1008 18:31:53.365395 4859 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-758b79db4c-ncfwx"] Oct 08 18:31:53 crc kubenswrapper[4859]: I1008 18:31:53.393342 4859 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-644597f84c-x8r6m"] Oct 08 18:31:53 crc kubenswrapper[4859]: I1008 18:31:53.395411 4859 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-644597f84c-x8r6m" Oct 08 18:31:53 crc kubenswrapper[4859]: I1008 18:31:53.402954 4859 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/834c27d2-782d-43a6-a955-e0c8b99d93a3-config\") pod \"dnsmasq-dns-644597f84c-x8r6m\" (UID: \"834c27d2-782d-43a6-a955-e0c8b99d93a3\") " pod="openstack/dnsmasq-dns-644597f84c-x8r6m" Oct 08 18:31:53 crc kubenswrapper[4859]: I1008 18:31:53.403042 4859 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/834c27d2-782d-43a6-a955-e0c8b99d93a3-dns-svc\") pod \"dnsmasq-dns-644597f84c-x8r6m\" (UID: \"834c27d2-782d-43a6-a955-e0c8b99d93a3\") " pod="openstack/dnsmasq-dns-644597f84c-x8r6m" Oct 08 18:31:53 crc kubenswrapper[4859]: I1008 18:31:53.403083 4859 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-rvnjq\" (UniqueName: \"kubernetes.io/projected/834c27d2-782d-43a6-a955-e0c8b99d93a3-kube-api-access-rvnjq\") pod \"dnsmasq-dns-644597f84c-x8r6m\" (UID: \"834c27d2-782d-43a6-a955-e0c8b99d93a3\") " pod="openstack/dnsmasq-dns-644597f84c-x8r6m" Oct 08 18:31:53 crc kubenswrapper[4859]: I1008 18:31:53.407321 4859 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-644597f84c-x8r6m"] Oct 08 18:31:53 crc kubenswrapper[4859]: I1008 18:31:53.518481 4859 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/834c27d2-782d-43a6-a955-e0c8b99d93a3-dns-svc\") pod \"dnsmasq-dns-644597f84c-x8r6m\" (UID: \"834c27d2-782d-43a6-a955-e0c8b99d93a3\") " pod="openstack/dnsmasq-dns-644597f84c-x8r6m" Oct 08 18:31:53 crc kubenswrapper[4859]: I1008 18:31:53.518560 4859 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rvnjq\" (UniqueName: \"kubernetes.io/projected/834c27d2-782d-43a6-a955-e0c8b99d93a3-kube-api-access-rvnjq\") pod \"dnsmasq-dns-644597f84c-x8r6m\" (UID: \"834c27d2-782d-43a6-a955-e0c8b99d93a3\") " pod="openstack/dnsmasq-dns-644597f84c-x8r6m" Oct 08 18:31:53 crc kubenswrapper[4859]: I1008 18:31:53.518604 4859 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/834c27d2-782d-43a6-a955-e0c8b99d93a3-config\") pod \"dnsmasq-dns-644597f84c-x8r6m\" (UID: \"834c27d2-782d-43a6-a955-e0c8b99d93a3\") " pod="openstack/dnsmasq-dns-644597f84c-x8r6m" Oct 08 18:31:53 crc kubenswrapper[4859]: I1008 18:31:53.520208 4859 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/834c27d2-782d-43a6-a955-e0c8b99d93a3-config\") pod \"dnsmasq-dns-644597f84c-x8r6m\" (UID: \"834c27d2-782d-43a6-a955-e0c8b99d93a3\") " pod="openstack/dnsmasq-dns-644597f84c-x8r6m" Oct 08 18:31:53 crc kubenswrapper[4859]: I1008 18:31:53.522032 4859 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/834c27d2-782d-43a6-a955-e0c8b99d93a3-dns-svc\") pod \"dnsmasq-dns-644597f84c-x8r6m\" (UID: \"834c27d2-782d-43a6-a955-e0c8b99d93a3\") " pod="openstack/dnsmasq-dns-644597f84c-x8r6m" Oct 08 18:31:53 crc kubenswrapper[4859]: I1008 18:31:53.548727 4859 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rvnjq\" (UniqueName: \"kubernetes.io/projected/834c27d2-782d-43a6-a955-e0c8b99d93a3-kube-api-access-rvnjq\") pod \"dnsmasq-dns-644597f84c-x8r6m\" (UID: \"834c27d2-782d-43a6-a955-e0c8b99d93a3\") " pod="openstack/dnsmasq-dns-644597f84c-x8r6m" Oct 08 18:31:53 crc kubenswrapper[4859]: I1008 18:31:53.685207 4859 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-7bfcb9d745-ntml8"] Oct 08 18:31:53 crc kubenswrapper[4859]: I1008 18:31:53.716382 4859 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-77597f887-q5jzd"] Oct 08 18:31:53 crc kubenswrapper[4859]: I1008 18:31:53.718728 4859 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-77597f887-q5jzd" Oct 08 18:31:53 crc kubenswrapper[4859]: I1008 18:31:53.731493 4859 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-77597f887-q5jzd"] Oct 08 18:31:53 crc kubenswrapper[4859]: I1008 18:31:53.783775 4859 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-644597f84c-x8r6m" Oct 08 18:31:53 crc kubenswrapper[4859]: I1008 18:31:53.830289 4859 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/e78ba437-04e5-4b40-af53-8b93921fb30f-config\") pod \"dnsmasq-dns-77597f887-q5jzd\" (UID: \"e78ba437-04e5-4b40-af53-8b93921fb30f\") " pod="openstack/dnsmasq-dns-77597f887-q5jzd" Oct 08 18:31:53 crc kubenswrapper[4859]: I1008 18:31:53.830376 4859 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-98hh5\" (UniqueName: \"kubernetes.io/projected/e78ba437-04e5-4b40-af53-8b93921fb30f-kube-api-access-98hh5\") pod \"dnsmasq-dns-77597f887-q5jzd\" (UID: \"e78ba437-04e5-4b40-af53-8b93921fb30f\") " pod="openstack/dnsmasq-dns-77597f887-q5jzd" Oct 08 18:31:53 crc kubenswrapper[4859]: I1008 18:31:53.830464 4859 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/e78ba437-04e5-4b40-af53-8b93921fb30f-dns-svc\") pod \"dnsmasq-dns-77597f887-q5jzd\" (UID: \"e78ba437-04e5-4b40-af53-8b93921fb30f\") " pod="openstack/dnsmasq-dns-77597f887-q5jzd" Oct 08 18:31:53 crc kubenswrapper[4859]: I1008 18:31:53.931607 4859 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/e78ba437-04e5-4b40-af53-8b93921fb30f-dns-svc\") pod \"dnsmasq-dns-77597f887-q5jzd\" (UID: \"e78ba437-04e5-4b40-af53-8b93921fb30f\") " pod="openstack/dnsmasq-dns-77597f887-q5jzd" Oct 08 18:31:53 crc kubenswrapper[4859]: I1008 18:31:53.931668 4859 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/e78ba437-04e5-4b40-af53-8b93921fb30f-config\") pod \"dnsmasq-dns-77597f887-q5jzd\" (UID: \"e78ba437-04e5-4b40-af53-8b93921fb30f\") " pod="openstack/dnsmasq-dns-77597f887-q5jzd" Oct 08 18:31:53 crc kubenswrapper[4859]: I1008 18:31:53.931742 4859 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-98hh5\" (UniqueName: \"kubernetes.io/projected/e78ba437-04e5-4b40-af53-8b93921fb30f-kube-api-access-98hh5\") pod \"dnsmasq-dns-77597f887-q5jzd\" (UID: \"e78ba437-04e5-4b40-af53-8b93921fb30f\") " pod="openstack/dnsmasq-dns-77597f887-q5jzd" Oct 08 18:31:53 crc kubenswrapper[4859]: I1008 18:31:53.932661 4859 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/e78ba437-04e5-4b40-af53-8b93921fb30f-config\") pod \"dnsmasq-dns-77597f887-q5jzd\" (UID: \"e78ba437-04e5-4b40-af53-8b93921fb30f\") " pod="openstack/dnsmasq-dns-77597f887-q5jzd" Oct 08 18:31:53 crc kubenswrapper[4859]: I1008 18:31:53.932657 4859 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/e78ba437-04e5-4b40-af53-8b93921fb30f-dns-svc\") pod \"dnsmasq-dns-77597f887-q5jzd\" (UID: \"e78ba437-04e5-4b40-af53-8b93921fb30f\") " pod="openstack/dnsmasq-dns-77597f887-q5jzd" Oct 08 18:31:53 crc kubenswrapper[4859]: I1008 18:31:53.955342 4859 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-98hh5\" (UniqueName: \"kubernetes.io/projected/e78ba437-04e5-4b40-af53-8b93921fb30f-kube-api-access-98hh5\") pod \"dnsmasq-dns-77597f887-q5jzd\" (UID: \"e78ba437-04e5-4b40-af53-8b93921fb30f\") " pod="openstack/dnsmasq-dns-77597f887-q5jzd" Oct 08 18:31:54 crc kubenswrapper[4859]: I1008 18:31:54.047595 4859 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-77597f887-q5jzd" Oct 08 18:31:54 crc kubenswrapper[4859]: I1008 18:31:54.373746 4859 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-644597f84c-x8r6m"] Oct 08 18:31:54 crc kubenswrapper[4859]: W1008 18:31:54.383040 4859 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod834c27d2_782d_43a6_a955_e0c8b99d93a3.slice/crio-12753bfe1617bad562af9bd2ee3c0adc6272a79a9598b983d12b4d78a5312a53 WatchSource:0}: Error finding container 12753bfe1617bad562af9bd2ee3c0adc6272a79a9598b983d12b4d78a5312a53: Status 404 returned error can't find the container with id 12753bfe1617bad562af9bd2ee3c0adc6272a79a9598b983d12b4d78a5312a53 Oct 08 18:31:54 crc kubenswrapper[4859]: I1008 18:31:54.569923 4859 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/rabbitmq-server-0"] Oct 08 18:31:54 crc kubenswrapper[4859]: I1008 18:31:54.571961 4859 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-server-0" Oct 08 18:31:54 crc kubenswrapper[4859]: I1008 18:31:54.585081 4859 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-server-conf" Oct 08 18:31:54 crc kubenswrapper[4859]: I1008 18:31:54.585156 4859 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-rabbitmq-svc" Oct 08 18:31:54 crc kubenswrapper[4859]: I1008 18:31:54.585228 4859 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-config-data" Oct 08 18:31:54 crc kubenswrapper[4859]: I1008 18:31:54.585081 4859 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-plugins-conf" Oct 08 18:31:54 crc kubenswrapper[4859]: I1008 18:31:54.585096 4859 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-default-user" Oct 08 18:31:54 crc kubenswrapper[4859]: I1008 18:31:54.585476 4859 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-server-dockercfg-rdvqq" Oct 08 18:31:54 crc kubenswrapper[4859]: I1008 18:31:54.586780 4859 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/rabbitmq-server-0"] Oct 08 18:31:54 crc kubenswrapper[4859]: I1008 18:31:54.588952 4859 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-erlang-cookie" Oct 08 18:31:54 crc kubenswrapper[4859]: I1008 18:31:54.605479 4859 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-77597f887-q5jzd"] Oct 08 18:31:54 crc kubenswrapper[4859]: W1008 18:31:54.629763 4859 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pode78ba437_04e5_4b40_af53_8b93921fb30f.slice/crio-9bf5c780ffaa2fe787596d6cce50246d54103e0ad7ee9dafe099a8fcfdc214ff WatchSource:0}: Error finding container 9bf5c780ffaa2fe787596d6cce50246d54103e0ad7ee9dafe099a8fcfdc214ff: Status 404 returned error can't find the container with id 9bf5c780ffaa2fe787596d6cce50246d54103e0ad7ee9dafe099a8fcfdc214ff Oct 08 18:31:54 crc kubenswrapper[4859]: I1008 18:31:54.643355 4859 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/ba5a78be-e7b3-401a-bdb9-ba31071d7075-server-conf\") pod \"rabbitmq-server-0\" (UID: \"ba5a78be-e7b3-401a-bdb9-ba31071d7075\") " pod="openstack/rabbitmq-server-0" Oct 08 18:31:54 crc kubenswrapper[4859]: I1008 18:31:54.643406 4859 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/ba5a78be-e7b3-401a-bdb9-ba31071d7075-rabbitmq-plugins\") pod \"rabbitmq-server-0\" (UID: \"ba5a78be-e7b3-401a-bdb9-ba31071d7075\") " pod="openstack/rabbitmq-server-0" Oct 08 18:31:54 crc kubenswrapper[4859]: I1008 18:31:54.643611 4859 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/ba5a78be-e7b3-401a-bdb9-ba31071d7075-plugins-conf\") pod \"rabbitmq-server-0\" (UID: \"ba5a78be-e7b3-401a-bdb9-ba31071d7075\") " pod="openstack/rabbitmq-server-0" Oct 08 18:31:54 crc kubenswrapper[4859]: I1008 18:31:54.643711 4859 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/ba5a78be-e7b3-401a-bdb9-ba31071d7075-pod-info\") pod \"rabbitmq-server-0\" (UID: \"ba5a78be-e7b3-401a-bdb9-ba31071d7075\") " pod="openstack/rabbitmq-server-0" Oct 08 18:31:54 crc kubenswrapper[4859]: I1008 18:31:54.643736 4859 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage08-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage08-crc\") pod \"rabbitmq-server-0\" (UID: \"ba5a78be-e7b3-401a-bdb9-ba31071d7075\") " pod="openstack/rabbitmq-server-0" Oct 08 18:31:54 crc kubenswrapper[4859]: I1008 18:31:54.643813 4859 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/ba5a78be-e7b3-401a-bdb9-ba31071d7075-rabbitmq-confd\") pod \"rabbitmq-server-0\" (UID: \"ba5a78be-e7b3-401a-bdb9-ba31071d7075\") " pod="openstack/rabbitmq-server-0" Oct 08 18:31:54 crc kubenswrapper[4859]: I1008 18:31:54.643896 4859 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/ba5a78be-e7b3-401a-bdb9-ba31071d7075-config-data\") pod \"rabbitmq-server-0\" (UID: \"ba5a78be-e7b3-401a-bdb9-ba31071d7075\") " pod="openstack/rabbitmq-server-0" Oct 08 18:31:54 crc kubenswrapper[4859]: I1008 18:31:54.643936 4859 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/ba5a78be-e7b3-401a-bdb9-ba31071d7075-rabbitmq-tls\") pod \"rabbitmq-server-0\" (UID: \"ba5a78be-e7b3-401a-bdb9-ba31071d7075\") " pod="openstack/rabbitmq-server-0" Oct 08 18:31:54 crc kubenswrapper[4859]: I1008 18:31:54.644034 4859 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/ba5a78be-e7b3-401a-bdb9-ba31071d7075-rabbitmq-erlang-cookie\") pod \"rabbitmq-server-0\" (UID: \"ba5a78be-e7b3-401a-bdb9-ba31071d7075\") " pod="openstack/rabbitmq-server-0" Oct 08 18:31:54 crc kubenswrapper[4859]: I1008 18:31:54.644057 4859 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-sbb5z\" (UniqueName: \"kubernetes.io/projected/ba5a78be-e7b3-401a-bdb9-ba31071d7075-kube-api-access-sbb5z\") pod \"rabbitmq-server-0\" (UID: \"ba5a78be-e7b3-401a-bdb9-ba31071d7075\") " pod="openstack/rabbitmq-server-0" Oct 08 18:31:54 crc kubenswrapper[4859]: I1008 18:31:54.644085 4859 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/ba5a78be-e7b3-401a-bdb9-ba31071d7075-erlang-cookie-secret\") pod \"rabbitmq-server-0\" (UID: \"ba5a78be-e7b3-401a-bdb9-ba31071d7075\") " pod="openstack/rabbitmq-server-0" Oct 08 18:31:54 crc kubenswrapper[4859]: I1008 18:31:54.746406 4859 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/ba5a78be-e7b3-401a-bdb9-ba31071d7075-plugins-conf\") pod \"rabbitmq-server-0\" (UID: \"ba5a78be-e7b3-401a-bdb9-ba31071d7075\") " pod="openstack/rabbitmq-server-0" Oct 08 18:31:54 crc kubenswrapper[4859]: I1008 18:31:54.746523 4859 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/ba5a78be-e7b3-401a-bdb9-ba31071d7075-pod-info\") pod \"rabbitmq-server-0\" (UID: \"ba5a78be-e7b3-401a-bdb9-ba31071d7075\") " pod="openstack/rabbitmq-server-0" Oct 08 18:31:54 crc kubenswrapper[4859]: I1008 18:31:54.746558 4859 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage08-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage08-crc\") pod \"rabbitmq-server-0\" (UID: \"ba5a78be-e7b3-401a-bdb9-ba31071d7075\") " pod="openstack/rabbitmq-server-0" Oct 08 18:31:54 crc kubenswrapper[4859]: I1008 18:31:54.746626 4859 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/ba5a78be-e7b3-401a-bdb9-ba31071d7075-rabbitmq-confd\") pod \"rabbitmq-server-0\" (UID: \"ba5a78be-e7b3-401a-bdb9-ba31071d7075\") " pod="openstack/rabbitmq-server-0" Oct 08 18:31:54 crc kubenswrapper[4859]: I1008 18:31:54.746698 4859 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/ba5a78be-e7b3-401a-bdb9-ba31071d7075-config-data\") pod \"rabbitmq-server-0\" (UID: \"ba5a78be-e7b3-401a-bdb9-ba31071d7075\") " pod="openstack/rabbitmq-server-0" Oct 08 18:31:54 crc kubenswrapper[4859]: I1008 18:31:54.746729 4859 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/ba5a78be-e7b3-401a-bdb9-ba31071d7075-rabbitmq-tls\") pod \"rabbitmq-server-0\" (UID: \"ba5a78be-e7b3-401a-bdb9-ba31071d7075\") " pod="openstack/rabbitmq-server-0" Oct 08 18:31:54 crc kubenswrapper[4859]: I1008 18:31:54.746801 4859 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/ba5a78be-e7b3-401a-bdb9-ba31071d7075-rabbitmq-erlang-cookie\") pod \"rabbitmq-server-0\" (UID: \"ba5a78be-e7b3-401a-bdb9-ba31071d7075\") " pod="openstack/rabbitmq-server-0" Oct 08 18:31:54 crc kubenswrapper[4859]: I1008 18:31:54.746833 4859 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-sbb5z\" (UniqueName: \"kubernetes.io/projected/ba5a78be-e7b3-401a-bdb9-ba31071d7075-kube-api-access-sbb5z\") pod \"rabbitmq-server-0\" (UID: \"ba5a78be-e7b3-401a-bdb9-ba31071d7075\") " pod="openstack/rabbitmq-server-0" Oct 08 18:31:54 crc kubenswrapper[4859]: I1008 18:31:54.746862 4859 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/ba5a78be-e7b3-401a-bdb9-ba31071d7075-erlang-cookie-secret\") pod \"rabbitmq-server-0\" (UID: \"ba5a78be-e7b3-401a-bdb9-ba31071d7075\") " pod="openstack/rabbitmq-server-0" Oct 08 18:31:54 crc kubenswrapper[4859]: I1008 18:31:54.746899 4859 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/ba5a78be-e7b3-401a-bdb9-ba31071d7075-server-conf\") pod \"rabbitmq-server-0\" (UID: \"ba5a78be-e7b3-401a-bdb9-ba31071d7075\") " pod="openstack/rabbitmq-server-0" Oct 08 18:31:54 crc kubenswrapper[4859]: I1008 18:31:54.746926 4859 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/ba5a78be-e7b3-401a-bdb9-ba31071d7075-rabbitmq-plugins\") pod \"rabbitmq-server-0\" (UID: \"ba5a78be-e7b3-401a-bdb9-ba31071d7075\") " pod="openstack/rabbitmq-server-0" Oct 08 18:31:54 crc kubenswrapper[4859]: I1008 18:31:54.747279 4859 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage08-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage08-crc\") pod \"rabbitmq-server-0\" (UID: \"ba5a78be-e7b3-401a-bdb9-ba31071d7075\") device mount path \"/mnt/openstack/pv08\"" pod="openstack/rabbitmq-server-0" Oct 08 18:31:54 crc kubenswrapper[4859]: I1008 18:31:54.748306 4859 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/ba5a78be-e7b3-401a-bdb9-ba31071d7075-plugins-conf\") pod \"rabbitmq-server-0\" (UID: \"ba5a78be-e7b3-401a-bdb9-ba31071d7075\") " pod="openstack/rabbitmq-server-0" Oct 08 18:31:54 crc kubenswrapper[4859]: I1008 18:31:54.748385 4859 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/ba5a78be-e7b3-401a-bdb9-ba31071d7075-config-data\") pod \"rabbitmq-server-0\" (UID: \"ba5a78be-e7b3-401a-bdb9-ba31071d7075\") " pod="openstack/rabbitmq-server-0" Oct 08 18:31:54 crc kubenswrapper[4859]: I1008 18:31:54.748553 4859 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/ba5a78be-e7b3-401a-bdb9-ba31071d7075-rabbitmq-erlang-cookie\") pod \"rabbitmq-server-0\" (UID: \"ba5a78be-e7b3-401a-bdb9-ba31071d7075\") " pod="openstack/rabbitmq-server-0" Oct 08 18:31:54 crc kubenswrapper[4859]: I1008 18:31:54.748824 4859 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/ba5a78be-e7b3-401a-bdb9-ba31071d7075-server-conf\") pod \"rabbitmq-server-0\" (UID: \"ba5a78be-e7b3-401a-bdb9-ba31071d7075\") " pod="openstack/rabbitmq-server-0" Oct 08 18:31:54 crc kubenswrapper[4859]: I1008 18:31:54.750651 4859 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/ba5a78be-e7b3-401a-bdb9-ba31071d7075-rabbitmq-plugins\") pod \"rabbitmq-server-0\" (UID: \"ba5a78be-e7b3-401a-bdb9-ba31071d7075\") " pod="openstack/rabbitmq-server-0" Oct 08 18:31:54 crc kubenswrapper[4859]: I1008 18:31:54.756463 4859 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/ba5a78be-e7b3-401a-bdb9-ba31071d7075-erlang-cookie-secret\") pod \"rabbitmq-server-0\" (UID: \"ba5a78be-e7b3-401a-bdb9-ba31071d7075\") " pod="openstack/rabbitmq-server-0" Oct 08 18:31:54 crc kubenswrapper[4859]: I1008 18:31:54.756998 4859 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/ba5a78be-e7b3-401a-bdb9-ba31071d7075-rabbitmq-tls\") pod \"rabbitmq-server-0\" (UID: \"ba5a78be-e7b3-401a-bdb9-ba31071d7075\") " pod="openstack/rabbitmq-server-0" Oct 08 18:31:54 crc kubenswrapper[4859]: I1008 18:31:54.757272 4859 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/ba5a78be-e7b3-401a-bdb9-ba31071d7075-rabbitmq-confd\") pod \"rabbitmq-server-0\" (UID: \"ba5a78be-e7b3-401a-bdb9-ba31071d7075\") " pod="openstack/rabbitmq-server-0" Oct 08 18:31:54 crc kubenswrapper[4859]: I1008 18:31:54.765003 4859 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/ba5a78be-e7b3-401a-bdb9-ba31071d7075-pod-info\") pod \"rabbitmq-server-0\" (UID: \"ba5a78be-e7b3-401a-bdb9-ba31071d7075\") " pod="openstack/rabbitmq-server-0" Oct 08 18:31:54 crc kubenswrapper[4859]: I1008 18:31:54.768497 4859 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-sbb5z\" (UniqueName: \"kubernetes.io/projected/ba5a78be-e7b3-401a-bdb9-ba31071d7075-kube-api-access-sbb5z\") pod \"rabbitmq-server-0\" (UID: \"ba5a78be-e7b3-401a-bdb9-ba31071d7075\") " pod="openstack/rabbitmq-server-0" Oct 08 18:31:54 crc kubenswrapper[4859]: I1008 18:31:54.771178 4859 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage08-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage08-crc\") pod \"rabbitmq-server-0\" (UID: \"ba5a78be-e7b3-401a-bdb9-ba31071d7075\") " pod="openstack/rabbitmq-server-0" Oct 08 18:31:54 crc kubenswrapper[4859]: I1008 18:31:54.846455 4859 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/rabbitmq-cell1-server-0"] Oct 08 18:31:54 crc kubenswrapper[4859]: I1008 18:31:54.848240 4859 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-cell1-server-0" Oct 08 18:31:54 crc kubenswrapper[4859]: I1008 18:31:54.850931 4859 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-cell1-erlang-cookie" Oct 08 18:31:54 crc kubenswrapper[4859]: I1008 18:31:54.856816 4859 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-rabbitmq-cell1-svc" Oct 08 18:31:54 crc kubenswrapper[4859]: I1008 18:31:54.857266 4859 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-cell1-default-user" Oct 08 18:31:54 crc kubenswrapper[4859]: I1008 18:31:54.858128 4859 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-cell1-config-data" Oct 08 18:31:54 crc kubenswrapper[4859]: I1008 18:31:54.859025 4859 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-cell1-plugins-conf" Oct 08 18:31:54 crc kubenswrapper[4859]: I1008 18:31:54.859231 4859 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-cell1-server-conf" Oct 08 18:31:54 crc kubenswrapper[4859]: I1008 18:31:54.863646 4859 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/rabbitmq-cell1-server-0"] Oct 08 18:31:54 crc kubenswrapper[4859]: I1008 18:31:54.870222 4859 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-cell1-server-dockercfg-rg8nn" Oct 08 18:31:54 crc kubenswrapper[4859]: I1008 18:31:54.903916 4859 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-server-0" Oct 08 18:31:54 crc kubenswrapper[4859]: I1008 18:31:54.949871 4859 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/3075d5c2-3e74-4a4c-829f-248e87f45f3b-pod-info\") pod \"rabbitmq-cell1-server-0\" (UID: \"3075d5c2-3e74-4a4c-829f-248e87f45f3b\") " pod="openstack/rabbitmq-cell1-server-0" Oct 08 18:31:54 crc kubenswrapper[4859]: I1008 18:31:54.949923 4859 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/3075d5c2-3e74-4a4c-829f-248e87f45f3b-plugins-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"3075d5c2-3e74-4a4c-829f-248e87f45f3b\") " pod="openstack/rabbitmq-cell1-server-0" Oct 08 18:31:54 crc kubenswrapper[4859]: I1008 18:31:54.949951 4859 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/3075d5c2-3e74-4a4c-829f-248e87f45f3b-rabbitmq-erlang-cookie\") pod \"rabbitmq-cell1-server-0\" (UID: \"3075d5c2-3e74-4a4c-829f-248e87f45f3b\") " pod="openstack/rabbitmq-cell1-server-0" Oct 08 18:31:54 crc kubenswrapper[4859]: I1008 18:31:54.949972 4859 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/3075d5c2-3e74-4a4c-829f-248e87f45f3b-rabbitmq-confd\") pod \"rabbitmq-cell1-server-0\" (UID: \"3075d5c2-3e74-4a4c-829f-248e87f45f3b\") " pod="openstack/rabbitmq-cell1-server-0" Oct 08 18:31:54 crc kubenswrapper[4859]: I1008 18:31:54.950140 4859 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/3075d5c2-3e74-4a4c-829f-248e87f45f3b-rabbitmq-plugins\") pod \"rabbitmq-cell1-server-0\" (UID: \"3075d5c2-3e74-4a4c-829f-248e87f45f3b\") " pod="openstack/rabbitmq-cell1-server-0" Oct 08 18:31:54 crc kubenswrapper[4859]: I1008 18:31:54.950229 4859 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage09-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage09-crc\") pod \"rabbitmq-cell1-server-0\" (UID: \"3075d5c2-3e74-4a4c-829f-248e87f45f3b\") " pod="openstack/rabbitmq-cell1-server-0" Oct 08 18:31:54 crc kubenswrapper[4859]: I1008 18:31:54.950266 4859 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-xvbcd\" (UniqueName: \"kubernetes.io/projected/3075d5c2-3e74-4a4c-829f-248e87f45f3b-kube-api-access-xvbcd\") pod \"rabbitmq-cell1-server-0\" (UID: \"3075d5c2-3e74-4a4c-829f-248e87f45f3b\") " pod="openstack/rabbitmq-cell1-server-0" Oct 08 18:31:54 crc kubenswrapper[4859]: I1008 18:31:54.950347 4859 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/3075d5c2-3e74-4a4c-829f-248e87f45f3b-rabbitmq-tls\") pod \"rabbitmq-cell1-server-0\" (UID: \"3075d5c2-3e74-4a4c-829f-248e87f45f3b\") " pod="openstack/rabbitmq-cell1-server-0" Oct 08 18:31:54 crc kubenswrapper[4859]: I1008 18:31:54.950374 4859 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/3075d5c2-3e74-4a4c-829f-248e87f45f3b-erlang-cookie-secret\") pod \"rabbitmq-cell1-server-0\" (UID: \"3075d5c2-3e74-4a4c-829f-248e87f45f3b\") " pod="openstack/rabbitmq-cell1-server-0" Oct 08 18:31:54 crc kubenswrapper[4859]: I1008 18:31:54.950445 4859 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/3075d5c2-3e74-4a4c-829f-248e87f45f3b-config-data\") pod \"rabbitmq-cell1-server-0\" (UID: \"3075d5c2-3e74-4a4c-829f-248e87f45f3b\") " pod="openstack/rabbitmq-cell1-server-0" Oct 08 18:31:54 crc kubenswrapper[4859]: I1008 18:31:54.950488 4859 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/3075d5c2-3e74-4a4c-829f-248e87f45f3b-server-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"3075d5c2-3e74-4a4c-829f-248e87f45f3b\") " pod="openstack/rabbitmq-cell1-server-0" Oct 08 18:31:55 crc kubenswrapper[4859]: I1008 18:31:55.052858 4859 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/3075d5c2-3e74-4a4c-829f-248e87f45f3b-config-data\") pod \"rabbitmq-cell1-server-0\" (UID: \"3075d5c2-3e74-4a4c-829f-248e87f45f3b\") " pod="openstack/rabbitmq-cell1-server-0" Oct 08 18:31:55 crc kubenswrapper[4859]: I1008 18:31:55.052916 4859 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/3075d5c2-3e74-4a4c-829f-248e87f45f3b-server-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"3075d5c2-3e74-4a4c-829f-248e87f45f3b\") " pod="openstack/rabbitmq-cell1-server-0" Oct 08 18:31:55 crc kubenswrapper[4859]: I1008 18:31:55.052958 4859 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/3075d5c2-3e74-4a4c-829f-248e87f45f3b-pod-info\") pod \"rabbitmq-cell1-server-0\" (UID: \"3075d5c2-3e74-4a4c-829f-248e87f45f3b\") " pod="openstack/rabbitmq-cell1-server-0" Oct 08 18:31:55 crc kubenswrapper[4859]: I1008 18:31:55.052985 4859 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/3075d5c2-3e74-4a4c-829f-248e87f45f3b-plugins-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"3075d5c2-3e74-4a4c-829f-248e87f45f3b\") " pod="openstack/rabbitmq-cell1-server-0" Oct 08 18:31:55 crc kubenswrapper[4859]: I1008 18:31:55.053153 4859 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/3075d5c2-3e74-4a4c-829f-248e87f45f3b-rabbitmq-erlang-cookie\") pod \"rabbitmq-cell1-server-0\" (UID: \"3075d5c2-3e74-4a4c-829f-248e87f45f3b\") " pod="openstack/rabbitmq-cell1-server-0" Oct 08 18:31:55 crc kubenswrapper[4859]: I1008 18:31:55.053270 4859 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/3075d5c2-3e74-4a4c-829f-248e87f45f3b-rabbitmq-confd\") pod \"rabbitmq-cell1-server-0\" (UID: \"3075d5c2-3e74-4a4c-829f-248e87f45f3b\") " pod="openstack/rabbitmq-cell1-server-0" Oct 08 18:31:55 crc kubenswrapper[4859]: I1008 18:31:55.053396 4859 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/3075d5c2-3e74-4a4c-829f-248e87f45f3b-rabbitmq-plugins\") pod \"rabbitmq-cell1-server-0\" (UID: \"3075d5c2-3e74-4a4c-829f-248e87f45f3b\") " pod="openstack/rabbitmq-cell1-server-0" Oct 08 18:31:55 crc kubenswrapper[4859]: I1008 18:31:55.053444 4859 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage09-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage09-crc\") pod \"rabbitmq-cell1-server-0\" (UID: \"3075d5c2-3e74-4a4c-829f-248e87f45f3b\") " pod="openstack/rabbitmq-cell1-server-0" Oct 08 18:31:55 crc kubenswrapper[4859]: I1008 18:31:55.053516 4859 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-xvbcd\" (UniqueName: \"kubernetes.io/projected/3075d5c2-3e74-4a4c-829f-248e87f45f3b-kube-api-access-xvbcd\") pod \"rabbitmq-cell1-server-0\" (UID: \"3075d5c2-3e74-4a4c-829f-248e87f45f3b\") " pod="openstack/rabbitmq-cell1-server-0" Oct 08 18:31:55 crc kubenswrapper[4859]: I1008 18:31:55.053817 4859 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/3075d5c2-3e74-4a4c-829f-248e87f45f3b-rabbitmq-tls\") pod \"rabbitmq-cell1-server-0\" (UID: \"3075d5c2-3e74-4a4c-829f-248e87f45f3b\") " pod="openstack/rabbitmq-cell1-server-0" Oct 08 18:31:55 crc kubenswrapper[4859]: I1008 18:31:55.053851 4859 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/3075d5c2-3e74-4a4c-829f-248e87f45f3b-erlang-cookie-secret\") pod \"rabbitmq-cell1-server-0\" (UID: \"3075d5c2-3e74-4a4c-829f-248e87f45f3b\") " pod="openstack/rabbitmq-cell1-server-0" Oct 08 18:31:55 crc kubenswrapper[4859]: I1008 18:31:55.053927 4859 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/3075d5c2-3e74-4a4c-829f-248e87f45f3b-rabbitmq-erlang-cookie\") pod \"rabbitmq-cell1-server-0\" (UID: \"3075d5c2-3e74-4a4c-829f-248e87f45f3b\") " pod="openstack/rabbitmq-cell1-server-0" Oct 08 18:31:55 crc kubenswrapper[4859]: I1008 18:31:55.054535 4859 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/3075d5c2-3e74-4a4c-829f-248e87f45f3b-rabbitmq-plugins\") pod \"rabbitmq-cell1-server-0\" (UID: \"3075d5c2-3e74-4a4c-829f-248e87f45f3b\") " pod="openstack/rabbitmq-cell1-server-0" Oct 08 18:31:55 crc kubenswrapper[4859]: I1008 18:31:55.054563 4859 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/3075d5c2-3e74-4a4c-829f-248e87f45f3b-plugins-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"3075d5c2-3e74-4a4c-829f-248e87f45f3b\") " pod="openstack/rabbitmq-cell1-server-0" Oct 08 18:31:55 crc kubenswrapper[4859]: I1008 18:31:55.054719 4859 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/3075d5c2-3e74-4a4c-829f-248e87f45f3b-server-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"3075d5c2-3e74-4a4c-829f-248e87f45f3b\") " pod="openstack/rabbitmq-cell1-server-0" Oct 08 18:31:55 crc kubenswrapper[4859]: I1008 18:31:55.054794 4859 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage09-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage09-crc\") pod \"rabbitmq-cell1-server-0\" (UID: \"3075d5c2-3e74-4a4c-829f-248e87f45f3b\") device mount path \"/mnt/openstack/pv09\"" pod="openstack/rabbitmq-cell1-server-0" Oct 08 18:31:55 crc kubenswrapper[4859]: I1008 18:31:55.055342 4859 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/3075d5c2-3e74-4a4c-829f-248e87f45f3b-config-data\") pod \"rabbitmq-cell1-server-0\" (UID: \"3075d5c2-3e74-4a4c-829f-248e87f45f3b\") " pod="openstack/rabbitmq-cell1-server-0" Oct 08 18:31:55 crc kubenswrapper[4859]: I1008 18:31:55.057388 4859 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/3075d5c2-3e74-4a4c-829f-248e87f45f3b-pod-info\") pod \"rabbitmq-cell1-server-0\" (UID: \"3075d5c2-3e74-4a4c-829f-248e87f45f3b\") " pod="openstack/rabbitmq-cell1-server-0" Oct 08 18:31:55 crc kubenswrapper[4859]: I1008 18:31:55.057529 4859 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/3075d5c2-3e74-4a4c-829f-248e87f45f3b-erlang-cookie-secret\") pod \"rabbitmq-cell1-server-0\" (UID: \"3075d5c2-3e74-4a4c-829f-248e87f45f3b\") " pod="openstack/rabbitmq-cell1-server-0" Oct 08 18:31:55 crc kubenswrapper[4859]: I1008 18:31:55.057929 4859 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/3075d5c2-3e74-4a4c-829f-248e87f45f3b-rabbitmq-tls\") pod \"rabbitmq-cell1-server-0\" (UID: \"3075d5c2-3e74-4a4c-829f-248e87f45f3b\") " pod="openstack/rabbitmq-cell1-server-0" Oct 08 18:31:55 crc kubenswrapper[4859]: I1008 18:31:55.064216 4859 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/3075d5c2-3e74-4a4c-829f-248e87f45f3b-rabbitmq-confd\") pod \"rabbitmq-cell1-server-0\" (UID: \"3075d5c2-3e74-4a4c-829f-248e87f45f3b\") " pod="openstack/rabbitmq-cell1-server-0" Oct 08 18:31:55 crc kubenswrapper[4859]: I1008 18:31:55.071659 4859 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-xvbcd\" (UniqueName: \"kubernetes.io/projected/3075d5c2-3e74-4a4c-829f-248e87f45f3b-kube-api-access-xvbcd\") pod \"rabbitmq-cell1-server-0\" (UID: \"3075d5c2-3e74-4a4c-829f-248e87f45f3b\") " pod="openstack/rabbitmq-cell1-server-0" Oct 08 18:31:55 crc kubenswrapper[4859]: I1008 18:31:55.076907 4859 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage09-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage09-crc\") pod \"rabbitmq-cell1-server-0\" (UID: \"3075d5c2-3e74-4a4c-829f-248e87f45f3b\") " pod="openstack/rabbitmq-cell1-server-0" Oct 08 18:31:55 crc kubenswrapper[4859]: I1008 18:31:55.188488 4859 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-cell1-server-0" Oct 08 18:31:55 crc kubenswrapper[4859]: I1008 18:31:55.201195 4859 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-644597f84c-x8r6m" event={"ID":"834c27d2-782d-43a6-a955-e0c8b99d93a3","Type":"ContainerStarted","Data":"12753bfe1617bad562af9bd2ee3c0adc6272a79a9598b983d12b4d78a5312a53"} Oct 08 18:31:55 crc kubenswrapper[4859]: I1008 18:31:55.202825 4859 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-77597f887-q5jzd" event={"ID":"e78ba437-04e5-4b40-af53-8b93921fb30f","Type":"ContainerStarted","Data":"9bf5c780ffaa2fe787596d6cce50246d54103e0ad7ee9dafe099a8fcfdc214ff"} Oct 08 18:31:57 crc kubenswrapper[4859]: I1008 18:31:57.225899 4859 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/openstack-galera-0"] Oct 08 18:31:57 crc kubenswrapper[4859]: I1008 18:31:57.227603 4859 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/openstack-galera-0" Oct 08 18:31:57 crc kubenswrapper[4859]: I1008 18:31:57.232932 4859 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"osp-secret" Oct 08 18:31:57 crc kubenswrapper[4859]: I1008 18:31:57.240787 4859 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-config-data" Oct 08 18:31:57 crc kubenswrapper[4859]: I1008 18:31:57.241106 4859 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"galera-openstack-dockercfg-bvbqr" Oct 08 18:31:57 crc kubenswrapper[4859]: I1008 18:31:57.243012 4859 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-galera-openstack-svc" Oct 08 18:31:57 crc kubenswrapper[4859]: I1008 18:31:57.243324 4859 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-scripts" Oct 08 18:31:57 crc kubenswrapper[4859]: I1008 18:31:57.245999 4859 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"combined-ca-bundle" Oct 08 18:31:57 crc kubenswrapper[4859]: I1008 18:31:57.249462 4859 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/openstack-galera-0"] Oct 08 18:31:57 crc kubenswrapper[4859]: I1008 18:31:57.293308 4859 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage07-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage07-crc\") pod \"openstack-galera-0\" (UID: \"f1828698-e403-4643-a9ab-aedf302aae70\") " pod="openstack/openstack-galera-0" Oct 08 18:31:57 crc kubenswrapper[4859]: I1008 18:31:57.293387 4859 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/f1828698-e403-4643-a9ab-aedf302aae70-operator-scripts\") pod \"openstack-galera-0\" (UID: \"f1828698-e403-4643-a9ab-aedf302aae70\") " pod="openstack/openstack-galera-0" Oct 08 18:31:57 crc kubenswrapper[4859]: I1008 18:31:57.293416 4859 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-p58hr\" (UniqueName: \"kubernetes.io/projected/f1828698-e403-4643-a9ab-aedf302aae70-kube-api-access-p58hr\") pod \"openstack-galera-0\" (UID: \"f1828698-e403-4643-a9ab-aedf302aae70\") " pod="openstack/openstack-galera-0" Oct 08 18:31:57 crc kubenswrapper[4859]: I1008 18:31:57.293443 4859 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/f1828698-e403-4643-a9ab-aedf302aae70-kolla-config\") pod \"openstack-galera-0\" (UID: \"f1828698-e403-4643-a9ab-aedf302aae70\") " pod="openstack/openstack-galera-0" Oct 08 18:31:57 crc kubenswrapper[4859]: I1008 18:31:57.293504 4859 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"galera-tls-certs\" (UniqueName: \"kubernetes.io/secret/f1828698-e403-4643-a9ab-aedf302aae70-galera-tls-certs\") pod \"openstack-galera-0\" (UID: \"f1828698-e403-4643-a9ab-aedf302aae70\") " pod="openstack/openstack-galera-0" Oct 08 18:31:57 crc kubenswrapper[4859]: I1008 18:31:57.293531 4859 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f1828698-e403-4643-a9ab-aedf302aae70-combined-ca-bundle\") pod \"openstack-galera-0\" (UID: \"f1828698-e403-4643-a9ab-aedf302aae70\") " pod="openstack/openstack-galera-0" Oct 08 18:31:57 crc kubenswrapper[4859]: I1008 18:31:57.293572 4859 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-default\" (UniqueName: \"kubernetes.io/configmap/f1828698-e403-4643-a9ab-aedf302aae70-config-data-default\") pod \"openstack-galera-0\" (UID: \"f1828698-e403-4643-a9ab-aedf302aae70\") " pod="openstack/openstack-galera-0" Oct 08 18:31:57 crc kubenswrapper[4859]: I1008 18:31:57.293595 4859 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-generated\" (UniqueName: \"kubernetes.io/empty-dir/f1828698-e403-4643-a9ab-aedf302aae70-config-data-generated\") pod \"openstack-galera-0\" (UID: \"f1828698-e403-4643-a9ab-aedf302aae70\") " pod="openstack/openstack-galera-0" Oct 08 18:31:57 crc kubenswrapper[4859]: I1008 18:31:57.293622 4859 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secrets\" (UniqueName: \"kubernetes.io/secret/f1828698-e403-4643-a9ab-aedf302aae70-secrets\") pod \"openstack-galera-0\" (UID: \"f1828698-e403-4643-a9ab-aedf302aae70\") " pod="openstack/openstack-galera-0" Oct 08 18:31:57 crc kubenswrapper[4859]: I1008 18:31:57.394844 4859 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/f1828698-e403-4643-a9ab-aedf302aae70-operator-scripts\") pod \"openstack-galera-0\" (UID: \"f1828698-e403-4643-a9ab-aedf302aae70\") " pod="openstack/openstack-galera-0" Oct 08 18:31:57 crc kubenswrapper[4859]: I1008 18:31:57.394914 4859 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-p58hr\" (UniqueName: \"kubernetes.io/projected/f1828698-e403-4643-a9ab-aedf302aae70-kube-api-access-p58hr\") pod \"openstack-galera-0\" (UID: \"f1828698-e403-4643-a9ab-aedf302aae70\") " pod="openstack/openstack-galera-0" Oct 08 18:31:57 crc kubenswrapper[4859]: I1008 18:31:57.394945 4859 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/f1828698-e403-4643-a9ab-aedf302aae70-kolla-config\") pod \"openstack-galera-0\" (UID: \"f1828698-e403-4643-a9ab-aedf302aae70\") " pod="openstack/openstack-galera-0" Oct 08 18:31:57 crc kubenswrapper[4859]: I1008 18:31:57.395009 4859 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"galera-tls-certs\" (UniqueName: \"kubernetes.io/secret/f1828698-e403-4643-a9ab-aedf302aae70-galera-tls-certs\") pod \"openstack-galera-0\" (UID: \"f1828698-e403-4643-a9ab-aedf302aae70\") " pod="openstack/openstack-galera-0" Oct 08 18:31:57 crc kubenswrapper[4859]: I1008 18:31:57.395037 4859 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f1828698-e403-4643-a9ab-aedf302aae70-combined-ca-bundle\") pod \"openstack-galera-0\" (UID: \"f1828698-e403-4643-a9ab-aedf302aae70\") " pod="openstack/openstack-galera-0" Oct 08 18:31:57 crc kubenswrapper[4859]: I1008 18:31:57.395080 4859 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-default\" (UniqueName: \"kubernetes.io/configmap/f1828698-e403-4643-a9ab-aedf302aae70-config-data-default\") pod \"openstack-galera-0\" (UID: \"f1828698-e403-4643-a9ab-aedf302aae70\") " pod="openstack/openstack-galera-0" Oct 08 18:31:57 crc kubenswrapper[4859]: I1008 18:31:57.395104 4859 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-generated\" (UniqueName: \"kubernetes.io/empty-dir/f1828698-e403-4643-a9ab-aedf302aae70-config-data-generated\") pod \"openstack-galera-0\" (UID: \"f1828698-e403-4643-a9ab-aedf302aae70\") " pod="openstack/openstack-galera-0" Oct 08 18:31:57 crc kubenswrapper[4859]: I1008 18:31:57.395126 4859 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"secrets\" (UniqueName: \"kubernetes.io/secret/f1828698-e403-4643-a9ab-aedf302aae70-secrets\") pod \"openstack-galera-0\" (UID: \"f1828698-e403-4643-a9ab-aedf302aae70\") " pod="openstack/openstack-galera-0" Oct 08 18:31:57 crc kubenswrapper[4859]: I1008 18:31:57.395173 4859 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage07-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage07-crc\") pod \"openstack-galera-0\" (UID: \"f1828698-e403-4643-a9ab-aedf302aae70\") " pod="openstack/openstack-galera-0" Oct 08 18:31:57 crc kubenswrapper[4859]: I1008 18:31:57.395469 4859 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage07-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage07-crc\") pod \"openstack-galera-0\" (UID: \"f1828698-e403-4643-a9ab-aedf302aae70\") device mount path \"/mnt/openstack/pv07\"" pod="openstack/openstack-galera-0" Oct 08 18:31:57 crc kubenswrapper[4859]: I1008 18:31:57.395814 4859 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-generated\" (UniqueName: \"kubernetes.io/empty-dir/f1828698-e403-4643-a9ab-aedf302aae70-config-data-generated\") pod \"openstack-galera-0\" (UID: \"f1828698-e403-4643-a9ab-aedf302aae70\") " pod="openstack/openstack-galera-0" Oct 08 18:31:57 crc kubenswrapper[4859]: I1008 18:31:57.396158 4859 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/f1828698-e403-4643-a9ab-aedf302aae70-kolla-config\") pod \"openstack-galera-0\" (UID: \"f1828698-e403-4643-a9ab-aedf302aae70\") " pod="openstack/openstack-galera-0" Oct 08 18:31:57 crc kubenswrapper[4859]: I1008 18:31:57.396405 4859 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/f1828698-e403-4643-a9ab-aedf302aae70-operator-scripts\") pod \"openstack-galera-0\" (UID: \"f1828698-e403-4643-a9ab-aedf302aae70\") " pod="openstack/openstack-galera-0" Oct 08 18:31:57 crc kubenswrapper[4859]: I1008 18:31:57.397309 4859 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-default\" (UniqueName: \"kubernetes.io/configmap/f1828698-e403-4643-a9ab-aedf302aae70-config-data-default\") pod \"openstack-galera-0\" (UID: \"f1828698-e403-4643-a9ab-aedf302aae70\") " pod="openstack/openstack-galera-0" Oct 08 18:31:57 crc kubenswrapper[4859]: I1008 18:31:57.401098 4859 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f1828698-e403-4643-a9ab-aedf302aae70-combined-ca-bundle\") pod \"openstack-galera-0\" (UID: \"f1828698-e403-4643-a9ab-aedf302aae70\") " pod="openstack/openstack-galera-0" Oct 08 18:31:57 crc kubenswrapper[4859]: I1008 18:31:57.404130 4859 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"secrets\" (UniqueName: \"kubernetes.io/secret/f1828698-e403-4643-a9ab-aedf302aae70-secrets\") pod \"openstack-galera-0\" (UID: \"f1828698-e403-4643-a9ab-aedf302aae70\") " pod="openstack/openstack-galera-0" Oct 08 18:31:57 crc kubenswrapper[4859]: I1008 18:31:57.414219 4859 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-p58hr\" (UniqueName: \"kubernetes.io/projected/f1828698-e403-4643-a9ab-aedf302aae70-kube-api-access-p58hr\") pod \"openstack-galera-0\" (UID: \"f1828698-e403-4643-a9ab-aedf302aae70\") " pod="openstack/openstack-galera-0" Oct 08 18:31:57 crc kubenswrapper[4859]: I1008 18:31:57.419924 4859 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"galera-tls-certs\" (UniqueName: \"kubernetes.io/secret/f1828698-e403-4643-a9ab-aedf302aae70-galera-tls-certs\") pod \"openstack-galera-0\" (UID: \"f1828698-e403-4643-a9ab-aedf302aae70\") " pod="openstack/openstack-galera-0" Oct 08 18:31:57 crc kubenswrapper[4859]: I1008 18:31:57.438326 4859 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage07-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage07-crc\") pod \"openstack-galera-0\" (UID: \"f1828698-e403-4643-a9ab-aedf302aae70\") " pod="openstack/openstack-galera-0" Oct 08 18:31:57 crc kubenswrapper[4859]: I1008 18:31:57.549848 4859 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/openstack-galera-0" Oct 08 18:31:57 crc kubenswrapper[4859]: I1008 18:31:57.654283 4859 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/openstack-cell1-galera-0"] Oct 08 18:31:57 crc kubenswrapper[4859]: I1008 18:31:57.655831 4859 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/openstack-cell1-galera-0" Oct 08 18:31:57 crc kubenswrapper[4859]: I1008 18:31:57.659442 4859 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"galera-openstack-cell1-dockercfg-f6qgt" Oct 08 18:31:57 crc kubenswrapper[4859]: I1008 18:31:57.659948 4859 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-cell1-config-data" Oct 08 18:31:57 crc kubenswrapper[4859]: I1008 18:31:57.660105 4859 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-galera-openstack-cell1-svc" Oct 08 18:31:57 crc kubenswrapper[4859]: I1008 18:31:57.662044 4859 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-cell1-scripts" Oct 08 18:31:57 crc kubenswrapper[4859]: I1008 18:31:57.666178 4859 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/openstack-cell1-galera-0"] Oct 08 18:31:57 crc kubenswrapper[4859]: I1008 18:31:57.700104 4859 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-default\" (UniqueName: \"kubernetes.io/configmap/a2c834dc-2b23-4081-8c71-e7ae462ca063-config-data-default\") pod \"openstack-cell1-galera-0\" (UID: \"a2c834dc-2b23-4081-8c71-e7ae462ca063\") " pod="openstack/openstack-cell1-galera-0" Oct 08 18:31:57 crc kubenswrapper[4859]: I1008 18:31:57.700152 4859 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"galera-tls-certs\" (UniqueName: \"kubernetes.io/secret/a2c834dc-2b23-4081-8c71-e7ae462ca063-galera-tls-certs\") pod \"openstack-cell1-galera-0\" (UID: \"a2c834dc-2b23-4081-8c71-e7ae462ca063\") " pod="openstack/openstack-cell1-galera-0" Oct 08 18:31:57 crc kubenswrapper[4859]: I1008 18:31:57.700173 4859 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secrets\" (UniqueName: \"kubernetes.io/secret/a2c834dc-2b23-4081-8c71-e7ae462ca063-secrets\") pod \"openstack-cell1-galera-0\" (UID: \"a2c834dc-2b23-4081-8c71-e7ae462ca063\") " pod="openstack/openstack-cell1-galera-0" Oct 08 18:31:57 crc kubenswrapper[4859]: I1008 18:31:57.700201 4859 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-generated\" (UniqueName: \"kubernetes.io/empty-dir/a2c834dc-2b23-4081-8c71-e7ae462ca063-config-data-generated\") pod \"openstack-cell1-galera-0\" (UID: \"a2c834dc-2b23-4081-8c71-e7ae462ca063\") " pod="openstack/openstack-cell1-galera-0" Oct 08 18:31:57 crc kubenswrapper[4859]: I1008 18:31:57.700228 4859 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/a2c834dc-2b23-4081-8c71-e7ae462ca063-operator-scripts\") pod \"openstack-cell1-galera-0\" (UID: \"a2c834dc-2b23-4081-8c71-e7ae462ca063\") " pod="openstack/openstack-cell1-galera-0" Oct 08 18:31:57 crc kubenswrapper[4859]: I1008 18:31:57.700256 4859 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-d6lhn\" (UniqueName: \"kubernetes.io/projected/a2c834dc-2b23-4081-8c71-e7ae462ca063-kube-api-access-d6lhn\") pod \"openstack-cell1-galera-0\" (UID: \"a2c834dc-2b23-4081-8c71-e7ae462ca063\") " pod="openstack/openstack-cell1-galera-0" Oct 08 18:31:57 crc kubenswrapper[4859]: I1008 18:31:57.700274 4859 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage04-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage04-crc\") pod \"openstack-cell1-galera-0\" (UID: \"a2c834dc-2b23-4081-8c71-e7ae462ca063\") " pod="openstack/openstack-cell1-galera-0" Oct 08 18:31:57 crc kubenswrapper[4859]: I1008 18:31:57.700295 4859 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a2c834dc-2b23-4081-8c71-e7ae462ca063-combined-ca-bundle\") pod \"openstack-cell1-galera-0\" (UID: \"a2c834dc-2b23-4081-8c71-e7ae462ca063\") " pod="openstack/openstack-cell1-galera-0" Oct 08 18:31:57 crc kubenswrapper[4859]: I1008 18:31:57.700310 4859 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/a2c834dc-2b23-4081-8c71-e7ae462ca063-kolla-config\") pod \"openstack-cell1-galera-0\" (UID: \"a2c834dc-2b23-4081-8c71-e7ae462ca063\") " pod="openstack/openstack-cell1-galera-0" Oct 08 18:31:57 crc kubenswrapper[4859]: I1008 18:31:57.801208 4859 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-generated\" (UniqueName: \"kubernetes.io/empty-dir/a2c834dc-2b23-4081-8c71-e7ae462ca063-config-data-generated\") pod \"openstack-cell1-galera-0\" (UID: \"a2c834dc-2b23-4081-8c71-e7ae462ca063\") " pod="openstack/openstack-cell1-galera-0" Oct 08 18:31:57 crc kubenswrapper[4859]: I1008 18:31:57.801656 4859 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/a2c834dc-2b23-4081-8c71-e7ae462ca063-operator-scripts\") pod \"openstack-cell1-galera-0\" (UID: \"a2c834dc-2b23-4081-8c71-e7ae462ca063\") " pod="openstack/openstack-cell1-galera-0" Oct 08 18:31:57 crc kubenswrapper[4859]: I1008 18:31:57.801712 4859 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-d6lhn\" (UniqueName: \"kubernetes.io/projected/a2c834dc-2b23-4081-8c71-e7ae462ca063-kube-api-access-d6lhn\") pod \"openstack-cell1-galera-0\" (UID: \"a2c834dc-2b23-4081-8c71-e7ae462ca063\") " pod="openstack/openstack-cell1-galera-0" Oct 08 18:31:57 crc kubenswrapper[4859]: I1008 18:31:57.801740 4859 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage04-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage04-crc\") pod \"openstack-cell1-galera-0\" (UID: \"a2c834dc-2b23-4081-8c71-e7ae462ca063\") " pod="openstack/openstack-cell1-galera-0" Oct 08 18:31:57 crc kubenswrapper[4859]: I1008 18:31:57.801775 4859 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a2c834dc-2b23-4081-8c71-e7ae462ca063-combined-ca-bundle\") pod \"openstack-cell1-galera-0\" (UID: \"a2c834dc-2b23-4081-8c71-e7ae462ca063\") " pod="openstack/openstack-cell1-galera-0" Oct 08 18:31:57 crc kubenswrapper[4859]: I1008 18:31:57.801796 4859 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/a2c834dc-2b23-4081-8c71-e7ae462ca063-kolla-config\") pod \"openstack-cell1-galera-0\" (UID: \"a2c834dc-2b23-4081-8c71-e7ae462ca063\") " pod="openstack/openstack-cell1-galera-0" Oct 08 18:31:57 crc kubenswrapper[4859]: I1008 18:31:57.801865 4859 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-default\" (UniqueName: \"kubernetes.io/configmap/a2c834dc-2b23-4081-8c71-e7ae462ca063-config-data-default\") pod \"openstack-cell1-galera-0\" (UID: \"a2c834dc-2b23-4081-8c71-e7ae462ca063\") " pod="openstack/openstack-cell1-galera-0" Oct 08 18:31:57 crc kubenswrapper[4859]: I1008 18:31:57.801891 4859 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"galera-tls-certs\" (UniqueName: \"kubernetes.io/secret/a2c834dc-2b23-4081-8c71-e7ae462ca063-galera-tls-certs\") pod \"openstack-cell1-galera-0\" (UID: \"a2c834dc-2b23-4081-8c71-e7ae462ca063\") " pod="openstack/openstack-cell1-galera-0" Oct 08 18:31:57 crc kubenswrapper[4859]: I1008 18:31:57.801914 4859 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"secrets\" (UniqueName: \"kubernetes.io/secret/a2c834dc-2b23-4081-8c71-e7ae462ca063-secrets\") pod \"openstack-cell1-galera-0\" (UID: \"a2c834dc-2b23-4081-8c71-e7ae462ca063\") " pod="openstack/openstack-cell1-galera-0" Oct 08 18:31:57 crc kubenswrapper[4859]: I1008 18:31:57.802015 4859 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-generated\" (UniqueName: \"kubernetes.io/empty-dir/a2c834dc-2b23-4081-8c71-e7ae462ca063-config-data-generated\") pod \"openstack-cell1-galera-0\" (UID: \"a2c834dc-2b23-4081-8c71-e7ae462ca063\") " pod="openstack/openstack-cell1-galera-0" Oct 08 18:31:57 crc kubenswrapper[4859]: I1008 18:31:57.809112 4859 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage04-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage04-crc\") pod \"openstack-cell1-galera-0\" (UID: \"a2c834dc-2b23-4081-8c71-e7ae462ca063\") device mount path \"/mnt/openstack/pv04\"" pod="openstack/openstack-cell1-galera-0" Oct 08 18:31:57 crc kubenswrapper[4859]: I1008 18:31:57.809584 4859 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/a2c834dc-2b23-4081-8c71-e7ae462ca063-operator-scripts\") pod \"openstack-cell1-galera-0\" (UID: \"a2c834dc-2b23-4081-8c71-e7ae462ca063\") " pod="openstack/openstack-cell1-galera-0" Oct 08 18:31:57 crc kubenswrapper[4859]: I1008 18:31:57.811223 4859 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a2c834dc-2b23-4081-8c71-e7ae462ca063-combined-ca-bundle\") pod \"openstack-cell1-galera-0\" (UID: \"a2c834dc-2b23-4081-8c71-e7ae462ca063\") " pod="openstack/openstack-cell1-galera-0" Oct 08 18:31:57 crc kubenswrapper[4859]: I1008 18:31:57.812911 4859 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-default\" (UniqueName: \"kubernetes.io/configmap/a2c834dc-2b23-4081-8c71-e7ae462ca063-config-data-default\") pod \"openstack-cell1-galera-0\" (UID: \"a2c834dc-2b23-4081-8c71-e7ae462ca063\") " pod="openstack/openstack-cell1-galera-0" Oct 08 18:31:57 crc kubenswrapper[4859]: I1008 18:31:57.812951 4859 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/a2c834dc-2b23-4081-8c71-e7ae462ca063-kolla-config\") pod \"openstack-cell1-galera-0\" (UID: \"a2c834dc-2b23-4081-8c71-e7ae462ca063\") " pod="openstack/openstack-cell1-galera-0" Oct 08 18:31:57 crc kubenswrapper[4859]: I1008 18:31:57.816589 4859 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"galera-tls-certs\" (UniqueName: \"kubernetes.io/secret/a2c834dc-2b23-4081-8c71-e7ae462ca063-galera-tls-certs\") pod \"openstack-cell1-galera-0\" (UID: \"a2c834dc-2b23-4081-8c71-e7ae462ca063\") " pod="openstack/openstack-cell1-galera-0" Oct 08 18:31:57 crc kubenswrapper[4859]: I1008 18:31:57.821789 4859 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"secrets\" (UniqueName: \"kubernetes.io/secret/a2c834dc-2b23-4081-8c71-e7ae462ca063-secrets\") pod \"openstack-cell1-galera-0\" (UID: \"a2c834dc-2b23-4081-8c71-e7ae462ca063\") " pod="openstack/openstack-cell1-galera-0" Oct 08 18:31:57 crc kubenswrapper[4859]: I1008 18:31:57.826770 4859 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-d6lhn\" (UniqueName: \"kubernetes.io/projected/a2c834dc-2b23-4081-8c71-e7ae462ca063-kube-api-access-d6lhn\") pod \"openstack-cell1-galera-0\" (UID: \"a2c834dc-2b23-4081-8c71-e7ae462ca063\") " pod="openstack/openstack-cell1-galera-0" Oct 08 18:31:57 crc kubenswrapper[4859]: I1008 18:31:57.845144 4859 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage04-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage04-crc\") pod \"openstack-cell1-galera-0\" (UID: \"a2c834dc-2b23-4081-8c71-e7ae462ca063\") " pod="openstack/openstack-cell1-galera-0" Oct 08 18:31:57 crc kubenswrapper[4859]: I1008 18:31:57.994857 4859 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/openstack-cell1-galera-0" Oct 08 18:31:58 crc kubenswrapper[4859]: I1008 18:31:58.039947 4859 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/memcached-0"] Oct 08 18:31:58 crc kubenswrapper[4859]: I1008 18:31:58.040977 4859 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/memcached-0" Oct 08 18:31:58 crc kubenswrapper[4859]: W1008 18:31:58.042884 4859 reflector.go:561] object-"openstack"/"memcached-config-data": failed to list *v1.ConfigMap: configmaps "memcached-config-data" is forbidden: User "system:node:crc" cannot list resource "configmaps" in API group "" in the namespace "openstack": no relationship found between node 'crc' and this object Oct 08 18:31:58 crc kubenswrapper[4859]: E1008 18:31:58.043013 4859 reflector.go:158] "Unhandled Error" err="object-\"openstack\"/\"memcached-config-data\": Failed to watch *v1.ConfigMap: failed to list *v1.ConfigMap: configmaps \"memcached-config-data\" is forbidden: User \"system:node:crc\" cannot list resource \"configmaps\" in API group \"\" in the namespace \"openstack\": no relationship found between node 'crc' and this object" logger="UnhandledError" Oct 08 18:31:58 crc kubenswrapper[4859]: W1008 18:31:58.043100 4859 reflector.go:561] object-"openstack"/"memcached-memcached-dockercfg-2tvbw": failed to list *v1.Secret: secrets "memcached-memcached-dockercfg-2tvbw" is forbidden: User "system:node:crc" cannot list resource "secrets" in API group "" in the namespace "openstack": no relationship found between node 'crc' and this object Oct 08 18:31:58 crc kubenswrapper[4859]: E1008 18:31:58.043155 4859 reflector.go:158] "Unhandled Error" err="object-\"openstack\"/\"memcached-memcached-dockercfg-2tvbw\": Failed to watch *v1.Secret: failed to list *v1.Secret: secrets \"memcached-memcached-dockercfg-2tvbw\" is forbidden: User \"system:node:crc\" cannot list resource \"secrets\" in API group \"\" in the namespace \"openstack\": no relationship found between node 'crc' and this object" logger="UnhandledError" Oct 08 18:31:58 crc kubenswrapper[4859]: W1008 18:31:58.044632 4859 reflector.go:561] object-"openstack"/"cert-memcached-svc": failed to list *v1.Secret: secrets "cert-memcached-svc" is forbidden: User "system:node:crc" cannot list resource "secrets" in API group "" in the namespace "openstack": no relationship found between node 'crc' and this object Oct 08 18:31:58 crc kubenswrapper[4859]: E1008 18:31:58.044672 4859 reflector.go:158] "Unhandled Error" err="object-\"openstack\"/\"cert-memcached-svc\": Failed to watch *v1.Secret: failed to list *v1.Secret: secrets \"cert-memcached-svc\" is forbidden: User \"system:node:crc\" cannot list resource \"secrets\" in API group \"\" in the namespace \"openstack\": no relationship found between node 'crc' and this object" logger="UnhandledError" Oct 08 18:31:58 crc kubenswrapper[4859]: I1008 18:31:58.073624 4859 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/memcached-0"] Oct 08 18:31:58 crc kubenswrapper[4859]: I1008 18:31:58.205907 4859 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"memcached-tls-certs\" (UniqueName: \"kubernetes.io/secret/6dc3fa5d-109b-4140-93aa-61b855c46df2-memcached-tls-certs\") pod \"memcached-0\" (UID: \"6dc3fa5d-109b-4140-93aa-61b855c46df2\") " pod="openstack/memcached-0" Oct 08 18:31:58 crc kubenswrapper[4859]: I1008 18:31:58.205975 4859 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6dc3fa5d-109b-4140-93aa-61b855c46df2-combined-ca-bundle\") pod \"memcached-0\" (UID: \"6dc3fa5d-109b-4140-93aa-61b855c46df2\") " pod="openstack/memcached-0" Oct 08 18:31:58 crc kubenswrapper[4859]: I1008 18:31:58.206019 4859 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/6dc3fa5d-109b-4140-93aa-61b855c46df2-config-data\") pod \"memcached-0\" (UID: \"6dc3fa5d-109b-4140-93aa-61b855c46df2\") " pod="openstack/memcached-0" Oct 08 18:31:58 crc kubenswrapper[4859]: I1008 18:31:58.206164 4859 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/6dc3fa5d-109b-4140-93aa-61b855c46df2-kolla-config\") pod \"memcached-0\" (UID: \"6dc3fa5d-109b-4140-93aa-61b855c46df2\") " pod="openstack/memcached-0" Oct 08 18:31:58 crc kubenswrapper[4859]: I1008 18:31:58.206214 4859 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-lj5g6\" (UniqueName: \"kubernetes.io/projected/6dc3fa5d-109b-4140-93aa-61b855c46df2-kube-api-access-lj5g6\") pod \"memcached-0\" (UID: \"6dc3fa5d-109b-4140-93aa-61b855c46df2\") " pod="openstack/memcached-0" Oct 08 18:31:58 crc kubenswrapper[4859]: I1008 18:31:58.307996 4859 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/6dc3fa5d-109b-4140-93aa-61b855c46df2-kolla-config\") pod \"memcached-0\" (UID: \"6dc3fa5d-109b-4140-93aa-61b855c46df2\") " pod="openstack/memcached-0" Oct 08 18:31:58 crc kubenswrapper[4859]: I1008 18:31:58.308082 4859 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-lj5g6\" (UniqueName: \"kubernetes.io/projected/6dc3fa5d-109b-4140-93aa-61b855c46df2-kube-api-access-lj5g6\") pod \"memcached-0\" (UID: \"6dc3fa5d-109b-4140-93aa-61b855c46df2\") " pod="openstack/memcached-0" Oct 08 18:31:58 crc kubenswrapper[4859]: I1008 18:31:58.308114 4859 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"memcached-tls-certs\" (UniqueName: \"kubernetes.io/secret/6dc3fa5d-109b-4140-93aa-61b855c46df2-memcached-tls-certs\") pod \"memcached-0\" (UID: \"6dc3fa5d-109b-4140-93aa-61b855c46df2\") " pod="openstack/memcached-0" Oct 08 18:31:58 crc kubenswrapper[4859]: I1008 18:31:58.308147 4859 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6dc3fa5d-109b-4140-93aa-61b855c46df2-combined-ca-bundle\") pod \"memcached-0\" (UID: \"6dc3fa5d-109b-4140-93aa-61b855c46df2\") " pod="openstack/memcached-0" Oct 08 18:31:58 crc kubenswrapper[4859]: I1008 18:31:58.308180 4859 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/6dc3fa5d-109b-4140-93aa-61b855c46df2-config-data\") pod \"memcached-0\" (UID: \"6dc3fa5d-109b-4140-93aa-61b855c46df2\") " pod="openstack/memcached-0" Oct 08 18:31:58 crc kubenswrapper[4859]: I1008 18:31:58.335356 4859 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6dc3fa5d-109b-4140-93aa-61b855c46df2-combined-ca-bundle\") pod \"memcached-0\" (UID: \"6dc3fa5d-109b-4140-93aa-61b855c46df2\") " pod="openstack/memcached-0" Oct 08 18:31:58 crc kubenswrapper[4859]: I1008 18:31:58.335713 4859 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-lj5g6\" (UniqueName: \"kubernetes.io/projected/6dc3fa5d-109b-4140-93aa-61b855c46df2-kube-api-access-lj5g6\") pod \"memcached-0\" (UID: \"6dc3fa5d-109b-4140-93aa-61b855c46df2\") " pod="openstack/memcached-0" Oct 08 18:31:58 crc kubenswrapper[4859]: I1008 18:31:58.978065 4859 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"memcached-memcached-dockercfg-2tvbw" Oct 08 18:31:59 crc kubenswrapper[4859]: I1008 18:31:59.014882 4859 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"memcached-config-data" Oct 08 18:31:59 crc kubenswrapper[4859]: I1008 18:31:59.019830 4859 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/6dc3fa5d-109b-4140-93aa-61b855c46df2-config-data\") pod \"memcached-0\" (UID: \"6dc3fa5d-109b-4140-93aa-61b855c46df2\") " pod="openstack/memcached-0" Oct 08 18:31:59 crc kubenswrapper[4859]: I1008 18:31:59.019874 4859 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/6dc3fa5d-109b-4140-93aa-61b855c46df2-kolla-config\") pod \"memcached-0\" (UID: \"6dc3fa5d-109b-4140-93aa-61b855c46df2\") " pod="openstack/memcached-0" Oct 08 18:31:59 crc kubenswrapper[4859]: E1008 18:31:59.314520 4859 secret.go:188] Couldn't get secret openstack/cert-memcached-svc: failed to sync secret cache: timed out waiting for the condition Oct 08 18:31:59 crc kubenswrapper[4859]: E1008 18:31:59.314644 4859 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/6dc3fa5d-109b-4140-93aa-61b855c46df2-memcached-tls-certs podName:6dc3fa5d-109b-4140-93aa-61b855c46df2 nodeName:}" failed. No retries permitted until 2025-10-08 18:31:59.814612027 +0000 UTC m=+890.061451406 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "memcached-tls-certs" (UniqueName: "kubernetes.io/secret/6dc3fa5d-109b-4140-93aa-61b855c46df2-memcached-tls-certs") pod "memcached-0" (UID: "6dc3fa5d-109b-4140-93aa-61b855c46df2") : failed to sync secret cache: timed out waiting for the condition Oct 08 18:31:59 crc kubenswrapper[4859]: I1008 18:31:59.459065 4859 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-memcached-svc" Oct 08 18:31:59 crc kubenswrapper[4859]: I1008 18:31:59.729765 4859 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/kube-state-metrics-0"] Oct 08 18:31:59 crc kubenswrapper[4859]: I1008 18:31:59.737444 4859 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/kube-state-metrics-0" Oct 08 18:31:59 crc kubenswrapper[4859]: I1008 18:31:59.739388 4859 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/kube-state-metrics-0"] Oct 08 18:31:59 crc kubenswrapper[4859]: I1008 18:31:59.748474 4859 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"telemetry-ceilometer-dockercfg-2bvs9" Oct 08 18:31:59 crc kubenswrapper[4859]: I1008 18:31:59.836569 4859 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"memcached-tls-certs\" (UniqueName: \"kubernetes.io/secret/6dc3fa5d-109b-4140-93aa-61b855c46df2-memcached-tls-certs\") pod \"memcached-0\" (UID: \"6dc3fa5d-109b-4140-93aa-61b855c46df2\") " pod="openstack/memcached-0" Oct 08 18:31:59 crc kubenswrapper[4859]: I1008 18:31:59.842000 4859 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"memcached-tls-certs\" (UniqueName: \"kubernetes.io/secret/6dc3fa5d-109b-4140-93aa-61b855c46df2-memcached-tls-certs\") pod \"memcached-0\" (UID: \"6dc3fa5d-109b-4140-93aa-61b855c46df2\") " pod="openstack/memcached-0" Oct 08 18:31:59 crc kubenswrapper[4859]: I1008 18:31:59.872089 4859 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/memcached-0" Oct 08 18:31:59 crc kubenswrapper[4859]: I1008 18:31:59.937872 4859 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-sh54v\" (UniqueName: \"kubernetes.io/projected/54254e29-9af5-4b28-8664-be427c4c4f3a-kube-api-access-sh54v\") pod \"kube-state-metrics-0\" (UID: \"54254e29-9af5-4b28-8664-be427c4c4f3a\") " pod="openstack/kube-state-metrics-0" Oct 08 18:32:00 crc kubenswrapper[4859]: I1008 18:32:00.040123 4859 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-sh54v\" (UniqueName: \"kubernetes.io/projected/54254e29-9af5-4b28-8664-be427c4c4f3a-kube-api-access-sh54v\") pod \"kube-state-metrics-0\" (UID: \"54254e29-9af5-4b28-8664-be427c4c4f3a\") " pod="openstack/kube-state-metrics-0" Oct 08 18:32:00 crc kubenswrapper[4859]: I1008 18:32:00.055899 4859 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-sh54v\" (UniqueName: \"kubernetes.io/projected/54254e29-9af5-4b28-8664-be427c4c4f3a-kube-api-access-sh54v\") pod \"kube-state-metrics-0\" (UID: \"54254e29-9af5-4b28-8664-be427c4c4f3a\") " pod="openstack/kube-state-metrics-0" Oct 08 18:32:00 crc kubenswrapper[4859]: I1008 18:32:00.063879 4859 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/kube-state-metrics-0" Oct 08 18:32:04 crc kubenswrapper[4859]: I1008 18:32:04.023382 4859 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ovn-controller-b89lv"] Oct 08 18:32:04 crc kubenswrapper[4859]: I1008 18:32:04.025390 4859 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-b89lv" Oct 08 18:32:04 crc kubenswrapper[4859]: I1008 18:32:04.029736 4859 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ovncontroller-ovncontroller-dockercfg-p6w59" Oct 08 18:32:04 crc kubenswrapper[4859]: I1008 18:32:04.029830 4859 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovncontroller-scripts" Oct 08 18:32:04 crc kubenswrapper[4859]: I1008 18:32:04.029736 4859 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-ovncontroller-ovndbs" Oct 08 18:32:04 crc kubenswrapper[4859]: I1008 18:32:04.038977 4859 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-controller-b89lv"] Oct 08 18:32:04 crc kubenswrapper[4859]: I1008 18:32:04.090005 4859 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ovn-controller-ovs-vthdk"] Oct 08 18:32:04 crc kubenswrapper[4859]: I1008 18:32:04.091544 4859 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-ovs-vthdk" Oct 08 18:32:04 crc kubenswrapper[4859]: I1008 18:32:04.104035 4859 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-controller-ovs-vthdk"] Oct 08 18:32:04 crc kubenswrapper[4859]: I1008 18:32:04.203933 4859 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-fxx7k\" (UniqueName: \"kubernetes.io/projected/ec20820a-066c-4606-8f1b-a26b172b7720-kube-api-access-fxx7k\") pod \"ovn-controller-b89lv\" (UID: \"ec20820a-066c-4606-8f1b-a26b172b7720\") " pod="openstack/ovn-controller-b89lv" Oct 08 18:32:04 crc kubenswrapper[4859]: I1008 18:32:04.204015 4859 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-ovs\" (UniqueName: \"kubernetes.io/host-path/eca01cc4-b7ef-429c-984d-1ccd1d95bcd2-etc-ovs\") pod \"ovn-controller-ovs-vthdk\" (UID: \"eca01cc4-b7ef-429c-984d-1ccd1d95bcd2\") " pod="openstack/ovn-controller-ovs-vthdk" Oct 08 18:32:04 crc kubenswrapper[4859]: I1008 18:32:04.204052 4859 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-run-ovn\" (UniqueName: \"kubernetes.io/host-path/ec20820a-066c-4606-8f1b-a26b172b7720-var-run-ovn\") pod \"ovn-controller-b89lv\" (UID: \"ec20820a-066c-4606-8f1b-a26b172b7720\") " pod="openstack/ovn-controller-b89lv" Oct 08 18:32:04 crc kubenswrapper[4859]: I1008 18:32:04.204069 4859 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-9z94f\" (UniqueName: \"kubernetes.io/projected/eca01cc4-b7ef-429c-984d-1ccd1d95bcd2-kube-api-access-9z94f\") pod \"ovn-controller-ovs-vthdk\" (UID: \"eca01cc4-b7ef-429c-984d-1ccd1d95bcd2\") " pod="openstack/ovn-controller-ovs-vthdk" Oct 08 18:32:04 crc kubenswrapper[4859]: I1008 18:32:04.204106 4859 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/eca01cc4-b7ef-429c-984d-1ccd1d95bcd2-var-run\") pod \"ovn-controller-ovs-vthdk\" (UID: \"eca01cc4-b7ef-429c-984d-1ccd1d95bcd2\") " pod="openstack/ovn-controller-ovs-vthdk" Oct 08 18:32:04 crc kubenswrapper[4859]: I1008 18:32:04.204173 4859 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/ec20820a-066c-4606-8f1b-a26b172b7720-scripts\") pod \"ovn-controller-b89lv\" (UID: \"ec20820a-066c-4606-8f1b-a26b172b7720\") " pod="openstack/ovn-controller-b89lv" Oct 08 18:32:04 crc kubenswrapper[4859]: I1008 18:32:04.204194 4859 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/eca01cc4-b7ef-429c-984d-1ccd1d95bcd2-scripts\") pod \"ovn-controller-ovs-vthdk\" (UID: \"eca01cc4-b7ef-429c-984d-1ccd1d95bcd2\") " pod="openstack/ovn-controller-ovs-vthdk" Oct 08 18:32:04 crc kubenswrapper[4859]: I1008 18:32:04.204225 4859 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/ec20820a-066c-4606-8f1b-a26b172b7720-var-run\") pod \"ovn-controller-b89lv\" (UID: \"ec20820a-066c-4606-8f1b-a26b172b7720\") " pod="openstack/ovn-controller-b89lv" Oct 08 18:32:04 crc kubenswrapper[4859]: I1008 18:32:04.204246 4859 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ec20820a-066c-4606-8f1b-a26b172b7720-combined-ca-bundle\") pod \"ovn-controller-b89lv\" (UID: \"ec20820a-066c-4606-8f1b-a26b172b7720\") " pod="openstack/ovn-controller-b89lv" Oct 08 18:32:04 crc kubenswrapper[4859]: I1008 18:32:04.204282 4859 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-log-ovn\" (UniqueName: \"kubernetes.io/host-path/ec20820a-066c-4606-8f1b-a26b172b7720-var-log-ovn\") pod \"ovn-controller-b89lv\" (UID: \"ec20820a-066c-4606-8f1b-a26b172b7720\") " pod="openstack/ovn-controller-b89lv" Oct 08 18:32:04 crc kubenswrapper[4859]: I1008 18:32:04.204301 4859 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-lib\" (UniqueName: \"kubernetes.io/host-path/eca01cc4-b7ef-429c-984d-1ccd1d95bcd2-var-lib\") pod \"ovn-controller-ovs-vthdk\" (UID: \"eca01cc4-b7ef-429c-984d-1ccd1d95bcd2\") " pod="openstack/ovn-controller-ovs-vthdk" Oct 08 18:32:04 crc kubenswrapper[4859]: I1008 18:32:04.204315 4859 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-log\" (UniqueName: \"kubernetes.io/host-path/eca01cc4-b7ef-429c-984d-1ccd1d95bcd2-var-log\") pod \"ovn-controller-ovs-vthdk\" (UID: \"eca01cc4-b7ef-429c-984d-1ccd1d95bcd2\") " pod="openstack/ovn-controller-ovs-vthdk" Oct 08 18:32:04 crc kubenswrapper[4859]: I1008 18:32:04.204333 4859 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovn-controller-tls-certs\" (UniqueName: \"kubernetes.io/secret/ec20820a-066c-4606-8f1b-a26b172b7720-ovn-controller-tls-certs\") pod \"ovn-controller-b89lv\" (UID: \"ec20820a-066c-4606-8f1b-a26b172b7720\") " pod="openstack/ovn-controller-b89lv" Oct 08 18:32:04 crc kubenswrapper[4859]: I1008 18:32:04.306095 4859 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/eca01cc4-b7ef-429c-984d-1ccd1d95bcd2-scripts\") pod \"ovn-controller-ovs-vthdk\" (UID: \"eca01cc4-b7ef-429c-984d-1ccd1d95bcd2\") " pod="openstack/ovn-controller-ovs-vthdk" Oct 08 18:32:04 crc kubenswrapper[4859]: I1008 18:32:04.306178 4859 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/ec20820a-066c-4606-8f1b-a26b172b7720-var-run\") pod \"ovn-controller-b89lv\" (UID: \"ec20820a-066c-4606-8f1b-a26b172b7720\") " pod="openstack/ovn-controller-b89lv" Oct 08 18:32:04 crc kubenswrapper[4859]: I1008 18:32:04.306203 4859 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ec20820a-066c-4606-8f1b-a26b172b7720-combined-ca-bundle\") pod \"ovn-controller-b89lv\" (UID: \"ec20820a-066c-4606-8f1b-a26b172b7720\") " pod="openstack/ovn-controller-b89lv" Oct 08 18:32:04 crc kubenswrapper[4859]: I1008 18:32:04.306249 4859 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-log-ovn\" (UniqueName: \"kubernetes.io/host-path/ec20820a-066c-4606-8f1b-a26b172b7720-var-log-ovn\") pod \"ovn-controller-b89lv\" (UID: \"ec20820a-066c-4606-8f1b-a26b172b7720\") " pod="openstack/ovn-controller-b89lv" Oct 08 18:32:04 crc kubenswrapper[4859]: I1008 18:32:04.306277 4859 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-lib\" (UniqueName: \"kubernetes.io/host-path/eca01cc4-b7ef-429c-984d-1ccd1d95bcd2-var-lib\") pod \"ovn-controller-ovs-vthdk\" (UID: \"eca01cc4-b7ef-429c-984d-1ccd1d95bcd2\") " pod="openstack/ovn-controller-ovs-vthdk" Oct 08 18:32:04 crc kubenswrapper[4859]: I1008 18:32:04.306290 4859 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-log\" (UniqueName: \"kubernetes.io/host-path/eca01cc4-b7ef-429c-984d-1ccd1d95bcd2-var-log\") pod \"ovn-controller-ovs-vthdk\" (UID: \"eca01cc4-b7ef-429c-984d-1ccd1d95bcd2\") " pod="openstack/ovn-controller-ovs-vthdk" Oct 08 18:32:04 crc kubenswrapper[4859]: I1008 18:32:04.306309 4859 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovn-controller-tls-certs\" (UniqueName: \"kubernetes.io/secret/ec20820a-066c-4606-8f1b-a26b172b7720-ovn-controller-tls-certs\") pod \"ovn-controller-b89lv\" (UID: \"ec20820a-066c-4606-8f1b-a26b172b7720\") " pod="openstack/ovn-controller-b89lv" Oct 08 18:32:04 crc kubenswrapper[4859]: I1008 18:32:04.306335 4859 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-fxx7k\" (UniqueName: \"kubernetes.io/projected/ec20820a-066c-4606-8f1b-a26b172b7720-kube-api-access-fxx7k\") pod \"ovn-controller-b89lv\" (UID: \"ec20820a-066c-4606-8f1b-a26b172b7720\") " pod="openstack/ovn-controller-b89lv" Oct 08 18:32:04 crc kubenswrapper[4859]: I1008 18:32:04.306368 4859 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-ovs\" (UniqueName: \"kubernetes.io/host-path/eca01cc4-b7ef-429c-984d-1ccd1d95bcd2-etc-ovs\") pod \"ovn-controller-ovs-vthdk\" (UID: \"eca01cc4-b7ef-429c-984d-1ccd1d95bcd2\") " pod="openstack/ovn-controller-ovs-vthdk" Oct 08 18:32:04 crc kubenswrapper[4859]: I1008 18:32:04.306394 4859 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-run-ovn\" (UniqueName: \"kubernetes.io/host-path/ec20820a-066c-4606-8f1b-a26b172b7720-var-run-ovn\") pod \"ovn-controller-b89lv\" (UID: \"ec20820a-066c-4606-8f1b-a26b172b7720\") " pod="openstack/ovn-controller-b89lv" Oct 08 18:32:04 crc kubenswrapper[4859]: I1008 18:32:04.306410 4859 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-9z94f\" (UniqueName: \"kubernetes.io/projected/eca01cc4-b7ef-429c-984d-1ccd1d95bcd2-kube-api-access-9z94f\") pod \"ovn-controller-ovs-vthdk\" (UID: \"eca01cc4-b7ef-429c-984d-1ccd1d95bcd2\") " pod="openstack/ovn-controller-ovs-vthdk" Oct 08 18:32:04 crc kubenswrapper[4859]: I1008 18:32:04.306436 4859 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/eca01cc4-b7ef-429c-984d-1ccd1d95bcd2-var-run\") pod \"ovn-controller-ovs-vthdk\" (UID: \"eca01cc4-b7ef-429c-984d-1ccd1d95bcd2\") " pod="openstack/ovn-controller-ovs-vthdk" Oct 08 18:32:04 crc kubenswrapper[4859]: I1008 18:32:04.306453 4859 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/ec20820a-066c-4606-8f1b-a26b172b7720-scripts\") pod \"ovn-controller-b89lv\" (UID: \"ec20820a-066c-4606-8f1b-a26b172b7720\") " pod="openstack/ovn-controller-b89lv" Oct 08 18:32:04 crc kubenswrapper[4859]: I1008 18:32:04.308425 4859 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/ec20820a-066c-4606-8f1b-a26b172b7720-scripts\") pod \"ovn-controller-b89lv\" (UID: \"ec20820a-066c-4606-8f1b-a26b172b7720\") " pod="openstack/ovn-controller-b89lv" Oct 08 18:32:04 crc kubenswrapper[4859]: I1008 18:32:04.308886 4859 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/ec20820a-066c-4606-8f1b-a26b172b7720-var-run\") pod \"ovn-controller-b89lv\" (UID: \"ec20820a-066c-4606-8f1b-a26b172b7720\") " pod="openstack/ovn-controller-b89lv" Oct 08 18:32:04 crc kubenswrapper[4859]: I1008 18:32:04.310007 4859 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-lib\" (UniqueName: \"kubernetes.io/host-path/eca01cc4-b7ef-429c-984d-1ccd1d95bcd2-var-lib\") pod \"ovn-controller-ovs-vthdk\" (UID: \"eca01cc4-b7ef-429c-984d-1ccd1d95bcd2\") " pod="openstack/ovn-controller-ovs-vthdk" Oct 08 18:32:04 crc kubenswrapper[4859]: I1008 18:32:04.310070 4859 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-log\" (UniqueName: \"kubernetes.io/host-path/eca01cc4-b7ef-429c-984d-1ccd1d95bcd2-var-log\") pod \"ovn-controller-ovs-vthdk\" (UID: \"eca01cc4-b7ef-429c-984d-1ccd1d95bcd2\") " pod="openstack/ovn-controller-ovs-vthdk" Oct 08 18:32:04 crc kubenswrapper[4859]: I1008 18:32:04.310158 4859 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-log-ovn\" (UniqueName: \"kubernetes.io/host-path/ec20820a-066c-4606-8f1b-a26b172b7720-var-log-ovn\") pod \"ovn-controller-b89lv\" (UID: \"ec20820a-066c-4606-8f1b-a26b172b7720\") " pod="openstack/ovn-controller-b89lv" Oct 08 18:32:04 crc kubenswrapper[4859]: I1008 18:32:04.310266 4859 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-run-ovn\" (UniqueName: \"kubernetes.io/host-path/ec20820a-066c-4606-8f1b-a26b172b7720-var-run-ovn\") pod \"ovn-controller-b89lv\" (UID: \"ec20820a-066c-4606-8f1b-a26b172b7720\") " pod="openstack/ovn-controller-b89lv" Oct 08 18:32:04 crc kubenswrapper[4859]: I1008 18:32:04.310414 4859 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-ovs\" (UniqueName: \"kubernetes.io/host-path/eca01cc4-b7ef-429c-984d-1ccd1d95bcd2-etc-ovs\") pod \"ovn-controller-ovs-vthdk\" (UID: \"eca01cc4-b7ef-429c-984d-1ccd1d95bcd2\") " pod="openstack/ovn-controller-ovs-vthdk" Oct 08 18:32:04 crc kubenswrapper[4859]: I1008 18:32:04.310828 4859 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/eca01cc4-b7ef-429c-984d-1ccd1d95bcd2-var-run\") pod \"ovn-controller-ovs-vthdk\" (UID: \"eca01cc4-b7ef-429c-984d-1ccd1d95bcd2\") " pod="openstack/ovn-controller-ovs-vthdk" Oct 08 18:32:04 crc kubenswrapper[4859]: I1008 18:32:04.315093 4859 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/eca01cc4-b7ef-429c-984d-1ccd1d95bcd2-scripts\") pod \"ovn-controller-ovs-vthdk\" (UID: \"eca01cc4-b7ef-429c-984d-1ccd1d95bcd2\") " pod="openstack/ovn-controller-ovs-vthdk" Oct 08 18:32:04 crc kubenswrapper[4859]: I1008 18:32:04.318471 4859 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovn-controller-tls-certs\" (UniqueName: \"kubernetes.io/secret/ec20820a-066c-4606-8f1b-a26b172b7720-ovn-controller-tls-certs\") pod \"ovn-controller-b89lv\" (UID: \"ec20820a-066c-4606-8f1b-a26b172b7720\") " pod="openstack/ovn-controller-b89lv" Oct 08 18:32:04 crc kubenswrapper[4859]: I1008 18:32:04.318556 4859 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ec20820a-066c-4606-8f1b-a26b172b7720-combined-ca-bundle\") pod \"ovn-controller-b89lv\" (UID: \"ec20820a-066c-4606-8f1b-a26b172b7720\") " pod="openstack/ovn-controller-b89lv" Oct 08 18:32:04 crc kubenswrapper[4859]: I1008 18:32:04.338449 4859 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-9z94f\" (UniqueName: \"kubernetes.io/projected/eca01cc4-b7ef-429c-984d-1ccd1d95bcd2-kube-api-access-9z94f\") pod \"ovn-controller-ovs-vthdk\" (UID: \"eca01cc4-b7ef-429c-984d-1ccd1d95bcd2\") " pod="openstack/ovn-controller-ovs-vthdk" Oct 08 18:32:04 crc kubenswrapper[4859]: I1008 18:32:04.339562 4859 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-fxx7k\" (UniqueName: \"kubernetes.io/projected/ec20820a-066c-4606-8f1b-a26b172b7720-kube-api-access-fxx7k\") pod \"ovn-controller-b89lv\" (UID: \"ec20820a-066c-4606-8f1b-a26b172b7720\") " pod="openstack/ovn-controller-b89lv" Oct 08 18:32:04 crc kubenswrapper[4859]: I1008 18:32:04.351215 4859 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-b89lv" Oct 08 18:32:04 crc kubenswrapper[4859]: I1008 18:32:04.410820 4859 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-ovs-vthdk" Oct 08 18:32:04 crc kubenswrapper[4859]: I1008 18:32:04.596975 4859 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ovsdbserver-nb-0"] Oct 08 18:32:04 crc kubenswrapper[4859]: I1008 18:32:04.598254 4859 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovsdbserver-nb-0" Oct 08 18:32:04 crc kubenswrapper[4859]: I1008 18:32:04.603046 4859 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-ovndbcluster-nb-ovndbs" Oct 08 18:32:04 crc kubenswrapper[4859]: I1008 18:32:04.603245 4859 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovndbcluster-nb-scripts" Oct 08 18:32:04 crc kubenswrapper[4859]: I1008 18:32:04.603393 4859 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-ovn-metrics" Oct 08 18:32:04 crc kubenswrapper[4859]: I1008 18:32:04.603538 4859 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ovncluster-ovndbcluster-nb-dockercfg-jcg5d" Oct 08 18:32:04 crc kubenswrapper[4859]: I1008 18:32:04.603738 4859 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovndbcluster-nb-config" Oct 08 18:32:04 crc kubenswrapper[4859]: I1008 18:32:04.619383 4859 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovsdbserver-nb-0"] Oct 08 18:32:04 crc kubenswrapper[4859]: I1008 18:32:04.722748 4859 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage02-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage02-crc\") pod \"ovsdbserver-nb-0\" (UID: \"9fc7c8d2-a890-43c0-8081-6049a57c7474\") " pod="openstack/ovsdbserver-nb-0" Oct 08 18:32:04 crc kubenswrapper[4859]: I1008 18:32:04.722821 4859 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb-tls-certs\" (UniqueName: \"kubernetes.io/secret/9fc7c8d2-a890-43c0-8081-6049a57c7474-ovsdbserver-nb-tls-certs\") pod \"ovsdbserver-nb-0\" (UID: \"9fc7c8d2-a890-43c0-8081-6049a57c7474\") " pod="openstack/ovsdbserver-nb-0" Oct 08 18:32:04 crc kubenswrapper[4859]: I1008 18:32:04.722852 4859 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9fc7c8d2-a890-43c0-8081-6049a57c7474-combined-ca-bundle\") pod \"ovsdbserver-nb-0\" (UID: \"9fc7c8d2-a890-43c0-8081-6049a57c7474\") " pod="openstack/ovsdbserver-nb-0" Oct 08 18:32:04 crc kubenswrapper[4859]: I1008 18:32:04.722890 4859 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/9fc7c8d2-a890-43c0-8081-6049a57c7474-scripts\") pod \"ovsdbserver-nb-0\" (UID: \"9fc7c8d2-a890-43c0-8081-6049a57c7474\") " pod="openstack/ovsdbserver-nb-0" Oct 08 18:32:04 crc kubenswrapper[4859]: I1008 18:32:04.722915 4859 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-5q775\" (UniqueName: \"kubernetes.io/projected/9fc7c8d2-a890-43c0-8081-6049a57c7474-kube-api-access-5q775\") pod \"ovsdbserver-nb-0\" (UID: \"9fc7c8d2-a890-43c0-8081-6049a57c7474\") " pod="openstack/ovsdbserver-nb-0" Oct 08 18:32:04 crc kubenswrapper[4859]: I1008 18:32:04.722943 4859 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdb-rundir\" (UniqueName: \"kubernetes.io/empty-dir/9fc7c8d2-a890-43c0-8081-6049a57c7474-ovsdb-rundir\") pod \"ovsdbserver-nb-0\" (UID: \"9fc7c8d2-a890-43c0-8081-6049a57c7474\") " pod="openstack/ovsdbserver-nb-0" Oct 08 18:32:04 crc kubenswrapper[4859]: I1008 18:32:04.723207 4859 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/9fc7c8d2-a890-43c0-8081-6049a57c7474-config\") pod \"ovsdbserver-nb-0\" (UID: \"9fc7c8d2-a890-43c0-8081-6049a57c7474\") " pod="openstack/ovsdbserver-nb-0" Oct 08 18:32:04 crc kubenswrapper[4859]: I1008 18:32:04.723347 4859 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/9fc7c8d2-a890-43c0-8081-6049a57c7474-metrics-certs-tls-certs\") pod \"ovsdbserver-nb-0\" (UID: \"9fc7c8d2-a890-43c0-8081-6049a57c7474\") " pod="openstack/ovsdbserver-nb-0" Oct 08 18:32:04 crc kubenswrapper[4859]: I1008 18:32:04.824827 4859 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdb-rundir\" (UniqueName: \"kubernetes.io/empty-dir/9fc7c8d2-a890-43c0-8081-6049a57c7474-ovsdb-rundir\") pod \"ovsdbserver-nb-0\" (UID: \"9fc7c8d2-a890-43c0-8081-6049a57c7474\") " pod="openstack/ovsdbserver-nb-0" Oct 08 18:32:04 crc kubenswrapper[4859]: I1008 18:32:04.824915 4859 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/9fc7c8d2-a890-43c0-8081-6049a57c7474-config\") pod \"ovsdbserver-nb-0\" (UID: \"9fc7c8d2-a890-43c0-8081-6049a57c7474\") " pod="openstack/ovsdbserver-nb-0" Oct 08 18:32:04 crc kubenswrapper[4859]: I1008 18:32:04.824997 4859 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/9fc7c8d2-a890-43c0-8081-6049a57c7474-metrics-certs-tls-certs\") pod \"ovsdbserver-nb-0\" (UID: \"9fc7c8d2-a890-43c0-8081-6049a57c7474\") " pod="openstack/ovsdbserver-nb-0" Oct 08 18:32:04 crc kubenswrapper[4859]: I1008 18:32:04.825041 4859 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage02-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage02-crc\") pod \"ovsdbserver-nb-0\" (UID: \"9fc7c8d2-a890-43c0-8081-6049a57c7474\") " pod="openstack/ovsdbserver-nb-0" Oct 08 18:32:04 crc kubenswrapper[4859]: I1008 18:32:04.825074 4859 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb-tls-certs\" (UniqueName: \"kubernetes.io/secret/9fc7c8d2-a890-43c0-8081-6049a57c7474-ovsdbserver-nb-tls-certs\") pod \"ovsdbserver-nb-0\" (UID: \"9fc7c8d2-a890-43c0-8081-6049a57c7474\") " pod="openstack/ovsdbserver-nb-0" Oct 08 18:32:04 crc kubenswrapper[4859]: I1008 18:32:04.825091 4859 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9fc7c8d2-a890-43c0-8081-6049a57c7474-combined-ca-bundle\") pod \"ovsdbserver-nb-0\" (UID: \"9fc7c8d2-a890-43c0-8081-6049a57c7474\") " pod="openstack/ovsdbserver-nb-0" Oct 08 18:32:04 crc kubenswrapper[4859]: I1008 18:32:04.825116 4859 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/9fc7c8d2-a890-43c0-8081-6049a57c7474-scripts\") pod \"ovsdbserver-nb-0\" (UID: \"9fc7c8d2-a890-43c0-8081-6049a57c7474\") " pod="openstack/ovsdbserver-nb-0" Oct 08 18:32:04 crc kubenswrapper[4859]: I1008 18:32:04.825133 4859 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-5q775\" (UniqueName: \"kubernetes.io/projected/9fc7c8d2-a890-43c0-8081-6049a57c7474-kube-api-access-5q775\") pod \"ovsdbserver-nb-0\" (UID: \"9fc7c8d2-a890-43c0-8081-6049a57c7474\") " pod="openstack/ovsdbserver-nb-0" Oct 08 18:32:04 crc kubenswrapper[4859]: I1008 18:32:04.825947 4859 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdb-rundir\" (UniqueName: \"kubernetes.io/empty-dir/9fc7c8d2-a890-43c0-8081-6049a57c7474-ovsdb-rundir\") pod \"ovsdbserver-nb-0\" (UID: \"9fc7c8d2-a890-43c0-8081-6049a57c7474\") " pod="openstack/ovsdbserver-nb-0" Oct 08 18:32:04 crc kubenswrapper[4859]: I1008 18:32:04.826556 4859 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/9fc7c8d2-a890-43c0-8081-6049a57c7474-config\") pod \"ovsdbserver-nb-0\" (UID: \"9fc7c8d2-a890-43c0-8081-6049a57c7474\") " pod="openstack/ovsdbserver-nb-0" Oct 08 18:32:04 crc kubenswrapper[4859]: I1008 18:32:04.827509 4859 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage02-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage02-crc\") pod \"ovsdbserver-nb-0\" (UID: \"9fc7c8d2-a890-43c0-8081-6049a57c7474\") device mount path \"/mnt/openstack/pv02\"" pod="openstack/ovsdbserver-nb-0" Oct 08 18:32:04 crc kubenswrapper[4859]: I1008 18:32:04.830083 4859 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/9fc7c8d2-a890-43c0-8081-6049a57c7474-scripts\") pod \"ovsdbserver-nb-0\" (UID: \"9fc7c8d2-a890-43c0-8081-6049a57c7474\") " pod="openstack/ovsdbserver-nb-0" Oct 08 18:32:04 crc kubenswrapper[4859]: I1008 18:32:04.832461 4859 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb-tls-certs\" (UniqueName: \"kubernetes.io/secret/9fc7c8d2-a890-43c0-8081-6049a57c7474-ovsdbserver-nb-tls-certs\") pod \"ovsdbserver-nb-0\" (UID: \"9fc7c8d2-a890-43c0-8081-6049a57c7474\") " pod="openstack/ovsdbserver-nb-0" Oct 08 18:32:04 crc kubenswrapper[4859]: I1008 18:32:04.833501 4859 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9fc7c8d2-a890-43c0-8081-6049a57c7474-combined-ca-bundle\") pod \"ovsdbserver-nb-0\" (UID: \"9fc7c8d2-a890-43c0-8081-6049a57c7474\") " pod="openstack/ovsdbserver-nb-0" Oct 08 18:32:04 crc kubenswrapper[4859]: I1008 18:32:04.837739 4859 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/9fc7c8d2-a890-43c0-8081-6049a57c7474-metrics-certs-tls-certs\") pod \"ovsdbserver-nb-0\" (UID: \"9fc7c8d2-a890-43c0-8081-6049a57c7474\") " pod="openstack/ovsdbserver-nb-0" Oct 08 18:32:04 crc kubenswrapper[4859]: I1008 18:32:04.843291 4859 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-5q775\" (UniqueName: \"kubernetes.io/projected/9fc7c8d2-a890-43c0-8081-6049a57c7474-kube-api-access-5q775\") pod \"ovsdbserver-nb-0\" (UID: \"9fc7c8d2-a890-43c0-8081-6049a57c7474\") " pod="openstack/ovsdbserver-nb-0" Oct 08 18:32:04 crc kubenswrapper[4859]: I1008 18:32:04.853040 4859 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage02-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage02-crc\") pod \"ovsdbserver-nb-0\" (UID: \"9fc7c8d2-a890-43c0-8081-6049a57c7474\") " pod="openstack/ovsdbserver-nb-0" Oct 08 18:32:04 crc kubenswrapper[4859]: I1008 18:32:04.939961 4859 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovsdbserver-nb-0" Oct 08 18:32:07 crc kubenswrapper[4859]: I1008 18:32:07.190297 4859 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ovsdbserver-sb-0"] Oct 08 18:32:07 crc kubenswrapper[4859]: I1008 18:32:07.192309 4859 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovsdbserver-sb-0" Oct 08 18:32:07 crc kubenswrapper[4859]: I1008 18:32:07.211528 4859 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovndbcluster-sb-scripts" Oct 08 18:32:07 crc kubenswrapper[4859]: I1008 18:32:07.211806 4859 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-ovndbcluster-sb-ovndbs" Oct 08 18:32:07 crc kubenswrapper[4859]: I1008 18:32:07.211895 4859 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ovncluster-ovndbcluster-sb-dockercfg-knhrj" Oct 08 18:32:07 crc kubenswrapper[4859]: I1008 18:32:07.227040 4859 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovsdbserver-sb-0"] Oct 08 18:32:07 crc kubenswrapper[4859]: I1008 18:32:07.227295 4859 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovndbcluster-sb-config" Oct 08 18:32:07 crc kubenswrapper[4859]: I1008 18:32:07.369081 4859 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/fb46e63a-80c0-4141-8e6a-ad49c95bc924-combined-ca-bundle\") pod \"ovsdbserver-sb-0\" (UID: \"fb46e63a-80c0-4141-8e6a-ad49c95bc924\") " pod="openstack/ovsdbserver-sb-0" Oct 08 18:32:07 crc kubenswrapper[4859]: I1008 18:32:07.369157 4859 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-vbl49\" (UniqueName: \"kubernetes.io/projected/fb46e63a-80c0-4141-8e6a-ad49c95bc924-kube-api-access-vbl49\") pod \"ovsdbserver-sb-0\" (UID: \"fb46e63a-80c0-4141-8e6a-ad49c95bc924\") " pod="openstack/ovsdbserver-sb-0" Oct 08 18:32:07 crc kubenswrapper[4859]: I1008 18:32:07.369208 4859 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage05-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage05-crc\") pod \"ovsdbserver-sb-0\" (UID: \"fb46e63a-80c0-4141-8e6a-ad49c95bc924\") " pod="openstack/ovsdbserver-sb-0" Oct 08 18:32:07 crc kubenswrapper[4859]: I1008 18:32:07.369274 4859 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdb-rundir\" (UniqueName: \"kubernetes.io/empty-dir/fb46e63a-80c0-4141-8e6a-ad49c95bc924-ovsdb-rundir\") pod \"ovsdbserver-sb-0\" (UID: \"fb46e63a-80c0-4141-8e6a-ad49c95bc924\") " pod="openstack/ovsdbserver-sb-0" Oct 08 18:32:07 crc kubenswrapper[4859]: I1008 18:32:07.369351 4859 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/fb46e63a-80c0-4141-8e6a-ad49c95bc924-metrics-certs-tls-certs\") pod \"ovsdbserver-sb-0\" (UID: \"fb46e63a-80c0-4141-8e6a-ad49c95bc924\") " pod="openstack/ovsdbserver-sb-0" Oct 08 18:32:07 crc kubenswrapper[4859]: I1008 18:32:07.369378 4859 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb-tls-certs\" (UniqueName: \"kubernetes.io/secret/fb46e63a-80c0-4141-8e6a-ad49c95bc924-ovsdbserver-sb-tls-certs\") pod \"ovsdbserver-sb-0\" (UID: \"fb46e63a-80c0-4141-8e6a-ad49c95bc924\") " pod="openstack/ovsdbserver-sb-0" Oct 08 18:32:07 crc kubenswrapper[4859]: I1008 18:32:07.369406 4859 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/fb46e63a-80c0-4141-8e6a-ad49c95bc924-scripts\") pod \"ovsdbserver-sb-0\" (UID: \"fb46e63a-80c0-4141-8e6a-ad49c95bc924\") " pod="openstack/ovsdbserver-sb-0" Oct 08 18:32:07 crc kubenswrapper[4859]: I1008 18:32:07.369441 4859 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/fb46e63a-80c0-4141-8e6a-ad49c95bc924-config\") pod \"ovsdbserver-sb-0\" (UID: \"fb46e63a-80c0-4141-8e6a-ad49c95bc924\") " pod="openstack/ovsdbserver-sb-0" Oct 08 18:32:07 crc kubenswrapper[4859]: I1008 18:32:07.470928 4859 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdb-rundir\" (UniqueName: \"kubernetes.io/empty-dir/fb46e63a-80c0-4141-8e6a-ad49c95bc924-ovsdb-rundir\") pod \"ovsdbserver-sb-0\" (UID: \"fb46e63a-80c0-4141-8e6a-ad49c95bc924\") " pod="openstack/ovsdbserver-sb-0" Oct 08 18:32:07 crc kubenswrapper[4859]: I1008 18:32:07.471050 4859 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/fb46e63a-80c0-4141-8e6a-ad49c95bc924-metrics-certs-tls-certs\") pod \"ovsdbserver-sb-0\" (UID: \"fb46e63a-80c0-4141-8e6a-ad49c95bc924\") " pod="openstack/ovsdbserver-sb-0" Oct 08 18:32:07 crc kubenswrapper[4859]: I1008 18:32:07.471089 4859 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb-tls-certs\" (UniqueName: \"kubernetes.io/secret/fb46e63a-80c0-4141-8e6a-ad49c95bc924-ovsdbserver-sb-tls-certs\") pod \"ovsdbserver-sb-0\" (UID: \"fb46e63a-80c0-4141-8e6a-ad49c95bc924\") " pod="openstack/ovsdbserver-sb-0" Oct 08 18:32:07 crc kubenswrapper[4859]: I1008 18:32:07.471143 4859 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/fb46e63a-80c0-4141-8e6a-ad49c95bc924-scripts\") pod \"ovsdbserver-sb-0\" (UID: \"fb46e63a-80c0-4141-8e6a-ad49c95bc924\") " pod="openstack/ovsdbserver-sb-0" Oct 08 18:32:07 crc kubenswrapper[4859]: I1008 18:32:07.471184 4859 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/fb46e63a-80c0-4141-8e6a-ad49c95bc924-config\") pod \"ovsdbserver-sb-0\" (UID: \"fb46e63a-80c0-4141-8e6a-ad49c95bc924\") " pod="openstack/ovsdbserver-sb-0" Oct 08 18:32:07 crc kubenswrapper[4859]: I1008 18:32:07.471234 4859 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/fb46e63a-80c0-4141-8e6a-ad49c95bc924-combined-ca-bundle\") pod \"ovsdbserver-sb-0\" (UID: \"fb46e63a-80c0-4141-8e6a-ad49c95bc924\") " pod="openstack/ovsdbserver-sb-0" Oct 08 18:32:07 crc kubenswrapper[4859]: I1008 18:32:07.471262 4859 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-vbl49\" (UniqueName: \"kubernetes.io/projected/fb46e63a-80c0-4141-8e6a-ad49c95bc924-kube-api-access-vbl49\") pod \"ovsdbserver-sb-0\" (UID: \"fb46e63a-80c0-4141-8e6a-ad49c95bc924\") " pod="openstack/ovsdbserver-sb-0" Oct 08 18:32:07 crc kubenswrapper[4859]: I1008 18:32:07.471301 4859 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage05-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage05-crc\") pod \"ovsdbserver-sb-0\" (UID: \"fb46e63a-80c0-4141-8e6a-ad49c95bc924\") " pod="openstack/ovsdbserver-sb-0" Oct 08 18:32:07 crc kubenswrapper[4859]: I1008 18:32:07.471666 4859 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage05-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage05-crc\") pod \"ovsdbserver-sb-0\" (UID: \"fb46e63a-80c0-4141-8e6a-ad49c95bc924\") device mount path \"/mnt/openstack/pv05\"" pod="openstack/ovsdbserver-sb-0" Oct 08 18:32:07 crc kubenswrapper[4859]: I1008 18:32:07.471761 4859 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdb-rundir\" (UniqueName: \"kubernetes.io/empty-dir/fb46e63a-80c0-4141-8e6a-ad49c95bc924-ovsdb-rundir\") pod \"ovsdbserver-sb-0\" (UID: \"fb46e63a-80c0-4141-8e6a-ad49c95bc924\") " pod="openstack/ovsdbserver-sb-0" Oct 08 18:32:07 crc kubenswrapper[4859]: I1008 18:32:07.472143 4859 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/fb46e63a-80c0-4141-8e6a-ad49c95bc924-config\") pod \"ovsdbserver-sb-0\" (UID: \"fb46e63a-80c0-4141-8e6a-ad49c95bc924\") " pod="openstack/ovsdbserver-sb-0" Oct 08 18:32:07 crc kubenswrapper[4859]: I1008 18:32:07.472900 4859 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/fb46e63a-80c0-4141-8e6a-ad49c95bc924-scripts\") pod \"ovsdbserver-sb-0\" (UID: \"fb46e63a-80c0-4141-8e6a-ad49c95bc924\") " pod="openstack/ovsdbserver-sb-0" Oct 08 18:32:07 crc kubenswrapper[4859]: I1008 18:32:07.477499 4859 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb-tls-certs\" (UniqueName: \"kubernetes.io/secret/fb46e63a-80c0-4141-8e6a-ad49c95bc924-ovsdbserver-sb-tls-certs\") pod \"ovsdbserver-sb-0\" (UID: \"fb46e63a-80c0-4141-8e6a-ad49c95bc924\") " pod="openstack/ovsdbserver-sb-0" Oct 08 18:32:07 crc kubenswrapper[4859]: I1008 18:32:07.488402 4859 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/fb46e63a-80c0-4141-8e6a-ad49c95bc924-combined-ca-bundle\") pod \"ovsdbserver-sb-0\" (UID: \"fb46e63a-80c0-4141-8e6a-ad49c95bc924\") " pod="openstack/ovsdbserver-sb-0" Oct 08 18:32:07 crc kubenswrapper[4859]: I1008 18:32:07.492450 4859 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/fb46e63a-80c0-4141-8e6a-ad49c95bc924-metrics-certs-tls-certs\") pod \"ovsdbserver-sb-0\" (UID: \"fb46e63a-80c0-4141-8e6a-ad49c95bc924\") " pod="openstack/ovsdbserver-sb-0" Oct 08 18:32:07 crc kubenswrapper[4859]: I1008 18:32:07.496024 4859 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-vbl49\" (UniqueName: \"kubernetes.io/projected/fb46e63a-80c0-4141-8e6a-ad49c95bc924-kube-api-access-vbl49\") pod \"ovsdbserver-sb-0\" (UID: \"fb46e63a-80c0-4141-8e6a-ad49c95bc924\") " pod="openstack/ovsdbserver-sb-0" Oct 08 18:32:07 crc kubenswrapper[4859]: I1008 18:32:07.498599 4859 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage05-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage05-crc\") pod \"ovsdbserver-sb-0\" (UID: \"fb46e63a-80c0-4141-8e6a-ad49c95bc924\") " pod="openstack/ovsdbserver-sb-0" Oct 08 18:32:07 crc kubenswrapper[4859]: I1008 18:32:07.569163 4859 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovsdbserver-sb-0" Oct 08 18:32:08 crc kubenswrapper[4859]: E1008 18:32:08.247372 4859 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/podified-antelope-centos9/openstack-neutron-server@sha256:c4e71b2158fd939dad8b8e705273493051d3023273d23b279f2699dce6db33df" Oct 08 18:32:08 crc kubenswrapper[4859]: E1008 18:32:08.247938 4859 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:init,Image:quay.io/podified-antelope-centos9/openstack-neutron-server@sha256:c4e71b2158fd939dad8b8e705273493051d3023273d23b279f2699dce6db33df,Command:[/bin/bash],Args:[-c dnsmasq --interface=* --conf-dir=/etc/dnsmasq.d --hostsdir=/etc/dnsmasq.d/hosts --keep-in-foreground --log-debug --bind-interfaces --listen-address=$(POD_IP) --port 5353 --log-facility=- --no-hosts --domain-needed --no-resolv --bogus-priv --log-queries --test],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:CONFIG_HASH,Value:nffh5bdhf4h5f8h79h55h77h58fh56dh7bh6fh578hbch55dh68h56bhd9h65dh57ch658hc9h566h666h688h58h65dh684h5d7h6ch575h5d6h88q,ValueFrom:nil,},EnvVar{Name:POD_IP,Value:,ValueFrom:&EnvVarSource{FieldRef:&ObjectFieldSelector{APIVersion:v1,FieldPath:status.podIP,},ResourceFieldRef:nil,ConfigMapKeyRef:nil,SecretKeyRef:nil,},},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:config,ReadOnly:true,MountPath:/etc/dnsmasq.d/config.cfg,SubPath:dns,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-plgwt,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000650000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:&SeccompProfile{Type:RuntimeDefault,LocalhostProfile:nil,},AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod dnsmasq-dns-7bfcb9d745-ntml8_openstack(e66f484b-21c2-45ff-9783-f73530d8ba73): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Oct 08 18:32:08 crc kubenswrapper[4859]: E1008 18:32:08.249436 4859 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"init\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack/dnsmasq-dns-7bfcb9d745-ntml8" podUID="e66f484b-21c2-45ff-9783-f73530d8ba73" Oct 08 18:32:08 crc kubenswrapper[4859]: E1008 18:32:08.338402 4859 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/podified-antelope-centos9/openstack-neutron-server@sha256:c4e71b2158fd939dad8b8e705273493051d3023273d23b279f2699dce6db33df" Oct 08 18:32:08 crc kubenswrapper[4859]: E1008 18:32:08.338874 4859 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:init,Image:quay.io/podified-antelope-centos9/openstack-neutron-server@sha256:c4e71b2158fd939dad8b8e705273493051d3023273d23b279f2699dce6db33df,Command:[/bin/bash],Args:[-c dnsmasq --interface=* --conf-dir=/etc/dnsmasq.d --hostsdir=/etc/dnsmasq.d/hosts --keep-in-foreground --log-debug --bind-interfaces --listen-address=$(POD_IP) --port 5353 --log-facility=- --no-hosts --domain-needed --no-resolv --bogus-priv --log-queries --test],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:CONFIG_HASH,Value:ndfhb5h667h568h584h5f9h58dh565h664h587h597h577h64bh5c4h66fh647hbdh68ch5c5h68dh686h5f7h64hd7hc6h55fh57bh98h57fh87h5fh57fq,ValueFrom:nil,},EnvVar{Name:POD_IP,Value:,ValueFrom:&EnvVarSource{FieldRef:&ObjectFieldSelector{APIVersion:v1,FieldPath:status.podIP,},ResourceFieldRef:nil,ConfigMapKeyRef:nil,SecretKeyRef:nil,},},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:config,ReadOnly:true,MountPath:/etc/dnsmasq.d/config.cfg,SubPath:dns,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:dns-svc,ReadOnly:true,MountPath:/etc/dnsmasq.d/hosts/dns-svc,SubPath:dns-svc,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-hgj4x,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000650000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:&SeccompProfile{Type:RuntimeDefault,LocalhostProfile:nil,},AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod dnsmasq-dns-758b79db4c-ncfwx_openstack(1b55c542-e6b1-4fd1-9459-f5f6555f8e21): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Oct 08 18:32:08 crc kubenswrapper[4859]: E1008 18:32:08.340527 4859 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"init\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack/dnsmasq-dns-758b79db4c-ncfwx" podUID="1b55c542-e6b1-4fd1-9459-f5f6555f8e21" Oct 08 18:32:08 crc kubenswrapper[4859]: I1008 18:32:08.874487 4859 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/rabbitmq-server-0"] Oct 08 18:32:08 crc kubenswrapper[4859]: I1008 18:32:08.973058 4859 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Oct 08 18:32:09 crc kubenswrapper[4859]: W1008 18:32:09.082847 4859 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod54254e29_9af5_4b28_8664_be427c4c4f3a.slice/crio-ab6edf00f86dbbad4da7ade8f9bcefe588dafcbb7c10cc34277b15e6639b3874 WatchSource:0}: Error finding container ab6edf00f86dbbad4da7ade8f9bcefe588dafcbb7c10cc34277b15e6639b3874: Status 404 returned error can't find the container with id ab6edf00f86dbbad4da7ade8f9bcefe588dafcbb7c10cc34277b15e6639b3874 Oct 08 18:32:09 crc kubenswrapper[4859]: I1008 18:32:09.083668 4859 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/kube-state-metrics-0"] Oct 08 18:32:09 crc kubenswrapper[4859]: I1008 18:32:09.089427 4859 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/rabbitmq-cell1-server-0"] Oct 08 18:32:09 crc kubenswrapper[4859]: W1008 18:32:09.093501 4859 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod3075d5c2_3e74_4a4c_829f_248e87f45f3b.slice/crio-2a8542e26bacc34037bb28e50535483d385fada430925298e7c6d4b368c6d9c1 WatchSource:0}: Error finding container 2a8542e26bacc34037bb28e50535483d385fada430925298e7c6d4b368c6d9c1: Status 404 returned error can't find the container with id 2a8542e26bacc34037bb28e50535483d385fada430925298e7c6d4b368c6d9c1 Oct 08 18:32:09 crc kubenswrapper[4859]: I1008 18:32:09.105545 4859 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/openstack-galera-0"] Oct 08 18:32:09 crc kubenswrapper[4859]: I1008 18:32:09.108993 4859 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-7bfcb9d745-ntml8" Oct 08 18:32:09 crc kubenswrapper[4859]: I1008 18:32:09.211763 4859 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-plgwt\" (UniqueName: \"kubernetes.io/projected/e66f484b-21c2-45ff-9783-f73530d8ba73-kube-api-access-plgwt\") pod \"e66f484b-21c2-45ff-9783-f73530d8ba73\" (UID: \"e66f484b-21c2-45ff-9783-f73530d8ba73\") " Oct 08 18:32:09 crc kubenswrapper[4859]: I1008 18:32:09.211968 4859 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/e66f484b-21c2-45ff-9783-f73530d8ba73-config\") pod \"e66f484b-21c2-45ff-9783-f73530d8ba73\" (UID: \"e66f484b-21c2-45ff-9783-f73530d8ba73\") " Oct 08 18:32:09 crc kubenswrapper[4859]: I1008 18:32:09.213179 4859 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/e66f484b-21c2-45ff-9783-f73530d8ba73-config" (OuterVolumeSpecName: "config") pod "e66f484b-21c2-45ff-9783-f73530d8ba73" (UID: "e66f484b-21c2-45ff-9783-f73530d8ba73"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 08 18:32:09 crc kubenswrapper[4859]: I1008 18:32:09.222734 4859 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/e66f484b-21c2-45ff-9783-f73530d8ba73-kube-api-access-plgwt" (OuterVolumeSpecName: "kube-api-access-plgwt") pod "e66f484b-21c2-45ff-9783-f73530d8ba73" (UID: "e66f484b-21c2-45ff-9783-f73530d8ba73"). InnerVolumeSpecName "kube-api-access-plgwt". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 08 18:32:09 crc kubenswrapper[4859]: I1008 18:32:09.314155 4859 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/e66f484b-21c2-45ff-9783-f73530d8ba73-config\") on node \"crc\" DevicePath \"\"" Oct 08 18:32:09 crc kubenswrapper[4859]: I1008 18:32:09.314201 4859 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-plgwt\" (UniqueName: \"kubernetes.io/projected/e66f484b-21c2-45ff-9783-f73530d8ba73-kube-api-access-plgwt\") on node \"crc\" DevicePath \"\"" Oct 08 18:32:09 crc kubenswrapper[4859]: I1008 18:32:09.344930 4859 generic.go:334] "Generic (PLEG): container finished" podID="e78ba437-04e5-4b40-af53-8b93921fb30f" containerID="9a36a11f754ff2f1288fc63ed49280399b579d6d8cc1a95aab5c7dbeed5c193b" exitCode=0 Oct 08 18:32:09 crc kubenswrapper[4859]: I1008 18:32:09.345007 4859 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-77597f887-q5jzd" event={"ID":"e78ba437-04e5-4b40-af53-8b93921fb30f","Type":"ContainerDied","Data":"9a36a11f754ff2f1288fc63ed49280399b579d6d8cc1a95aab5c7dbeed5c193b"} Oct 08 18:32:09 crc kubenswrapper[4859]: I1008 18:32:09.346500 4859 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstack-galera-0" event={"ID":"f1828698-e403-4643-a9ab-aedf302aae70","Type":"ContainerStarted","Data":"9b44346f6f65116f6595b4f30b51f854eddc955824e054150c2ce88f197750ea"} Oct 08 18:32:09 crc kubenswrapper[4859]: I1008 18:32:09.357078 4859 generic.go:334] "Generic (PLEG): container finished" podID="834c27d2-782d-43a6-a955-e0c8b99d93a3" containerID="e91b3aae5d1d3741897c2cd6293709241da5add80bdae9705b74dd18568b6c1b" exitCode=0 Oct 08 18:32:09 crc kubenswrapper[4859]: I1008 18:32:09.357170 4859 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-644597f84c-x8r6m" event={"ID":"834c27d2-782d-43a6-a955-e0c8b99d93a3","Type":"ContainerDied","Data":"e91b3aae5d1d3741897c2cd6293709241da5add80bdae9705b74dd18568b6c1b"} Oct 08 18:32:09 crc kubenswrapper[4859]: I1008 18:32:09.359101 4859 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-cell1-server-0" event={"ID":"3075d5c2-3e74-4a4c-829f-248e87f45f3b","Type":"ContainerStarted","Data":"2a8542e26bacc34037bb28e50535483d385fada430925298e7c6d4b368c6d9c1"} Oct 08 18:32:09 crc kubenswrapper[4859]: I1008 18:32:09.360838 4859 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/kube-state-metrics-0" event={"ID":"54254e29-9af5-4b28-8664-be427c4c4f3a","Type":"ContainerStarted","Data":"ab6edf00f86dbbad4da7ade8f9bcefe588dafcbb7c10cc34277b15e6639b3874"} Oct 08 18:32:09 crc kubenswrapper[4859]: I1008 18:32:09.362605 4859 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-7bfcb9d745-ntml8" event={"ID":"e66f484b-21c2-45ff-9783-f73530d8ba73","Type":"ContainerDied","Data":"32ac335953cb8ee65305bcaaadbeaf862c442a1a5c5ad7da82a336877b0a933f"} Oct 08 18:32:09 crc kubenswrapper[4859]: I1008 18:32:09.362681 4859 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-7bfcb9d745-ntml8" Oct 08 18:32:09 crc kubenswrapper[4859]: I1008 18:32:09.367895 4859 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-server-0" event={"ID":"ba5a78be-e7b3-401a-bdb9-ba31071d7075","Type":"ContainerStarted","Data":"b85c8c35dbecff3d9510cdce48f70961ca8f93abedbf900186ddf7c96916969a"} Oct 08 18:32:09 crc kubenswrapper[4859]: I1008 18:32:09.472218 4859 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-7bfcb9d745-ntml8"] Oct 08 18:32:09 crc kubenswrapper[4859]: I1008 18:32:09.478420 4859 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-7bfcb9d745-ntml8"] Oct 08 18:32:09 crc kubenswrapper[4859]: I1008 18:32:09.529941 4859 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/memcached-0"] Oct 08 18:32:09 crc kubenswrapper[4859]: I1008 18:32:09.583432 4859 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-controller-b89lv"] Oct 08 18:32:09 crc kubenswrapper[4859]: I1008 18:32:09.603172 4859 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/openstack-cell1-galera-0"] Oct 08 18:32:09 crc kubenswrapper[4859]: W1008 18:32:09.634248 4859 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod6dc3fa5d_109b_4140_93aa_61b855c46df2.slice/crio-a94882a6f32e0ecfbed38a29edcc430cd10e6432187b04f04ac8fd178efd2b21 WatchSource:0}: Error finding container a94882a6f32e0ecfbed38a29edcc430cd10e6432187b04f04ac8fd178efd2b21: Status 404 returned error can't find the container with id a94882a6f32e0ecfbed38a29edcc430cd10e6432187b04f04ac8fd178efd2b21 Oct 08 18:32:09 crc kubenswrapper[4859]: I1008 18:32:09.644023 4859 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovsdbserver-sb-0"] Oct 08 18:32:09 crc kubenswrapper[4859]: I1008 18:32:09.770530 4859 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-controller-ovs-vthdk"] Oct 08 18:32:09 crc kubenswrapper[4859]: W1008 18:32:09.788192 4859 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podeca01cc4_b7ef_429c_984d_1ccd1d95bcd2.slice/crio-e35b3f1a6f451a6b1ae50ed8f703c988e56670201b248f23deee62fd5db96f44 WatchSource:0}: Error finding container e35b3f1a6f451a6b1ae50ed8f703c988e56670201b248f23deee62fd5db96f44: Status 404 returned error can't find the container with id e35b3f1a6f451a6b1ae50ed8f703c988e56670201b248f23deee62fd5db96f44 Oct 08 18:32:09 crc kubenswrapper[4859]: E1008 18:32:09.791845 4859 log.go:32] "CreateContainer in sandbox from runtime service failed" err=< Oct 08 18:32:09 crc kubenswrapper[4859]: rpc error: code = Unknown desc = container create failed: mount `/var/lib/kubelet/pods/834c27d2-782d-43a6-a955-e0c8b99d93a3/volume-subpaths/dns-svc/dnsmasq-dns/1` to `etc/dnsmasq.d/hosts/dns-svc`: No such file or directory Oct 08 18:32:09 crc kubenswrapper[4859]: > podSandboxID="12753bfe1617bad562af9bd2ee3c0adc6272a79a9598b983d12b4d78a5312a53" Oct 08 18:32:09 crc kubenswrapper[4859]: E1008 18:32:09.792044 4859 kuberuntime_manager.go:1274] "Unhandled Error" err=< Oct 08 18:32:09 crc kubenswrapper[4859]: container &Container{Name:dnsmasq-dns,Image:quay.io/podified-antelope-centos9/openstack-neutron-server@sha256:c4e71b2158fd939dad8b8e705273493051d3023273d23b279f2699dce6db33df,Command:[/bin/bash],Args:[-c dnsmasq --interface=* --conf-dir=/etc/dnsmasq.d --hostsdir=/etc/dnsmasq.d/hosts --keep-in-foreground --log-debug --bind-interfaces --listen-address=$(POD_IP) --port 5353 --log-facility=- --no-hosts --domain-needed --no-resolv --bogus-priv --log-queries],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:CONFIG_HASH,Value:n68chd6h679hbfh55fhc6h5ffh5d8h94h56ch589hb4hc5h57bh677hcdh655h8dh667h675h654h66ch567h8fh659h5b4h675h566h55bh54h67dh6dq,ValueFrom:nil,},EnvVar{Name:POD_IP,Value:,ValueFrom:&EnvVarSource{FieldRef:&ObjectFieldSelector{APIVersion:v1,FieldPath:status.podIP,},ResourceFieldRef:nil,ConfigMapKeyRef:nil,SecretKeyRef:nil,},},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:config,ReadOnly:true,MountPath:/etc/dnsmasq.d/config.cfg,SubPath:dns,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:dns-svc,ReadOnly:true,MountPath:/etc/dnsmasq.d/hosts/dns-svc,SubPath:dns-svc,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-rvnjq,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:nil,TCPSocket:&TCPSocketAction{Port:{0 5353 },Host:,},GRPC:nil,},InitialDelaySeconds:3,TimeoutSeconds:5,PeriodSeconds:3,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:nil,TCPSocket:&TCPSocketAction{Port:{0 5353 },Host:,},GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:5,PeriodSeconds:5,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000650000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:&SeccompProfile{Type:RuntimeDefault,LocalhostProfile:nil,},AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod dnsmasq-dns-644597f84c-x8r6m_openstack(834c27d2-782d-43a6-a955-e0c8b99d93a3): CreateContainerError: container create failed: mount `/var/lib/kubelet/pods/834c27d2-782d-43a6-a955-e0c8b99d93a3/volume-subpaths/dns-svc/dnsmasq-dns/1` to `etc/dnsmasq.d/hosts/dns-svc`: No such file or directory Oct 08 18:32:09 crc kubenswrapper[4859]: > logger="UnhandledError" Oct 08 18:32:09 crc kubenswrapper[4859]: E1008 18:32:09.793209 4859 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"dnsmasq-dns\" with CreateContainerError: \"container create failed: mount `/var/lib/kubelet/pods/834c27d2-782d-43a6-a955-e0c8b99d93a3/volume-subpaths/dns-svc/dnsmasq-dns/1` to `etc/dnsmasq.d/hosts/dns-svc`: No such file or directory\\n\"" pod="openstack/dnsmasq-dns-644597f84c-x8r6m" podUID="834c27d2-782d-43a6-a955-e0c8b99d93a3" Oct 08 18:32:09 crc kubenswrapper[4859]: I1008 18:32:09.929536 4859 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-758b79db4c-ncfwx" Oct 08 18:32:10 crc kubenswrapper[4859]: I1008 18:32:10.041791 4859 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/1b55c542-e6b1-4fd1-9459-f5f6555f8e21-dns-svc\") pod \"1b55c542-e6b1-4fd1-9459-f5f6555f8e21\" (UID: \"1b55c542-e6b1-4fd1-9459-f5f6555f8e21\") " Oct 08 18:32:10 crc kubenswrapper[4859]: I1008 18:32:10.041893 4859 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/1b55c542-e6b1-4fd1-9459-f5f6555f8e21-config\") pod \"1b55c542-e6b1-4fd1-9459-f5f6555f8e21\" (UID: \"1b55c542-e6b1-4fd1-9459-f5f6555f8e21\") " Oct 08 18:32:10 crc kubenswrapper[4859]: I1008 18:32:10.041931 4859 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-hgj4x\" (UniqueName: \"kubernetes.io/projected/1b55c542-e6b1-4fd1-9459-f5f6555f8e21-kube-api-access-hgj4x\") pod \"1b55c542-e6b1-4fd1-9459-f5f6555f8e21\" (UID: \"1b55c542-e6b1-4fd1-9459-f5f6555f8e21\") " Oct 08 18:32:10 crc kubenswrapper[4859]: I1008 18:32:10.044176 4859 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1b55c542-e6b1-4fd1-9459-f5f6555f8e21-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "1b55c542-e6b1-4fd1-9459-f5f6555f8e21" (UID: "1b55c542-e6b1-4fd1-9459-f5f6555f8e21"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 08 18:32:10 crc kubenswrapper[4859]: I1008 18:32:10.044408 4859 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1b55c542-e6b1-4fd1-9459-f5f6555f8e21-config" (OuterVolumeSpecName: "config") pod "1b55c542-e6b1-4fd1-9459-f5f6555f8e21" (UID: "1b55c542-e6b1-4fd1-9459-f5f6555f8e21"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 08 18:32:10 crc kubenswrapper[4859]: I1008 18:32:10.049641 4859 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/1b55c542-e6b1-4fd1-9459-f5f6555f8e21-kube-api-access-hgj4x" (OuterVolumeSpecName: "kube-api-access-hgj4x") pod "1b55c542-e6b1-4fd1-9459-f5f6555f8e21" (UID: "1b55c542-e6b1-4fd1-9459-f5f6555f8e21"). InnerVolumeSpecName "kube-api-access-hgj4x". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 08 18:32:10 crc kubenswrapper[4859]: I1008 18:32:10.144060 4859 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/1b55c542-e6b1-4fd1-9459-f5f6555f8e21-dns-svc\") on node \"crc\" DevicePath \"\"" Oct 08 18:32:10 crc kubenswrapper[4859]: I1008 18:32:10.144088 4859 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/1b55c542-e6b1-4fd1-9459-f5f6555f8e21-config\") on node \"crc\" DevicePath \"\"" Oct 08 18:32:10 crc kubenswrapper[4859]: I1008 18:32:10.144101 4859 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-hgj4x\" (UniqueName: \"kubernetes.io/projected/1b55c542-e6b1-4fd1-9459-f5f6555f8e21-kube-api-access-hgj4x\") on node \"crc\" DevicePath \"\"" Oct 08 18:32:10 crc kubenswrapper[4859]: I1008 18:32:10.335597 4859 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovsdbserver-nb-0"] Oct 08 18:32:10 crc kubenswrapper[4859]: I1008 18:32:10.379942 4859 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/memcached-0" event={"ID":"6dc3fa5d-109b-4140-93aa-61b855c46df2","Type":"ContainerStarted","Data":"a94882a6f32e0ecfbed38a29edcc430cd10e6432187b04f04ac8fd178efd2b21"} Oct 08 18:32:10 crc kubenswrapper[4859]: I1008 18:32:10.382890 4859 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-77597f887-q5jzd" event={"ID":"e78ba437-04e5-4b40-af53-8b93921fb30f","Type":"ContainerStarted","Data":"fb1976088f65191d06b2f0f31dffcf3d48a999de694d3b33c8208687385a83a9"} Oct 08 18:32:10 crc kubenswrapper[4859]: I1008 18:32:10.382962 4859 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-77597f887-q5jzd" Oct 08 18:32:10 crc kubenswrapper[4859]: I1008 18:32:10.384156 4859 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstack-cell1-galera-0" event={"ID":"a2c834dc-2b23-4081-8c71-e7ae462ca063","Type":"ContainerStarted","Data":"38cd6ab7b0d8b1dc308f9b53ff305f699136a22a86c34859bb86e912ab8254e6"} Oct 08 18:32:10 crc kubenswrapper[4859]: I1008 18:32:10.386121 4859 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovsdbserver-sb-0" event={"ID":"fb46e63a-80c0-4141-8e6a-ad49c95bc924","Type":"ContainerStarted","Data":"61ad851ac1ac6d12b198191c4cf0471c5d373f27ba66d0e0cabf24d9cf908083"} Oct 08 18:32:10 crc kubenswrapper[4859]: I1008 18:32:10.393063 4859 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-758b79db4c-ncfwx" Oct 08 18:32:10 crc kubenswrapper[4859]: I1008 18:32:10.393061 4859 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-758b79db4c-ncfwx" event={"ID":"1b55c542-e6b1-4fd1-9459-f5f6555f8e21","Type":"ContainerDied","Data":"2a163aafcd54edb49185c5e4b31adcbe370d9f9b9e038d4c4477f5eb92aea65f"} Oct 08 18:32:10 crc kubenswrapper[4859]: I1008 18:32:10.402940 4859 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-ovs-vthdk" event={"ID":"eca01cc4-b7ef-429c-984d-1ccd1d95bcd2","Type":"ContainerStarted","Data":"e35b3f1a6f451a6b1ae50ed8f703c988e56670201b248f23deee62fd5db96f44"} Oct 08 18:32:10 crc kubenswrapper[4859]: I1008 18:32:10.406439 4859 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-77597f887-q5jzd" podStartSLOduration=3.528790053 podStartE2EDuration="17.406417446s" podCreationTimestamp="2025-10-08 18:31:53 +0000 UTC" firstStartedPulling="2025-10-08 18:31:54.633954913 +0000 UTC m=+884.880794292" lastFinishedPulling="2025-10-08 18:32:08.511582306 +0000 UTC m=+898.758421685" observedRunningTime="2025-10-08 18:32:10.401468331 +0000 UTC m=+900.648307720" watchObservedRunningTime="2025-10-08 18:32:10.406417446 +0000 UTC m=+900.653256815" Oct 08 18:32:10 crc kubenswrapper[4859]: I1008 18:32:10.406962 4859 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-b89lv" event={"ID":"ec20820a-066c-4606-8f1b-a26b172b7720","Type":"ContainerStarted","Data":"30785daebecc7d48e5f775ad006523c3d56e6aee509c099659b780217454a62b"} Oct 08 18:32:10 crc kubenswrapper[4859]: I1008 18:32:10.489449 4859 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="e66f484b-21c2-45ff-9783-f73530d8ba73" path="/var/lib/kubelet/pods/e66f484b-21c2-45ff-9783-f73530d8ba73/volumes" Oct 08 18:32:10 crc kubenswrapper[4859]: I1008 18:32:10.513588 4859 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-758b79db4c-ncfwx"] Oct 08 18:32:10 crc kubenswrapper[4859]: I1008 18:32:10.532606 4859 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-758b79db4c-ncfwx"] Oct 08 18:32:11 crc kubenswrapper[4859]: W1008 18:32:11.232892 4859 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod9fc7c8d2_a890_43c0_8081_6049a57c7474.slice/crio-8950652999a4d877945826c8e9cec5e0c359e9b4360239b8d3f728ee91e160fb WatchSource:0}: Error finding container 8950652999a4d877945826c8e9cec5e0c359e9b4360239b8d3f728ee91e160fb: Status 404 returned error can't find the container with id 8950652999a4d877945826c8e9cec5e0c359e9b4360239b8d3f728ee91e160fb Oct 08 18:32:11 crc kubenswrapper[4859]: I1008 18:32:11.415475 4859 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovsdbserver-nb-0" event={"ID":"9fc7c8d2-a890-43c0-8081-6049a57c7474","Type":"ContainerStarted","Data":"8950652999a4d877945826c8e9cec5e0c359e9b4360239b8d3f728ee91e160fb"} Oct 08 18:32:12 crc kubenswrapper[4859]: I1008 18:32:12.483136 4859 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="1b55c542-e6b1-4fd1-9459-f5f6555f8e21" path="/var/lib/kubelet/pods/1b55c542-e6b1-4fd1-9459-f5f6555f8e21/volumes" Oct 08 18:32:14 crc kubenswrapper[4859]: I1008 18:32:14.048975 4859 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-77597f887-q5jzd" Oct 08 18:32:14 crc kubenswrapper[4859]: I1008 18:32:14.106837 4859 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-644597f84c-x8r6m"] Oct 08 18:32:20 crc kubenswrapper[4859]: I1008 18:32:20.515322 4859 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/kube-state-metrics-0" event={"ID":"54254e29-9af5-4b28-8664-be427c4c4f3a","Type":"ContainerStarted","Data":"428e4f197e15b371774b895f204f030f0983940198b04559e4433b1c1cf15bb3"} Oct 08 18:32:20 crc kubenswrapper[4859]: I1008 18:32:20.516061 4859 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/kube-state-metrics-0" Oct 08 18:32:20 crc kubenswrapper[4859]: I1008 18:32:20.519161 4859 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/memcached-0" event={"ID":"6dc3fa5d-109b-4140-93aa-61b855c46df2","Type":"ContainerStarted","Data":"881d3ccffe45656185a6456dd0a7c96f3a99a5075224e6080439da6a9559b650"} Oct 08 18:32:20 crc kubenswrapper[4859]: I1008 18:32:20.519272 4859 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/memcached-0" Oct 08 18:32:20 crc kubenswrapper[4859]: I1008 18:32:20.521577 4859 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovsdbserver-nb-0" event={"ID":"9fc7c8d2-a890-43c0-8081-6049a57c7474","Type":"ContainerStarted","Data":"228579249d641134fe9ebeb90def7096b20b7ca4a7dc92f642070c96415fa3b8"} Oct 08 18:32:20 crc kubenswrapper[4859]: I1008 18:32:20.523898 4859 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-644597f84c-x8r6m" event={"ID":"834c27d2-782d-43a6-a955-e0c8b99d93a3","Type":"ContainerStarted","Data":"64de3e763e0d977fac34a2f269c986141cb6b7725ad1d158ac25a95c6bcf0cf9"} Oct 08 18:32:20 crc kubenswrapper[4859]: I1008 18:32:20.524108 4859 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-644597f84c-x8r6m" podUID="834c27d2-782d-43a6-a955-e0c8b99d93a3" containerName="dnsmasq-dns" containerID="cri-o://64de3e763e0d977fac34a2f269c986141cb6b7725ad1d158ac25a95c6bcf0cf9" gracePeriod=10 Oct 08 18:32:20 crc kubenswrapper[4859]: I1008 18:32:20.524209 4859 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-644597f84c-x8r6m" Oct 08 18:32:20 crc kubenswrapper[4859]: I1008 18:32:20.539129 4859 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/kube-state-metrics-0" podStartSLOduration=11.141636586 podStartE2EDuration="21.539111754s" podCreationTimestamp="2025-10-08 18:31:59 +0000 UTC" firstStartedPulling="2025-10-08 18:32:09.086512158 +0000 UTC m=+899.333351527" lastFinishedPulling="2025-10-08 18:32:19.483987296 +0000 UTC m=+909.730826695" observedRunningTime="2025-10-08 18:32:20.534849228 +0000 UTC m=+910.781688607" watchObservedRunningTime="2025-10-08 18:32:20.539111754 +0000 UTC m=+910.785951133" Oct 08 18:32:20 crc kubenswrapper[4859]: I1008 18:32:20.563731 4859 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/memcached-0" podStartSLOduration=12.931231805 podStartE2EDuration="22.563680069s" podCreationTimestamp="2025-10-08 18:31:58 +0000 UTC" firstStartedPulling="2025-10-08 18:32:09.645809122 +0000 UTC m=+899.892648501" lastFinishedPulling="2025-10-08 18:32:19.278257386 +0000 UTC m=+909.525096765" observedRunningTime="2025-10-08 18:32:20.559690761 +0000 UTC m=+910.806530140" watchObservedRunningTime="2025-10-08 18:32:20.563680069 +0000 UTC m=+910.810519458" Oct 08 18:32:20 crc kubenswrapper[4859]: I1008 18:32:20.583302 4859 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-644597f84c-x8r6m" podStartSLOduration=13.486929393 podStartE2EDuration="27.583283937s" podCreationTimestamp="2025-10-08 18:31:53 +0000 UTC" firstStartedPulling="2025-10-08 18:31:54.385784479 +0000 UTC m=+884.632623858" lastFinishedPulling="2025-10-08 18:32:08.482139023 +0000 UTC m=+898.728978402" observedRunningTime="2025-10-08 18:32:20.576853667 +0000 UTC m=+910.823693066" watchObservedRunningTime="2025-10-08 18:32:20.583283937 +0000 UTC m=+910.830123326" Oct 08 18:32:21 crc kubenswrapper[4859]: I1008 18:32:21.538274 4859 generic.go:334] "Generic (PLEG): container finished" podID="834c27d2-782d-43a6-a955-e0c8b99d93a3" containerID="64de3e763e0d977fac34a2f269c986141cb6b7725ad1d158ac25a95c6bcf0cf9" exitCode=0 Oct 08 18:32:21 crc kubenswrapper[4859]: I1008 18:32:21.538376 4859 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-644597f84c-x8r6m" event={"ID":"834c27d2-782d-43a6-a955-e0c8b99d93a3","Type":"ContainerDied","Data":"64de3e763e0d977fac34a2f269c986141cb6b7725ad1d158ac25a95c6bcf0cf9"} Oct 08 18:32:24 crc kubenswrapper[4859]: I1008 18:32:24.580295 4859 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-server-0" event={"ID":"ba5a78be-e7b3-401a-bdb9-ba31071d7075","Type":"ContainerStarted","Data":"6e2de3cebf597d7492590e385a7bc16bfc073ff1e91df7b11869d97d8960f7dd"} Oct 08 18:32:24 crc kubenswrapper[4859]: I1008 18:32:24.587473 4859 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstack-cell1-galera-0" event={"ID":"a2c834dc-2b23-4081-8c71-e7ae462ca063","Type":"ContainerStarted","Data":"e0b5b460528dbdc99b56a6418321df42cf289d42107a5c3d0d87c7c8c8121671"} Oct 08 18:32:24 crc kubenswrapper[4859]: I1008 18:32:24.590358 4859 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstack-galera-0" event={"ID":"f1828698-e403-4643-a9ab-aedf302aae70","Type":"ContainerStarted","Data":"bea379aa5e1102a1a24186068dde40b8e41b27b104f98ddf880333cd0cfc98b6"} Oct 08 18:32:24 crc kubenswrapper[4859]: I1008 18:32:24.592864 4859 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovsdbserver-sb-0" event={"ID":"fb46e63a-80c0-4141-8e6a-ad49c95bc924","Type":"ContainerStarted","Data":"9e4d89765c64f79e2b3516f930996d104df3b9ad574638f1502d2043994e2b8b"} Oct 08 18:32:24 crc kubenswrapper[4859]: I1008 18:32:24.595669 4859 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-ovs-vthdk" event={"ID":"eca01cc4-b7ef-429c-984d-1ccd1d95bcd2","Type":"ContainerStarted","Data":"a6d39459bd9f030900d334eac56ce02725d08eb705029163c8c4188e6a9ce43d"} Oct 08 18:32:24 crc kubenswrapper[4859]: I1008 18:32:24.600089 4859 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-b89lv" event={"ID":"ec20820a-066c-4606-8f1b-a26b172b7720","Type":"ContainerStarted","Data":"a0a60043860efc78999d6ff3c9046a78d0fd08b67512be4daff42555d6a83d37"} Oct 08 18:32:24 crc kubenswrapper[4859]: I1008 18:32:24.600512 4859 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ovn-controller-b89lv" Oct 08 18:32:24 crc kubenswrapper[4859]: I1008 18:32:24.611390 4859 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-cell1-server-0" event={"ID":"3075d5c2-3e74-4a4c-829f-248e87f45f3b","Type":"ContainerStarted","Data":"fecccb4aab6e4695efa36ed5ba960994a10727b9aac69914dd84867213d10810"} Oct 08 18:32:24 crc kubenswrapper[4859]: I1008 18:32:24.731404 4859 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ovn-controller-b89lv" podStartSLOduration=11.629035229 podStartE2EDuration="20.731374865s" podCreationTimestamp="2025-10-08 18:32:04 +0000 UTC" firstStartedPulling="2025-10-08 18:32:09.635788858 +0000 UTC m=+899.882628237" lastFinishedPulling="2025-10-08 18:32:18.738128494 +0000 UTC m=+908.984967873" observedRunningTime="2025-10-08 18:32:24.728370816 +0000 UTC m=+914.975210205" watchObservedRunningTime="2025-10-08 18:32:24.731374865 +0000 UTC m=+914.978214244" Oct 08 18:32:24 crc kubenswrapper[4859]: I1008 18:32:24.809642 4859 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-644597f84c-x8r6m" Oct 08 18:32:24 crc kubenswrapper[4859]: I1008 18:32:24.956977 4859 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/834c27d2-782d-43a6-a955-e0c8b99d93a3-config\") pod \"834c27d2-782d-43a6-a955-e0c8b99d93a3\" (UID: \"834c27d2-782d-43a6-a955-e0c8b99d93a3\") " Oct 08 18:32:24 crc kubenswrapper[4859]: I1008 18:32:24.957082 4859 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-rvnjq\" (UniqueName: \"kubernetes.io/projected/834c27d2-782d-43a6-a955-e0c8b99d93a3-kube-api-access-rvnjq\") pod \"834c27d2-782d-43a6-a955-e0c8b99d93a3\" (UID: \"834c27d2-782d-43a6-a955-e0c8b99d93a3\") " Oct 08 18:32:24 crc kubenswrapper[4859]: I1008 18:32:24.957162 4859 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/834c27d2-782d-43a6-a955-e0c8b99d93a3-dns-svc\") pod \"834c27d2-782d-43a6-a955-e0c8b99d93a3\" (UID: \"834c27d2-782d-43a6-a955-e0c8b99d93a3\") " Oct 08 18:32:24 crc kubenswrapper[4859]: I1008 18:32:24.978662 4859 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/834c27d2-782d-43a6-a955-e0c8b99d93a3-kube-api-access-rvnjq" (OuterVolumeSpecName: "kube-api-access-rvnjq") pod "834c27d2-782d-43a6-a955-e0c8b99d93a3" (UID: "834c27d2-782d-43a6-a955-e0c8b99d93a3"). InnerVolumeSpecName "kube-api-access-rvnjq". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 08 18:32:25 crc kubenswrapper[4859]: I1008 18:32:25.000494 4859 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/834c27d2-782d-43a6-a955-e0c8b99d93a3-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "834c27d2-782d-43a6-a955-e0c8b99d93a3" (UID: "834c27d2-782d-43a6-a955-e0c8b99d93a3"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 08 18:32:25 crc kubenswrapper[4859]: I1008 18:32:25.001512 4859 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/834c27d2-782d-43a6-a955-e0c8b99d93a3-config" (OuterVolumeSpecName: "config") pod "834c27d2-782d-43a6-a955-e0c8b99d93a3" (UID: "834c27d2-782d-43a6-a955-e0c8b99d93a3"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 08 18:32:25 crc kubenswrapper[4859]: I1008 18:32:25.059282 4859 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/834c27d2-782d-43a6-a955-e0c8b99d93a3-config\") on node \"crc\" DevicePath \"\"" Oct 08 18:32:25 crc kubenswrapper[4859]: I1008 18:32:25.059588 4859 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-rvnjq\" (UniqueName: \"kubernetes.io/projected/834c27d2-782d-43a6-a955-e0c8b99d93a3-kube-api-access-rvnjq\") on node \"crc\" DevicePath \"\"" Oct 08 18:32:25 crc kubenswrapper[4859]: I1008 18:32:25.059673 4859 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/834c27d2-782d-43a6-a955-e0c8b99d93a3-dns-svc\") on node \"crc\" DevicePath \"\"" Oct 08 18:32:25 crc kubenswrapper[4859]: I1008 18:32:25.639630 4859 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-644597f84c-x8r6m" event={"ID":"834c27d2-782d-43a6-a955-e0c8b99d93a3","Type":"ContainerDied","Data":"12753bfe1617bad562af9bd2ee3c0adc6272a79a9598b983d12b4d78a5312a53"} Oct 08 18:32:25 crc kubenswrapper[4859]: I1008 18:32:25.640005 4859 scope.go:117] "RemoveContainer" containerID="64de3e763e0d977fac34a2f269c986141cb6b7725ad1d158ac25a95c6bcf0cf9" Oct 08 18:32:25 crc kubenswrapper[4859]: I1008 18:32:25.639648 4859 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-644597f84c-x8r6m" Oct 08 18:32:25 crc kubenswrapper[4859]: I1008 18:32:25.643568 4859 generic.go:334] "Generic (PLEG): container finished" podID="eca01cc4-b7ef-429c-984d-1ccd1d95bcd2" containerID="a6d39459bd9f030900d334eac56ce02725d08eb705029163c8c4188e6a9ce43d" exitCode=0 Oct 08 18:32:25 crc kubenswrapper[4859]: I1008 18:32:25.645395 4859 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-ovs-vthdk" event={"ID":"eca01cc4-b7ef-429c-984d-1ccd1d95bcd2","Type":"ContainerDied","Data":"a6d39459bd9f030900d334eac56ce02725d08eb705029163c8c4188e6a9ce43d"} Oct 08 18:32:25 crc kubenswrapper[4859]: I1008 18:32:25.688820 4859 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-644597f84c-x8r6m"] Oct 08 18:32:25 crc kubenswrapper[4859]: I1008 18:32:25.696788 4859 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-644597f84c-x8r6m"] Oct 08 18:32:26 crc kubenswrapper[4859]: I1008 18:32:26.483811 4859 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="834c27d2-782d-43a6-a955-e0c8b99d93a3" path="/var/lib/kubelet/pods/834c27d2-782d-43a6-a955-e0c8b99d93a3/volumes" Oct 08 18:32:26 crc kubenswrapper[4859]: I1008 18:32:26.976150 4859 scope.go:117] "RemoveContainer" containerID="e91b3aae5d1d3741897c2cd6293709241da5add80bdae9705b74dd18568b6c1b" Oct 08 18:32:27 crc kubenswrapper[4859]: I1008 18:32:27.524721 4859 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ovn-controller-metrics-bjb7d"] Oct 08 18:32:27 crc kubenswrapper[4859]: E1008 18:32:27.525947 4859 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="834c27d2-782d-43a6-a955-e0c8b99d93a3" containerName="dnsmasq-dns" Oct 08 18:32:27 crc kubenswrapper[4859]: I1008 18:32:27.525971 4859 state_mem.go:107] "Deleted CPUSet assignment" podUID="834c27d2-782d-43a6-a955-e0c8b99d93a3" containerName="dnsmasq-dns" Oct 08 18:32:27 crc kubenswrapper[4859]: E1008 18:32:27.526012 4859 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="834c27d2-782d-43a6-a955-e0c8b99d93a3" containerName="init" Oct 08 18:32:27 crc kubenswrapper[4859]: I1008 18:32:27.526020 4859 state_mem.go:107] "Deleted CPUSet assignment" podUID="834c27d2-782d-43a6-a955-e0c8b99d93a3" containerName="init" Oct 08 18:32:27 crc kubenswrapper[4859]: I1008 18:32:27.526167 4859 memory_manager.go:354] "RemoveStaleState removing state" podUID="834c27d2-782d-43a6-a955-e0c8b99d93a3" containerName="dnsmasq-dns" Oct 08 18:32:27 crc kubenswrapper[4859]: I1008 18:32:27.527120 4859 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-metrics-bjb7d" Oct 08 18:32:27 crc kubenswrapper[4859]: I1008 18:32:27.531893 4859 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovncontroller-metrics-config" Oct 08 18:32:27 crc kubenswrapper[4859]: I1008 18:32:27.532435 4859 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-controller-metrics-bjb7d"] Oct 08 18:32:27 crc kubenswrapper[4859]: I1008 18:32:27.603367 4859 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/f29e1ba4-3f25-45b7-ba3a-3592ce322840-config\") pod \"ovn-controller-metrics-bjb7d\" (UID: \"f29e1ba4-3f25-45b7-ba3a-3592ce322840\") " pod="openstack/ovn-controller-metrics-bjb7d" Oct 08 18:32:27 crc kubenswrapper[4859]: I1008 18:32:27.603490 4859 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/f29e1ba4-3f25-45b7-ba3a-3592ce322840-metrics-certs-tls-certs\") pod \"ovn-controller-metrics-bjb7d\" (UID: \"f29e1ba4-3f25-45b7-ba3a-3592ce322840\") " pod="openstack/ovn-controller-metrics-bjb7d" Oct 08 18:32:27 crc kubenswrapper[4859]: I1008 18:32:27.603559 4859 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovs-rundir\" (UniqueName: \"kubernetes.io/host-path/f29e1ba4-3f25-45b7-ba3a-3592ce322840-ovs-rundir\") pod \"ovn-controller-metrics-bjb7d\" (UID: \"f29e1ba4-3f25-45b7-ba3a-3592ce322840\") " pod="openstack/ovn-controller-metrics-bjb7d" Oct 08 18:32:27 crc kubenswrapper[4859]: I1008 18:32:27.603613 4859 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovn-rundir\" (UniqueName: \"kubernetes.io/host-path/f29e1ba4-3f25-45b7-ba3a-3592ce322840-ovn-rundir\") pod \"ovn-controller-metrics-bjb7d\" (UID: \"f29e1ba4-3f25-45b7-ba3a-3592ce322840\") " pod="openstack/ovn-controller-metrics-bjb7d" Oct 08 18:32:27 crc kubenswrapper[4859]: I1008 18:32:27.603992 4859 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f29e1ba4-3f25-45b7-ba3a-3592ce322840-combined-ca-bundle\") pod \"ovn-controller-metrics-bjb7d\" (UID: \"f29e1ba4-3f25-45b7-ba3a-3592ce322840\") " pod="openstack/ovn-controller-metrics-bjb7d" Oct 08 18:32:27 crc kubenswrapper[4859]: I1008 18:32:27.604062 4859 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-5dgld\" (UniqueName: \"kubernetes.io/projected/f29e1ba4-3f25-45b7-ba3a-3592ce322840-kube-api-access-5dgld\") pod \"ovn-controller-metrics-bjb7d\" (UID: \"f29e1ba4-3f25-45b7-ba3a-3592ce322840\") " pod="openstack/ovn-controller-metrics-bjb7d" Oct 08 18:32:27 crc kubenswrapper[4859]: I1008 18:32:27.662627 4859 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovsdbserver-sb-0" event={"ID":"fb46e63a-80c0-4141-8e6a-ad49c95bc924","Type":"ContainerStarted","Data":"3ce6f455584f11874b1ecb7c7d3cd9de3dc4389e8f09700cf010b7570a0ccd3a"} Oct 08 18:32:27 crc kubenswrapper[4859]: I1008 18:32:27.667108 4859 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-ovs-vthdk" event={"ID":"eca01cc4-b7ef-429c-984d-1ccd1d95bcd2","Type":"ContainerStarted","Data":"b2bbf81e80bde44dc9ff7972dd054011390fc634534df642dd4e88325591b170"} Oct 08 18:32:27 crc kubenswrapper[4859]: I1008 18:32:27.667175 4859 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-ovs-vthdk" event={"ID":"eca01cc4-b7ef-429c-984d-1ccd1d95bcd2","Type":"ContainerStarted","Data":"955c0254e2051dd433558b02c4746621c4b0acc665de49e4a0db513994df50fa"} Oct 08 18:32:27 crc kubenswrapper[4859]: I1008 18:32:27.667229 4859 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ovn-controller-ovs-vthdk" Oct 08 18:32:27 crc kubenswrapper[4859]: I1008 18:32:27.667262 4859 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ovn-controller-ovs-vthdk" Oct 08 18:32:27 crc kubenswrapper[4859]: I1008 18:32:27.669930 4859 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovsdbserver-nb-0" event={"ID":"9fc7c8d2-a890-43c0-8081-6049a57c7474","Type":"ContainerStarted","Data":"fe7a64a5ad16ca624c03107ff7be09208f2a77069a48111429dbbcc56d630eb7"} Oct 08 18:32:27 crc kubenswrapper[4859]: I1008 18:32:27.697208 4859 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ovsdbserver-sb-0" podStartSLOduration=4.280393613 podStartE2EDuration="21.697178177s" podCreationTimestamp="2025-10-08 18:32:06 +0000 UTC" firstStartedPulling="2025-10-08 18:32:09.664138759 +0000 UTC m=+899.910978138" lastFinishedPulling="2025-10-08 18:32:27.080923323 +0000 UTC m=+917.327762702" observedRunningTime="2025-10-08 18:32:27.693469047 +0000 UTC m=+917.940308446" watchObservedRunningTime="2025-10-08 18:32:27.697178177 +0000 UTC m=+917.944017556" Oct 08 18:32:27 crc kubenswrapper[4859]: I1008 18:32:27.706168 4859 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovn-rundir\" (UniqueName: \"kubernetes.io/host-path/f29e1ba4-3f25-45b7-ba3a-3592ce322840-ovn-rundir\") pod \"ovn-controller-metrics-bjb7d\" (UID: \"f29e1ba4-3f25-45b7-ba3a-3592ce322840\") " pod="openstack/ovn-controller-metrics-bjb7d" Oct 08 18:32:27 crc kubenswrapper[4859]: I1008 18:32:27.706344 4859 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f29e1ba4-3f25-45b7-ba3a-3592ce322840-combined-ca-bundle\") pod \"ovn-controller-metrics-bjb7d\" (UID: \"f29e1ba4-3f25-45b7-ba3a-3592ce322840\") " pod="openstack/ovn-controller-metrics-bjb7d" Oct 08 18:32:27 crc kubenswrapper[4859]: I1008 18:32:27.706392 4859 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-5dgld\" (UniqueName: \"kubernetes.io/projected/f29e1ba4-3f25-45b7-ba3a-3592ce322840-kube-api-access-5dgld\") pod \"ovn-controller-metrics-bjb7d\" (UID: \"f29e1ba4-3f25-45b7-ba3a-3592ce322840\") " pod="openstack/ovn-controller-metrics-bjb7d" Oct 08 18:32:27 crc kubenswrapper[4859]: I1008 18:32:27.706516 4859 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/f29e1ba4-3f25-45b7-ba3a-3592ce322840-config\") pod \"ovn-controller-metrics-bjb7d\" (UID: \"f29e1ba4-3f25-45b7-ba3a-3592ce322840\") " pod="openstack/ovn-controller-metrics-bjb7d" Oct 08 18:32:27 crc kubenswrapper[4859]: I1008 18:32:27.706560 4859 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/f29e1ba4-3f25-45b7-ba3a-3592ce322840-metrics-certs-tls-certs\") pod \"ovn-controller-metrics-bjb7d\" (UID: \"f29e1ba4-3f25-45b7-ba3a-3592ce322840\") " pod="openstack/ovn-controller-metrics-bjb7d" Oct 08 18:32:27 crc kubenswrapper[4859]: I1008 18:32:27.706659 4859 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovs-rundir\" (UniqueName: \"kubernetes.io/host-path/f29e1ba4-3f25-45b7-ba3a-3592ce322840-ovs-rundir\") pod \"ovn-controller-metrics-bjb7d\" (UID: \"f29e1ba4-3f25-45b7-ba3a-3592ce322840\") " pod="openstack/ovn-controller-metrics-bjb7d" Oct 08 18:32:27 crc kubenswrapper[4859]: I1008 18:32:27.707118 4859 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovs-rundir\" (UniqueName: \"kubernetes.io/host-path/f29e1ba4-3f25-45b7-ba3a-3592ce322840-ovs-rundir\") pod \"ovn-controller-metrics-bjb7d\" (UID: \"f29e1ba4-3f25-45b7-ba3a-3592ce322840\") " pod="openstack/ovn-controller-metrics-bjb7d" Oct 08 18:32:27 crc kubenswrapper[4859]: I1008 18:32:27.707980 4859 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovn-rundir\" (UniqueName: \"kubernetes.io/host-path/f29e1ba4-3f25-45b7-ba3a-3592ce322840-ovn-rundir\") pod \"ovn-controller-metrics-bjb7d\" (UID: \"f29e1ba4-3f25-45b7-ba3a-3592ce322840\") " pod="openstack/ovn-controller-metrics-bjb7d" Oct 08 18:32:27 crc kubenswrapper[4859]: I1008 18:32:27.710126 4859 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/f29e1ba4-3f25-45b7-ba3a-3592ce322840-config\") pod \"ovn-controller-metrics-bjb7d\" (UID: \"f29e1ba4-3f25-45b7-ba3a-3592ce322840\") " pod="openstack/ovn-controller-metrics-bjb7d" Oct 08 18:32:27 crc kubenswrapper[4859]: I1008 18:32:27.715627 4859 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f29e1ba4-3f25-45b7-ba3a-3592ce322840-combined-ca-bundle\") pod \"ovn-controller-metrics-bjb7d\" (UID: \"f29e1ba4-3f25-45b7-ba3a-3592ce322840\") " pod="openstack/ovn-controller-metrics-bjb7d" Oct 08 18:32:27 crc kubenswrapper[4859]: I1008 18:32:27.728410 4859 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/f29e1ba4-3f25-45b7-ba3a-3592ce322840-metrics-certs-tls-certs\") pod \"ovn-controller-metrics-bjb7d\" (UID: \"f29e1ba4-3f25-45b7-ba3a-3592ce322840\") " pod="openstack/ovn-controller-metrics-bjb7d" Oct 08 18:32:27 crc kubenswrapper[4859]: I1008 18:32:27.737779 4859 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-f6b595d95-k798k"] Oct 08 18:32:27 crc kubenswrapper[4859]: I1008 18:32:27.742423 4859 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-f6b595d95-k798k" Oct 08 18:32:27 crc kubenswrapper[4859]: I1008 18:32:27.750120 4859 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovsdbserver-nb" Oct 08 18:32:27 crc kubenswrapper[4859]: I1008 18:32:27.754060 4859 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ovn-controller-ovs-vthdk" podStartSLOduration=14.815070647 podStartE2EDuration="23.754031594s" podCreationTimestamp="2025-10-08 18:32:04 +0000 UTC" firstStartedPulling="2025-10-08 18:32:09.794220382 +0000 UTC m=+900.041059761" lastFinishedPulling="2025-10-08 18:32:18.733181319 +0000 UTC m=+908.980020708" observedRunningTime="2025-10-08 18:32:27.726478421 +0000 UTC m=+917.973317810" watchObservedRunningTime="2025-10-08 18:32:27.754031594 +0000 UTC m=+918.000870983" Oct 08 18:32:27 crc kubenswrapper[4859]: I1008 18:32:27.775519 4859 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-5dgld\" (UniqueName: \"kubernetes.io/projected/f29e1ba4-3f25-45b7-ba3a-3592ce322840-kube-api-access-5dgld\") pod \"ovn-controller-metrics-bjb7d\" (UID: \"f29e1ba4-3f25-45b7-ba3a-3592ce322840\") " pod="openstack/ovn-controller-metrics-bjb7d" Oct 08 18:32:27 crc kubenswrapper[4859]: I1008 18:32:27.794060 4859 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-f6b595d95-k798k"] Oct 08 18:32:27 crc kubenswrapper[4859]: I1008 18:32:27.806534 4859 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ovsdbserver-nb-0" podStartSLOduration=8.978052754 podStartE2EDuration="24.806512461s" podCreationTimestamp="2025-10-08 18:32:03 +0000 UTC" firstStartedPulling="2025-10-08 18:32:11.23596336 +0000 UTC m=+901.482802739" lastFinishedPulling="2025-10-08 18:32:27.064423067 +0000 UTC m=+917.311262446" observedRunningTime="2025-10-08 18:32:27.761199515 +0000 UTC m=+918.008038894" watchObservedRunningTime="2025-10-08 18:32:27.806512461 +0000 UTC m=+918.053351840" Oct 08 18:32:27 crc kubenswrapper[4859]: I1008 18:32:27.811788 4859 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/4a8089d5-f64f-4e08-bbef-e929819bb0b0-dns-svc\") pod \"dnsmasq-dns-f6b595d95-k798k\" (UID: \"4a8089d5-f64f-4e08-bbef-e929819bb0b0\") " pod="openstack/dnsmasq-dns-f6b595d95-k798k" Oct 08 18:32:27 crc kubenswrapper[4859]: I1008 18:32:27.812130 4859 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/4a8089d5-f64f-4e08-bbef-e929819bb0b0-ovsdbserver-nb\") pod \"dnsmasq-dns-f6b595d95-k798k\" (UID: \"4a8089d5-f64f-4e08-bbef-e929819bb0b0\") " pod="openstack/dnsmasq-dns-f6b595d95-k798k" Oct 08 18:32:27 crc kubenswrapper[4859]: I1008 18:32:27.812313 4859 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-c6vkw\" (UniqueName: \"kubernetes.io/projected/4a8089d5-f64f-4e08-bbef-e929819bb0b0-kube-api-access-c6vkw\") pod \"dnsmasq-dns-f6b595d95-k798k\" (UID: \"4a8089d5-f64f-4e08-bbef-e929819bb0b0\") " pod="openstack/dnsmasq-dns-f6b595d95-k798k" Oct 08 18:32:27 crc kubenswrapper[4859]: I1008 18:32:27.812459 4859 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/4a8089d5-f64f-4e08-bbef-e929819bb0b0-config\") pod \"dnsmasq-dns-f6b595d95-k798k\" (UID: \"4a8089d5-f64f-4e08-bbef-e929819bb0b0\") " pod="openstack/dnsmasq-dns-f6b595d95-k798k" Oct 08 18:32:27 crc kubenswrapper[4859]: I1008 18:32:27.844412 4859 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-metrics-bjb7d" Oct 08 18:32:27 crc kubenswrapper[4859]: I1008 18:32:27.914996 4859 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/4a8089d5-f64f-4e08-bbef-e929819bb0b0-dns-svc\") pod \"dnsmasq-dns-f6b595d95-k798k\" (UID: \"4a8089d5-f64f-4e08-bbef-e929819bb0b0\") " pod="openstack/dnsmasq-dns-f6b595d95-k798k" Oct 08 18:32:27 crc kubenswrapper[4859]: I1008 18:32:27.916035 4859 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/4a8089d5-f64f-4e08-bbef-e929819bb0b0-ovsdbserver-nb\") pod \"dnsmasq-dns-f6b595d95-k798k\" (UID: \"4a8089d5-f64f-4e08-bbef-e929819bb0b0\") " pod="openstack/dnsmasq-dns-f6b595d95-k798k" Oct 08 18:32:27 crc kubenswrapper[4859]: I1008 18:32:27.915798 4859 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-f6b595d95-k798k"] Oct 08 18:32:27 crc kubenswrapper[4859]: I1008 18:32:27.915987 4859 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/4a8089d5-f64f-4e08-bbef-e929819bb0b0-dns-svc\") pod \"dnsmasq-dns-f6b595d95-k798k\" (UID: \"4a8089d5-f64f-4e08-bbef-e929819bb0b0\") " pod="openstack/dnsmasq-dns-f6b595d95-k798k" Oct 08 18:32:27 crc kubenswrapper[4859]: E1008 18:32:27.916310 4859 pod_workers.go:1301] "Error syncing pod, skipping" err="unmounted volumes=[config dns-svc kube-api-access-c6vkw ovsdbserver-nb], unattached volumes=[], failed to process volumes=[]: context canceled" pod="openstack/dnsmasq-dns-f6b595d95-k798k" podUID="4a8089d5-f64f-4e08-bbef-e929819bb0b0" Oct 08 18:32:27 crc kubenswrapper[4859]: I1008 18:32:27.916393 4859 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-c6vkw\" (UniqueName: \"kubernetes.io/projected/4a8089d5-f64f-4e08-bbef-e929819bb0b0-kube-api-access-c6vkw\") pod \"dnsmasq-dns-f6b595d95-k798k\" (UID: \"4a8089d5-f64f-4e08-bbef-e929819bb0b0\") " pod="openstack/dnsmasq-dns-f6b595d95-k798k" Oct 08 18:32:27 crc kubenswrapper[4859]: I1008 18:32:27.916755 4859 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/4a8089d5-f64f-4e08-bbef-e929819bb0b0-config\") pod \"dnsmasq-dns-f6b595d95-k798k\" (UID: \"4a8089d5-f64f-4e08-bbef-e929819bb0b0\") " pod="openstack/dnsmasq-dns-f6b595d95-k798k" Oct 08 18:32:27 crc kubenswrapper[4859]: I1008 18:32:27.916866 4859 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/4a8089d5-f64f-4e08-bbef-e929819bb0b0-ovsdbserver-nb\") pod \"dnsmasq-dns-f6b595d95-k798k\" (UID: \"4a8089d5-f64f-4e08-bbef-e929819bb0b0\") " pod="openstack/dnsmasq-dns-f6b595d95-k798k" Oct 08 18:32:27 crc kubenswrapper[4859]: I1008 18:32:27.917308 4859 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/4a8089d5-f64f-4e08-bbef-e929819bb0b0-config\") pod \"dnsmasq-dns-f6b595d95-k798k\" (UID: \"4a8089d5-f64f-4e08-bbef-e929819bb0b0\") " pod="openstack/dnsmasq-dns-f6b595d95-k798k" Oct 08 18:32:27 crc kubenswrapper[4859]: I1008 18:32:27.940196 4859 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-c6vkw\" (UniqueName: \"kubernetes.io/projected/4a8089d5-f64f-4e08-bbef-e929819bb0b0-kube-api-access-c6vkw\") pod \"dnsmasq-dns-f6b595d95-k798k\" (UID: \"4a8089d5-f64f-4e08-bbef-e929819bb0b0\") " pod="openstack/dnsmasq-dns-f6b595d95-k798k" Oct 08 18:32:27 crc kubenswrapper[4859]: I1008 18:32:27.940610 4859 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-dc9d58d7-tdt5s"] Oct 08 18:32:27 crc kubenswrapper[4859]: I1008 18:32:27.942666 4859 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-dc9d58d7-tdt5s" Oct 08 18:32:27 crc kubenswrapper[4859]: I1008 18:32:27.945495 4859 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovsdbserver-sb" Oct 08 18:32:27 crc kubenswrapper[4859]: I1008 18:32:27.973330 4859 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-dc9d58d7-tdt5s"] Oct 08 18:32:28 crc kubenswrapper[4859]: I1008 18:32:28.019530 4859 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/6b7754e8-93c9-4b3a-983c-1127f73459b6-dns-svc\") pod \"dnsmasq-dns-dc9d58d7-tdt5s\" (UID: \"6b7754e8-93c9-4b3a-983c-1127f73459b6\") " pod="openstack/dnsmasq-dns-dc9d58d7-tdt5s" Oct 08 18:32:28 crc kubenswrapper[4859]: I1008 18:32:28.019601 4859 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/6b7754e8-93c9-4b3a-983c-1127f73459b6-ovsdbserver-nb\") pod \"dnsmasq-dns-dc9d58d7-tdt5s\" (UID: \"6b7754e8-93c9-4b3a-983c-1127f73459b6\") " pod="openstack/dnsmasq-dns-dc9d58d7-tdt5s" Oct 08 18:32:28 crc kubenswrapper[4859]: I1008 18:32:28.019635 4859 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-5qjcb\" (UniqueName: \"kubernetes.io/projected/6b7754e8-93c9-4b3a-983c-1127f73459b6-kube-api-access-5qjcb\") pod \"dnsmasq-dns-dc9d58d7-tdt5s\" (UID: \"6b7754e8-93c9-4b3a-983c-1127f73459b6\") " pod="openstack/dnsmasq-dns-dc9d58d7-tdt5s" Oct 08 18:32:28 crc kubenswrapper[4859]: I1008 18:32:28.019660 4859 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/6b7754e8-93c9-4b3a-983c-1127f73459b6-config\") pod \"dnsmasq-dns-dc9d58d7-tdt5s\" (UID: \"6b7754e8-93c9-4b3a-983c-1127f73459b6\") " pod="openstack/dnsmasq-dns-dc9d58d7-tdt5s" Oct 08 18:32:28 crc kubenswrapper[4859]: I1008 18:32:28.019728 4859 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/6b7754e8-93c9-4b3a-983c-1127f73459b6-ovsdbserver-sb\") pod \"dnsmasq-dns-dc9d58d7-tdt5s\" (UID: \"6b7754e8-93c9-4b3a-983c-1127f73459b6\") " pod="openstack/dnsmasq-dns-dc9d58d7-tdt5s" Oct 08 18:32:28 crc kubenswrapper[4859]: I1008 18:32:28.121514 4859 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/6b7754e8-93c9-4b3a-983c-1127f73459b6-dns-svc\") pod \"dnsmasq-dns-dc9d58d7-tdt5s\" (UID: \"6b7754e8-93c9-4b3a-983c-1127f73459b6\") " pod="openstack/dnsmasq-dns-dc9d58d7-tdt5s" Oct 08 18:32:28 crc kubenswrapper[4859]: I1008 18:32:28.121569 4859 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/6b7754e8-93c9-4b3a-983c-1127f73459b6-ovsdbserver-nb\") pod \"dnsmasq-dns-dc9d58d7-tdt5s\" (UID: \"6b7754e8-93c9-4b3a-983c-1127f73459b6\") " pod="openstack/dnsmasq-dns-dc9d58d7-tdt5s" Oct 08 18:32:28 crc kubenswrapper[4859]: I1008 18:32:28.121587 4859 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-5qjcb\" (UniqueName: \"kubernetes.io/projected/6b7754e8-93c9-4b3a-983c-1127f73459b6-kube-api-access-5qjcb\") pod \"dnsmasq-dns-dc9d58d7-tdt5s\" (UID: \"6b7754e8-93c9-4b3a-983c-1127f73459b6\") " pod="openstack/dnsmasq-dns-dc9d58d7-tdt5s" Oct 08 18:32:28 crc kubenswrapper[4859]: I1008 18:32:28.121606 4859 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/6b7754e8-93c9-4b3a-983c-1127f73459b6-config\") pod \"dnsmasq-dns-dc9d58d7-tdt5s\" (UID: \"6b7754e8-93c9-4b3a-983c-1127f73459b6\") " pod="openstack/dnsmasq-dns-dc9d58d7-tdt5s" Oct 08 18:32:28 crc kubenswrapper[4859]: I1008 18:32:28.121636 4859 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/6b7754e8-93c9-4b3a-983c-1127f73459b6-ovsdbserver-sb\") pod \"dnsmasq-dns-dc9d58d7-tdt5s\" (UID: \"6b7754e8-93c9-4b3a-983c-1127f73459b6\") " pod="openstack/dnsmasq-dns-dc9d58d7-tdt5s" Oct 08 18:32:28 crc kubenswrapper[4859]: I1008 18:32:28.123388 4859 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/6b7754e8-93c9-4b3a-983c-1127f73459b6-ovsdbserver-nb\") pod \"dnsmasq-dns-dc9d58d7-tdt5s\" (UID: \"6b7754e8-93c9-4b3a-983c-1127f73459b6\") " pod="openstack/dnsmasq-dns-dc9d58d7-tdt5s" Oct 08 18:32:28 crc kubenswrapper[4859]: I1008 18:32:28.123597 4859 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/6b7754e8-93c9-4b3a-983c-1127f73459b6-config\") pod \"dnsmasq-dns-dc9d58d7-tdt5s\" (UID: \"6b7754e8-93c9-4b3a-983c-1127f73459b6\") " pod="openstack/dnsmasq-dns-dc9d58d7-tdt5s" Oct 08 18:32:28 crc kubenswrapper[4859]: I1008 18:32:28.123607 4859 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/6b7754e8-93c9-4b3a-983c-1127f73459b6-dns-svc\") pod \"dnsmasq-dns-dc9d58d7-tdt5s\" (UID: \"6b7754e8-93c9-4b3a-983c-1127f73459b6\") " pod="openstack/dnsmasq-dns-dc9d58d7-tdt5s" Oct 08 18:32:28 crc kubenswrapper[4859]: I1008 18:32:28.124117 4859 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/6b7754e8-93c9-4b3a-983c-1127f73459b6-ovsdbserver-sb\") pod \"dnsmasq-dns-dc9d58d7-tdt5s\" (UID: \"6b7754e8-93c9-4b3a-983c-1127f73459b6\") " pod="openstack/dnsmasq-dns-dc9d58d7-tdt5s" Oct 08 18:32:28 crc kubenswrapper[4859]: I1008 18:32:28.153110 4859 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-5qjcb\" (UniqueName: \"kubernetes.io/projected/6b7754e8-93c9-4b3a-983c-1127f73459b6-kube-api-access-5qjcb\") pod \"dnsmasq-dns-dc9d58d7-tdt5s\" (UID: \"6b7754e8-93c9-4b3a-983c-1127f73459b6\") " pod="openstack/dnsmasq-dns-dc9d58d7-tdt5s" Oct 08 18:32:28 crc kubenswrapper[4859]: I1008 18:32:28.265548 4859 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-dc9d58d7-tdt5s" Oct 08 18:32:28 crc kubenswrapper[4859]: I1008 18:32:28.390103 4859 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-controller-metrics-bjb7d"] Oct 08 18:32:28 crc kubenswrapper[4859]: I1008 18:32:28.569889 4859 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/ovsdbserver-sb-0" Oct 08 18:32:28 crc kubenswrapper[4859]: I1008 18:32:28.619772 4859 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/ovsdbserver-sb-0" Oct 08 18:32:28 crc kubenswrapper[4859]: I1008 18:32:28.678706 4859 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-metrics-bjb7d" event={"ID":"f29e1ba4-3f25-45b7-ba3a-3592ce322840","Type":"ContainerStarted","Data":"9bdaa2f93020fa4ddf213ff7bdbfbcd6ef84e0695807b2ef264d9733eb44eeb9"} Oct 08 18:32:28 crc kubenswrapper[4859]: I1008 18:32:28.678763 4859 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-metrics-bjb7d" event={"ID":"f29e1ba4-3f25-45b7-ba3a-3592ce322840","Type":"ContainerStarted","Data":"306cf8b330904e3af7d5ff64f32709f8b45a105f200ee36f77439de46c72db89"} Oct 08 18:32:28 crc kubenswrapper[4859]: I1008 18:32:28.680481 4859 generic.go:334] "Generic (PLEG): container finished" podID="a2c834dc-2b23-4081-8c71-e7ae462ca063" containerID="e0b5b460528dbdc99b56a6418321df42cf289d42107a5c3d0d87c7c8c8121671" exitCode=0 Oct 08 18:32:28 crc kubenswrapper[4859]: I1008 18:32:28.680547 4859 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstack-cell1-galera-0" event={"ID":"a2c834dc-2b23-4081-8c71-e7ae462ca063","Type":"ContainerDied","Data":"e0b5b460528dbdc99b56a6418321df42cf289d42107a5c3d0d87c7c8c8121671"} Oct 08 18:32:28 crc kubenswrapper[4859]: I1008 18:32:28.681824 4859 generic.go:334] "Generic (PLEG): container finished" podID="f1828698-e403-4643-a9ab-aedf302aae70" containerID="bea379aa5e1102a1a24186068dde40b8e41b27b104f98ddf880333cd0cfc98b6" exitCode=0 Oct 08 18:32:28 crc kubenswrapper[4859]: I1008 18:32:28.682280 4859 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstack-galera-0" event={"ID":"f1828698-e403-4643-a9ab-aedf302aae70","Type":"ContainerDied","Data":"bea379aa5e1102a1a24186068dde40b8e41b27b104f98ddf880333cd0cfc98b6"} Oct 08 18:32:28 crc kubenswrapper[4859]: I1008 18:32:28.682500 4859 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-f6b595d95-k798k" Oct 08 18:32:28 crc kubenswrapper[4859]: I1008 18:32:28.682813 4859 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ovsdbserver-sb-0" Oct 08 18:32:28 crc kubenswrapper[4859]: I1008 18:32:28.700863 4859 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ovn-controller-metrics-bjb7d" podStartSLOduration=1.7008370959999999 podStartE2EDuration="1.700837096s" podCreationTimestamp="2025-10-08 18:32:27 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-08 18:32:28.697459706 +0000 UTC m=+918.944299085" watchObservedRunningTime="2025-10-08 18:32:28.700837096 +0000 UTC m=+918.947676475" Oct 08 18:32:28 crc kubenswrapper[4859]: I1008 18:32:28.701675 4859 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-f6b595d95-k798k" Oct 08 18:32:28 crc kubenswrapper[4859]: I1008 18:32:28.726345 4859 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-dc9d58d7-tdt5s"] Oct 08 18:32:28 crc kubenswrapper[4859]: I1008 18:32:28.832555 4859 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/4a8089d5-f64f-4e08-bbef-e929819bb0b0-dns-svc\") pod \"4a8089d5-f64f-4e08-bbef-e929819bb0b0\" (UID: \"4a8089d5-f64f-4e08-bbef-e929819bb0b0\") " Oct 08 18:32:28 crc kubenswrapper[4859]: I1008 18:32:28.832632 4859 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/4a8089d5-f64f-4e08-bbef-e929819bb0b0-config\") pod \"4a8089d5-f64f-4e08-bbef-e929819bb0b0\" (UID: \"4a8089d5-f64f-4e08-bbef-e929819bb0b0\") " Oct 08 18:32:28 crc kubenswrapper[4859]: I1008 18:32:28.832679 4859 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-c6vkw\" (UniqueName: \"kubernetes.io/projected/4a8089d5-f64f-4e08-bbef-e929819bb0b0-kube-api-access-c6vkw\") pod \"4a8089d5-f64f-4e08-bbef-e929819bb0b0\" (UID: \"4a8089d5-f64f-4e08-bbef-e929819bb0b0\") " Oct 08 18:32:28 crc kubenswrapper[4859]: I1008 18:32:28.832799 4859 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/4a8089d5-f64f-4e08-bbef-e929819bb0b0-ovsdbserver-nb\") pod \"4a8089d5-f64f-4e08-bbef-e929819bb0b0\" (UID: \"4a8089d5-f64f-4e08-bbef-e929819bb0b0\") " Oct 08 18:32:28 crc kubenswrapper[4859]: I1008 18:32:28.833127 4859 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/4a8089d5-f64f-4e08-bbef-e929819bb0b0-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "4a8089d5-f64f-4e08-bbef-e929819bb0b0" (UID: "4a8089d5-f64f-4e08-bbef-e929819bb0b0"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 08 18:32:28 crc kubenswrapper[4859]: I1008 18:32:28.833139 4859 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/4a8089d5-f64f-4e08-bbef-e929819bb0b0-config" (OuterVolumeSpecName: "config") pod "4a8089d5-f64f-4e08-bbef-e929819bb0b0" (UID: "4a8089d5-f64f-4e08-bbef-e929819bb0b0"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 08 18:32:28 crc kubenswrapper[4859]: I1008 18:32:28.833352 4859 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/4a8089d5-f64f-4e08-bbef-e929819bb0b0-dns-svc\") on node \"crc\" DevicePath \"\"" Oct 08 18:32:28 crc kubenswrapper[4859]: I1008 18:32:28.833647 4859 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/4a8089d5-f64f-4e08-bbef-e929819bb0b0-config\") on node \"crc\" DevicePath \"\"" Oct 08 18:32:28 crc kubenswrapper[4859]: I1008 18:32:28.833896 4859 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/4a8089d5-f64f-4e08-bbef-e929819bb0b0-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "4a8089d5-f64f-4e08-bbef-e929819bb0b0" (UID: "4a8089d5-f64f-4e08-bbef-e929819bb0b0"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 08 18:32:28 crc kubenswrapper[4859]: I1008 18:32:28.840702 4859 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/4a8089d5-f64f-4e08-bbef-e929819bb0b0-kube-api-access-c6vkw" (OuterVolumeSpecName: "kube-api-access-c6vkw") pod "4a8089d5-f64f-4e08-bbef-e929819bb0b0" (UID: "4a8089d5-f64f-4e08-bbef-e929819bb0b0"). InnerVolumeSpecName "kube-api-access-c6vkw". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 08 18:32:28 crc kubenswrapper[4859]: I1008 18:32:28.935073 4859 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/4a8089d5-f64f-4e08-bbef-e929819bb0b0-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Oct 08 18:32:28 crc kubenswrapper[4859]: I1008 18:32:28.935450 4859 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-c6vkw\" (UniqueName: \"kubernetes.io/projected/4a8089d5-f64f-4e08-bbef-e929819bb0b0-kube-api-access-c6vkw\") on node \"crc\" DevicePath \"\"" Oct 08 18:32:28 crc kubenswrapper[4859]: I1008 18:32:28.940384 4859 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/ovsdbserver-nb-0" Oct 08 18:32:28 crc kubenswrapper[4859]: I1008 18:32:28.976681 4859 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/ovsdbserver-nb-0" Oct 08 18:32:29 crc kubenswrapper[4859]: I1008 18:32:29.692053 4859 generic.go:334] "Generic (PLEG): container finished" podID="6b7754e8-93c9-4b3a-983c-1127f73459b6" containerID="6405e0cc748bc4ff427676070a865dceab5700fd5524c4a4eb073ddf9a15f340" exitCode=0 Oct 08 18:32:29 crc kubenswrapper[4859]: I1008 18:32:29.692135 4859 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-dc9d58d7-tdt5s" event={"ID":"6b7754e8-93c9-4b3a-983c-1127f73459b6","Type":"ContainerDied","Data":"6405e0cc748bc4ff427676070a865dceab5700fd5524c4a4eb073ddf9a15f340"} Oct 08 18:32:29 crc kubenswrapper[4859]: I1008 18:32:29.692189 4859 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-dc9d58d7-tdt5s" event={"ID":"6b7754e8-93c9-4b3a-983c-1127f73459b6","Type":"ContainerStarted","Data":"e13f728449ca1682232a1ba8c3c50ec1f4b97b099ebd8d0a9272b3dd169cb997"} Oct 08 18:32:29 crc kubenswrapper[4859]: I1008 18:32:29.694585 4859 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstack-cell1-galera-0" event={"ID":"a2c834dc-2b23-4081-8c71-e7ae462ca063","Type":"ContainerStarted","Data":"a3722214652f53d0d075a3e95392683551d728193e9e37308ed7781033c1161b"} Oct 08 18:32:29 crc kubenswrapper[4859]: I1008 18:32:29.696798 4859 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-f6b595d95-k798k" Oct 08 18:32:29 crc kubenswrapper[4859]: I1008 18:32:29.696865 4859 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstack-galera-0" event={"ID":"f1828698-e403-4643-a9ab-aedf302aae70","Type":"ContainerStarted","Data":"f8de0ba12d4440830d922654ccdddda6a7baa9aac4849dfc21a22cd482cbce7e"} Oct 08 18:32:29 crc kubenswrapper[4859]: I1008 18:32:29.700216 4859 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ovsdbserver-nb-0" Oct 08 18:32:29 crc kubenswrapper[4859]: I1008 18:32:29.760528 4859 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/ovsdbserver-nb-0" Oct 08 18:32:29 crc kubenswrapper[4859]: I1008 18:32:29.761143 4859 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/ovsdbserver-sb-0" Oct 08 18:32:29 crc kubenswrapper[4859]: I1008 18:32:29.777061 4859 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/openstack-galera-0" podStartSLOduration=23.619665394 podStartE2EDuration="33.777038273s" podCreationTimestamp="2025-10-08 18:31:56 +0000 UTC" firstStartedPulling="2025-10-08 18:32:09.106110623 +0000 UTC m=+899.352950012" lastFinishedPulling="2025-10-08 18:32:19.263483512 +0000 UTC m=+909.510322891" observedRunningTime="2025-10-08 18:32:29.765514363 +0000 UTC m=+920.012353762" watchObservedRunningTime="2025-10-08 18:32:29.777038273 +0000 UTC m=+920.023877652" Oct 08 18:32:29 crc kubenswrapper[4859]: I1008 18:32:29.806497 4859 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/openstack-cell1-galera-0" podStartSLOduration=24.705907017 podStartE2EDuration="33.806476751s" podCreationTimestamp="2025-10-08 18:31:56 +0000 UTC" firstStartedPulling="2025-10-08 18:32:09.621455808 +0000 UTC m=+899.868295187" lastFinishedPulling="2025-10-08 18:32:18.722025542 +0000 UTC m=+908.968864921" observedRunningTime="2025-10-08 18:32:29.795092495 +0000 UTC m=+920.041931874" watchObservedRunningTime="2025-10-08 18:32:29.806476751 +0000 UTC m=+920.053316130" Oct 08 18:32:29 crc kubenswrapper[4859]: I1008 18:32:29.883085 4859 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/memcached-0" Oct 08 18:32:29 crc kubenswrapper[4859]: I1008 18:32:29.940057 4859 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-f6b595d95-k798k"] Oct 08 18:32:29 crc kubenswrapper[4859]: I1008 18:32:29.945322 4859 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-f6b595d95-k798k"] Oct 08 18:32:30 crc kubenswrapper[4859]: I1008 18:32:30.067984 4859 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/kube-state-metrics-0" Oct 08 18:32:30 crc kubenswrapper[4859]: I1008 18:32:30.182201 4859 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ovn-northd-0"] Oct 08 18:32:30 crc kubenswrapper[4859]: I1008 18:32:30.183726 4859 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-northd-0" Oct 08 18:32:30 crc kubenswrapper[4859]: I1008 18:32:30.192135 4859 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovnnorthd-scripts" Oct 08 18:32:30 crc kubenswrapper[4859]: I1008 18:32:30.192326 4859 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-ovnnorthd-ovndbs" Oct 08 18:32:30 crc kubenswrapper[4859]: I1008 18:32:30.192493 4859 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovnnorthd-config" Oct 08 18:32:30 crc kubenswrapper[4859]: I1008 18:32:30.192521 4859 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ovnnorthd-ovnnorthd-dockercfg-llr4p" Oct 08 18:32:30 crc kubenswrapper[4859]: I1008 18:32:30.206574 4859 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-northd-0"] Oct 08 18:32:30 crc kubenswrapper[4859]: I1008 18:32:30.262153 4859 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3d31b043-3252-458e-8f23-afd68b624849-combined-ca-bundle\") pod \"ovn-northd-0\" (UID: \"3d31b043-3252-458e-8f23-afd68b624849\") " pod="openstack/ovn-northd-0" Oct 08 18:32:30 crc kubenswrapper[4859]: I1008 18:32:30.262295 4859 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/3d31b043-3252-458e-8f23-afd68b624849-metrics-certs-tls-certs\") pod \"ovn-northd-0\" (UID: \"3d31b043-3252-458e-8f23-afd68b624849\") " pod="openstack/ovn-northd-0" Oct 08 18:32:30 crc kubenswrapper[4859]: I1008 18:32:30.262330 4859 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/3d31b043-3252-458e-8f23-afd68b624849-config\") pod \"ovn-northd-0\" (UID: \"3d31b043-3252-458e-8f23-afd68b624849\") " pod="openstack/ovn-northd-0" Oct 08 18:32:30 crc kubenswrapper[4859]: I1008 18:32:30.262473 4859 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovn-northd-tls-certs\" (UniqueName: \"kubernetes.io/secret/3d31b043-3252-458e-8f23-afd68b624849-ovn-northd-tls-certs\") pod \"ovn-northd-0\" (UID: \"3d31b043-3252-458e-8f23-afd68b624849\") " pod="openstack/ovn-northd-0" Oct 08 18:32:30 crc kubenswrapper[4859]: I1008 18:32:30.262548 4859 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovn-rundir\" (UniqueName: \"kubernetes.io/empty-dir/3d31b043-3252-458e-8f23-afd68b624849-ovn-rundir\") pod \"ovn-northd-0\" (UID: \"3d31b043-3252-458e-8f23-afd68b624849\") " pod="openstack/ovn-northd-0" Oct 08 18:32:30 crc kubenswrapper[4859]: I1008 18:32:30.262573 4859 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/3d31b043-3252-458e-8f23-afd68b624849-scripts\") pod \"ovn-northd-0\" (UID: \"3d31b043-3252-458e-8f23-afd68b624849\") " pod="openstack/ovn-northd-0" Oct 08 18:32:30 crc kubenswrapper[4859]: I1008 18:32:30.262644 4859 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-9nhbr\" (UniqueName: \"kubernetes.io/projected/3d31b043-3252-458e-8f23-afd68b624849-kube-api-access-9nhbr\") pod \"ovn-northd-0\" (UID: \"3d31b043-3252-458e-8f23-afd68b624849\") " pod="openstack/ovn-northd-0" Oct 08 18:32:30 crc kubenswrapper[4859]: I1008 18:32:30.364246 4859 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovn-rundir\" (UniqueName: \"kubernetes.io/empty-dir/3d31b043-3252-458e-8f23-afd68b624849-ovn-rundir\") pod \"ovn-northd-0\" (UID: \"3d31b043-3252-458e-8f23-afd68b624849\") " pod="openstack/ovn-northd-0" Oct 08 18:32:30 crc kubenswrapper[4859]: I1008 18:32:30.364303 4859 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/3d31b043-3252-458e-8f23-afd68b624849-scripts\") pod \"ovn-northd-0\" (UID: \"3d31b043-3252-458e-8f23-afd68b624849\") " pod="openstack/ovn-northd-0" Oct 08 18:32:30 crc kubenswrapper[4859]: I1008 18:32:30.364338 4859 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-9nhbr\" (UniqueName: \"kubernetes.io/projected/3d31b043-3252-458e-8f23-afd68b624849-kube-api-access-9nhbr\") pod \"ovn-northd-0\" (UID: \"3d31b043-3252-458e-8f23-afd68b624849\") " pod="openstack/ovn-northd-0" Oct 08 18:32:30 crc kubenswrapper[4859]: I1008 18:32:30.365235 4859 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovn-rundir\" (UniqueName: \"kubernetes.io/empty-dir/3d31b043-3252-458e-8f23-afd68b624849-ovn-rundir\") pod \"ovn-northd-0\" (UID: \"3d31b043-3252-458e-8f23-afd68b624849\") " pod="openstack/ovn-northd-0" Oct 08 18:32:30 crc kubenswrapper[4859]: I1008 18:32:30.365293 4859 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/3d31b043-3252-458e-8f23-afd68b624849-scripts\") pod \"ovn-northd-0\" (UID: \"3d31b043-3252-458e-8f23-afd68b624849\") " pod="openstack/ovn-northd-0" Oct 08 18:32:30 crc kubenswrapper[4859]: I1008 18:32:30.365418 4859 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3d31b043-3252-458e-8f23-afd68b624849-combined-ca-bundle\") pod \"ovn-northd-0\" (UID: \"3d31b043-3252-458e-8f23-afd68b624849\") " pod="openstack/ovn-northd-0" Oct 08 18:32:30 crc kubenswrapper[4859]: I1008 18:32:30.365455 4859 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/3d31b043-3252-458e-8f23-afd68b624849-metrics-certs-tls-certs\") pod \"ovn-northd-0\" (UID: \"3d31b043-3252-458e-8f23-afd68b624849\") " pod="openstack/ovn-northd-0" Oct 08 18:32:30 crc kubenswrapper[4859]: I1008 18:32:30.366447 4859 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/3d31b043-3252-458e-8f23-afd68b624849-config\") pod \"ovn-northd-0\" (UID: \"3d31b043-3252-458e-8f23-afd68b624849\") " pod="openstack/ovn-northd-0" Oct 08 18:32:30 crc kubenswrapper[4859]: I1008 18:32:30.366537 4859 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovn-northd-tls-certs\" (UniqueName: \"kubernetes.io/secret/3d31b043-3252-458e-8f23-afd68b624849-ovn-northd-tls-certs\") pod \"ovn-northd-0\" (UID: \"3d31b043-3252-458e-8f23-afd68b624849\") " pod="openstack/ovn-northd-0" Oct 08 18:32:30 crc kubenswrapper[4859]: I1008 18:32:30.367344 4859 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/3d31b043-3252-458e-8f23-afd68b624849-config\") pod \"ovn-northd-0\" (UID: \"3d31b043-3252-458e-8f23-afd68b624849\") " pod="openstack/ovn-northd-0" Oct 08 18:32:30 crc kubenswrapper[4859]: I1008 18:32:30.372541 4859 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3d31b043-3252-458e-8f23-afd68b624849-combined-ca-bundle\") pod \"ovn-northd-0\" (UID: \"3d31b043-3252-458e-8f23-afd68b624849\") " pod="openstack/ovn-northd-0" Oct 08 18:32:30 crc kubenswrapper[4859]: I1008 18:32:30.373080 4859 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovn-northd-tls-certs\" (UniqueName: \"kubernetes.io/secret/3d31b043-3252-458e-8f23-afd68b624849-ovn-northd-tls-certs\") pod \"ovn-northd-0\" (UID: \"3d31b043-3252-458e-8f23-afd68b624849\") " pod="openstack/ovn-northd-0" Oct 08 18:32:30 crc kubenswrapper[4859]: I1008 18:32:30.379828 4859 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/3d31b043-3252-458e-8f23-afd68b624849-metrics-certs-tls-certs\") pod \"ovn-northd-0\" (UID: \"3d31b043-3252-458e-8f23-afd68b624849\") " pod="openstack/ovn-northd-0" Oct 08 18:32:30 crc kubenswrapper[4859]: I1008 18:32:30.385239 4859 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-9nhbr\" (UniqueName: \"kubernetes.io/projected/3d31b043-3252-458e-8f23-afd68b624849-kube-api-access-9nhbr\") pod \"ovn-northd-0\" (UID: \"3d31b043-3252-458e-8f23-afd68b624849\") " pod="openstack/ovn-northd-0" Oct 08 18:32:30 crc kubenswrapper[4859]: I1008 18:32:30.484827 4859 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="4a8089d5-f64f-4e08-bbef-e929819bb0b0" path="/var/lib/kubelet/pods/4a8089d5-f64f-4e08-bbef-e929819bb0b0/volumes" Oct 08 18:32:30 crc kubenswrapper[4859]: I1008 18:32:30.507942 4859 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-northd-0" Oct 08 18:32:30 crc kubenswrapper[4859]: I1008 18:32:30.708053 4859 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-dc9d58d7-tdt5s" event={"ID":"6b7754e8-93c9-4b3a-983c-1127f73459b6","Type":"ContainerStarted","Data":"1a3ccfc530fa04517677d969892cf026028ca37a8d1773bc7122e74690e2aca6"} Oct 08 18:32:30 crc kubenswrapper[4859]: I1008 18:32:30.708967 4859 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-dc9d58d7-tdt5s" Oct 08 18:32:30 crc kubenswrapper[4859]: I1008 18:32:30.736006 4859 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-dc9d58d7-tdt5s" podStartSLOduration=3.735987233 podStartE2EDuration="3.735987233s" podCreationTimestamp="2025-10-08 18:32:27 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-08 18:32:30.726482162 +0000 UTC m=+920.973321541" watchObservedRunningTime="2025-10-08 18:32:30.735987233 +0000 UTC m=+920.982826602" Oct 08 18:32:30 crc kubenswrapper[4859]: I1008 18:32:30.942221 4859 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-northd-0"] Oct 08 18:32:30 crc kubenswrapper[4859]: W1008 18:32:30.949420 4859 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod3d31b043_3252_458e_8f23_afd68b624849.slice/crio-f99356a6ebd5508bfd3a9ba2368ac161d2787fb1780c674e260d2cc95e33c480 WatchSource:0}: Error finding container f99356a6ebd5508bfd3a9ba2368ac161d2787fb1780c674e260d2cc95e33c480: Status 404 returned error can't find the container with id f99356a6ebd5508bfd3a9ba2368ac161d2787fb1780c674e260d2cc95e33c480 Oct 08 18:32:31 crc kubenswrapper[4859]: I1008 18:32:31.717582 4859 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-northd-0" event={"ID":"3d31b043-3252-458e-8f23-afd68b624849","Type":"ContainerStarted","Data":"f99356a6ebd5508bfd3a9ba2368ac161d2787fb1780c674e260d2cc95e33c480"} Oct 08 18:32:32 crc kubenswrapper[4859]: I1008 18:32:32.729009 4859 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-northd-0" event={"ID":"3d31b043-3252-458e-8f23-afd68b624849","Type":"ContainerStarted","Data":"fde8d0bdb9e1f000aeaddf9e29a08081c9cee076bd22f490f2831d341a9210b1"} Oct 08 18:32:32 crc kubenswrapper[4859]: I1008 18:32:32.729654 4859 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ovn-northd-0" Oct 08 18:32:32 crc kubenswrapper[4859]: I1008 18:32:32.729668 4859 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-northd-0" event={"ID":"3d31b043-3252-458e-8f23-afd68b624849","Type":"ContainerStarted","Data":"d53a78206aacdd269b63d68b0cf6596f750ce128608da671efb5c18aa6997fb5"} Oct 08 18:32:32 crc kubenswrapper[4859]: I1008 18:32:32.756010 4859 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ovn-northd-0" podStartSLOduration=1.673101989 podStartE2EDuration="2.755983403s" podCreationTimestamp="2025-10-08 18:32:30 +0000 UTC" firstStartedPulling="2025-10-08 18:32:30.954869948 +0000 UTC m=+921.201709327" lastFinishedPulling="2025-10-08 18:32:32.037751362 +0000 UTC m=+922.284590741" observedRunningTime="2025-10-08 18:32:32.750757988 +0000 UTC m=+922.997597387" watchObservedRunningTime="2025-10-08 18:32:32.755983403 +0000 UTC m=+923.002822792" Oct 08 18:32:37 crc kubenswrapper[4859]: I1008 18:32:37.550053 4859 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/openstack-galera-0" Oct 08 18:32:37 crc kubenswrapper[4859]: I1008 18:32:37.550602 4859 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/openstack-galera-0" Oct 08 18:32:37 crc kubenswrapper[4859]: I1008 18:32:37.614898 4859 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/openstack-galera-0" Oct 08 18:32:37 crc kubenswrapper[4859]: I1008 18:32:37.822898 4859 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/openstack-galera-0" Oct 08 18:32:37 crc kubenswrapper[4859]: I1008 18:32:37.995709 4859 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/openstack-cell1-galera-0" Oct 08 18:32:37 crc kubenswrapper[4859]: I1008 18:32:37.996026 4859 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/openstack-cell1-galera-0" Oct 08 18:32:38 crc kubenswrapper[4859]: I1008 18:32:38.061174 4859 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/openstack-cell1-galera-0" Oct 08 18:32:38 crc kubenswrapper[4859]: I1008 18:32:38.266885 4859 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-dc9d58d7-tdt5s" Oct 08 18:32:38 crc kubenswrapper[4859]: I1008 18:32:38.329131 4859 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-77597f887-q5jzd"] Oct 08 18:32:38 crc kubenswrapper[4859]: I1008 18:32:38.329419 4859 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-77597f887-q5jzd" podUID="e78ba437-04e5-4b40-af53-8b93921fb30f" containerName="dnsmasq-dns" containerID="cri-o://fb1976088f65191d06b2f0f31dffcf3d48a999de694d3b33c8208687385a83a9" gracePeriod=10 Oct 08 18:32:38 crc kubenswrapper[4859]: I1008 18:32:38.388155 4859 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/placement-db-create-vrx2g"] Oct 08 18:32:38 crc kubenswrapper[4859]: I1008 18:32:38.392424 4859 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/placement-db-create-vrx2g" Oct 08 18:32:38 crc kubenswrapper[4859]: I1008 18:32:38.421004 4859 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-sx28r\" (UniqueName: \"kubernetes.io/projected/5330b621-b843-410c-b41a-ffeeb10ea9e4-kube-api-access-sx28r\") pod \"placement-db-create-vrx2g\" (UID: \"5330b621-b843-410c-b41a-ffeeb10ea9e4\") " pod="openstack/placement-db-create-vrx2g" Oct 08 18:32:38 crc kubenswrapper[4859]: I1008 18:32:38.452904 4859 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/placement-db-create-vrx2g"] Oct 08 18:32:38 crc kubenswrapper[4859]: I1008 18:32:38.526570 4859 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-sx28r\" (UniqueName: \"kubernetes.io/projected/5330b621-b843-410c-b41a-ffeeb10ea9e4-kube-api-access-sx28r\") pod \"placement-db-create-vrx2g\" (UID: \"5330b621-b843-410c-b41a-ffeeb10ea9e4\") " pod="openstack/placement-db-create-vrx2g" Oct 08 18:32:38 crc kubenswrapper[4859]: I1008 18:32:38.552185 4859 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-sx28r\" (UniqueName: \"kubernetes.io/projected/5330b621-b843-410c-b41a-ffeeb10ea9e4-kube-api-access-sx28r\") pod \"placement-db-create-vrx2g\" (UID: \"5330b621-b843-410c-b41a-ffeeb10ea9e4\") " pod="openstack/placement-db-create-vrx2g" Oct 08 18:32:38 crc kubenswrapper[4859]: I1008 18:32:38.734226 4859 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/placement-db-create-vrx2g" Oct 08 18:32:38 crc kubenswrapper[4859]: I1008 18:32:38.782934 4859 generic.go:334] "Generic (PLEG): container finished" podID="e78ba437-04e5-4b40-af53-8b93921fb30f" containerID="fb1976088f65191d06b2f0f31dffcf3d48a999de694d3b33c8208687385a83a9" exitCode=0 Oct 08 18:32:38 crc kubenswrapper[4859]: I1008 18:32:38.784035 4859 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-77597f887-q5jzd" event={"ID":"e78ba437-04e5-4b40-af53-8b93921fb30f","Type":"ContainerDied","Data":"fb1976088f65191d06b2f0f31dffcf3d48a999de694d3b33c8208687385a83a9"} Oct 08 18:32:38 crc kubenswrapper[4859]: I1008 18:32:38.784073 4859 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-77597f887-q5jzd" event={"ID":"e78ba437-04e5-4b40-af53-8b93921fb30f","Type":"ContainerDied","Data":"9bf5c780ffaa2fe787596d6cce50246d54103e0ad7ee9dafe099a8fcfdc214ff"} Oct 08 18:32:38 crc kubenswrapper[4859]: I1008 18:32:38.784085 4859 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="9bf5c780ffaa2fe787596d6cce50246d54103e0ad7ee9dafe099a8fcfdc214ff" Oct 08 18:32:38 crc kubenswrapper[4859]: I1008 18:32:38.803251 4859 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-77597f887-q5jzd" Oct 08 18:32:38 crc kubenswrapper[4859]: I1008 18:32:38.830496 4859 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-98hh5\" (UniqueName: \"kubernetes.io/projected/e78ba437-04e5-4b40-af53-8b93921fb30f-kube-api-access-98hh5\") pod \"e78ba437-04e5-4b40-af53-8b93921fb30f\" (UID: \"e78ba437-04e5-4b40-af53-8b93921fb30f\") " Oct 08 18:32:38 crc kubenswrapper[4859]: I1008 18:32:38.831001 4859 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/e78ba437-04e5-4b40-af53-8b93921fb30f-config\") pod \"e78ba437-04e5-4b40-af53-8b93921fb30f\" (UID: \"e78ba437-04e5-4b40-af53-8b93921fb30f\") " Oct 08 18:32:38 crc kubenswrapper[4859]: I1008 18:32:38.831111 4859 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/e78ba437-04e5-4b40-af53-8b93921fb30f-dns-svc\") pod \"e78ba437-04e5-4b40-af53-8b93921fb30f\" (UID: \"e78ba437-04e5-4b40-af53-8b93921fb30f\") " Oct 08 18:32:38 crc kubenswrapper[4859]: I1008 18:32:38.848297 4859 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/e78ba437-04e5-4b40-af53-8b93921fb30f-kube-api-access-98hh5" (OuterVolumeSpecName: "kube-api-access-98hh5") pod "e78ba437-04e5-4b40-af53-8b93921fb30f" (UID: "e78ba437-04e5-4b40-af53-8b93921fb30f"). InnerVolumeSpecName "kube-api-access-98hh5". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 08 18:32:38 crc kubenswrapper[4859]: I1008 18:32:38.875058 4859 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/openstack-cell1-galera-0" Oct 08 18:32:38 crc kubenswrapper[4859]: I1008 18:32:38.915353 4859 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/e78ba437-04e5-4b40-af53-8b93921fb30f-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "e78ba437-04e5-4b40-af53-8b93921fb30f" (UID: "e78ba437-04e5-4b40-af53-8b93921fb30f"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 08 18:32:38 crc kubenswrapper[4859]: I1008 18:32:38.933232 4859 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/e78ba437-04e5-4b40-af53-8b93921fb30f-dns-svc\") on node \"crc\" DevicePath \"\"" Oct 08 18:32:38 crc kubenswrapper[4859]: I1008 18:32:38.933279 4859 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-98hh5\" (UniqueName: \"kubernetes.io/projected/e78ba437-04e5-4b40-af53-8b93921fb30f-kube-api-access-98hh5\") on node \"crc\" DevicePath \"\"" Oct 08 18:32:38 crc kubenswrapper[4859]: I1008 18:32:38.940289 4859 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/e78ba437-04e5-4b40-af53-8b93921fb30f-config" (OuterVolumeSpecName: "config") pod "e78ba437-04e5-4b40-af53-8b93921fb30f" (UID: "e78ba437-04e5-4b40-af53-8b93921fb30f"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 08 18:32:39 crc kubenswrapper[4859]: I1008 18:32:39.035140 4859 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/e78ba437-04e5-4b40-af53-8b93921fb30f-config\") on node \"crc\" DevicePath \"\"" Oct 08 18:32:39 crc kubenswrapper[4859]: I1008 18:32:39.196938 4859 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/placement-db-create-vrx2g"] Oct 08 18:32:39 crc kubenswrapper[4859]: I1008 18:32:39.794794 4859 generic.go:334] "Generic (PLEG): container finished" podID="5330b621-b843-410c-b41a-ffeeb10ea9e4" containerID="be9e6049d5bf7c698d69ee258fe38046579bf0fe9d4b9edec6903e4286455a14" exitCode=0 Oct 08 18:32:39 crc kubenswrapper[4859]: I1008 18:32:39.794921 4859 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-db-create-vrx2g" event={"ID":"5330b621-b843-410c-b41a-ffeeb10ea9e4","Type":"ContainerDied","Data":"be9e6049d5bf7c698d69ee258fe38046579bf0fe9d4b9edec6903e4286455a14"} Oct 08 18:32:39 crc kubenswrapper[4859]: I1008 18:32:39.795192 4859 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-db-create-vrx2g" event={"ID":"5330b621-b843-410c-b41a-ffeeb10ea9e4","Type":"ContainerStarted","Data":"7cba4a264e29e476b1e4bbbecf31204288891eaf871b0ba2c9cb4be1e4c93da6"} Oct 08 18:32:39 crc kubenswrapper[4859]: I1008 18:32:39.795244 4859 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-77597f887-q5jzd" Oct 08 18:32:39 crc kubenswrapper[4859]: I1008 18:32:39.839597 4859 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-77597f887-q5jzd"] Oct 08 18:32:39 crc kubenswrapper[4859]: I1008 18:32:39.847035 4859 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-77597f887-q5jzd"] Oct 08 18:32:40 crc kubenswrapper[4859]: I1008 18:32:40.074451 4859 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/glance-db-create-cgx9g"] Oct 08 18:32:40 crc kubenswrapper[4859]: E1008 18:32:40.074913 4859 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e78ba437-04e5-4b40-af53-8b93921fb30f" containerName="init" Oct 08 18:32:40 crc kubenswrapper[4859]: I1008 18:32:40.074933 4859 state_mem.go:107] "Deleted CPUSet assignment" podUID="e78ba437-04e5-4b40-af53-8b93921fb30f" containerName="init" Oct 08 18:32:40 crc kubenswrapper[4859]: E1008 18:32:40.074965 4859 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e78ba437-04e5-4b40-af53-8b93921fb30f" containerName="dnsmasq-dns" Oct 08 18:32:40 crc kubenswrapper[4859]: I1008 18:32:40.074974 4859 state_mem.go:107] "Deleted CPUSet assignment" podUID="e78ba437-04e5-4b40-af53-8b93921fb30f" containerName="dnsmasq-dns" Oct 08 18:32:40 crc kubenswrapper[4859]: I1008 18:32:40.075181 4859 memory_manager.go:354] "RemoveStaleState removing state" podUID="e78ba437-04e5-4b40-af53-8b93921fb30f" containerName="dnsmasq-dns" Oct 08 18:32:40 crc kubenswrapper[4859]: I1008 18:32:40.075927 4859 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-db-create-cgx9g" Oct 08 18:32:40 crc kubenswrapper[4859]: I1008 18:32:40.103761 4859 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-db-create-cgx9g"] Oct 08 18:32:40 crc kubenswrapper[4859]: I1008 18:32:40.117507 4859 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-7b587f8db7-lm5t5"] Oct 08 18:32:40 crc kubenswrapper[4859]: I1008 18:32:40.119325 4859 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-7b587f8db7-lm5t5" Oct 08 18:32:40 crc kubenswrapper[4859]: I1008 18:32:40.141026 4859 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-7b587f8db7-lm5t5"] Oct 08 18:32:40 crc kubenswrapper[4859]: I1008 18:32:40.256651 4859 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/528e6076-58d3-4fcd-ac6e-4fc5d32dd85f-ovsdbserver-nb\") pod \"dnsmasq-dns-7b587f8db7-lm5t5\" (UID: \"528e6076-58d3-4fcd-ac6e-4fc5d32dd85f\") " pod="openstack/dnsmasq-dns-7b587f8db7-lm5t5" Oct 08 18:32:40 crc kubenswrapper[4859]: I1008 18:32:40.257002 4859 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/528e6076-58d3-4fcd-ac6e-4fc5d32dd85f-config\") pod \"dnsmasq-dns-7b587f8db7-lm5t5\" (UID: \"528e6076-58d3-4fcd-ac6e-4fc5d32dd85f\") " pod="openstack/dnsmasq-dns-7b587f8db7-lm5t5" Oct 08 18:32:40 crc kubenswrapper[4859]: I1008 18:32:40.257042 4859 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-mxlls\" (UniqueName: \"kubernetes.io/projected/528e6076-58d3-4fcd-ac6e-4fc5d32dd85f-kube-api-access-mxlls\") pod \"dnsmasq-dns-7b587f8db7-lm5t5\" (UID: \"528e6076-58d3-4fcd-ac6e-4fc5d32dd85f\") " pod="openstack/dnsmasq-dns-7b587f8db7-lm5t5" Oct 08 18:32:40 crc kubenswrapper[4859]: I1008 18:32:40.257080 4859 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-wzrjr\" (UniqueName: \"kubernetes.io/projected/a380400a-453c-4cc9-b24c-540f762e1404-kube-api-access-wzrjr\") pod \"glance-db-create-cgx9g\" (UID: \"a380400a-453c-4cc9-b24c-540f762e1404\") " pod="openstack/glance-db-create-cgx9g" Oct 08 18:32:40 crc kubenswrapper[4859]: I1008 18:32:40.257344 4859 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/528e6076-58d3-4fcd-ac6e-4fc5d32dd85f-ovsdbserver-sb\") pod \"dnsmasq-dns-7b587f8db7-lm5t5\" (UID: \"528e6076-58d3-4fcd-ac6e-4fc5d32dd85f\") " pod="openstack/dnsmasq-dns-7b587f8db7-lm5t5" Oct 08 18:32:40 crc kubenswrapper[4859]: I1008 18:32:40.257406 4859 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/528e6076-58d3-4fcd-ac6e-4fc5d32dd85f-dns-svc\") pod \"dnsmasq-dns-7b587f8db7-lm5t5\" (UID: \"528e6076-58d3-4fcd-ac6e-4fc5d32dd85f\") " pod="openstack/dnsmasq-dns-7b587f8db7-lm5t5" Oct 08 18:32:40 crc kubenswrapper[4859]: I1008 18:32:40.358627 4859 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/528e6076-58d3-4fcd-ac6e-4fc5d32dd85f-ovsdbserver-sb\") pod \"dnsmasq-dns-7b587f8db7-lm5t5\" (UID: \"528e6076-58d3-4fcd-ac6e-4fc5d32dd85f\") " pod="openstack/dnsmasq-dns-7b587f8db7-lm5t5" Oct 08 18:32:40 crc kubenswrapper[4859]: I1008 18:32:40.358721 4859 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/528e6076-58d3-4fcd-ac6e-4fc5d32dd85f-dns-svc\") pod \"dnsmasq-dns-7b587f8db7-lm5t5\" (UID: \"528e6076-58d3-4fcd-ac6e-4fc5d32dd85f\") " pod="openstack/dnsmasq-dns-7b587f8db7-lm5t5" Oct 08 18:32:40 crc kubenswrapper[4859]: I1008 18:32:40.358828 4859 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/528e6076-58d3-4fcd-ac6e-4fc5d32dd85f-ovsdbserver-nb\") pod \"dnsmasq-dns-7b587f8db7-lm5t5\" (UID: \"528e6076-58d3-4fcd-ac6e-4fc5d32dd85f\") " pod="openstack/dnsmasq-dns-7b587f8db7-lm5t5" Oct 08 18:32:40 crc kubenswrapper[4859]: I1008 18:32:40.358863 4859 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/528e6076-58d3-4fcd-ac6e-4fc5d32dd85f-config\") pod \"dnsmasq-dns-7b587f8db7-lm5t5\" (UID: \"528e6076-58d3-4fcd-ac6e-4fc5d32dd85f\") " pod="openstack/dnsmasq-dns-7b587f8db7-lm5t5" Oct 08 18:32:40 crc kubenswrapper[4859]: I1008 18:32:40.359848 4859 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/528e6076-58d3-4fcd-ac6e-4fc5d32dd85f-ovsdbserver-sb\") pod \"dnsmasq-dns-7b587f8db7-lm5t5\" (UID: \"528e6076-58d3-4fcd-ac6e-4fc5d32dd85f\") " pod="openstack/dnsmasq-dns-7b587f8db7-lm5t5" Oct 08 18:32:40 crc kubenswrapper[4859]: I1008 18:32:40.360096 4859 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/528e6076-58d3-4fcd-ac6e-4fc5d32dd85f-ovsdbserver-nb\") pod \"dnsmasq-dns-7b587f8db7-lm5t5\" (UID: \"528e6076-58d3-4fcd-ac6e-4fc5d32dd85f\") " pod="openstack/dnsmasq-dns-7b587f8db7-lm5t5" Oct 08 18:32:40 crc kubenswrapper[4859]: I1008 18:32:40.360311 4859 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/528e6076-58d3-4fcd-ac6e-4fc5d32dd85f-dns-svc\") pod \"dnsmasq-dns-7b587f8db7-lm5t5\" (UID: \"528e6076-58d3-4fcd-ac6e-4fc5d32dd85f\") " pod="openstack/dnsmasq-dns-7b587f8db7-lm5t5" Oct 08 18:32:40 crc kubenswrapper[4859]: I1008 18:32:40.360311 4859 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/528e6076-58d3-4fcd-ac6e-4fc5d32dd85f-config\") pod \"dnsmasq-dns-7b587f8db7-lm5t5\" (UID: \"528e6076-58d3-4fcd-ac6e-4fc5d32dd85f\") " pod="openstack/dnsmasq-dns-7b587f8db7-lm5t5" Oct 08 18:32:40 crc kubenswrapper[4859]: I1008 18:32:40.360739 4859 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-mxlls\" (UniqueName: \"kubernetes.io/projected/528e6076-58d3-4fcd-ac6e-4fc5d32dd85f-kube-api-access-mxlls\") pod \"dnsmasq-dns-7b587f8db7-lm5t5\" (UID: \"528e6076-58d3-4fcd-ac6e-4fc5d32dd85f\") " pod="openstack/dnsmasq-dns-7b587f8db7-lm5t5" Oct 08 18:32:40 crc kubenswrapper[4859]: I1008 18:32:40.360814 4859 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-wzrjr\" (UniqueName: \"kubernetes.io/projected/a380400a-453c-4cc9-b24c-540f762e1404-kube-api-access-wzrjr\") pod \"glance-db-create-cgx9g\" (UID: \"a380400a-453c-4cc9-b24c-540f762e1404\") " pod="openstack/glance-db-create-cgx9g" Oct 08 18:32:40 crc kubenswrapper[4859]: I1008 18:32:40.380744 4859 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-wzrjr\" (UniqueName: \"kubernetes.io/projected/a380400a-453c-4cc9-b24c-540f762e1404-kube-api-access-wzrjr\") pod \"glance-db-create-cgx9g\" (UID: \"a380400a-453c-4cc9-b24c-540f762e1404\") " pod="openstack/glance-db-create-cgx9g" Oct 08 18:32:40 crc kubenswrapper[4859]: I1008 18:32:40.380760 4859 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-mxlls\" (UniqueName: \"kubernetes.io/projected/528e6076-58d3-4fcd-ac6e-4fc5d32dd85f-kube-api-access-mxlls\") pod \"dnsmasq-dns-7b587f8db7-lm5t5\" (UID: \"528e6076-58d3-4fcd-ac6e-4fc5d32dd85f\") " pod="openstack/dnsmasq-dns-7b587f8db7-lm5t5" Oct 08 18:32:40 crc kubenswrapper[4859]: I1008 18:32:40.420132 4859 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-db-create-cgx9g" Oct 08 18:32:40 crc kubenswrapper[4859]: I1008 18:32:40.450497 4859 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-7b587f8db7-lm5t5" Oct 08 18:32:40 crc kubenswrapper[4859]: I1008 18:32:40.494542 4859 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="e78ba437-04e5-4b40-af53-8b93921fb30f" path="/var/lib/kubelet/pods/e78ba437-04e5-4b40-af53-8b93921fb30f/volumes" Oct 08 18:32:40 crc kubenswrapper[4859]: I1008 18:32:40.973178 4859 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-db-create-cgx9g"] Oct 08 18:32:41 crc kubenswrapper[4859]: I1008 18:32:41.089232 4859 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-7b587f8db7-lm5t5"] Oct 08 18:32:41 crc kubenswrapper[4859]: W1008 18:32:41.104336 4859 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod528e6076_58d3_4fcd_ac6e_4fc5d32dd85f.slice/crio-178b91783a6ecda6b48b4d222dd123d392ae794e7e329ae3c6e12c430eca7ece WatchSource:0}: Error finding container 178b91783a6ecda6b48b4d222dd123d392ae794e7e329ae3c6e12c430eca7ece: Status 404 returned error can't find the container with id 178b91783a6ecda6b48b4d222dd123d392ae794e7e329ae3c6e12c430eca7ece Oct 08 18:32:41 crc kubenswrapper[4859]: I1008 18:32:41.107793 4859 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/placement-db-create-vrx2g" Oct 08 18:32:41 crc kubenswrapper[4859]: I1008 18:32:41.218769 4859 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-sx28r\" (UniqueName: \"kubernetes.io/projected/5330b621-b843-410c-b41a-ffeeb10ea9e4-kube-api-access-sx28r\") pod \"5330b621-b843-410c-b41a-ffeeb10ea9e4\" (UID: \"5330b621-b843-410c-b41a-ffeeb10ea9e4\") " Oct 08 18:32:41 crc kubenswrapper[4859]: I1008 18:32:41.223178 4859 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5330b621-b843-410c-b41a-ffeeb10ea9e4-kube-api-access-sx28r" (OuterVolumeSpecName: "kube-api-access-sx28r") pod "5330b621-b843-410c-b41a-ffeeb10ea9e4" (UID: "5330b621-b843-410c-b41a-ffeeb10ea9e4"). InnerVolumeSpecName "kube-api-access-sx28r". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 08 18:32:41 crc kubenswrapper[4859]: I1008 18:32:41.280494 4859 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/swift-storage-0"] Oct 08 18:32:41 crc kubenswrapper[4859]: E1008 18:32:41.281832 4859 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5330b621-b843-410c-b41a-ffeeb10ea9e4" containerName="mariadb-database-create" Oct 08 18:32:41 crc kubenswrapper[4859]: I1008 18:32:41.281860 4859 state_mem.go:107] "Deleted CPUSet assignment" podUID="5330b621-b843-410c-b41a-ffeeb10ea9e4" containerName="mariadb-database-create" Oct 08 18:32:41 crc kubenswrapper[4859]: I1008 18:32:41.282092 4859 memory_manager.go:354] "RemoveStaleState removing state" podUID="5330b621-b843-410c-b41a-ffeeb10ea9e4" containerName="mariadb-database-create" Oct 08 18:32:41 crc kubenswrapper[4859]: I1008 18:32:41.287313 4859 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/swift-storage-0" Oct 08 18:32:41 crc kubenswrapper[4859]: I1008 18:32:41.289789 4859 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"swift-swift-dockercfg-rd4rn" Oct 08 18:32:41 crc kubenswrapper[4859]: I1008 18:32:41.289876 4859 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"swift-storage-config-data" Oct 08 18:32:41 crc kubenswrapper[4859]: I1008 18:32:41.290033 4859 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"swift-ring-files" Oct 08 18:32:41 crc kubenswrapper[4859]: I1008 18:32:41.290209 4859 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"swift-conf" Oct 08 18:32:41 crc kubenswrapper[4859]: I1008 18:32:41.300714 4859 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/swift-storage-0"] Oct 08 18:32:41 crc kubenswrapper[4859]: I1008 18:32:41.321419 4859 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-sx28r\" (UniqueName: \"kubernetes.io/projected/5330b621-b843-410c-b41a-ffeeb10ea9e4-kube-api-access-sx28r\") on node \"crc\" DevicePath \"\"" Oct 08 18:32:41 crc kubenswrapper[4859]: I1008 18:32:41.422652 4859 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage03-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage03-crc\") pod \"swift-storage-0\" (UID: \"c860866f-d309-4824-a31c-e88c5d820712\") " pod="openstack/swift-storage-0" Oct 08 18:32:41 crc kubenswrapper[4859]: I1008 18:32:41.422737 4859 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cache\" (UniqueName: \"kubernetes.io/empty-dir/c860866f-d309-4824-a31c-e88c5d820712-cache\") pod \"swift-storage-0\" (UID: \"c860866f-d309-4824-a31c-e88c5d820712\") " pod="openstack/swift-storage-0" Oct 08 18:32:41 crc kubenswrapper[4859]: I1008 18:32:41.422776 4859 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"lock\" (UniqueName: \"kubernetes.io/empty-dir/c860866f-d309-4824-a31c-e88c5d820712-lock\") pod \"swift-storage-0\" (UID: \"c860866f-d309-4824-a31c-e88c5d820712\") " pod="openstack/swift-storage-0" Oct 08 18:32:41 crc kubenswrapper[4859]: I1008 18:32:41.422827 4859 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/c860866f-d309-4824-a31c-e88c5d820712-etc-swift\") pod \"swift-storage-0\" (UID: \"c860866f-d309-4824-a31c-e88c5d820712\") " pod="openstack/swift-storage-0" Oct 08 18:32:41 crc kubenswrapper[4859]: I1008 18:32:41.422846 4859 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-pkz9n\" (UniqueName: \"kubernetes.io/projected/c860866f-d309-4824-a31c-e88c5d820712-kube-api-access-pkz9n\") pod \"swift-storage-0\" (UID: \"c860866f-d309-4824-a31c-e88c5d820712\") " pod="openstack/swift-storage-0" Oct 08 18:32:41 crc kubenswrapper[4859]: I1008 18:32:41.524352 4859 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"lock\" (UniqueName: \"kubernetes.io/empty-dir/c860866f-d309-4824-a31c-e88c5d820712-lock\") pod \"swift-storage-0\" (UID: \"c860866f-d309-4824-a31c-e88c5d820712\") " pod="openstack/swift-storage-0" Oct 08 18:32:41 crc kubenswrapper[4859]: I1008 18:32:41.524440 4859 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/c860866f-d309-4824-a31c-e88c5d820712-etc-swift\") pod \"swift-storage-0\" (UID: \"c860866f-d309-4824-a31c-e88c5d820712\") " pod="openstack/swift-storage-0" Oct 08 18:32:41 crc kubenswrapper[4859]: I1008 18:32:41.524471 4859 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-pkz9n\" (UniqueName: \"kubernetes.io/projected/c860866f-d309-4824-a31c-e88c5d820712-kube-api-access-pkz9n\") pod \"swift-storage-0\" (UID: \"c860866f-d309-4824-a31c-e88c5d820712\") " pod="openstack/swift-storage-0" Oct 08 18:32:41 crc kubenswrapper[4859]: I1008 18:32:41.524581 4859 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage03-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage03-crc\") pod \"swift-storage-0\" (UID: \"c860866f-d309-4824-a31c-e88c5d820712\") " pod="openstack/swift-storage-0" Oct 08 18:32:41 crc kubenswrapper[4859]: I1008 18:32:41.524603 4859 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cache\" (UniqueName: \"kubernetes.io/empty-dir/c860866f-d309-4824-a31c-e88c5d820712-cache\") pod \"swift-storage-0\" (UID: \"c860866f-d309-4824-a31c-e88c5d820712\") " pod="openstack/swift-storage-0" Oct 08 18:32:41 crc kubenswrapper[4859]: E1008 18:32:41.524842 4859 projected.go:288] Couldn't get configMap openstack/swift-ring-files: configmap "swift-ring-files" not found Oct 08 18:32:41 crc kubenswrapper[4859]: E1008 18:32:41.524903 4859 projected.go:194] Error preparing data for projected volume etc-swift for pod openstack/swift-storage-0: configmap "swift-ring-files" not found Oct 08 18:32:41 crc kubenswrapper[4859]: E1008 18:32:41.524999 4859 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/c860866f-d309-4824-a31c-e88c5d820712-etc-swift podName:c860866f-d309-4824-a31c-e88c5d820712 nodeName:}" failed. No retries permitted until 2025-10-08 18:32:42.024970263 +0000 UTC m=+932.271809642 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "etc-swift" (UniqueName: "kubernetes.io/projected/c860866f-d309-4824-a31c-e88c5d820712-etc-swift") pod "swift-storage-0" (UID: "c860866f-d309-4824-a31c-e88c5d820712") : configmap "swift-ring-files" not found Oct 08 18:32:41 crc kubenswrapper[4859]: I1008 18:32:41.525077 4859 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"lock\" (UniqueName: \"kubernetes.io/empty-dir/c860866f-d309-4824-a31c-e88c5d820712-lock\") pod \"swift-storage-0\" (UID: \"c860866f-d309-4824-a31c-e88c5d820712\") " pod="openstack/swift-storage-0" Oct 08 18:32:41 crc kubenswrapper[4859]: I1008 18:32:41.525143 4859 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage03-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage03-crc\") pod \"swift-storage-0\" (UID: \"c860866f-d309-4824-a31c-e88c5d820712\") device mount path \"/mnt/openstack/pv03\"" pod="openstack/swift-storage-0" Oct 08 18:32:41 crc kubenswrapper[4859]: I1008 18:32:41.525175 4859 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cache\" (UniqueName: \"kubernetes.io/empty-dir/c860866f-d309-4824-a31c-e88c5d820712-cache\") pod \"swift-storage-0\" (UID: \"c860866f-d309-4824-a31c-e88c5d820712\") " pod="openstack/swift-storage-0" Oct 08 18:32:41 crc kubenswrapper[4859]: I1008 18:32:41.552612 4859 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage03-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage03-crc\") pod \"swift-storage-0\" (UID: \"c860866f-d309-4824-a31c-e88c5d820712\") " pod="openstack/swift-storage-0" Oct 08 18:32:41 crc kubenswrapper[4859]: I1008 18:32:41.553805 4859 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-pkz9n\" (UniqueName: \"kubernetes.io/projected/c860866f-d309-4824-a31c-e88c5d820712-kube-api-access-pkz9n\") pod \"swift-storage-0\" (UID: \"c860866f-d309-4824-a31c-e88c5d820712\") " pod="openstack/swift-storage-0" Oct 08 18:32:41 crc kubenswrapper[4859]: I1008 18:32:41.816911 4859 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-db-create-vrx2g" event={"ID":"5330b621-b843-410c-b41a-ffeeb10ea9e4","Type":"ContainerDied","Data":"7cba4a264e29e476b1e4bbbecf31204288891eaf871b0ba2c9cb4be1e4c93da6"} Oct 08 18:32:41 crc kubenswrapper[4859]: I1008 18:32:41.816961 4859 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="7cba4a264e29e476b1e4bbbecf31204288891eaf871b0ba2c9cb4be1e4c93da6" Oct 08 18:32:41 crc kubenswrapper[4859]: I1008 18:32:41.817006 4859 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/placement-db-create-vrx2g" Oct 08 18:32:41 crc kubenswrapper[4859]: I1008 18:32:41.819387 4859 generic.go:334] "Generic (PLEG): container finished" podID="528e6076-58d3-4fcd-ac6e-4fc5d32dd85f" containerID="313d9ef521bf8874aff0eb11214820ac2b18d10f20ca005846085c0cb3d3a769" exitCode=0 Oct 08 18:32:41 crc kubenswrapper[4859]: I1008 18:32:41.819452 4859 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-7b587f8db7-lm5t5" event={"ID":"528e6076-58d3-4fcd-ac6e-4fc5d32dd85f","Type":"ContainerDied","Data":"313d9ef521bf8874aff0eb11214820ac2b18d10f20ca005846085c0cb3d3a769"} Oct 08 18:32:41 crc kubenswrapper[4859]: I1008 18:32:41.820238 4859 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-7b587f8db7-lm5t5" event={"ID":"528e6076-58d3-4fcd-ac6e-4fc5d32dd85f","Type":"ContainerStarted","Data":"178b91783a6ecda6b48b4d222dd123d392ae794e7e329ae3c6e12c430eca7ece"} Oct 08 18:32:41 crc kubenswrapper[4859]: I1008 18:32:41.822415 4859 generic.go:334] "Generic (PLEG): container finished" podID="a380400a-453c-4cc9-b24c-540f762e1404" containerID="166ef925709e4e3e1b6f80efc26c123ca3d703657de04aacb1d2659ec9dd5886" exitCode=0 Oct 08 18:32:41 crc kubenswrapper[4859]: I1008 18:32:41.822482 4859 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-db-create-cgx9g" event={"ID":"a380400a-453c-4cc9-b24c-540f762e1404","Type":"ContainerDied","Data":"166ef925709e4e3e1b6f80efc26c123ca3d703657de04aacb1d2659ec9dd5886"} Oct 08 18:32:41 crc kubenswrapper[4859]: I1008 18:32:41.822523 4859 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-db-create-cgx9g" event={"ID":"a380400a-453c-4cc9-b24c-540f762e1404","Type":"ContainerStarted","Data":"36aaaaed0d16f66d2db1933833f98edabedd3f1b0025482fe495067a616c027c"} Oct 08 18:32:42 crc kubenswrapper[4859]: I1008 18:32:42.036374 4859 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/c860866f-d309-4824-a31c-e88c5d820712-etc-swift\") pod \"swift-storage-0\" (UID: \"c860866f-d309-4824-a31c-e88c5d820712\") " pod="openstack/swift-storage-0" Oct 08 18:32:42 crc kubenswrapper[4859]: E1008 18:32:42.036606 4859 projected.go:288] Couldn't get configMap openstack/swift-ring-files: configmap "swift-ring-files" not found Oct 08 18:32:42 crc kubenswrapper[4859]: E1008 18:32:42.036640 4859 projected.go:194] Error preparing data for projected volume etc-swift for pod openstack/swift-storage-0: configmap "swift-ring-files" not found Oct 08 18:32:42 crc kubenswrapper[4859]: E1008 18:32:42.036729 4859 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/c860866f-d309-4824-a31c-e88c5d820712-etc-swift podName:c860866f-d309-4824-a31c-e88c5d820712 nodeName:}" failed. No retries permitted until 2025-10-08 18:32:43.036706795 +0000 UTC m=+933.283546174 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "etc-swift" (UniqueName: "kubernetes.io/projected/c860866f-d309-4824-a31c-e88c5d820712-etc-swift") pod "swift-storage-0" (UID: "c860866f-d309-4824-a31c-e88c5d820712") : configmap "swift-ring-files" not found Oct 08 18:32:42 crc kubenswrapper[4859]: I1008 18:32:42.832417 4859 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-7b587f8db7-lm5t5" event={"ID":"528e6076-58d3-4fcd-ac6e-4fc5d32dd85f","Type":"ContainerStarted","Data":"4277ff351d5cc7e7b03ffd72a0f5d7bb37123fc379df3099fb15b5d20798ff85"} Oct 08 18:32:43 crc kubenswrapper[4859]: I1008 18:32:43.057119 4859 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/c860866f-d309-4824-a31c-e88c5d820712-etc-swift\") pod \"swift-storage-0\" (UID: \"c860866f-d309-4824-a31c-e88c5d820712\") " pod="openstack/swift-storage-0" Oct 08 18:32:43 crc kubenswrapper[4859]: E1008 18:32:43.057385 4859 projected.go:288] Couldn't get configMap openstack/swift-ring-files: configmap "swift-ring-files" not found Oct 08 18:32:43 crc kubenswrapper[4859]: E1008 18:32:43.057502 4859 projected.go:194] Error preparing data for projected volume etc-swift for pod openstack/swift-storage-0: configmap "swift-ring-files" not found Oct 08 18:32:43 crc kubenswrapper[4859]: E1008 18:32:43.057559 4859 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/c860866f-d309-4824-a31c-e88c5d820712-etc-swift podName:c860866f-d309-4824-a31c-e88c5d820712 nodeName:}" failed. No retries permitted until 2025-10-08 18:32:45.05753843 +0000 UTC m=+935.304377809 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "etc-swift" (UniqueName: "kubernetes.io/projected/c860866f-d309-4824-a31c-e88c5d820712-etc-swift") pod "swift-storage-0" (UID: "c860866f-d309-4824-a31c-e88c5d820712") : configmap "swift-ring-files" not found Oct 08 18:32:43 crc kubenswrapper[4859]: I1008 18:32:43.230125 4859 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-db-create-cgx9g" Oct 08 18:32:43 crc kubenswrapper[4859]: I1008 18:32:43.246643 4859 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-7b587f8db7-lm5t5" podStartSLOduration=3.246614746 podStartE2EDuration="3.246614746s" podCreationTimestamp="2025-10-08 18:32:40 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-08 18:32:42.851488163 +0000 UTC m=+933.098327552" watchObservedRunningTime="2025-10-08 18:32:43.246614746 +0000 UTC m=+933.493454165" Oct 08 18:32:43 crc kubenswrapper[4859]: I1008 18:32:43.362285 4859 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-wzrjr\" (UniqueName: \"kubernetes.io/projected/a380400a-453c-4cc9-b24c-540f762e1404-kube-api-access-wzrjr\") pod \"a380400a-453c-4cc9-b24c-540f762e1404\" (UID: \"a380400a-453c-4cc9-b24c-540f762e1404\") " Oct 08 18:32:43 crc kubenswrapper[4859]: I1008 18:32:43.377585 4859 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a380400a-453c-4cc9-b24c-540f762e1404-kube-api-access-wzrjr" (OuterVolumeSpecName: "kube-api-access-wzrjr") pod "a380400a-453c-4cc9-b24c-540f762e1404" (UID: "a380400a-453c-4cc9-b24c-540f762e1404"). InnerVolumeSpecName "kube-api-access-wzrjr". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 08 18:32:43 crc kubenswrapper[4859]: I1008 18:32:43.463730 4859 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-wzrjr\" (UniqueName: \"kubernetes.io/projected/a380400a-453c-4cc9-b24c-540f762e1404-kube-api-access-wzrjr\") on node \"crc\" DevicePath \"\"" Oct 08 18:32:43 crc kubenswrapper[4859]: I1008 18:32:43.843040 4859 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-db-create-cgx9g" Oct 08 18:32:43 crc kubenswrapper[4859]: I1008 18:32:43.843038 4859 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-db-create-cgx9g" event={"ID":"a380400a-453c-4cc9-b24c-540f762e1404","Type":"ContainerDied","Data":"36aaaaed0d16f66d2db1933833f98edabedd3f1b0025482fe495067a616c027c"} Oct 08 18:32:43 crc kubenswrapper[4859]: I1008 18:32:43.843102 4859 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="36aaaaed0d16f66d2db1933833f98edabedd3f1b0025482fe495067a616c027c" Oct 08 18:32:43 crc kubenswrapper[4859]: I1008 18:32:43.843248 4859 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-7b587f8db7-lm5t5" Oct 08 18:32:45 crc kubenswrapper[4859]: I1008 18:32:45.141098 4859 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/c860866f-d309-4824-a31c-e88c5d820712-etc-swift\") pod \"swift-storage-0\" (UID: \"c860866f-d309-4824-a31c-e88c5d820712\") " pod="openstack/swift-storage-0" Oct 08 18:32:45 crc kubenswrapper[4859]: E1008 18:32:45.141437 4859 projected.go:288] Couldn't get configMap openstack/swift-ring-files: configmap "swift-ring-files" not found Oct 08 18:32:45 crc kubenswrapper[4859]: E1008 18:32:45.141783 4859 projected.go:194] Error preparing data for projected volume etc-swift for pod openstack/swift-storage-0: configmap "swift-ring-files" not found Oct 08 18:32:45 crc kubenswrapper[4859]: E1008 18:32:45.141903 4859 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/c860866f-d309-4824-a31c-e88c5d820712-etc-swift podName:c860866f-d309-4824-a31c-e88c5d820712 nodeName:}" failed. No retries permitted until 2025-10-08 18:32:49.141868757 +0000 UTC m=+939.388708176 (durationBeforeRetry 4s). Error: MountVolume.SetUp failed for volume "etc-swift" (UniqueName: "kubernetes.io/projected/c860866f-d309-4824-a31c-e88c5d820712-etc-swift") pod "swift-storage-0" (UID: "c860866f-d309-4824-a31c-e88c5d820712") : configmap "swift-ring-files" not found Oct 08 18:32:45 crc kubenswrapper[4859]: I1008 18:32:45.237182 4859 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/swift-ring-rebalance-qrpd6"] Oct 08 18:32:45 crc kubenswrapper[4859]: E1008 18:32:45.237927 4859 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a380400a-453c-4cc9-b24c-540f762e1404" containerName="mariadb-database-create" Oct 08 18:32:45 crc kubenswrapper[4859]: I1008 18:32:45.237969 4859 state_mem.go:107] "Deleted CPUSet assignment" podUID="a380400a-453c-4cc9-b24c-540f762e1404" containerName="mariadb-database-create" Oct 08 18:32:45 crc kubenswrapper[4859]: I1008 18:32:45.238301 4859 memory_manager.go:354] "RemoveStaleState removing state" podUID="a380400a-453c-4cc9-b24c-540f762e1404" containerName="mariadb-database-create" Oct 08 18:32:45 crc kubenswrapper[4859]: I1008 18:32:45.239596 4859 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/swift-ring-rebalance-qrpd6" Oct 08 18:32:45 crc kubenswrapper[4859]: I1008 18:32:45.242846 4859 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"swift-proxy-config-data" Oct 08 18:32:45 crc kubenswrapper[4859]: I1008 18:32:45.242901 4859 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"swift-ring-scripts" Oct 08 18:32:45 crc kubenswrapper[4859]: I1008 18:32:45.243832 4859 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/431b4e6e-415d-4c14-8783-f94409691ae3-ring-data-devices\") pod \"swift-ring-rebalance-qrpd6\" (UID: \"431b4e6e-415d-4c14-8783-f94409691ae3\") " pod="openstack/swift-ring-rebalance-qrpd6" Oct 08 18:32:45 crc kubenswrapper[4859]: I1008 18:32:45.243885 4859 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/431b4e6e-415d-4c14-8783-f94409691ae3-scripts\") pod \"swift-ring-rebalance-qrpd6\" (UID: \"431b4e6e-415d-4c14-8783-f94409691ae3\") " pod="openstack/swift-ring-rebalance-qrpd6" Oct 08 18:32:45 crc kubenswrapper[4859]: I1008 18:32:45.243940 4859 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/431b4e6e-415d-4c14-8783-f94409691ae3-etc-swift\") pod \"swift-ring-rebalance-qrpd6\" (UID: \"431b4e6e-415d-4c14-8783-f94409691ae3\") " pod="openstack/swift-ring-rebalance-qrpd6" Oct 08 18:32:45 crc kubenswrapper[4859]: I1008 18:32:45.243981 4859 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/431b4e6e-415d-4c14-8783-f94409691ae3-dispersionconf\") pod \"swift-ring-rebalance-qrpd6\" (UID: \"431b4e6e-415d-4c14-8783-f94409691ae3\") " pod="openstack/swift-ring-rebalance-qrpd6" Oct 08 18:32:45 crc kubenswrapper[4859]: I1008 18:32:45.244047 4859 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/431b4e6e-415d-4c14-8783-f94409691ae3-swiftconf\") pod \"swift-ring-rebalance-qrpd6\" (UID: \"431b4e6e-415d-4c14-8783-f94409691ae3\") " pod="openstack/swift-ring-rebalance-qrpd6" Oct 08 18:32:45 crc kubenswrapper[4859]: I1008 18:32:45.244162 4859 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/431b4e6e-415d-4c14-8783-f94409691ae3-combined-ca-bundle\") pod \"swift-ring-rebalance-qrpd6\" (UID: \"431b4e6e-415d-4c14-8783-f94409691ae3\") " pod="openstack/swift-ring-rebalance-qrpd6" Oct 08 18:32:45 crc kubenswrapper[4859]: I1008 18:32:45.244370 4859 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-twhsd\" (UniqueName: \"kubernetes.io/projected/431b4e6e-415d-4c14-8783-f94409691ae3-kube-api-access-twhsd\") pod \"swift-ring-rebalance-qrpd6\" (UID: \"431b4e6e-415d-4c14-8783-f94409691ae3\") " pod="openstack/swift-ring-rebalance-qrpd6" Oct 08 18:32:45 crc kubenswrapper[4859]: I1008 18:32:45.246435 4859 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"swift-ring-config-data" Oct 08 18:32:45 crc kubenswrapper[4859]: I1008 18:32:45.256710 4859 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/swift-ring-rebalance-qrpd6"] Oct 08 18:32:45 crc kubenswrapper[4859]: I1008 18:32:45.347042 4859 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-twhsd\" (UniqueName: \"kubernetes.io/projected/431b4e6e-415d-4c14-8783-f94409691ae3-kube-api-access-twhsd\") pod \"swift-ring-rebalance-qrpd6\" (UID: \"431b4e6e-415d-4c14-8783-f94409691ae3\") " pod="openstack/swift-ring-rebalance-qrpd6" Oct 08 18:32:45 crc kubenswrapper[4859]: I1008 18:32:45.347154 4859 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/431b4e6e-415d-4c14-8783-f94409691ae3-ring-data-devices\") pod \"swift-ring-rebalance-qrpd6\" (UID: \"431b4e6e-415d-4c14-8783-f94409691ae3\") " pod="openstack/swift-ring-rebalance-qrpd6" Oct 08 18:32:45 crc kubenswrapper[4859]: I1008 18:32:45.347190 4859 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/431b4e6e-415d-4c14-8783-f94409691ae3-scripts\") pod \"swift-ring-rebalance-qrpd6\" (UID: \"431b4e6e-415d-4c14-8783-f94409691ae3\") " pod="openstack/swift-ring-rebalance-qrpd6" Oct 08 18:32:45 crc kubenswrapper[4859]: I1008 18:32:45.347212 4859 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/431b4e6e-415d-4c14-8783-f94409691ae3-etc-swift\") pod \"swift-ring-rebalance-qrpd6\" (UID: \"431b4e6e-415d-4c14-8783-f94409691ae3\") " pod="openstack/swift-ring-rebalance-qrpd6" Oct 08 18:32:45 crc kubenswrapper[4859]: I1008 18:32:45.347271 4859 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/431b4e6e-415d-4c14-8783-f94409691ae3-dispersionconf\") pod \"swift-ring-rebalance-qrpd6\" (UID: \"431b4e6e-415d-4c14-8783-f94409691ae3\") " pod="openstack/swift-ring-rebalance-qrpd6" Oct 08 18:32:45 crc kubenswrapper[4859]: I1008 18:32:45.347308 4859 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/431b4e6e-415d-4c14-8783-f94409691ae3-swiftconf\") pod \"swift-ring-rebalance-qrpd6\" (UID: \"431b4e6e-415d-4c14-8783-f94409691ae3\") " pod="openstack/swift-ring-rebalance-qrpd6" Oct 08 18:32:45 crc kubenswrapper[4859]: I1008 18:32:45.347366 4859 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/431b4e6e-415d-4c14-8783-f94409691ae3-combined-ca-bundle\") pod \"swift-ring-rebalance-qrpd6\" (UID: \"431b4e6e-415d-4c14-8783-f94409691ae3\") " pod="openstack/swift-ring-rebalance-qrpd6" Oct 08 18:32:45 crc kubenswrapper[4859]: I1008 18:32:45.348042 4859 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/431b4e6e-415d-4c14-8783-f94409691ae3-etc-swift\") pod \"swift-ring-rebalance-qrpd6\" (UID: \"431b4e6e-415d-4c14-8783-f94409691ae3\") " pod="openstack/swift-ring-rebalance-qrpd6" Oct 08 18:32:45 crc kubenswrapper[4859]: I1008 18:32:45.348343 4859 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/431b4e6e-415d-4c14-8783-f94409691ae3-ring-data-devices\") pod \"swift-ring-rebalance-qrpd6\" (UID: \"431b4e6e-415d-4c14-8783-f94409691ae3\") " pod="openstack/swift-ring-rebalance-qrpd6" Oct 08 18:32:45 crc kubenswrapper[4859]: I1008 18:32:45.349042 4859 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/431b4e6e-415d-4c14-8783-f94409691ae3-scripts\") pod \"swift-ring-rebalance-qrpd6\" (UID: \"431b4e6e-415d-4c14-8783-f94409691ae3\") " pod="openstack/swift-ring-rebalance-qrpd6" Oct 08 18:32:45 crc kubenswrapper[4859]: I1008 18:32:45.352845 4859 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/431b4e6e-415d-4c14-8783-f94409691ae3-dispersionconf\") pod \"swift-ring-rebalance-qrpd6\" (UID: \"431b4e6e-415d-4c14-8783-f94409691ae3\") " pod="openstack/swift-ring-rebalance-qrpd6" Oct 08 18:32:45 crc kubenswrapper[4859]: I1008 18:32:45.355787 4859 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/431b4e6e-415d-4c14-8783-f94409691ae3-swiftconf\") pod \"swift-ring-rebalance-qrpd6\" (UID: \"431b4e6e-415d-4c14-8783-f94409691ae3\") " pod="openstack/swift-ring-rebalance-qrpd6" Oct 08 18:32:45 crc kubenswrapper[4859]: I1008 18:32:45.361253 4859 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/431b4e6e-415d-4c14-8783-f94409691ae3-combined-ca-bundle\") pod \"swift-ring-rebalance-qrpd6\" (UID: \"431b4e6e-415d-4c14-8783-f94409691ae3\") " pod="openstack/swift-ring-rebalance-qrpd6" Oct 08 18:32:45 crc kubenswrapper[4859]: I1008 18:32:45.368376 4859 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-twhsd\" (UniqueName: \"kubernetes.io/projected/431b4e6e-415d-4c14-8783-f94409691ae3-kube-api-access-twhsd\") pod \"swift-ring-rebalance-qrpd6\" (UID: \"431b4e6e-415d-4c14-8783-f94409691ae3\") " pod="openstack/swift-ring-rebalance-qrpd6" Oct 08 18:32:45 crc kubenswrapper[4859]: I1008 18:32:45.563212 4859 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/swift-ring-rebalance-qrpd6" Oct 08 18:32:45 crc kubenswrapper[4859]: I1008 18:32:45.580932 4859 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/ovn-northd-0" Oct 08 18:32:46 crc kubenswrapper[4859]: I1008 18:32:46.007998 4859 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/swift-ring-rebalance-qrpd6"] Oct 08 18:32:46 crc kubenswrapper[4859]: W1008 18:32:46.014314 4859 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod431b4e6e_415d_4c14_8783_f94409691ae3.slice/crio-539152ac4a4d7a20de6eea2fe06d75749ae2e36da39b01c442f930ad8fabd528 WatchSource:0}: Error finding container 539152ac4a4d7a20de6eea2fe06d75749ae2e36da39b01c442f930ad8fabd528: Status 404 returned error can't find the container with id 539152ac4a4d7a20de6eea2fe06d75749ae2e36da39b01c442f930ad8fabd528 Oct 08 18:32:46 crc kubenswrapper[4859]: I1008 18:32:46.877929 4859 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-ring-rebalance-qrpd6" event={"ID":"431b4e6e-415d-4c14-8783-f94409691ae3","Type":"ContainerStarted","Data":"539152ac4a4d7a20de6eea2fe06d75749ae2e36da39b01c442f930ad8fabd528"} Oct 08 18:32:47 crc kubenswrapper[4859]: I1008 18:32:47.943316 4859 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/keystone-db-create-9sdlj"] Oct 08 18:32:47 crc kubenswrapper[4859]: I1008 18:32:47.944726 4859 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-db-create-9sdlj" Oct 08 18:32:47 crc kubenswrapper[4859]: I1008 18:32:47.954801 4859 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-db-create-9sdlj"] Oct 08 18:32:48 crc kubenswrapper[4859]: I1008 18:32:48.005302 4859 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-lt4jl\" (UniqueName: \"kubernetes.io/projected/1e1f9094-586a-4f98-a7f6-32ca4816fa8b-kube-api-access-lt4jl\") pod \"keystone-db-create-9sdlj\" (UID: \"1e1f9094-586a-4f98-a7f6-32ca4816fa8b\") " pod="openstack/keystone-db-create-9sdlj" Oct 08 18:32:48 crc kubenswrapper[4859]: I1008 18:32:48.107491 4859 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-lt4jl\" (UniqueName: \"kubernetes.io/projected/1e1f9094-586a-4f98-a7f6-32ca4816fa8b-kube-api-access-lt4jl\") pod \"keystone-db-create-9sdlj\" (UID: \"1e1f9094-586a-4f98-a7f6-32ca4816fa8b\") " pod="openstack/keystone-db-create-9sdlj" Oct 08 18:32:48 crc kubenswrapper[4859]: I1008 18:32:48.131200 4859 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-lt4jl\" (UniqueName: \"kubernetes.io/projected/1e1f9094-586a-4f98-a7f6-32ca4816fa8b-kube-api-access-lt4jl\") pod \"keystone-db-create-9sdlj\" (UID: \"1e1f9094-586a-4f98-a7f6-32ca4816fa8b\") " pod="openstack/keystone-db-create-9sdlj" Oct 08 18:32:48 crc kubenswrapper[4859]: I1008 18:32:48.264843 4859 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-db-create-9sdlj" Oct 08 18:32:48 crc kubenswrapper[4859]: I1008 18:32:48.490247 4859 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/placement-b31f-account-create-r48bf"] Oct 08 18:32:48 crc kubenswrapper[4859]: I1008 18:32:48.492710 4859 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/placement-b31f-account-create-r48bf"] Oct 08 18:32:48 crc kubenswrapper[4859]: I1008 18:32:48.492994 4859 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/placement-b31f-account-create-r48bf" Oct 08 18:32:48 crc kubenswrapper[4859]: I1008 18:32:48.496386 4859 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"placement-db-secret" Oct 08 18:32:48 crc kubenswrapper[4859]: I1008 18:32:48.515673 4859 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-sn7l9\" (UniqueName: \"kubernetes.io/projected/823f174d-4981-4db7-97cb-18d5ec2eead5-kube-api-access-sn7l9\") pod \"placement-b31f-account-create-r48bf\" (UID: \"823f174d-4981-4db7-97cb-18d5ec2eead5\") " pod="openstack/placement-b31f-account-create-r48bf" Oct 08 18:32:48 crc kubenswrapper[4859]: I1008 18:32:48.617089 4859 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-sn7l9\" (UniqueName: \"kubernetes.io/projected/823f174d-4981-4db7-97cb-18d5ec2eead5-kube-api-access-sn7l9\") pod \"placement-b31f-account-create-r48bf\" (UID: \"823f174d-4981-4db7-97cb-18d5ec2eead5\") " pod="openstack/placement-b31f-account-create-r48bf" Oct 08 18:32:48 crc kubenswrapper[4859]: I1008 18:32:48.637714 4859 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-sn7l9\" (UniqueName: \"kubernetes.io/projected/823f174d-4981-4db7-97cb-18d5ec2eead5-kube-api-access-sn7l9\") pod \"placement-b31f-account-create-r48bf\" (UID: \"823f174d-4981-4db7-97cb-18d5ec2eead5\") " pod="openstack/placement-b31f-account-create-r48bf" Oct 08 18:32:48 crc kubenswrapper[4859]: I1008 18:32:48.766370 4859 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-db-create-9sdlj"] Oct 08 18:32:48 crc kubenswrapper[4859]: I1008 18:32:48.818394 4859 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/placement-b31f-account-create-r48bf" Oct 08 18:32:48 crc kubenswrapper[4859]: I1008 18:32:48.902878 4859 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-db-create-9sdlj" event={"ID":"1e1f9094-586a-4f98-a7f6-32ca4816fa8b","Type":"ContainerStarted","Data":"44ec86d631d85935d1577c7b9c68a358987e643ce558d4a4244a8276302b304e"} Oct 08 18:32:49 crc kubenswrapper[4859]: I1008 18:32:49.240274 4859 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/c860866f-d309-4824-a31c-e88c5d820712-etc-swift\") pod \"swift-storage-0\" (UID: \"c860866f-d309-4824-a31c-e88c5d820712\") " pod="openstack/swift-storage-0" Oct 08 18:32:49 crc kubenswrapper[4859]: E1008 18:32:49.240590 4859 projected.go:288] Couldn't get configMap openstack/swift-ring-files: configmap "swift-ring-files" not found Oct 08 18:32:49 crc kubenswrapper[4859]: E1008 18:32:49.240964 4859 projected.go:194] Error preparing data for projected volume etc-swift for pod openstack/swift-storage-0: configmap "swift-ring-files" not found Oct 08 18:32:49 crc kubenswrapper[4859]: E1008 18:32:49.241113 4859 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/c860866f-d309-4824-a31c-e88c5d820712-etc-swift podName:c860866f-d309-4824-a31c-e88c5d820712 nodeName:}" failed. No retries permitted until 2025-10-08 18:32:57.241076843 +0000 UTC m=+947.487916222 (durationBeforeRetry 8s). Error: MountVolume.SetUp failed for volume "etc-swift" (UniqueName: "kubernetes.io/projected/c860866f-d309-4824-a31c-e88c5d820712-etc-swift") pod "swift-storage-0" (UID: "c860866f-d309-4824-a31c-e88c5d820712") : configmap "swift-ring-files" not found Oct 08 18:32:49 crc kubenswrapper[4859]: I1008 18:32:49.256653 4859 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/placement-b31f-account-create-r48bf"] Oct 08 18:32:49 crc kubenswrapper[4859]: W1008 18:32:49.268338 4859 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod823f174d_4981_4db7_97cb_18d5ec2eead5.slice/crio-9a0d933e40e1e9c70f63c8b2a4d967ba99d7204fe411e3d3d9a25c678ddd990f WatchSource:0}: Error finding container 9a0d933e40e1e9c70f63c8b2a4d967ba99d7204fe411e3d3d9a25c678ddd990f: Status 404 returned error can't find the container with id 9a0d933e40e1e9c70f63c8b2a4d967ba99d7204fe411e3d3d9a25c678ddd990f Oct 08 18:32:49 crc kubenswrapper[4859]: I1008 18:32:49.913955 4859 generic.go:334] "Generic (PLEG): container finished" podID="823f174d-4981-4db7-97cb-18d5ec2eead5" containerID="dca6842589b2acb69ce6753cc593071dc1f48239b1c6a978c540c3ac3a1cb9a3" exitCode=0 Oct 08 18:32:49 crc kubenswrapper[4859]: I1008 18:32:49.914033 4859 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-b31f-account-create-r48bf" event={"ID":"823f174d-4981-4db7-97cb-18d5ec2eead5","Type":"ContainerDied","Data":"dca6842589b2acb69ce6753cc593071dc1f48239b1c6a978c540c3ac3a1cb9a3"} Oct 08 18:32:49 crc kubenswrapper[4859]: I1008 18:32:49.914349 4859 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-b31f-account-create-r48bf" event={"ID":"823f174d-4981-4db7-97cb-18d5ec2eead5","Type":"ContainerStarted","Data":"9a0d933e40e1e9c70f63c8b2a4d967ba99d7204fe411e3d3d9a25c678ddd990f"} Oct 08 18:32:49 crc kubenswrapper[4859]: I1008 18:32:49.917155 4859 generic.go:334] "Generic (PLEG): container finished" podID="1e1f9094-586a-4f98-a7f6-32ca4816fa8b" containerID="b402f3bb9b51f6d60caa8e68d4522dc322ffd9e059d829c1c745f190aaee7b39" exitCode=0 Oct 08 18:32:49 crc kubenswrapper[4859]: I1008 18:32:49.917195 4859 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-db-create-9sdlj" event={"ID":"1e1f9094-586a-4f98-a7f6-32ca4816fa8b","Type":"ContainerDied","Data":"b402f3bb9b51f6d60caa8e68d4522dc322ffd9e059d829c1c745f190aaee7b39"} Oct 08 18:32:50 crc kubenswrapper[4859]: I1008 18:32:50.262258 4859 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/glance-f62d-account-create-pcklr"] Oct 08 18:32:50 crc kubenswrapper[4859]: I1008 18:32:50.263846 4859 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-f62d-account-create-pcklr" Oct 08 18:32:50 crc kubenswrapper[4859]: I1008 18:32:50.274048 4859 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-db-secret" Oct 08 18:32:50 crc kubenswrapper[4859]: I1008 18:32:50.293121 4859 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-f62d-account-create-pcklr"] Oct 08 18:32:50 crc kubenswrapper[4859]: I1008 18:32:50.362835 4859 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-bmq48\" (UniqueName: \"kubernetes.io/projected/c3879c1b-a548-4090-8709-777af70d9578-kube-api-access-bmq48\") pod \"glance-f62d-account-create-pcklr\" (UID: \"c3879c1b-a548-4090-8709-777af70d9578\") " pod="openstack/glance-f62d-account-create-pcklr" Oct 08 18:32:50 crc kubenswrapper[4859]: I1008 18:32:50.452842 4859 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-7b587f8db7-lm5t5" Oct 08 18:32:50 crc kubenswrapper[4859]: I1008 18:32:50.464364 4859 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-bmq48\" (UniqueName: \"kubernetes.io/projected/c3879c1b-a548-4090-8709-777af70d9578-kube-api-access-bmq48\") pod \"glance-f62d-account-create-pcklr\" (UID: \"c3879c1b-a548-4090-8709-777af70d9578\") " pod="openstack/glance-f62d-account-create-pcklr" Oct 08 18:32:50 crc kubenswrapper[4859]: I1008 18:32:50.509713 4859 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-bmq48\" (UniqueName: \"kubernetes.io/projected/c3879c1b-a548-4090-8709-777af70d9578-kube-api-access-bmq48\") pod \"glance-f62d-account-create-pcklr\" (UID: \"c3879c1b-a548-4090-8709-777af70d9578\") " pod="openstack/glance-f62d-account-create-pcklr" Oct 08 18:32:50 crc kubenswrapper[4859]: I1008 18:32:50.525532 4859 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-dc9d58d7-tdt5s"] Oct 08 18:32:50 crc kubenswrapper[4859]: I1008 18:32:50.526279 4859 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-dc9d58d7-tdt5s" podUID="6b7754e8-93c9-4b3a-983c-1127f73459b6" containerName="dnsmasq-dns" containerID="cri-o://1a3ccfc530fa04517677d969892cf026028ca37a8d1773bc7122e74690e2aca6" gracePeriod=10 Oct 08 18:32:50 crc kubenswrapper[4859]: I1008 18:32:50.596253 4859 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-f62d-account-create-pcklr" Oct 08 18:32:50 crc kubenswrapper[4859]: I1008 18:32:50.928583 4859 generic.go:334] "Generic (PLEG): container finished" podID="6b7754e8-93c9-4b3a-983c-1127f73459b6" containerID="1a3ccfc530fa04517677d969892cf026028ca37a8d1773bc7122e74690e2aca6" exitCode=0 Oct 08 18:32:50 crc kubenswrapper[4859]: I1008 18:32:50.928662 4859 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-dc9d58d7-tdt5s" event={"ID":"6b7754e8-93c9-4b3a-983c-1127f73459b6","Type":"ContainerDied","Data":"1a3ccfc530fa04517677d969892cf026028ca37a8d1773bc7122e74690e2aca6"} Oct 08 18:32:51 crc kubenswrapper[4859]: I1008 18:32:51.655418 4859 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-db-create-9sdlj" Oct 08 18:32:51 crc kubenswrapper[4859]: I1008 18:32:51.698785 4859 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/placement-b31f-account-create-r48bf" Oct 08 18:32:51 crc kubenswrapper[4859]: I1008 18:32:51.790195 4859 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-lt4jl\" (UniqueName: \"kubernetes.io/projected/1e1f9094-586a-4f98-a7f6-32ca4816fa8b-kube-api-access-lt4jl\") pod \"1e1f9094-586a-4f98-a7f6-32ca4816fa8b\" (UID: \"1e1f9094-586a-4f98-a7f6-32ca4816fa8b\") " Oct 08 18:32:51 crc kubenswrapper[4859]: I1008 18:32:51.795533 4859 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/1e1f9094-586a-4f98-a7f6-32ca4816fa8b-kube-api-access-lt4jl" (OuterVolumeSpecName: "kube-api-access-lt4jl") pod "1e1f9094-586a-4f98-a7f6-32ca4816fa8b" (UID: "1e1f9094-586a-4f98-a7f6-32ca4816fa8b"). InnerVolumeSpecName "kube-api-access-lt4jl". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 08 18:32:51 crc kubenswrapper[4859]: I1008 18:32:51.826908 4859 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-dc9d58d7-tdt5s" Oct 08 18:32:51 crc kubenswrapper[4859]: I1008 18:32:51.892362 4859 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-sn7l9\" (UniqueName: \"kubernetes.io/projected/823f174d-4981-4db7-97cb-18d5ec2eead5-kube-api-access-sn7l9\") pod \"823f174d-4981-4db7-97cb-18d5ec2eead5\" (UID: \"823f174d-4981-4db7-97cb-18d5ec2eead5\") " Oct 08 18:32:51 crc kubenswrapper[4859]: I1008 18:32:51.892879 4859 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-lt4jl\" (UniqueName: \"kubernetes.io/projected/1e1f9094-586a-4f98-a7f6-32ca4816fa8b-kube-api-access-lt4jl\") on node \"crc\" DevicePath \"\"" Oct 08 18:32:51 crc kubenswrapper[4859]: I1008 18:32:51.900015 4859 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/823f174d-4981-4db7-97cb-18d5ec2eead5-kube-api-access-sn7l9" (OuterVolumeSpecName: "kube-api-access-sn7l9") pod "823f174d-4981-4db7-97cb-18d5ec2eead5" (UID: "823f174d-4981-4db7-97cb-18d5ec2eead5"). InnerVolumeSpecName "kube-api-access-sn7l9". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 08 18:32:51 crc kubenswrapper[4859]: I1008 18:32:51.937425 4859 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-ring-rebalance-qrpd6" event={"ID":"431b4e6e-415d-4c14-8783-f94409691ae3","Type":"ContainerStarted","Data":"8ed73c1ee2d4ba6368087ff364642bdd69d9bbb3ce048cdec6ce2242428874bc"} Oct 08 18:32:51 crc kubenswrapper[4859]: I1008 18:32:51.938916 4859 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-b31f-account-create-r48bf" event={"ID":"823f174d-4981-4db7-97cb-18d5ec2eead5","Type":"ContainerDied","Data":"9a0d933e40e1e9c70f63c8b2a4d967ba99d7204fe411e3d3d9a25c678ddd990f"} Oct 08 18:32:51 crc kubenswrapper[4859]: I1008 18:32:51.938942 4859 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="9a0d933e40e1e9c70f63c8b2a4d967ba99d7204fe411e3d3d9a25c678ddd990f" Oct 08 18:32:51 crc kubenswrapper[4859]: I1008 18:32:51.938939 4859 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/placement-b31f-account-create-r48bf" Oct 08 18:32:51 crc kubenswrapper[4859]: I1008 18:32:51.940730 4859 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-db-create-9sdlj" event={"ID":"1e1f9094-586a-4f98-a7f6-32ca4816fa8b","Type":"ContainerDied","Data":"44ec86d631d85935d1577c7b9c68a358987e643ce558d4a4244a8276302b304e"} Oct 08 18:32:51 crc kubenswrapper[4859]: I1008 18:32:51.940799 4859 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="44ec86d631d85935d1577c7b9c68a358987e643ce558d4a4244a8276302b304e" Oct 08 18:32:51 crc kubenswrapper[4859]: I1008 18:32:51.940840 4859 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-db-create-9sdlj" Oct 08 18:32:51 crc kubenswrapper[4859]: I1008 18:32:51.947426 4859 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-dc9d58d7-tdt5s" event={"ID":"6b7754e8-93c9-4b3a-983c-1127f73459b6","Type":"ContainerDied","Data":"e13f728449ca1682232a1ba8c3c50ec1f4b97b099ebd8d0a9272b3dd169cb997"} Oct 08 18:32:51 crc kubenswrapper[4859]: I1008 18:32:51.947454 4859 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-dc9d58d7-tdt5s" Oct 08 18:32:51 crc kubenswrapper[4859]: I1008 18:32:51.947493 4859 scope.go:117] "RemoveContainer" containerID="1a3ccfc530fa04517677d969892cf026028ca37a8d1773bc7122e74690e2aca6" Oct 08 18:32:51 crc kubenswrapper[4859]: I1008 18:32:51.963591 4859 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/swift-ring-rebalance-qrpd6" podStartSLOduration=1.456332051 podStartE2EDuration="6.96356873s" podCreationTimestamp="2025-10-08 18:32:45 +0000 UTC" firstStartedPulling="2025-10-08 18:32:46.017278023 +0000 UTC m=+936.264117402" lastFinishedPulling="2025-10-08 18:32:51.524514692 +0000 UTC m=+941.771354081" observedRunningTime="2025-10-08 18:32:51.957352307 +0000 UTC m=+942.204191686" watchObservedRunningTime="2025-10-08 18:32:51.96356873 +0000 UTC m=+942.210408109" Oct 08 18:32:51 crc kubenswrapper[4859]: I1008 18:32:51.968886 4859 scope.go:117] "RemoveContainer" containerID="6405e0cc748bc4ff427676070a865dceab5700fd5524c4a4eb073ddf9a15f340" Oct 08 18:32:51 crc kubenswrapper[4859]: I1008 18:32:51.993800 4859 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/6b7754e8-93c9-4b3a-983c-1127f73459b6-ovsdbserver-nb\") pod \"6b7754e8-93c9-4b3a-983c-1127f73459b6\" (UID: \"6b7754e8-93c9-4b3a-983c-1127f73459b6\") " Oct 08 18:32:51 crc kubenswrapper[4859]: I1008 18:32:51.993916 4859 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/6b7754e8-93c9-4b3a-983c-1127f73459b6-ovsdbserver-sb\") pod \"6b7754e8-93c9-4b3a-983c-1127f73459b6\" (UID: \"6b7754e8-93c9-4b3a-983c-1127f73459b6\") " Oct 08 18:32:51 crc kubenswrapper[4859]: I1008 18:32:51.993994 4859 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-5qjcb\" (UniqueName: \"kubernetes.io/projected/6b7754e8-93c9-4b3a-983c-1127f73459b6-kube-api-access-5qjcb\") pod \"6b7754e8-93c9-4b3a-983c-1127f73459b6\" (UID: \"6b7754e8-93c9-4b3a-983c-1127f73459b6\") " Oct 08 18:32:51 crc kubenswrapper[4859]: I1008 18:32:51.994101 4859 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/6b7754e8-93c9-4b3a-983c-1127f73459b6-dns-svc\") pod \"6b7754e8-93c9-4b3a-983c-1127f73459b6\" (UID: \"6b7754e8-93c9-4b3a-983c-1127f73459b6\") " Oct 08 18:32:51 crc kubenswrapper[4859]: I1008 18:32:51.994146 4859 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/6b7754e8-93c9-4b3a-983c-1127f73459b6-config\") pod \"6b7754e8-93c9-4b3a-983c-1127f73459b6\" (UID: \"6b7754e8-93c9-4b3a-983c-1127f73459b6\") " Oct 08 18:32:51 crc kubenswrapper[4859]: I1008 18:32:51.994592 4859 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-sn7l9\" (UniqueName: \"kubernetes.io/projected/823f174d-4981-4db7-97cb-18d5ec2eead5-kube-api-access-sn7l9\") on node \"crc\" DevicePath \"\"" Oct 08 18:32:52 crc kubenswrapper[4859]: I1008 18:32:52.000192 4859 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6b7754e8-93c9-4b3a-983c-1127f73459b6-kube-api-access-5qjcb" (OuterVolumeSpecName: "kube-api-access-5qjcb") pod "6b7754e8-93c9-4b3a-983c-1127f73459b6" (UID: "6b7754e8-93c9-4b3a-983c-1127f73459b6"). InnerVolumeSpecName "kube-api-access-5qjcb". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 08 18:32:52 crc kubenswrapper[4859]: I1008 18:32:52.033233 4859 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6b7754e8-93c9-4b3a-983c-1127f73459b6-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "6b7754e8-93c9-4b3a-983c-1127f73459b6" (UID: "6b7754e8-93c9-4b3a-983c-1127f73459b6"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 08 18:32:52 crc kubenswrapper[4859]: I1008 18:32:52.040773 4859 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6b7754e8-93c9-4b3a-983c-1127f73459b6-config" (OuterVolumeSpecName: "config") pod "6b7754e8-93c9-4b3a-983c-1127f73459b6" (UID: "6b7754e8-93c9-4b3a-983c-1127f73459b6"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 08 18:32:52 crc kubenswrapper[4859]: I1008 18:32:52.044276 4859 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6b7754e8-93c9-4b3a-983c-1127f73459b6-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "6b7754e8-93c9-4b3a-983c-1127f73459b6" (UID: "6b7754e8-93c9-4b3a-983c-1127f73459b6"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 08 18:32:52 crc kubenswrapper[4859]: I1008 18:32:52.046456 4859 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-f62d-account-create-pcklr"] Oct 08 18:32:52 crc kubenswrapper[4859]: I1008 18:32:52.046895 4859 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6b7754e8-93c9-4b3a-983c-1127f73459b6-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "6b7754e8-93c9-4b3a-983c-1127f73459b6" (UID: "6b7754e8-93c9-4b3a-983c-1127f73459b6"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 08 18:32:52 crc kubenswrapper[4859]: W1008 18:32:52.050560 4859 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podc3879c1b_a548_4090_8709_777af70d9578.slice/crio-df517a18cbf1e784f086b8c612d7ed200e44279a68b64d89bbcf7cbe877f3584 WatchSource:0}: Error finding container df517a18cbf1e784f086b8c612d7ed200e44279a68b64d89bbcf7cbe877f3584: Status 404 returned error can't find the container with id df517a18cbf1e784f086b8c612d7ed200e44279a68b64d89bbcf7cbe877f3584 Oct 08 18:32:52 crc kubenswrapper[4859]: I1008 18:32:52.097475 4859 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/6b7754e8-93c9-4b3a-983c-1127f73459b6-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Oct 08 18:32:52 crc kubenswrapper[4859]: I1008 18:32:52.097563 4859 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/6b7754e8-93c9-4b3a-983c-1127f73459b6-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Oct 08 18:32:52 crc kubenswrapper[4859]: I1008 18:32:52.097579 4859 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-5qjcb\" (UniqueName: \"kubernetes.io/projected/6b7754e8-93c9-4b3a-983c-1127f73459b6-kube-api-access-5qjcb\") on node \"crc\" DevicePath \"\"" Oct 08 18:32:52 crc kubenswrapper[4859]: I1008 18:32:52.097595 4859 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/6b7754e8-93c9-4b3a-983c-1127f73459b6-dns-svc\") on node \"crc\" DevicePath \"\"" Oct 08 18:32:52 crc kubenswrapper[4859]: I1008 18:32:52.097608 4859 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/6b7754e8-93c9-4b3a-983c-1127f73459b6-config\") on node \"crc\" DevicePath \"\"" Oct 08 18:32:52 crc kubenswrapper[4859]: I1008 18:32:52.325620 4859 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-dc9d58d7-tdt5s"] Oct 08 18:32:52 crc kubenswrapper[4859]: I1008 18:32:52.333448 4859 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-dc9d58d7-tdt5s"] Oct 08 18:32:52 crc kubenswrapper[4859]: I1008 18:32:52.483330 4859 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="6b7754e8-93c9-4b3a-983c-1127f73459b6" path="/var/lib/kubelet/pods/6b7754e8-93c9-4b3a-983c-1127f73459b6/volumes" Oct 08 18:32:52 crc kubenswrapper[4859]: I1008 18:32:52.957627 4859 generic.go:334] "Generic (PLEG): container finished" podID="c3879c1b-a548-4090-8709-777af70d9578" containerID="fa34d069118c0da8f3d0adada988ca9e9482254ede94e90a33bd0b93333db806" exitCode=0 Oct 08 18:32:52 crc kubenswrapper[4859]: I1008 18:32:52.957742 4859 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-f62d-account-create-pcklr" event={"ID":"c3879c1b-a548-4090-8709-777af70d9578","Type":"ContainerDied","Data":"fa34d069118c0da8f3d0adada988ca9e9482254ede94e90a33bd0b93333db806"} Oct 08 18:32:52 crc kubenswrapper[4859]: I1008 18:32:52.957780 4859 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-f62d-account-create-pcklr" event={"ID":"c3879c1b-a548-4090-8709-777af70d9578","Type":"ContainerStarted","Data":"df517a18cbf1e784f086b8c612d7ed200e44279a68b64d89bbcf7cbe877f3584"} Oct 08 18:32:54 crc kubenswrapper[4859]: I1008 18:32:54.311052 4859 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-f62d-account-create-pcklr" Oct 08 18:32:54 crc kubenswrapper[4859]: I1008 18:32:54.394979 4859 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/ovn-controller-b89lv" podUID="ec20820a-066c-4606-8f1b-a26b172b7720" containerName="ovn-controller" probeResult="failure" output=< Oct 08 18:32:54 crc kubenswrapper[4859]: ERROR - ovn-controller connection status is 'not connected', expecting 'connected' status Oct 08 18:32:54 crc kubenswrapper[4859]: > Oct 08 18:32:54 crc kubenswrapper[4859]: I1008 18:32:54.444910 4859 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-bmq48\" (UniqueName: \"kubernetes.io/projected/c3879c1b-a548-4090-8709-777af70d9578-kube-api-access-bmq48\") pod \"c3879c1b-a548-4090-8709-777af70d9578\" (UID: \"c3879c1b-a548-4090-8709-777af70d9578\") " Oct 08 18:32:54 crc kubenswrapper[4859]: I1008 18:32:54.455921 4859 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/c3879c1b-a548-4090-8709-777af70d9578-kube-api-access-bmq48" (OuterVolumeSpecName: "kube-api-access-bmq48") pod "c3879c1b-a548-4090-8709-777af70d9578" (UID: "c3879c1b-a548-4090-8709-777af70d9578"). InnerVolumeSpecName "kube-api-access-bmq48". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 08 18:32:54 crc kubenswrapper[4859]: I1008 18:32:54.547277 4859 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-bmq48\" (UniqueName: \"kubernetes.io/projected/c3879c1b-a548-4090-8709-777af70d9578-kube-api-access-bmq48\") on node \"crc\" DevicePath \"\"" Oct 08 18:32:54 crc kubenswrapper[4859]: I1008 18:32:54.985262 4859 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-f62d-account-create-pcklr" event={"ID":"c3879c1b-a548-4090-8709-777af70d9578","Type":"ContainerDied","Data":"df517a18cbf1e784f086b8c612d7ed200e44279a68b64d89bbcf7cbe877f3584"} Oct 08 18:32:54 crc kubenswrapper[4859]: I1008 18:32:54.985326 4859 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="df517a18cbf1e784f086b8c612d7ed200e44279a68b64d89bbcf7cbe877f3584" Oct 08 18:32:54 crc kubenswrapper[4859]: I1008 18:32:54.985404 4859 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-f62d-account-create-pcklr" Oct 08 18:32:57 crc kubenswrapper[4859]: I1008 18:32:57.007134 4859 generic.go:334] "Generic (PLEG): container finished" podID="ba5a78be-e7b3-401a-bdb9-ba31071d7075" containerID="6e2de3cebf597d7492590e385a7bc16bfc073ff1e91df7b11869d97d8960f7dd" exitCode=0 Oct 08 18:32:57 crc kubenswrapper[4859]: I1008 18:32:57.007241 4859 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-server-0" event={"ID":"ba5a78be-e7b3-401a-bdb9-ba31071d7075","Type":"ContainerDied","Data":"6e2de3cebf597d7492590e385a7bc16bfc073ff1e91df7b11869d97d8960f7dd"} Oct 08 18:32:57 crc kubenswrapper[4859]: I1008 18:32:57.011292 4859 generic.go:334] "Generic (PLEG): container finished" podID="3075d5c2-3e74-4a4c-829f-248e87f45f3b" containerID="fecccb4aab6e4695efa36ed5ba960994a10727b9aac69914dd84867213d10810" exitCode=0 Oct 08 18:32:57 crc kubenswrapper[4859]: I1008 18:32:57.011363 4859 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-cell1-server-0" event={"ID":"3075d5c2-3e74-4a4c-829f-248e87f45f3b","Type":"ContainerDied","Data":"fecccb4aab6e4695efa36ed5ba960994a10727b9aac69914dd84867213d10810"} Oct 08 18:32:57 crc kubenswrapper[4859]: I1008 18:32:57.304843 4859 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/c860866f-d309-4824-a31c-e88c5d820712-etc-swift\") pod \"swift-storage-0\" (UID: \"c860866f-d309-4824-a31c-e88c5d820712\") " pod="openstack/swift-storage-0" Oct 08 18:32:57 crc kubenswrapper[4859]: E1008 18:32:57.305145 4859 projected.go:288] Couldn't get configMap openstack/swift-ring-files: configmap "swift-ring-files" not found Oct 08 18:32:57 crc kubenswrapper[4859]: E1008 18:32:57.305386 4859 projected.go:194] Error preparing data for projected volume etc-swift for pod openstack/swift-storage-0: configmap "swift-ring-files" not found Oct 08 18:32:57 crc kubenswrapper[4859]: E1008 18:32:57.305460 4859 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/c860866f-d309-4824-a31c-e88c5d820712-etc-swift podName:c860866f-d309-4824-a31c-e88c5d820712 nodeName:}" failed. No retries permitted until 2025-10-08 18:33:13.305431493 +0000 UTC m=+963.552270872 (durationBeforeRetry 16s). Error: MountVolume.SetUp failed for volume "etc-swift" (UniqueName: "kubernetes.io/projected/c860866f-d309-4824-a31c-e88c5d820712-etc-swift") pod "swift-storage-0" (UID: "c860866f-d309-4824-a31c-e88c5d820712") : configmap "swift-ring-files" not found Oct 08 18:32:58 crc kubenswrapper[4859]: I1008 18:32:58.024194 4859 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-cell1-server-0" event={"ID":"3075d5c2-3e74-4a4c-829f-248e87f45f3b","Type":"ContainerStarted","Data":"56c96709e82513e92c3d9367fe8c571724d57b12903fa91ea473fc776eee2ec6"} Oct 08 18:32:58 crc kubenswrapper[4859]: I1008 18:32:58.024760 4859 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/rabbitmq-cell1-server-0" Oct 08 18:32:58 crc kubenswrapper[4859]: I1008 18:32:58.027100 4859 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-server-0" event={"ID":"ba5a78be-e7b3-401a-bdb9-ba31071d7075","Type":"ContainerStarted","Data":"1ff56b52d2d31d1f67ef0e984e2373890024eeaf7a31c66677cb07168c88da23"} Oct 08 18:32:58 crc kubenswrapper[4859]: I1008 18:32:58.027569 4859 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/rabbitmq-server-0" Oct 08 18:32:58 crc kubenswrapper[4859]: I1008 18:32:58.058462 4859 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/rabbitmq-cell1-server-0" podStartSLOduration=55.867809056 podStartE2EDuration="1m5.058430759s" podCreationTimestamp="2025-10-08 18:31:53 +0000 UTC" firstStartedPulling="2025-10-08 18:32:09.107852484 +0000 UTC m=+899.354691863" lastFinishedPulling="2025-10-08 18:32:18.298474187 +0000 UTC m=+908.545313566" observedRunningTime="2025-10-08 18:32:58.04967295 +0000 UTC m=+948.296512359" watchObservedRunningTime="2025-10-08 18:32:58.058430759 +0000 UTC m=+948.305270138" Oct 08 18:32:58 crc kubenswrapper[4859]: I1008 18:32:58.091177 4859 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/keystone-f7b9-account-create-47kvw"] Oct 08 18:32:58 crc kubenswrapper[4859]: E1008 18:32:58.091639 4859 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c3879c1b-a548-4090-8709-777af70d9578" containerName="mariadb-account-create" Oct 08 18:32:58 crc kubenswrapper[4859]: I1008 18:32:58.091667 4859 state_mem.go:107] "Deleted CPUSet assignment" podUID="c3879c1b-a548-4090-8709-777af70d9578" containerName="mariadb-account-create" Oct 08 18:32:58 crc kubenswrapper[4859]: E1008 18:32:58.091708 4859 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="6b7754e8-93c9-4b3a-983c-1127f73459b6" containerName="init" Oct 08 18:32:58 crc kubenswrapper[4859]: I1008 18:32:58.091717 4859 state_mem.go:107] "Deleted CPUSet assignment" podUID="6b7754e8-93c9-4b3a-983c-1127f73459b6" containerName="init" Oct 08 18:32:58 crc kubenswrapper[4859]: E1008 18:32:58.091735 4859 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="6b7754e8-93c9-4b3a-983c-1127f73459b6" containerName="dnsmasq-dns" Oct 08 18:32:58 crc kubenswrapper[4859]: I1008 18:32:58.091743 4859 state_mem.go:107] "Deleted CPUSet assignment" podUID="6b7754e8-93c9-4b3a-983c-1127f73459b6" containerName="dnsmasq-dns" Oct 08 18:32:58 crc kubenswrapper[4859]: E1008 18:32:58.091760 4859 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1e1f9094-586a-4f98-a7f6-32ca4816fa8b" containerName="mariadb-database-create" Oct 08 18:32:58 crc kubenswrapper[4859]: I1008 18:32:58.091771 4859 state_mem.go:107] "Deleted CPUSet assignment" podUID="1e1f9094-586a-4f98-a7f6-32ca4816fa8b" containerName="mariadb-database-create" Oct 08 18:32:58 crc kubenswrapper[4859]: E1008 18:32:58.091786 4859 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="823f174d-4981-4db7-97cb-18d5ec2eead5" containerName="mariadb-account-create" Oct 08 18:32:58 crc kubenswrapper[4859]: I1008 18:32:58.091793 4859 state_mem.go:107] "Deleted CPUSet assignment" podUID="823f174d-4981-4db7-97cb-18d5ec2eead5" containerName="mariadb-account-create" Oct 08 18:32:58 crc kubenswrapper[4859]: I1008 18:32:58.091985 4859 memory_manager.go:354] "RemoveStaleState removing state" podUID="823f174d-4981-4db7-97cb-18d5ec2eead5" containerName="mariadb-account-create" Oct 08 18:32:58 crc kubenswrapper[4859]: I1008 18:32:58.092001 4859 memory_manager.go:354] "RemoveStaleState removing state" podUID="6b7754e8-93c9-4b3a-983c-1127f73459b6" containerName="dnsmasq-dns" Oct 08 18:32:58 crc kubenswrapper[4859]: I1008 18:32:58.092016 4859 memory_manager.go:354] "RemoveStaleState removing state" podUID="c3879c1b-a548-4090-8709-777af70d9578" containerName="mariadb-account-create" Oct 08 18:32:58 crc kubenswrapper[4859]: I1008 18:32:58.092032 4859 memory_manager.go:354] "RemoveStaleState removing state" podUID="1e1f9094-586a-4f98-a7f6-32ca4816fa8b" containerName="mariadb-database-create" Oct 08 18:32:58 crc kubenswrapper[4859]: I1008 18:32:58.092785 4859 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-f7b9-account-create-47kvw" Oct 08 18:32:58 crc kubenswrapper[4859]: I1008 18:32:58.094056 4859 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/rabbitmq-server-0" podStartSLOduration=55.179552139 podStartE2EDuration="1m5.094030839s" podCreationTimestamp="2025-10-08 18:31:53 +0000 UTC" firstStartedPulling="2025-10-08 18:32:08.972744604 +0000 UTC m=+899.219583973" lastFinishedPulling="2025-10-08 18:32:18.887223304 +0000 UTC m=+909.134062673" observedRunningTime="2025-10-08 18:32:58.09136386 +0000 UTC m=+948.338203239" watchObservedRunningTime="2025-10-08 18:32:58.094030839 +0000 UTC m=+948.340870218" Oct 08 18:32:58 crc kubenswrapper[4859]: I1008 18:32:58.107715 4859 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-db-secret" Oct 08 18:32:58 crc kubenswrapper[4859]: I1008 18:32:58.113925 4859 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-f7b9-account-create-47kvw"] Oct 08 18:32:58 crc kubenswrapper[4859]: I1008 18:32:58.220996 4859 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-fbl4k\" (UniqueName: \"kubernetes.io/projected/b44043d2-d505-4e7f-8431-f42a3738ca8a-kube-api-access-fbl4k\") pod \"keystone-f7b9-account-create-47kvw\" (UID: \"b44043d2-d505-4e7f-8431-f42a3738ca8a\") " pod="openstack/keystone-f7b9-account-create-47kvw" Oct 08 18:32:58 crc kubenswrapper[4859]: I1008 18:32:58.324117 4859 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-fbl4k\" (UniqueName: \"kubernetes.io/projected/b44043d2-d505-4e7f-8431-f42a3738ca8a-kube-api-access-fbl4k\") pod \"keystone-f7b9-account-create-47kvw\" (UID: \"b44043d2-d505-4e7f-8431-f42a3738ca8a\") " pod="openstack/keystone-f7b9-account-create-47kvw" Oct 08 18:32:58 crc kubenswrapper[4859]: I1008 18:32:58.351643 4859 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-fbl4k\" (UniqueName: \"kubernetes.io/projected/b44043d2-d505-4e7f-8431-f42a3738ca8a-kube-api-access-fbl4k\") pod \"keystone-f7b9-account-create-47kvw\" (UID: \"b44043d2-d505-4e7f-8431-f42a3738ca8a\") " pod="openstack/keystone-f7b9-account-create-47kvw" Oct 08 18:32:58 crc kubenswrapper[4859]: I1008 18:32:58.412071 4859 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-f7b9-account-create-47kvw" Oct 08 18:32:58 crc kubenswrapper[4859]: I1008 18:32:58.887338 4859 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-f7b9-account-create-47kvw"] Oct 08 18:32:58 crc kubenswrapper[4859]: W1008 18:32:58.893841 4859 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podb44043d2_d505_4e7f_8431_f42a3738ca8a.slice/crio-a2076720ad9786d004b3aca83e14e20c24d14cd08b295395ed844a23c43895bf WatchSource:0}: Error finding container a2076720ad9786d004b3aca83e14e20c24d14cd08b295395ed844a23c43895bf: Status 404 returned error can't find the container with id a2076720ad9786d004b3aca83e14e20c24d14cd08b295395ed844a23c43895bf Oct 08 18:32:59 crc kubenswrapper[4859]: I1008 18:32:59.035825 4859 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-f7b9-account-create-47kvw" event={"ID":"b44043d2-d505-4e7f-8431-f42a3738ca8a","Type":"ContainerStarted","Data":"a2076720ad9786d004b3aca83e14e20c24d14cd08b295395ed844a23c43895bf"} Oct 08 18:32:59 crc kubenswrapper[4859]: I1008 18:32:59.400628 4859 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/ovn-controller-b89lv" podUID="ec20820a-066c-4606-8f1b-a26b172b7720" containerName="ovn-controller" probeResult="failure" output=< Oct 08 18:32:59 crc kubenswrapper[4859]: ERROR - ovn-controller connection status is 'not connected', expecting 'connected' status Oct 08 18:32:59 crc kubenswrapper[4859]: > Oct 08 18:32:59 crc kubenswrapper[4859]: I1008 18:32:59.457944 4859 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/ovn-controller-ovs-vthdk" Oct 08 18:32:59 crc kubenswrapper[4859]: I1008 18:32:59.464586 4859 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/ovn-controller-ovs-vthdk" Oct 08 18:32:59 crc kubenswrapper[4859]: I1008 18:32:59.701306 4859 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ovn-controller-b89lv-config-mq7sr"] Oct 08 18:32:59 crc kubenswrapper[4859]: I1008 18:32:59.702514 4859 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-b89lv-config-mq7sr" Oct 08 18:32:59 crc kubenswrapper[4859]: I1008 18:32:59.704964 4859 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovncontroller-extra-scripts" Oct 08 18:32:59 crc kubenswrapper[4859]: I1008 18:32:59.734149 4859 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-controller-b89lv-config-mq7sr"] Oct 08 18:32:59 crc kubenswrapper[4859]: I1008 18:32:59.851930 4859 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-log-ovn\" (UniqueName: \"kubernetes.io/host-path/bb094b68-378d-4c33-81db-d31d58bdeadf-var-log-ovn\") pod \"ovn-controller-b89lv-config-mq7sr\" (UID: \"bb094b68-378d-4c33-81db-d31d58bdeadf\") " pod="openstack/ovn-controller-b89lv-config-mq7sr" Oct 08 18:32:59 crc kubenswrapper[4859]: I1008 18:32:59.851995 4859 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-l55bl\" (UniqueName: \"kubernetes.io/projected/bb094b68-378d-4c33-81db-d31d58bdeadf-kube-api-access-l55bl\") pod \"ovn-controller-b89lv-config-mq7sr\" (UID: \"bb094b68-378d-4c33-81db-d31d58bdeadf\") " pod="openstack/ovn-controller-b89lv-config-mq7sr" Oct 08 18:32:59 crc kubenswrapper[4859]: I1008 18:32:59.852033 4859 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/bb094b68-378d-4c33-81db-d31d58bdeadf-var-run\") pod \"ovn-controller-b89lv-config-mq7sr\" (UID: \"bb094b68-378d-4c33-81db-d31d58bdeadf\") " pod="openstack/ovn-controller-b89lv-config-mq7sr" Oct 08 18:32:59 crc kubenswrapper[4859]: I1008 18:32:59.852053 4859 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"additional-scripts\" (UniqueName: \"kubernetes.io/configmap/bb094b68-378d-4c33-81db-d31d58bdeadf-additional-scripts\") pod \"ovn-controller-b89lv-config-mq7sr\" (UID: \"bb094b68-378d-4c33-81db-d31d58bdeadf\") " pod="openstack/ovn-controller-b89lv-config-mq7sr" Oct 08 18:32:59 crc kubenswrapper[4859]: I1008 18:32:59.852139 4859 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-run-ovn\" (UniqueName: \"kubernetes.io/host-path/bb094b68-378d-4c33-81db-d31d58bdeadf-var-run-ovn\") pod \"ovn-controller-b89lv-config-mq7sr\" (UID: \"bb094b68-378d-4c33-81db-d31d58bdeadf\") " pod="openstack/ovn-controller-b89lv-config-mq7sr" Oct 08 18:32:59 crc kubenswrapper[4859]: I1008 18:32:59.852264 4859 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/bb094b68-378d-4c33-81db-d31d58bdeadf-scripts\") pod \"ovn-controller-b89lv-config-mq7sr\" (UID: \"bb094b68-378d-4c33-81db-d31d58bdeadf\") " pod="openstack/ovn-controller-b89lv-config-mq7sr" Oct 08 18:32:59 crc kubenswrapper[4859]: I1008 18:32:59.953972 4859 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-l55bl\" (UniqueName: \"kubernetes.io/projected/bb094b68-378d-4c33-81db-d31d58bdeadf-kube-api-access-l55bl\") pod \"ovn-controller-b89lv-config-mq7sr\" (UID: \"bb094b68-378d-4c33-81db-d31d58bdeadf\") " pod="openstack/ovn-controller-b89lv-config-mq7sr" Oct 08 18:32:59 crc kubenswrapper[4859]: I1008 18:32:59.954058 4859 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/bb094b68-378d-4c33-81db-d31d58bdeadf-var-run\") pod \"ovn-controller-b89lv-config-mq7sr\" (UID: \"bb094b68-378d-4c33-81db-d31d58bdeadf\") " pod="openstack/ovn-controller-b89lv-config-mq7sr" Oct 08 18:32:59 crc kubenswrapper[4859]: I1008 18:32:59.954082 4859 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"additional-scripts\" (UniqueName: \"kubernetes.io/configmap/bb094b68-378d-4c33-81db-d31d58bdeadf-additional-scripts\") pod \"ovn-controller-b89lv-config-mq7sr\" (UID: \"bb094b68-378d-4c33-81db-d31d58bdeadf\") " pod="openstack/ovn-controller-b89lv-config-mq7sr" Oct 08 18:32:59 crc kubenswrapper[4859]: I1008 18:32:59.954165 4859 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-run-ovn\" (UniqueName: \"kubernetes.io/host-path/bb094b68-378d-4c33-81db-d31d58bdeadf-var-run-ovn\") pod \"ovn-controller-b89lv-config-mq7sr\" (UID: \"bb094b68-378d-4c33-81db-d31d58bdeadf\") " pod="openstack/ovn-controller-b89lv-config-mq7sr" Oct 08 18:32:59 crc kubenswrapper[4859]: I1008 18:32:59.954220 4859 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/bb094b68-378d-4c33-81db-d31d58bdeadf-scripts\") pod \"ovn-controller-b89lv-config-mq7sr\" (UID: \"bb094b68-378d-4c33-81db-d31d58bdeadf\") " pod="openstack/ovn-controller-b89lv-config-mq7sr" Oct 08 18:32:59 crc kubenswrapper[4859]: I1008 18:32:59.954287 4859 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-log-ovn\" (UniqueName: \"kubernetes.io/host-path/bb094b68-378d-4c33-81db-d31d58bdeadf-var-log-ovn\") pod \"ovn-controller-b89lv-config-mq7sr\" (UID: \"bb094b68-378d-4c33-81db-d31d58bdeadf\") " pod="openstack/ovn-controller-b89lv-config-mq7sr" Oct 08 18:32:59 crc kubenswrapper[4859]: I1008 18:32:59.954659 4859 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-log-ovn\" (UniqueName: \"kubernetes.io/host-path/bb094b68-378d-4c33-81db-d31d58bdeadf-var-log-ovn\") pod \"ovn-controller-b89lv-config-mq7sr\" (UID: \"bb094b68-378d-4c33-81db-d31d58bdeadf\") " pod="openstack/ovn-controller-b89lv-config-mq7sr" Oct 08 18:32:59 crc kubenswrapper[4859]: I1008 18:32:59.955132 4859 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/bb094b68-378d-4c33-81db-d31d58bdeadf-var-run\") pod \"ovn-controller-b89lv-config-mq7sr\" (UID: \"bb094b68-378d-4c33-81db-d31d58bdeadf\") " pod="openstack/ovn-controller-b89lv-config-mq7sr" Oct 08 18:32:59 crc kubenswrapper[4859]: I1008 18:32:59.955861 4859 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"additional-scripts\" (UniqueName: \"kubernetes.io/configmap/bb094b68-378d-4c33-81db-d31d58bdeadf-additional-scripts\") pod \"ovn-controller-b89lv-config-mq7sr\" (UID: \"bb094b68-378d-4c33-81db-d31d58bdeadf\") " pod="openstack/ovn-controller-b89lv-config-mq7sr" Oct 08 18:32:59 crc kubenswrapper[4859]: I1008 18:32:59.955915 4859 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-run-ovn\" (UniqueName: \"kubernetes.io/host-path/bb094b68-378d-4c33-81db-d31d58bdeadf-var-run-ovn\") pod \"ovn-controller-b89lv-config-mq7sr\" (UID: \"bb094b68-378d-4c33-81db-d31d58bdeadf\") " pod="openstack/ovn-controller-b89lv-config-mq7sr" Oct 08 18:32:59 crc kubenswrapper[4859]: I1008 18:32:59.958113 4859 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/bb094b68-378d-4c33-81db-d31d58bdeadf-scripts\") pod \"ovn-controller-b89lv-config-mq7sr\" (UID: \"bb094b68-378d-4c33-81db-d31d58bdeadf\") " pod="openstack/ovn-controller-b89lv-config-mq7sr" Oct 08 18:32:59 crc kubenswrapper[4859]: I1008 18:32:59.978390 4859 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-l55bl\" (UniqueName: \"kubernetes.io/projected/bb094b68-378d-4c33-81db-d31d58bdeadf-kube-api-access-l55bl\") pod \"ovn-controller-b89lv-config-mq7sr\" (UID: \"bb094b68-378d-4c33-81db-d31d58bdeadf\") " pod="openstack/ovn-controller-b89lv-config-mq7sr" Oct 08 18:33:00 crc kubenswrapper[4859]: I1008 18:33:00.028153 4859 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-b89lv-config-mq7sr" Oct 08 18:33:00 crc kubenswrapper[4859]: I1008 18:33:00.046129 4859 generic.go:334] "Generic (PLEG): container finished" podID="b44043d2-d505-4e7f-8431-f42a3738ca8a" containerID="74719fdf7321a3335282b9fa039ba753eb5cb2640408902b649f9ee2b83c0268" exitCode=0 Oct 08 18:33:00 crc kubenswrapper[4859]: I1008 18:33:00.046238 4859 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-f7b9-account-create-47kvw" event={"ID":"b44043d2-d505-4e7f-8431-f42a3738ca8a","Type":"ContainerDied","Data":"74719fdf7321a3335282b9fa039ba753eb5cb2640408902b649f9ee2b83c0268"} Oct 08 18:33:00 crc kubenswrapper[4859]: I1008 18:33:00.048396 4859 generic.go:334] "Generic (PLEG): container finished" podID="431b4e6e-415d-4c14-8783-f94409691ae3" containerID="8ed73c1ee2d4ba6368087ff364642bdd69d9bbb3ce048cdec6ce2242428874bc" exitCode=0 Oct 08 18:33:00 crc kubenswrapper[4859]: I1008 18:33:00.048547 4859 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-ring-rebalance-qrpd6" event={"ID":"431b4e6e-415d-4c14-8783-f94409691ae3","Type":"ContainerDied","Data":"8ed73c1ee2d4ba6368087ff364642bdd69d9bbb3ce048cdec6ce2242428874bc"} Oct 08 18:33:00 crc kubenswrapper[4859]: I1008 18:33:00.496157 4859 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-controller-b89lv-config-mq7sr"] Oct 08 18:33:00 crc kubenswrapper[4859]: I1008 18:33:00.524386 4859 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/glance-db-sync-xk8qg"] Oct 08 18:33:00 crc kubenswrapper[4859]: I1008 18:33:00.526628 4859 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-db-sync-xk8qg" Oct 08 18:33:00 crc kubenswrapper[4859]: I1008 18:33:00.535842 4859 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-config-data" Oct 08 18:33:00 crc kubenswrapper[4859]: I1008 18:33:00.536338 4859 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-glance-dockercfg-2fnvk" Oct 08 18:33:00 crc kubenswrapper[4859]: I1008 18:33:00.543001 4859 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-db-sync-xk8qg"] Oct 08 18:33:00 crc kubenswrapper[4859]: I1008 18:33:00.667486 4859 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-lczsb\" (UniqueName: \"kubernetes.io/projected/9c2ff226-ee9a-4cb3-b66f-9df0c3497748-kube-api-access-lczsb\") pod \"glance-db-sync-xk8qg\" (UID: \"9c2ff226-ee9a-4cb3-b66f-9df0c3497748\") " pod="openstack/glance-db-sync-xk8qg" Oct 08 18:33:00 crc kubenswrapper[4859]: I1008 18:33:00.667595 4859 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9c2ff226-ee9a-4cb3-b66f-9df0c3497748-combined-ca-bundle\") pod \"glance-db-sync-xk8qg\" (UID: \"9c2ff226-ee9a-4cb3-b66f-9df0c3497748\") " pod="openstack/glance-db-sync-xk8qg" Oct 08 18:33:00 crc kubenswrapper[4859]: I1008 18:33:00.667671 4859 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/9c2ff226-ee9a-4cb3-b66f-9df0c3497748-config-data\") pod \"glance-db-sync-xk8qg\" (UID: \"9c2ff226-ee9a-4cb3-b66f-9df0c3497748\") " pod="openstack/glance-db-sync-xk8qg" Oct 08 18:33:00 crc kubenswrapper[4859]: I1008 18:33:00.667894 4859 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/9c2ff226-ee9a-4cb3-b66f-9df0c3497748-db-sync-config-data\") pod \"glance-db-sync-xk8qg\" (UID: \"9c2ff226-ee9a-4cb3-b66f-9df0c3497748\") " pod="openstack/glance-db-sync-xk8qg" Oct 08 18:33:00 crc kubenswrapper[4859]: I1008 18:33:00.770121 4859 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-lczsb\" (UniqueName: \"kubernetes.io/projected/9c2ff226-ee9a-4cb3-b66f-9df0c3497748-kube-api-access-lczsb\") pod \"glance-db-sync-xk8qg\" (UID: \"9c2ff226-ee9a-4cb3-b66f-9df0c3497748\") " pod="openstack/glance-db-sync-xk8qg" Oct 08 18:33:00 crc kubenswrapper[4859]: I1008 18:33:00.771769 4859 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9c2ff226-ee9a-4cb3-b66f-9df0c3497748-combined-ca-bundle\") pod \"glance-db-sync-xk8qg\" (UID: \"9c2ff226-ee9a-4cb3-b66f-9df0c3497748\") " pod="openstack/glance-db-sync-xk8qg" Oct 08 18:33:00 crc kubenswrapper[4859]: I1008 18:33:00.772896 4859 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/9c2ff226-ee9a-4cb3-b66f-9df0c3497748-config-data\") pod \"glance-db-sync-xk8qg\" (UID: \"9c2ff226-ee9a-4cb3-b66f-9df0c3497748\") " pod="openstack/glance-db-sync-xk8qg" Oct 08 18:33:00 crc kubenswrapper[4859]: I1008 18:33:00.772982 4859 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/9c2ff226-ee9a-4cb3-b66f-9df0c3497748-db-sync-config-data\") pod \"glance-db-sync-xk8qg\" (UID: \"9c2ff226-ee9a-4cb3-b66f-9df0c3497748\") " pod="openstack/glance-db-sync-xk8qg" Oct 08 18:33:00 crc kubenswrapper[4859]: I1008 18:33:00.778351 4859 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/9c2ff226-ee9a-4cb3-b66f-9df0c3497748-db-sync-config-data\") pod \"glance-db-sync-xk8qg\" (UID: \"9c2ff226-ee9a-4cb3-b66f-9df0c3497748\") " pod="openstack/glance-db-sync-xk8qg" Oct 08 18:33:00 crc kubenswrapper[4859]: I1008 18:33:00.778401 4859 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9c2ff226-ee9a-4cb3-b66f-9df0c3497748-combined-ca-bundle\") pod \"glance-db-sync-xk8qg\" (UID: \"9c2ff226-ee9a-4cb3-b66f-9df0c3497748\") " pod="openstack/glance-db-sync-xk8qg" Oct 08 18:33:00 crc kubenswrapper[4859]: I1008 18:33:00.783674 4859 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/9c2ff226-ee9a-4cb3-b66f-9df0c3497748-config-data\") pod \"glance-db-sync-xk8qg\" (UID: \"9c2ff226-ee9a-4cb3-b66f-9df0c3497748\") " pod="openstack/glance-db-sync-xk8qg" Oct 08 18:33:00 crc kubenswrapper[4859]: I1008 18:33:00.794170 4859 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-lczsb\" (UniqueName: \"kubernetes.io/projected/9c2ff226-ee9a-4cb3-b66f-9df0c3497748-kube-api-access-lczsb\") pod \"glance-db-sync-xk8qg\" (UID: \"9c2ff226-ee9a-4cb3-b66f-9df0c3497748\") " pod="openstack/glance-db-sync-xk8qg" Oct 08 18:33:00 crc kubenswrapper[4859]: I1008 18:33:00.863784 4859 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-db-sync-xk8qg" Oct 08 18:33:01 crc kubenswrapper[4859]: I1008 18:33:01.062899 4859 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-b89lv-config-mq7sr" event={"ID":"bb094b68-378d-4c33-81db-d31d58bdeadf","Type":"ContainerStarted","Data":"f6db0e633c22c7dd66148f73868bba2294e2402d3003b5e69e1a17c426ad5c1c"} Oct 08 18:33:01 crc kubenswrapper[4859]: I1008 18:33:01.063308 4859 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-b89lv-config-mq7sr" event={"ID":"bb094b68-378d-4c33-81db-d31d58bdeadf","Type":"ContainerStarted","Data":"095713e095feaed839034eb8d205264126b24881a274d3e24a13e217c8fb1017"} Oct 08 18:33:01 crc kubenswrapper[4859]: I1008 18:33:01.092964 4859 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ovn-controller-b89lv-config-mq7sr" podStartSLOduration=2.092937639 podStartE2EDuration="2.092937639s" podCreationTimestamp="2025-10-08 18:32:59 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-08 18:33:01.084872811 +0000 UTC m=+951.331712200" watchObservedRunningTime="2025-10-08 18:33:01.092937639 +0000 UTC m=+951.339777018" Oct 08 18:33:01 crc kubenswrapper[4859]: I1008 18:33:01.444252 4859 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-f7b9-account-create-47kvw" Oct 08 18:33:01 crc kubenswrapper[4859]: I1008 18:33:01.503917 4859 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-db-sync-xk8qg"] Oct 08 18:33:01 crc kubenswrapper[4859]: W1008 18:33:01.513274 4859 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod9c2ff226_ee9a_4cb3_b66f_9df0c3497748.slice/crio-024d3bf433e6715d5b3a98f9acdd357677562ec6274a20d769a481bdb033f266 WatchSource:0}: Error finding container 024d3bf433e6715d5b3a98f9acdd357677562ec6274a20d769a481bdb033f266: Status 404 returned error can't find the container with id 024d3bf433e6715d5b3a98f9acdd357677562ec6274a20d769a481bdb033f266 Oct 08 18:33:01 crc kubenswrapper[4859]: I1008 18:33:01.560945 4859 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/swift-ring-rebalance-qrpd6" Oct 08 18:33:01 crc kubenswrapper[4859]: I1008 18:33:01.590141 4859 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-fbl4k\" (UniqueName: \"kubernetes.io/projected/b44043d2-d505-4e7f-8431-f42a3738ca8a-kube-api-access-fbl4k\") pod \"b44043d2-d505-4e7f-8431-f42a3738ca8a\" (UID: \"b44043d2-d505-4e7f-8431-f42a3738ca8a\") " Oct 08 18:33:01 crc kubenswrapper[4859]: I1008 18:33:01.603295 4859 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b44043d2-d505-4e7f-8431-f42a3738ca8a-kube-api-access-fbl4k" (OuterVolumeSpecName: "kube-api-access-fbl4k") pod "b44043d2-d505-4e7f-8431-f42a3738ca8a" (UID: "b44043d2-d505-4e7f-8431-f42a3738ca8a"). InnerVolumeSpecName "kube-api-access-fbl4k". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 08 18:33:01 crc kubenswrapper[4859]: I1008 18:33:01.692248 4859 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/431b4e6e-415d-4c14-8783-f94409691ae3-scripts\") pod \"431b4e6e-415d-4c14-8783-f94409691ae3\" (UID: \"431b4e6e-415d-4c14-8783-f94409691ae3\") " Oct 08 18:33:01 crc kubenswrapper[4859]: I1008 18:33:01.692594 4859 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-twhsd\" (UniqueName: \"kubernetes.io/projected/431b4e6e-415d-4c14-8783-f94409691ae3-kube-api-access-twhsd\") pod \"431b4e6e-415d-4c14-8783-f94409691ae3\" (UID: \"431b4e6e-415d-4c14-8783-f94409691ae3\") " Oct 08 18:33:01 crc kubenswrapper[4859]: I1008 18:33:01.692767 4859 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/431b4e6e-415d-4c14-8783-f94409691ae3-combined-ca-bundle\") pod \"431b4e6e-415d-4c14-8783-f94409691ae3\" (UID: \"431b4e6e-415d-4c14-8783-f94409691ae3\") " Oct 08 18:33:01 crc kubenswrapper[4859]: I1008 18:33:01.693293 4859 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/431b4e6e-415d-4c14-8783-f94409691ae3-ring-data-devices\") pod \"431b4e6e-415d-4c14-8783-f94409691ae3\" (UID: \"431b4e6e-415d-4c14-8783-f94409691ae3\") " Oct 08 18:33:01 crc kubenswrapper[4859]: I1008 18:33:01.693383 4859 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/431b4e6e-415d-4c14-8783-f94409691ae3-swiftconf\") pod \"431b4e6e-415d-4c14-8783-f94409691ae3\" (UID: \"431b4e6e-415d-4c14-8783-f94409691ae3\") " Oct 08 18:33:01 crc kubenswrapper[4859]: I1008 18:33:01.693526 4859 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/431b4e6e-415d-4c14-8783-f94409691ae3-etc-swift\") pod \"431b4e6e-415d-4c14-8783-f94409691ae3\" (UID: \"431b4e6e-415d-4c14-8783-f94409691ae3\") " Oct 08 18:33:01 crc kubenswrapper[4859]: I1008 18:33:01.693614 4859 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/431b4e6e-415d-4c14-8783-f94409691ae3-dispersionconf\") pod \"431b4e6e-415d-4c14-8783-f94409691ae3\" (UID: \"431b4e6e-415d-4c14-8783-f94409691ae3\") " Oct 08 18:33:01 crc kubenswrapper[4859]: I1008 18:33:01.694242 4859 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/431b4e6e-415d-4c14-8783-f94409691ae3-ring-data-devices" (OuterVolumeSpecName: "ring-data-devices") pod "431b4e6e-415d-4c14-8783-f94409691ae3" (UID: "431b4e6e-415d-4c14-8783-f94409691ae3"). InnerVolumeSpecName "ring-data-devices". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 08 18:33:01 crc kubenswrapper[4859]: I1008 18:33:01.694292 4859 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-fbl4k\" (UniqueName: \"kubernetes.io/projected/b44043d2-d505-4e7f-8431-f42a3738ca8a-kube-api-access-fbl4k\") on node \"crc\" DevicePath \"\"" Oct 08 18:33:01 crc kubenswrapper[4859]: I1008 18:33:01.694452 4859 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/431b4e6e-415d-4c14-8783-f94409691ae3-etc-swift" (OuterVolumeSpecName: "etc-swift") pod "431b4e6e-415d-4c14-8783-f94409691ae3" (UID: "431b4e6e-415d-4c14-8783-f94409691ae3"). InnerVolumeSpecName "etc-swift". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 08 18:33:01 crc kubenswrapper[4859]: I1008 18:33:01.697919 4859 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/431b4e6e-415d-4c14-8783-f94409691ae3-kube-api-access-twhsd" (OuterVolumeSpecName: "kube-api-access-twhsd") pod "431b4e6e-415d-4c14-8783-f94409691ae3" (UID: "431b4e6e-415d-4c14-8783-f94409691ae3"). InnerVolumeSpecName "kube-api-access-twhsd". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 08 18:33:01 crc kubenswrapper[4859]: I1008 18:33:01.700581 4859 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/431b4e6e-415d-4c14-8783-f94409691ae3-dispersionconf" (OuterVolumeSpecName: "dispersionconf") pod "431b4e6e-415d-4c14-8783-f94409691ae3" (UID: "431b4e6e-415d-4c14-8783-f94409691ae3"). InnerVolumeSpecName "dispersionconf". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 08 18:33:01 crc kubenswrapper[4859]: I1008 18:33:01.719011 4859 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/431b4e6e-415d-4c14-8783-f94409691ae3-scripts" (OuterVolumeSpecName: "scripts") pod "431b4e6e-415d-4c14-8783-f94409691ae3" (UID: "431b4e6e-415d-4c14-8783-f94409691ae3"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 08 18:33:01 crc kubenswrapper[4859]: I1008 18:33:01.719463 4859 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/431b4e6e-415d-4c14-8783-f94409691ae3-swiftconf" (OuterVolumeSpecName: "swiftconf") pod "431b4e6e-415d-4c14-8783-f94409691ae3" (UID: "431b4e6e-415d-4c14-8783-f94409691ae3"). InnerVolumeSpecName "swiftconf". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 08 18:33:01 crc kubenswrapper[4859]: I1008 18:33:01.721743 4859 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/431b4e6e-415d-4c14-8783-f94409691ae3-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "431b4e6e-415d-4c14-8783-f94409691ae3" (UID: "431b4e6e-415d-4c14-8783-f94409691ae3"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 08 18:33:01 crc kubenswrapper[4859]: I1008 18:33:01.796378 4859 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/431b4e6e-415d-4c14-8783-f94409691ae3-scripts\") on node \"crc\" DevicePath \"\"" Oct 08 18:33:01 crc kubenswrapper[4859]: I1008 18:33:01.796427 4859 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-twhsd\" (UniqueName: \"kubernetes.io/projected/431b4e6e-415d-4c14-8783-f94409691ae3-kube-api-access-twhsd\") on node \"crc\" DevicePath \"\"" Oct 08 18:33:01 crc kubenswrapper[4859]: I1008 18:33:01.796442 4859 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/431b4e6e-415d-4c14-8783-f94409691ae3-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 08 18:33:01 crc kubenswrapper[4859]: I1008 18:33:01.796453 4859 reconciler_common.go:293] "Volume detached for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/431b4e6e-415d-4c14-8783-f94409691ae3-ring-data-devices\") on node \"crc\" DevicePath \"\"" Oct 08 18:33:01 crc kubenswrapper[4859]: I1008 18:33:01.796462 4859 reconciler_common.go:293] "Volume detached for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/431b4e6e-415d-4c14-8783-f94409691ae3-swiftconf\") on node \"crc\" DevicePath \"\"" Oct 08 18:33:01 crc kubenswrapper[4859]: I1008 18:33:01.796476 4859 reconciler_common.go:293] "Volume detached for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/431b4e6e-415d-4c14-8783-f94409691ae3-etc-swift\") on node \"crc\" DevicePath \"\"" Oct 08 18:33:01 crc kubenswrapper[4859]: I1008 18:33:01.796485 4859 reconciler_common.go:293] "Volume detached for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/431b4e6e-415d-4c14-8783-f94409691ae3-dispersionconf\") on node \"crc\" DevicePath \"\"" Oct 08 18:33:02 crc kubenswrapper[4859]: I1008 18:33:02.072651 4859 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-f7b9-account-create-47kvw" event={"ID":"b44043d2-d505-4e7f-8431-f42a3738ca8a","Type":"ContainerDied","Data":"a2076720ad9786d004b3aca83e14e20c24d14cd08b295395ed844a23c43895bf"} Oct 08 18:33:02 crc kubenswrapper[4859]: I1008 18:33:02.074253 4859 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="a2076720ad9786d004b3aca83e14e20c24d14cd08b295395ed844a23c43895bf" Oct 08 18:33:02 crc kubenswrapper[4859]: I1008 18:33:02.074450 4859 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-f7b9-account-create-47kvw" Oct 08 18:33:02 crc kubenswrapper[4859]: I1008 18:33:02.077776 4859 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-ring-rebalance-qrpd6" event={"ID":"431b4e6e-415d-4c14-8783-f94409691ae3","Type":"ContainerDied","Data":"539152ac4a4d7a20de6eea2fe06d75749ae2e36da39b01c442f930ad8fabd528"} Oct 08 18:33:02 crc kubenswrapper[4859]: I1008 18:33:02.077817 4859 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="539152ac4a4d7a20de6eea2fe06d75749ae2e36da39b01c442f930ad8fabd528" Oct 08 18:33:02 crc kubenswrapper[4859]: I1008 18:33:02.077858 4859 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/swift-ring-rebalance-qrpd6" Oct 08 18:33:02 crc kubenswrapper[4859]: I1008 18:33:02.084411 4859 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-db-sync-xk8qg" event={"ID":"9c2ff226-ee9a-4cb3-b66f-9df0c3497748","Type":"ContainerStarted","Data":"024d3bf433e6715d5b3a98f9acdd357677562ec6274a20d769a481bdb033f266"} Oct 08 18:33:02 crc kubenswrapper[4859]: I1008 18:33:02.086501 4859 generic.go:334] "Generic (PLEG): container finished" podID="bb094b68-378d-4c33-81db-d31d58bdeadf" containerID="f6db0e633c22c7dd66148f73868bba2294e2402d3003b5e69e1a17c426ad5c1c" exitCode=0 Oct 08 18:33:02 crc kubenswrapper[4859]: I1008 18:33:02.086558 4859 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-b89lv-config-mq7sr" event={"ID":"bb094b68-378d-4c33-81db-d31d58bdeadf","Type":"ContainerDied","Data":"f6db0e633c22c7dd66148f73868bba2294e2402d3003b5e69e1a17c426ad5c1c"} Oct 08 18:33:03 crc kubenswrapper[4859]: I1008 18:33:03.458608 4859 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-b89lv-config-mq7sr" Oct 08 18:33:03 crc kubenswrapper[4859]: I1008 18:33:03.629635 4859 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/bb094b68-378d-4c33-81db-d31d58bdeadf-scripts\") pod \"bb094b68-378d-4c33-81db-d31d58bdeadf\" (UID: \"bb094b68-378d-4c33-81db-d31d58bdeadf\") " Oct 08 18:33:03 crc kubenswrapper[4859]: I1008 18:33:03.629763 4859 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"var-log-ovn\" (UniqueName: \"kubernetes.io/host-path/bb094b68-378d-4c33-81db-d31d58bdeadf-var-log-ovn\") pod \"bb094b68-378d-4c33-81db-d31d58bdeadf\" (UID: \"bb094b68-378d-4c33-81db-d31d58bdeadf\") " Oct 08 18:33:03 crc kubenswrapper[4859]: I1008 18:33:03.629810 4859 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/bb094b68-378d-4c33-81db-d31d58bdeadf-var-log-ovn" (OuterVolumeSpecName: "var-log-ovn") pod "bb094b68-378d-4c33-81db-d31d58bdeadf" (UID: "bb094b68-378d-4c33-81db-d31d58bdeadf"). InnerVolumeSpecName "var-log-ovn". PluginName "kubernetes.io/host-path", VolumeGidValue "" Oct 08 18:33:03 crc kubenswrapper[4859]: I1008 18:33:03.629869 4859 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"var-run-ovn\" (UniqueName: \"kubernetes.io/host-path/bb094b68-378d-4c33-81db-d31d58bdeadf-var-run-ovn\") pod \"bb094b68-378d-4c33-81db-d31d58bdeadf\" (UID: \"bb094b68-378d-4c33-81db-d31d58bdeadf\") " Oct 08 18:33:03 crc kubenswrapper[4859]: I1008 18:33:03.629915 4859 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"additional-scripts\" (UniqueName: \"kubernetes.io/configmap/bb094b68-378d-4c33-81db-d31d58bdeadf-additional-scripts\") pod \"bb094b68-378d-4c33-81db-d31d58bdeadf\" (UID: \"bb094b68-378d-4c33-81db-d31d58bdeadf\") " Oct 08 18:33:03 crc kubenswrapper[4859]: I1008 18:33:03.629990 4859 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-l55bl\" (UniqueName: \"kubernetes.io/projected/bb094b68-378d-4c33-81db-d31d58bdeadf-kube-api-access-l55bl\") pod \"bb094b68-378d-4c33-81db-d31d58bdeadf\" (UID: \"bb094b68-378d-4c33-81db-d31d58bdeadf\") " Oct 08 18:33:03 crc kubenswrapper[4859]: I1008 18:33:03.630048 4859 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/bb094b68-378d-4c33-81db-d31d58bdeadf-var-run\") pod \"bb094b68-378d-4c33-81db-d31d58bdeadf\" (UID: \"bb094b68-378d-4c33-81db-d31d58bdeadf\") " Oct 08 18:33:03 crc kubenswrapper[4859]: I1008 18:33:03.630277 4859 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/bb094b68-378d-4c33-81db-d31d58bdeadf-var-run-ovn" (OuterVolumeSpecName: "var-run-ovn") pod "bb094b68-378d-4c33-81db-d31d58bdeadf" (UID: "bb094b68-378d-4c33-81db-d31d58bdeadf"). InnerVolumeSpecName "var-run-ovn". PluginName "kubernetes.io/host-path", VolumeGidValue "" Oct 08 18:33:03 crc kubenswrapper[4859]: I1008 18:33:03.630456 4859 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/bb094b68-378d-4c33-81db-d31d58bdeadf-var-run" (OuterVolumeSpecName: "var-run") pod "bb094b68-378d-4c33-81db-d31d58bdeadf" (UID: "bb094b68-378d-4c33-81db-d31d58bdeadf"). InnerVolumeSpecName "var-run". PluginName "kubernetes.io/host-path", VolumeGidValue "" Oct 08 18:33:03 crc kubenswrapper[4859]: I1008 18:33:03.630861 4859 reconciler_common.go:293] "Volume detached for volume \"var-log-ovn\" (UniqueName: \"kubernetes.io/host-path/bb094b68-378d-4c33-81db-d31d58bdeadf-var-log-ovn\") on node \"crc\" DevicePath \"\"" Oct 08 18:33:03 crc kubenswrapper[4859]: I1008 18:33:03.630882 4859 reconciler_common.go:293] "Volume detached for volume \"var-run-ovn\" (UniqueName: \"kubernetes.io/host-path/bb094b68-378d-4c33-81db-d31d58bdeadf-var-run-ovn\") on node \"crc\" DevicePath \"\"" Oct 08 18:33:03 crc kubenswrapper[4859]: I1008 18:33:03.630875 4859 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/bb094b68-378d-4c33-81db-d31d58bdeadf-additional-scripts" (OuterVolumeSpecName: "additional-scripts") pod "bb094b68-378d-4c33-81db-d31d58bdeadf" (UID: "bb094b68-378d-4c33-81db-d31d58bdeadf"). InnerVolumeSpecName "additional-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 08 18:33:03 crc kubenswrapper[4859]: I1008 18:33:03.630895 4859 reconciler_common.go:293] "Volume detached for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/bb094b68-378d-4c33-81db-d31d58bdeadf-var-run\") on node \"crc\" DevicePath \"\"" Oct 08 18:33:03 crc kubenswrapper[4859]: I1008 18:33:03.631765 4859 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/bb094b68-378d-4c33-81db-d31d58bdeadf-scripts" (OuterVolumeSpecName: "scripts") pod "bb094b68-378d-4c33-81db-d31d58bdeadf" (UID: "bb094b68-378d-4c33-81db-d31d58bdeadf"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 08 18:33:03 crc kubenswrapper[4859]: I1008 18:33:03.640570 4859 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/bb094b68-378d-4c33-81db-d31d58bdeadf-kube-api-access-l55bl" (OuterVolumeSpecName: "kube-api-access-l55bl") pod "bb094b68-378d-4c33-81db-d31d58bdeadf" (UID: "bb094b68-378d-4c33-81db-d31d58bdeadf"). InnerVolumeSpecName "kube-api-access-l55bl". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 08 18:33:03 crc kubenswrapper[4859]: I1008 18:33:03.733187 4859 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/bb094b68-378d-4c33-81db-d31d58bdeadf-scripts\") on node \"crc\" DevicePath \"\"" Oct 08 18:33:03 crc kubenswrapper[4859]: I1008 18:33:03.733255 4859 reconciler_common.go:293] "Volume detached for volume \"additional-scripts\" (UniqueName: \"kubernetes.io/configmap/bb094b68-378d-4c33-81db-d31d58bdeadf-additional-scripts\") on node \"crc\" DevicePath \"\"" Oct 08 18:33:03 crc kubenswrapper[4859]: I1008 18:33:03.733275 4859 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-l55bl\" (UniqueName: \"kubernetes.io/projected/bb094b68-378d-4c33-81db-d31d58bdeadf-kube-api-access-l55bl\") on node \"crc\" DevicePath \"\"" Oct 08 18:33:04 crc kubenswrapper[4859]: I1008 18:33:04.123056 4859 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-b89lv-config-mq7sr" event={"ID":"bb094b68-378d-4c33-81db-d31d58bdeadf","Type":"ContainerDied","Data":"095713e095feaed839034eb8d205264126b24881a274d3e24a13e217c8fb1017"} Oct 08 18:33:04 crc kubenswrapper[4859]: I1008 18:33:04.123112 4859 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="095713e095feaed839034eb8d205264126b24881a274d3e24a13e217c8fb1017" Oct 08 18:33:04 crc kubenswrapper[4859]: I1008 18:33:04.123157 4859 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-b89lv-config-mq7sr" Oct 08 18:33:04 crc kubenswrapper[4859]: I1008 18:33:04.197962 4859 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ovn-controller-b89lv-config-mq7sr"] Oct 08 18:33:04 crc kubenswrapper[4859]: I1008 18:33:04.211227 4859 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/ovn-controller-b89lv-config-mq7sr"] Oct 08 18:33:04 crc kubenswrapper[4859]: I1008 18:33:04.402617 4859 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/ovn-controller-b89lv" Oct 08 18:33:04 crc kubenswrapper[4859]: I1008 18:33:04.480482 4859 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="bb094b68-378d-4c33-81db-d31d58bdeadf" path="/var/lib/kubelet/pods/bb094b68-378d-4c33-81db-d31d58bdeadf/volumes" Oct 08 18:33:13 crc kubenswrapper[4859]: I1008 18:33:13.330508 4859 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/c860866f-d309-4824-a31c-e88c5d820712-etc-swift\") pod \"swift-storage-0\" (UID: \"c860866f-d309-4824-a31c-e88c5d820712\") " pod="openstack/swift-storage-0" Oct 08 18:33:13 crc kubenswrapper[4859]: I1008 18:33:13.338212 4859 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/c860866f-d309-4824-a31c-e88c5d820712-etc-swift\") pod \"swift-storage-0\" (UID: \"c860866f-d309-4824-a31c-e88c5d820712\") " pod="openstack/swift-storage-0" Oct 08 18:33:13 crc kubenswrapper[4859]: I1008 18:33:13.484636 4859 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/swift-storage-0" Oct 08 18:33:14 crc kubenswrapper[4859]: I1008 18:33:14.333597 4859 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/swift-storage-0"] Oct 08 18:33:14 crc kubenswrapper[4859]: I1008 18:33:14.906928 4859 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/rabbitmq-server-0" Oct 08 18:33:15 crc kubenswrapper[4859]: I1008 18:33:15.198900 4859 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/rabbitmq-cell1-server-0" Oct 08 18:33:15 crc kubenswrapper[4859]: I1008 18:33:15.251095 4859 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"c860866f-d309-4824-a31c-e88c5d820712","Type":"ContainerStarted","Data":"c1a58b6d578330bfc3921b12a2a3319cd1332b6266ca08bbe90d88e06eebdfa6"} Oct 08 18:33:15 crc kubenswrapper[4859]: I1008 18:33:15.257901 4859 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-db-sync-xk8qg" event={"ID":"9c2ff226-ee9a-4cb3-b66f-9df0c3497748","Type":"ContainerStarted","Data":"ba60ab3185c0a8ffefed44fd91a99acbf6a00486ec248ab2c7753e30b2f4a4a9"} Oct 08 18:33:15 crc kubenswrapper[4859]: I1008 18:33:15.291814 4859 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/cinder-db-create-wczl5"] Oct 08 18:33:15 crc kubenswrapper[4859]: E1008 18:33:15.292149 4859 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="431b4e6e-415d-4c14-8783-f94409691ae3" containerName="swift-ring-rebalance" Oct 08 18:33:15 crc kubenswrapper[4859]: I1008 18:33:15.292169 4859 state_mem.go:107] "Deleted CPUSet assignment" podUID="431b4e6e-415d-4c14-8783-f94409691ae3" containerName="swift-ring-rebalance" Oct 08 18:33:15 crc kubenswrapper[4859]: E1008 18:33:15.292202 4859 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="bb094b68-378d-4c33-81db-d31d58bdeadf" containerName="ovn-config" Oct 08 18:33:15 crc kubenswrapper[4859]: I1008 18:33:15.292211 4859 state_mem.go:107] "Deleted CPUSet assignment" podUID="bb094b68-378d-4c33-81db-d31d58bdeadf" containerName="ovn-config" Oct 08 18:33:15 crc kubenswrapper[4859]: E1008 18:33:15.292235 4859 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b44043d2-d505-4e7f-8431-f42a3738ca8a" containerName="mariadb-account-create" Oct 08 18:33:15 crc kubenswrapper[4859]: I1008 18:33:15.292243 4859 state_mem.go:107] "Deleted CPUSet assignment" podUID="b44043d2-d505-4e7f-8431-f42a3738ca8a" containerName="mariadb-account-create" Oct 08 18:33:15 crc kubenswrapper[4859]: I1008 18:33:15.292428 4859 memory_manager.go:354] "RemoveStaleState removing state" podUID="431b4e6e-415d-4c14-8783-f94409691ae3" containerName="swift-ring-rebalance" Oct 08 18:33:15 crc kubenswrapper[4859]: I1008 18:33:15.292454 4859 memory_manager.go:354] "RemoveStaleState removing state" podUID="b44043d2-d505-4e7f-8431-f42a3738ca8a" containerName="mariadb-account-create" Oct 08 18:33:15 crc kubenswrapper[4859]: I1008 18:33:15.292465 4859 memory_manager.go:354] "RemoveStaleState removing state" podUID="bb094b68-378d-4c33-81db-d31d58bdeadf" containerName="ovn-config" Oct 08 18:33:15 crc kubenswrapper[4859]: I1008 18:33:15.293001 4859 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-db-create-wczl5" Oct 08 18:33:15 crc kubenswrapper[4859]: I1008 18:33:15.306126 4859 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-db-create-wczl5"] Oct 08 18:33:15 crc kubenswrapper[4859]: I1008 18:33:15.313352 4859 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/glance-db-sync-xk8qg" podStartSLOduration=2.903456841 podStartE2EDuration="15.313331812s" podCreationTimestamp="2025-10-08 18:33:00 +0000 UTC" firstStartedPulling="2025-10-08 18:33:01.51583743 +0000 UTC m=+951.762676809" lastFinishedPulling="2025-10-08 18:33:13.925712401 +0000 UTC m=+964.172551780" observedRunningTime="2025-10-08 18:33:15.294858227 +0000 UTC m=+965.541697616" watchObservedRunningTime="2025-10-08 18:33:15.313331812 +0000 UTC m=+965.560171191" Oct 08 18:33:15 crc kubenswrapper[4859]: I1008 18:33:15.374177 4859 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-lfr57\" (UniqueName: \"kubernetes.io/projected/84999f2e-5fbf-46d7-8334-436d1f9b12d7-kube-api-access-lfr57\") pod \"cinder-db-create-wczl5\" (UID: \"84999f2e-5fbf-46d7-8334-436d1f9b12d7\") " pod="openstack/cinder-db-create-wczl5" Oct 08 18:33:15 crc kubenswrapper[4859]: I1008 18:33:15.467589 4859 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/barbican-db-create-dbftt"] Oct 08 18:33:15 crc kubenswrapper[4859]: I1008 18:33:15.468603 4859 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-db-create-dbftt" Oct 08 18:33:15 crc kubenswrapper[4859]: I1008 18:33:15.476060 4859 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-lfr57\" (UniqueName: \"kubernetes.io/projected/84999f2e-5fbf-46d7-8334-436d1f9b12d7-kube-api-access-lfr57\") pod \"cinder-db-create-wczl5\" (UID: \"84999f2e-5fbf-46d7-8334-436d1f9b12d7\") " pod="openstack/cinder-db-create-wczl5" Oct 08 18:33:15 crc kubenswrapper[4859]: I1008 18:33:15.481992 4859 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-db-create-dbftt"] Oct 08 18:33:15 crc kubenswrapper[4859]: I1008 18:33:15.499356 4859 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-lfr57\" (UniqueName: \"kubernetes.io/projected/84999f2e-5fbf-46d7-8334-436d1f9b12d7-kube-api-access-lfr57\") pod \"cinder-db-create-wczl5\" (UID: \"84999f2e-5fbf-46d7-8334-436d1f9b12d7\") " pod="openstack/cinder-db-create-wczl5" Oct 08 18:33:15 crc kubenswrapper[4859]: I1008 18:33:15.577770 4859 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-m2fm4\" (UniqueName: \"kubernetes.io/projected/b51059c4-e80f-4c66-be2d-c4ec6b9deaf9-kube-api-access-m2fm4\") pod \"barbican-db-create-dbftt\" (UID: \"b51059c4-e80f-4c66-be2d-c4ec6b9deaf9\") " pod="openstack/barbican-db-create-dbftt" Oct 08 18:33:15 crc kubenswrapper[4859]: I1008 18:33:15.613894 4859 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-db-create-wczl5" Oct 08 18:33:15 crc kubenswrapper[4859]: I1008 18:33:15.641584 4859 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/neutron-db-create-p8mb2"] Oct 08 18:33:15 crc kubenswrapper[4859]: I1008 18:33:15.642586 4859 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-db-create-p8mb2" Oct 08 18:33:15 crc kubenswrapper[4859]: I1008 18:33:15.666922 4859 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-db-create-p8mb2"] Oct 08 18:33:15 crc kubenswrapper[4859]: I1008 18:33:15.690879 4859 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-m2fm4\" (UniqueName: \"kubernetes.io/projected/b51059c4-e80f-4c66-be2d-c4ec6b9deaf9-kube-api-access-m2fm4\") pod \"barbican-db-create-dbftt\" (UID: \"b51059c4-e80f-4c66-be2d-c4ec6b9deaf9\") " pod="openstack/barbican-db-create-dbftt" Oct 08 18:33:15 crc kubenswrapper[4859]: I1008 18:33:15.717538 4859 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-m2fm4\" (UniqueName: \"kubernetes.io/projected/b51059c4-e80f-4c66-be2d-c4ec6b9deaf9-kube-api-access-m2fm4\") pod \"barbican-db-create-dbftt\" (UID: \"b51059c4-e80f-4c66-be2d-c4ec6b9deaf9\") " pod="openstack/barbican-db-create-dbftt" Oct 08 18:33:15 crc kubenswrapper[4859]: I1008 18:33:15.784494 4859 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-db-create-dbftt" Oct 08 18:33:15 crc kubenswrapper[4859]: I1008 18:33:15.792956 4859 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-lkq7m\" (UniqueName: \"kubernetes.io/projected/c0da13c2-e693-4da8-83fa-ae3032d30ec5-kube-api-access-lkq7m\") pod \"neutron-db-create-p8mb2\" (UID: \"c0da13c2-e693-4da8-83fa-ae3032d30ec5\") " pod="openstack/neutron-db-create-p8mb2" Oct 08 18:33:15 crc kubenswrapper[4859]: I1008 18:33:15.828641 4859 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/keystone-db-sync-gsq5w"] Oct 08 18:33:15 crc kubenswrapper[4859]: I1008 18:33:15.830763 4859 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-db-sync-gsq5w" Oct 08 18:33:15 crc kubenswrapper[4859]: I1008 18:33:15.840922 4859 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-scripts" Oct 08 18:33:15 crc kubenswrapper[4859]: I1008 18:33:15.840930 4859 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone" Oct 08 18:33:15 crc kubenswrapper[4859]: I1008 18:33:15.840973 4859 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-config-data" Oct 08 18:33:15 crc kubenswrapper[4859]: I1008 18:33:15.841363 4859 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-keystone-dockercfg-jrkml" Oct 08 18:33:15 crc kubenswrapper[4859]: I1008 18:33:15.858133 4859 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-db-sync-gsq5w"] Oct 08 18:33:15 crc kubenswrapper[4859]: I1008 18:33:15.897058 4859 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-lkq7m\" (UniqueName: \"kubernetes.io/projected/c0da13c2-e693-4da8-83fa-ae3032d30ec5-kube-api-access-lkq7m\") pod \"neutron-db-create-p8mb2\" (UID: \"c0da13c2-e693-4da8-83fa-ae3032d30ec5\") " pod="openstack/neutron-db-create-p8mb2" Oct 08 18:33:15 crc kubenswrapper[4859]: I1008 18:33:15.919540 4859 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-lkq7m\" (UniqueName: \"kubernetes.io/projected/c0da13c2-e693-4da8-83fa-ae3032d30ec5-kube-api-access-lkq7m\") pod \"neutron-db-create-p8mb2\" (UID: \"c0da13c2-e693-4da8-83fa-ae3032d30ec5\") " pod="openstack/neutron-db-create-p8mb2" Oct 08 18:33:15 crc kubenswrapper[4859]: I1008 18:33:15.999274 4859 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-4x9vr\" (UniqueName: \"kubernetes.io/projected/498d1fd9-4fd1-4dce-a122-6f79d4274882-kube-api-access-4x9vr\") pod \"keystone-db-sync-gsq5w\" (UID: \"498d1fd9-4fd1-4dce-a122-6f79d4274882\") " pod="openstack/keystone-db-sync-gsq5w" Oct 08 18:33:15 crc kubenswrapper[4859]: I1008 18:33:15.999344 4859 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/498d1fd9-4fd1-4dce-a122-6f79d4274882-config-data\") pod \"keystone-db-sync-gsq5w\" (UID: \"498d1fd9-4fd1-4dce-a122-6f79d4274882\") " pod="openstack/keystone-db-sync-gsq5w" Oct 08 18:33:15 crc kubenswrapper[4859]: I1008 18:33:15.999416 4859 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/498d1fd9-4fd1-4dce-a122-6f79d4274882-combined-ca-bundle\") pod \"keystone-db-sync-gsq5w\" (UID: \"498d1fd9-4fd1-4dce-a122-6f79d4274882\") " pod="openstack/keystone-db-sync-gsq5w" Oct 08 18:33:16 crc kubenswrapper[4859]: I1008 18:33:16.054386 4859 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-db-create-p8mb2" Oct 08 18:33:16 crc kubenswrapper[4859]: I1008 18:33:16.103338 4859 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-db-create-dbftt"] Oct 08 18:33:16 crc kubenswrapper[4859]: I1008 18:33:16.103769 4859 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/498d1fd9-4fd1-4dce-a122-6f79d4274882-combined-ca-bundle\") pod \"keystone-db-sync-gsq5w\" (UID: \"498d1fd9-4fd1-4dce-a122-6f79d4274882\") " pod="openstack/keystone-db-sync-gsq5w" Oct 08 18:33:16 crc kubenswrapper[4859]: I1008 18:33:16.103935 4859 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-4x9vr\" (UniqueName: \"kubernetes.io/projected/498d1fd9-4fd1-4dce-a122-6f79d4274882-kube-api-access-4x9vr\") pod \"keystone-db-sync-gsq5w\" (UID: \"498d1fd9-4fd1-4dce-a122-6f79d4274882\") " pod="openstack/keystone-db-sync-gsq5w" Oct 08 18:33:16 crc kubenswrapper[4859]: I1008 18:33:16.103968 4859 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/498d1fd9-4fd1-4dce-a122-6f79d4274882-config-data\") pod \"keystone-db-sync-gsq5w\" (UID: \"498d1fd9-4fd1-4dce-a122-6f79d4274882\") " pod="openstack/keystone-db-sync-gsq5w" Oct 08 18:33:16 crc kubenswrapper[4859]: I1008 18:33:16.109485 4859 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/498d1fd9-4fd1-4dce-a122-6f79d4274882-config-data\") pod \"keystone-db-sync-gsq5w\" (UID: \"498d1fd9-4fd1-4dce-a122-6f79d4274882\") " pod="openstack/keystone-db-sync-gsq5w" Oct 08 18:33:16 crc kubenswrapper[4859]: I1008 18:33:16.109970 4859 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/498d1fd9-4fd1-4dce-a122-6f79d4274882-combined-ca-bundle\") pod \"keystone-db-sync-gsq5w\" (UID: \"498d1fd9-4fd1-4dce-a122-6f79d4274882\") " pod="openstack/keystone-db-sync-gsq5w" Oct 08 18:33:16 crc kubenswrapper[4859]: I1008 18:33:16.126266 4859 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-4x9vr\" (UniqueName: \"kubernetes.io/projected/498d1fd9-4fd1-4dce-a122-6f79d4274882-kube-api-access-4x9vr\") pod \"keystone-db-sync-gsq5w\" (UID: \"498d1fd9-4fd1-4dce-a122-6f79d4274882\") " pod="openstack/keystone-db-sync-gsq5w" Oct 08 18:33:16 crc kubenswrapper[4859]: I1008 18:33:16.180535 4859 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-db-sync-gsq5w" Oct 08 18:33:16 crc kubenswrapper[4859]: I1008 18:33:16.253131 4859 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-db-create-wczl5"] Oct 08 18:33:16 crc kubenswrapper[4859]: I1008 18:33:16.275065 4859 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-db-create-dbftt" event={"ID":"b51059c4-e80f-4c66-be2d-c4ec6b9deaf9","Type":"ContainerStarted","Data":"c832b74f2d4a625f9f3f0f938efb5f19947364a7b5b35172450b3c88bc76e9f2"} Oct 08 18:33:16 crc kubenswrapper[4859]: I1008 18:33:16.352091 4859 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-db-create-p8mb2"] Oct 08 18:33:16 crc kubenswrapper[4859]: I1008 18:33:16.694926 4859 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-db-sync-gsq5w"] Oct 08 18:33:16 crc kubenswrapper[4859]: W1008 18:33:16.834669 4859 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod498d1fd9_4fd1_4dce_a122_6f79d4274882.slice/crio-56183b00ad1b371d8528e2f952cf250c9f5f256f000435c1a78b61ee2fad014d WatchSource:0}: Error finding container 56183b00ad1b371d8528e2f952cf250c9f5f256f000435c1a78b61ee2fad014d: Status 404 returned error can't find the container with id 56183b00ad1b371d8528e2f952cf250c9f5f256f000435c1a78b61ee2fad014d Oct 08 18:33:17 crc kubenswrapper[4859]: I1008 18:33:17.283584 4859 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-db-sync-gsq5w" event={"ID":"498d1fd9-4fd1-4dce-a122-6f79d4274882","Type":"ContainerStarted","Data":"56183b00ad1b371d8528e2f952cf250c9f5f256f000435c1a78b61ee2fad014d"} Oct 08 18:33:17 crc kubenswrapper[4859]: I1008 18:33:17.284896 4859 generic.go:334] "Generic (PLEG): container finished" podID="b51059c4-e80f-4c66-be2d-c4ec6b9deaf9" containerID="bbdcd7b0e44c3f12f0238a79fe0cc08a70a78cfa68fbe75279c10fe65fab927e" exitCode=0 Oct 08 18:33:17 crc kubenswrapper[4859]: I1008 18:33:17.284960 4859 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-db-create-dbftt" event={"ID":"b51059c4-e80f-4c66-be2d-c4ec6b9deaf9","Type":"ContainerDied","Data":"bbdcd7b0e44c3f12f0238a79fe0cc08a70a78cfa68fbe75279c10fe65fab927e"} Oct 08 18:33:17 crc kubenswrapper[4859]: I1008 18:33:17.287200 4859 generic.go:334] "Generic (PLEG): container finished" podID="84999f2e-5fbf-46d7-8334-436d1f9b12d7" containerID="cea203469b8abcb099d7c42979e0bdaf41e94673bba18642f9ca7349d2deaf3e" exitCode=0 Oct 08 18:33:17 crc kubenswrapper[4859]: I1008 18:33:17.287292 4859 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-db-create-wczl5" event={"ID":"84999f2e-5fbf-46d7-8334-436d1f9b12d7","Type":"ContainerDied","Data":"cea203469b8abcb099d7c42979e0bdaf41e94673bba18642f9ca7349d2deaf3e"} Oct 08 18:33:17 crc kubenswrapper[4859]: I1008 18:33:17.287338 4859 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-db-create-wczl5" event={"ID":"84999f2e-5fbf-46d7-8334-436d1f9b12d7","Type":"ContainerStarted","Data":"6da942b565fb4748f4beb762b5cc8bc4195db8b54f8a72296570d42b069b8de2"} Oct 08 18:33:17 crc kubenswrapper[4859]: I1008 18:33:17.290123 4859 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"c860866f-d309-4824-a31c-e88c5d820712","Type":"ContainerStarted","Data":"dc8c2228e843a323901d3de7b37419c4a0abe64533ed62cfe853cae3832ad385"} Oct 08 18:33:17 crc kubenswrapper[4859]: I1008 18:33:17.290172 4859 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"c860866f-d309-4824-a31c-e88c5d820712","Type":"ContainerStarted","Data":"d42ab1a92a4fa33f36cc6995fba4bfe4af56b03b566a15f019d5d56aba916e04"} Oct 08 18:33:17 crc kubenswrapper[4859]: I1008 18:33:17.291998 4859 generic.go:334] "Generic (PLEG): container finished" podID="c0da13c2-e693-4da8-83fa-ae3032d30ec5" containerID="2b8b1c1c7faf850e99249b3d02dba4845cbd57dde84a18c9f93e8f54883c8d9e" exitCode=0 Oct 08 18:33:17 crc kubenswrapper[4859]: I1008 18:33:17.292030 4859 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-db-create-p8mb2" event={"ID":"c0da13c2-e693-4da8-83fa-ae3032d30ec5","Type":"ContainerDied","Data":"2b8b1c1c7faf850e99249b3d02dba4845cbd57dde84a18c9f93e8f54883c8d9e"} Oct 08 18:33:17 crc kubenswrapper[4859]: I1008 18:33:17.292062 4859 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-db-create-p8mb2" event={"ID":"c0da13c2-e693-4da8-83fa-ae3032d30ec5","Type":"ContainerStarted","Data":"4c51fd64a89830b0dcdda4d92886998c04519e109732bd20186725d9cab36cd7"} Oct 08 18:33:18 crc kubenswrapper[4859]: I1008 18:33:18.304105 4859 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"c860866f-d309-4824-a31c-e88c5d820712","Type":"ContainerStarted","Data":"5bb03e02124a01370d8a8a64313128cc6059f34171fca44bfcb077b87bba0acb"} Oct 08 18:33:18 crc kubenswrapper[4859]: I1008 18:33:18.304445 4859 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"c860866f-d309-4824-a31c-e88c5d820712","Type":"ContainerStarted","Data":"23dbd1973e4d2048ad002840b2877df6579083489c88da0302147a700b779dc6"} Oct 08 18:33:18 crc kubenswrapper[4859]: I1008 18:33:18.673188 4859 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-db-create-wczl5" Oct 08 18:33:18 crc kubenswrapper[4859]: I1008 18:33:18.828928 4859 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-db-create-dbftt" Oct 08 18:33:18 crc kubenswrapper[4859]: I1008 18:33:18.834642 4859 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-db-create-p8mb2" Oct 08 18:33:18 crc kubenswrapper[4859]: I1008 18:33:18.858144 4859 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-lfr57\" (UniqueName: \"kubernetes.io/projected/84999f2e-5fbf-46d7-8334-436d1f9b12d7-kube-api-access-lfr57\") pod \"84999f2e-5fbf-46d7-8334-436d1f9b12d7\" (UID: \"84999f2e-5fbf-46d7-8334-436d1f9b12d7\") " Oct 08 18:33:18 crc kubenswrapper[4859]: I1008 18:33:18.894540 4859 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/84999f2e-5fbf-46d7-8334-436d1f9b12d7-kube-api-access-lfr57" (OuterVolumeSpecName: "kube-api-access-lfr57") pod "84999f2e-5fbf-46d7-8334-436d1f9b12d7" (UID: "84999f2e-5fbf-46d7-8334-436d1f9b12d7"). InnerVolumeSpecName "kube-api-access-lfr57". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 08 18:33:18 crc kubenswrapper[4859]: I1008 18:33:18.959381 4859 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-lkq7m\" (UniqueName: \"kubernetes.io/projected/c0da13c2-e693-4da8-83fa-ae3032d30ec5-kube-api-access-lkq7m\") pod \"c0da13c2-e693-4da8-83fa-ae3032d30ec5\" (UID: \"c0da13c2-e693-4da8-83fa-ae3032d30ec5\") " Oct 08 18:33:18 crc kubenswrapper[4859]: I1008 18:33:18.959573 4859 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-m2fm4\" (UniqueName: \"kubernetes.io/projected/b51059c4-e80f-4c66-be2d-c4ec6b9deaf9-kube-api-access-m2fm4\") pod \"b51059c4-e80f-4c66-be2d-c4ec6b9deaf9\" (UID: \"b51059c4-e80f-4c66-be2d-c4ec6b9deaf9\") " Oct 08 18:33:18 crc kubenswrapper[4859]: I1008 18:33:18.959969 4859 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-lfr57\" (UniqueName: \"kubernetes.io/projected/84999f2e-5fbf-46d7-8334-436d1f9b12d7-kube-api-access-lfr57\") on node \"crc\" DevicePath \"\"" Oct 08 18:33:18 crc kubenswrapper[4859]: I1008 18:33:18.963536 4859 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/c0da13c2-e693-4da8-83fa-ae3032d30ec5-kube-api-access-lkq7m" (OuterVolumeSpecName: "kube-api-access-lkq7m") pod "c0da13c2-e693-4da8-83fa-ae3032d30ec5" (UID: "c0da13c2-e693-4da8-83fa-ae3032d30ec5"). InnerVolumeSpecName "kube-api-access-lkq7m". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 08 18:33:18 crc kubenswrapper[4859]: I1008 18:33:18.963927 4859 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b51059c4-e80f-4c66-be2d-c4ec6b9deaf9-kube-api-access-m2fm4" (OuterVolumeSpecName: "kube-api-access-m2fm4") pod "b51059c4-e80f-4c66-be2d-c4ec6b9deaf9" (UID: "b51059c4-e80f-4c66-be2d-c4ec6b9deaf9"). InnerVolumeSpecName "kube-api-access-m2fm4". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 08 18:33:19 crc kubenswrapper[4859]: I1008 18:33:19.062673 4859 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-m2fm4\" (UniqueName: \"kubernetes.io/projected/b51059c4-e80f-4c66-be2d-c4ec6b9deaf9-kube-api-access-m2fm4\") on node \"crc\" DevicePath \"\"" Oct 08 18:33:19 crc kubenswrapper[4859]: I1008 18:33:19.062727 4859 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-lkq7m\" (UniqueName: \"kubernetes.io/projected/c0da13c2-e693-4da8-83fa-ae3032d30ec5-kube-api-access-lkq7m\") on node \"crc\" DevicePath \"\"" Oct 08 18:33:19 crc kubenswrapper[4859]: I1008 18:33:19.315845 4859 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-db-create-p8mb2" event={"ID":"c0da13c2-e693-4da8-83fa-ae3032d30ec5","Type":"ContainerDied","Data":"4c51fd64a89830b0dcdda4d92886998c04519e109732bd20186725d9cab36cd7"} Oct 08 18:33:19 crc kubenswrapper[4859]: I1008 18:33:19.316121 4859 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="4c51fd64a89830b0dcdda4d92886998c04519e109732bd20186725d9cab36cd7" Oct 08 18:33:19 crc kubenswrapper[4859]: I1008 18:33:19.315852 4859 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-db-create-p8mb2" Oct 08 18:33:19 crc kubenswrapper[4859]: I1008 18:33:19.317411 4859 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-db-create-dbftt" event={"ID":"b51059c4-e80f-4c66-be2d-c4ec6b9deaf9","Type":"ContainerDied","Data":"c832b74f2d4a625f9f3f0f938efb5f19947364a7b5b35172450b3c88bc76e9f2"} Oct 08 18:33:19 crc kubenswrapper[4859]: I1008 18:33:19.317455 4859 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="c832b74f2d4a625f9f3f0f938efb5f19947364a7b5b35172450b3c88bc76e9f2" Oct 08 18:33:19 crc kubenswrapper[4859]: I1008 18:33:19.317428 4859 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-db-create-dbftt" Oct 08 18:33:19 crc kubenswrapper[4859]: I1008 18:33:19.319221 4859 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-db-create-wczl5" event={"ID":"84999f2e-5fbf-46d7-8334-436d1f9b12d7","Type":"ContainerDied","Data":"6da942b565fb4748f4beb762b5cc8bc4195db8b54f8a72296570d42b069b8de2"} Oct 08 18:33:19 crc kubenswrapper[4859]: I1008 18:33:19.319265 4859 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="6da942b565fb4748f4beb762b5cc8bc4195db8b54f8a72296570d42b069b8de2" Oct 08 18:33:19 crc kubenswrapper[4859]: I1008 18:33:19.319331 4859 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-db-create-wczl5" Oct 08 18:33:20 crc kubenswrapper[4859]: I1008 18:33:20.327824 4859 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"c860866f-d309-4824-a31c-e88c5d820712","Type":"ContainerStarted","Data":"7a078cf2d6c90edb59614c7398947c5d326635c42cc20b2f6e998f6682ad0dad"} Oct 08 18:33:20 crc kubenswrapper[4859]: I1008 18:33:20.328135 4859 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"c860866f-d309-4824-a31c-e88c5d820712","Type":"ContainerStarted","Data":"3aa82c5b93ec166984aa4772b33659cdfc0808560e9ef48aa4f086d03ab37db0"} Oct 08 18:33:20 crc kubenswrapper[4859]: I1008 18:33:20.328148 4859 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"c860866f-d309-4824-a31c-e88c5d820712","Type":"ContainerStarted","Data":"1fe50b0f0d568b64880ab9fe9cceb650057b17a3af10b1c2ebcbb9b3f2db8426"} Oct 08 18:33:20 crc kubenswrapper[4859]: I1008 18:33:20.328155 4859 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"c860866f-d309-4824-a31c-e88c5d820712","Type":"ContainerStarted","Data":"5bc3366ff398704c00002136223a31f73c7ec451c4c60e83a7584c20a608d50c"} Oct 08 18:33:22 crc kubenswrapper[4859]: I1008 18:33:22.351765 4859 generic.go:334] "Generic (PLEG): container finished" podID="9c2ff226-ee9a-4cb3-b66f-9df0c3497748" containerID="ba60ab3185c0a8ffefed44fd91a99acbf6a00486ec248ab2c7753e30b2f4a4a9" exitCode=0 Oct 08 18:33:22 crc kubenswrapper[4859]: I1008 18:33:22.351917 4859 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-db-sync-xk8qg" event={"ID":"9c2ff226-ee9a-4cb3-b66f-9df0c3497748","Type":"ContainerDied","Data":"ba60ab3185c0a8ffefed44fd91a99acbf6a00486ec248ab2c7753e30b2f4a4a9"} Oct 08 18:33:23 crc kubenswrapper[4859]: I1008 18:33:23.363553 4859 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-db-sync-gsq5w" event={"ID":"498d1fd9-4fd1-4dce-a122-6f79d4274882","Type":"ContainerStarted","Data":"0188f80e29f014fdf763c155e5438e9a2bebbf2b89f5e7918a56107b4fe05af2"} Oct 08 18:33:23 crc kubenswrapper[4859]: I1008 18:33:23.392324 4859 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/keystone-db-sync-gsq5w" podStartSLOduration=2.589254948 podStartE2EDuration="8.392301182s" podCreationTimestamp="2025-10-08 18:33:15 +0000 UTC" firstStartedPulling="2025-10-08 18:33:16.836862761 +0000 UTC m=+967.083702140" lastFinishedPulling="2025-10-08 18:33:22.639908995 +0000 UTC m=+972.886748374" observedRunningTime="2025-10-08 18:33:23.384846012 +0000 UTC m=+973.631685421" watchObservedRunningTime="2025-10-08 18:33:23.392301182 +0000 UTC m=+973.639140571" Oct 08 18:33:23 crc kubenswrapper[4859]: I1008 18:33:23.807450 4859 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-db-sync-xk8qg" Oct 08 18:33:23 crc kubenswrapper[4859]: I1008 18:33:23.943967 4859 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-lczsb\" (UniqueName: \"kubernetes.io/projected/9c2ff226-ee9a-4cb3-b66f-9df0c3497748-kube-api-access-lczsb\") pod \"9c2ff226-ee9a-4cb3-b66f-9df0c3497748\" (UID: \"9c2ff226-ee9a-4cb3-b66f-9df0c3497748\") " Oct 08 18:33:23 crc kubenswrapper[4859]: I1008 18:33:23.944068 4859 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9c2ff226-ee9a-4cb3-b66f-9df0c3497748-combined-ca-bundle\") pod \"9c2ff226-ee9a-4cb3-b66f-9df0c3497748\" (UID: \"9c2ff226-ee9a-4cb3-b66f-9df0c3497748\") " Oct 08 18:33:23 crc kubenswrapper[4859]: I1008 18:33:23.944105 4859 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/9c2ff226-ee9a-4cb3-b66f-9df0c3497748-db-sync-config-data\") pod \"9c2ff226-ee9a-4cb3-b66f-9df0c3497748\" (UID: \"9c2ff226-ee9a-4cb3-b66f-9df0c3497748\") " Oct 08 18:33:23 crc kubenswrapper[4859]: I1008 18:33:23.944191 4859 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/9c2ff226-ee9a-4cb3-b66f-9df0c3497748-config-data\") pod \"9c2ff226-ee9a-4cb3-b66f-9df0c3497748\" (UID: \"9c2ff226-ee9a-4cb3-b66f-9df0c3497748\") " Oct 08 18:33:23 crc kubenswrapper[4859]: I1008 18:33:23.958020 4859 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/9c2ff226-ee9a-4cb3-b66f-9df0c3497748-db-sync-config-data" (OuterVolumeSpecName: "db-sync-config-data") pod "9c2ff226-ee9a-4cb3-b66f-9df0c3497748" (UID: "9c2ff226-ee9a-4cb3-b66f-9df0c3497748"). InnerVolumeSpecName "db-sync-config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 08 18:33:23 crc kubenswrapper[4859]: I1008 18:33:23.958144 4859 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/9c2ff226-ee9a-4cb3-b66f-9df0c3497748-kube-api-access-lczsb" (OuterVolumeSpecName: "kube-api-access-lczsb") pod "9c2ff226-ee9a-4cb3-b66f-9df0c3497748" (UID: "9c2ff226-ee9a-4cb3-b66f-9df0c3497748"). InnerVolumeSpecName "kube-api-access-lczsb". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 08 18:33:23 crc kubenswrapper[4859]: I1008 18:33:23.987432 4859 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/9c2ff226-ee9a-4cb3-b66f-9df0c3497748-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "9c2ff226-ee9a-4cb3-b66f-9df0c3497748" (UID: "9c2ff226-ee9a-4cb3-b66f-9df0c3497748"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 08 18:33:24 crc kubenswrapper[4859]: I1008 18:33:24.000886 4859 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/9c2ff226-ee9a-4cb3-b66f-9df0c3497748-config-data" (OuterVolumeSpecName: "config-data") pod "9c2ff226-ee9a-4cb3-b66f-9df0c3497748" (UID: "9c2ff226-ee9a-4cb3-b66f-9df0c3497748"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 08 18:33:24 crc kubenswrapper[4859]: I1008 18:33:24.046615 4859 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-lczsb\" (UniqueName: \"kubernetes.io/projected/9c2ff226-ee9a-4cb3-b66f-9df0c3497748-kube-api-access-lczsb\") on node \"crc\" DevicePath \"\"" Oct 08 18:33:24 crc kubenswrapper[4859]: I1008 18:33:24.046650 4859 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9c2ff226-ee9a-4cb3-b66f-9df0c3497748-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 08 18:33:24 crc kubenswrapper[4859]: I1008 18:33:24.046661 4859 reconciler_common.go:293] "Volume detached for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/9c2ff226-ee9a-4cb3-b66f-9df0c3497748-db-sync-config-data\") on node \"crc\" DevicePath \"\"" Oct 08 18:33:24 crc kubenswrapper[4859]: I1008 18:33:24.046673 4859 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/9c2ff226-ee9a-4cb3-b66f-9df0c3497748-config-data\") on node \"crc\" DevicePath \"\"" Oct 08 18:33:24 crc kubenswrapper[4859]: I1008 18:33:24.381180 4859 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-db-sync-xk8qg" Oct 08 18:33:24 crc kubenswrapper[4859]: I1008 18:33:24.382182 4859 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-db-sync-xk8qg" event={"ID":"9c2ff226-ee9a-4cb3-b66f-9df0c3497748","Type":"ContainerDied","Data":"024d3bf433e6715d5b3a98f9acdd357677562ec6274a20d769a481bdb033f266"} Oct 08 18:33:24 crc kubenswrapper[4859]: I1008 18:33:24.382221 4859 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="024d3bf433e6715d5b3a98f9acdd357677562ec6274a20d769a481bdb033f266" Oct 08 18:33:24 crc kubenswrapper[4859]: I1008 18:33:24.404748 4859 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"c860866f-d309-4824-a31c-e88c5d820712","Type":"ContainerStarted","Data":"3de9fc8cfa83faa3f614486954f249ecb46ed7bf24b2bfa05034cf950fecb718"} Oct 08 18:33:24 crc kubenswrapper[4859]: I1008 18:33:24.404791 4859 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"c860866f-d309-4824-a31c-e88c5d820712","Type":"ContainerStarted","Data":"099ee665fd5a58b587c3544b7a3d115ea81bae812a9f9bb6a18b10bd16ed4a4d"} Oct 08 18:33:24 crc kubenswrapper[4859]: I1008 18:33:24.404808 4859 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"c860866f-d309-4824-a31c-e88c5d820712","Type":"ContainerStarted","Data":"b4c0daba33632ad91db08415560954d2caa7c0a8cdc4987990899a7fe36da337"} Oct 08 18:33:24 crc kubenswrapper[4859]: I1008 18:33:24.404817 4859 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"c860866f-d309-4824-a31c-e88c5d820712","Type":"ContainerStarted","Data":"b5b5f340bdb31d9b0f9d5d353000f2b35a201f6c05523eaaa7aca09fe2567da2"} Oct 08 18:33:24 crc kubenswrapper[4859]: I1008 18:33:24.404826 4859 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"c860866f-d309-4824-a31c-e88c5d820712","Type":"ContainerStarted","Data":"e144109a2817fe661c2515f0bf737b81fe8744c4a86cb528934cdf024ae05510"} Oct 08 18:33:24 crc kubenswrapper[4859]: I1008 18:33:24.803337 4859 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-7c456cb699-8ngvp"] Oct 08 18:33:24 crc kubenswrapper[4859]: E1008 18:33:24.804244 4859 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="84999f2e-5fbf-46d7-8334-436d1f9b12d7" containerName="mariadb-database-create" Oct 08 18:33:24 crc kubenswrapper[4859]: I1008 18:33:24.804261 4859 state_mem.go:107] "Deleted CPUSet assignment" podUID="84999f2e-5fbf-46d7-8334-436d1f9b12d7" containerName="mariadb-database-create" Oct 08 18:33:24 crc kubenswrapper[4859]: E1008 18:33:24.804272 4859 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c0da13c2-e693-4da8-83fa-ae3032d30ec5" containerName="mariadb-database-create" Oct 08 18:33:24 crc kubenswrapper[4859]: I1008 18:33:24.804278 4859 state_mem.go:107] "Deleted CPUSet assignment" podUID="c0da13c2-e693-4da8-83fa-ae3032d30ec5" containerName="mariadb-database-create" Oct 08 18:33:24 crc kubenswrapper[4859]: E1008 18:33:24.804296 4859 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9c2ff226-ee9a-4cb3-b66f-9df0c3497748" containerName="glance-db-sync" Oct 08 18:33:24 crc kubenswrapper[4859]: I1008 18:33:24.804303 4859 state_mem.go:107] "Deleted CPUSet assignment" podUID="9c2ff226-ee9a-4cb3-b66f-9df0c3497748" containerName="glance-db-sync" Oct 08 18:33:24 crc kubenswrapper[4859]: E1008 18:33:24.804321 4859 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b51059c4-e80f-4c66-be2d-c4ec6b9deaf9" containerName="mariadb-database-create" Oct 08 18:33:24 crc kubenswrapper[4859]: I1008 18:33:24.804327 4859 state_mem.go:107] "Deleted CPUSet assignment" podUID="b51059c4-e80f-4c66-be2d-c4ec6b9deaf9" containerName="mariadb-database-create" Oct 08 18:33:24 crc kubenswrapper[4859]: I1008 18:33:24.804472 4859 memory_manager.go:354] "RemoveStaleState removing state" podUID="84999f2e-5fbf-46d7-8334-436d1f9b12d7" containerName="mariadb-database-create" Oct 08 18:33:24 crc kubenswrapper[4859]: I1008 18:33:24.804486 4859 memory_manager.go:354] "RemoveStaleState removing state" podUID="c0da13c2-e693-4da8-83fa-ae3032d30ec5" containerName="mariadb-database-create" Oct 08 18:33:24 crc kubenswrapper[4859]: I1008 18:33:24.804506 4859 memory_manager.go:354] "RemoveStaleState removing state" podUID="b51059c4-e80f-4c66-be2d-c4ec6b9deaf9" containerName="mariadb-database-create" Oct 08 18:33:24 crc kubenswrapper[4859]: I1008 18:33:24.804524 4859 memory_manager.go:354] "RemoveStaleState removing state" podUID="9c2ff226-ee9a-4cb3-b66f-9df0c3497748" containerName="glance-db-sync" Oct 08 18:33:24 crc kubenswrapper[4859]: I1008 18:33:24.805359 4859 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-7c456cb699-8ngvp" Oct 08 18:33:24 crc kubenswrapper[4859]: I1008 18:33:24.829667 4859 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-7c456cb699-8ngvp"] Oct 08 18:33:24 crc kubenswrapper[4859]: I1008 18:33:24.971774 4859 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-gk46v\" (UniqueName: \"kubernetes.io/projected/ac58cb73-6174-4cdd-ab31-66641a15bfc9-kube-api-access-gk46v\") pod \"dnsmasq-dns-7c456cb699-8ngvp\" (UID: \"ac58cb73-6174-4cdd-ab31-66641a15bfc9\") " pod="openstack/dnsmasq-dns-7c456cb699-8ngvp" Oct 08 18:33:24 crc kubenswrapper[4859]: I1008 18:33:24.971868 4859 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/ac58cb73-6174-4cdd-ab31-66641a15bfc9-config\") pod \"dnsmasq-dns-7c456cb699-8ngvp\" (UID: \"ac58cb73-6174-4cdd-ab31-66641a15bfc9\") " pod="openstack/dnsmasq-dns-7c456cb699-8ngvp" Oct 08 18:33:24 crc kubenswrapper[4859]: I1008 18:33:24.971901 4859 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/ac58cb73-6174-4cdd-ab31-66641a15bfc9-ovsdbserver-nb\") pod \"dnsmasq-dns-7c456cb699-8ngvp\" (UID: \"ac58cb73-6174-4cdd-ab31-66641a15bfc9\") " pod="openstack/dnsmasq-dns-7c456cb699-8ngvp" Oct 08 18:33:24 crc kubenswrapper[4859]: I1008 18:33:24.972076 4859 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/ac58cb73-6174-4cdd-ab31-66641a15bfc9-dns-svc\") pod \"dnsmasq-dns-7c456cb699-8ngvp\" (UID: \"ac58cb73-6174-4cdd-ab31-66641a15bfc9\") " pod="openstack/dnsmasq-dns-7c456cb699-8ngvp" Oct 08 18:33:24 crc kubenswrapper[4859]: I1008 18:33:24.972157 4859 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/ac58cb73-6174-4cdd-ab31-66641a15bfc9-ovsdbserver-sb\") pod \"dnsmasq-dns-7c456cb699-8ngvp\" (UID: \"ac58cb73-6174-4cdd-ab31-66641a15bfc9\") " pod="openstack/dnsmasq-dns-7c456cb699-8ngvp" Oct 08 18:33:25 crc kubenswrapper[4859]: I1008 18:33:25.073875 4859 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-gk46v\" (UniqueName: \"kubernetes.io/projected/ac58cb73-6174-4cdd-ab31-66641a15bfc9-kube-api-access-gk46v\") pod \"dnsmasq-dns-7c456cb699-8ngvp\" (UID: \"ac58cb73-6174-4cdd-ab31-66641a15bfc9\") " pod="openstack/dnsmasq-dns-7c456cb699-8ngvp" Oct 08 18:33:25 crc kubenswrapper[4859]: I1008 18:33:25.073978 4859 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/ac58cb73-6174-4cdd-ab31-66641a15bfc9-config\") pod \"dnsmasq-dns-7c456cb699-8ngvp\" (UID: \"ac58cb73-6174-4cdd-ab31-66641a15bfc9\") " pod="openstack/dnsmasq-dns-7c456cb699-8ngvp" Oct 08 18:33:25 crc kubenswrapper[4859]: I1008 18:33:25.074009 4859 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/ac58cb73-6174-4cdd-ab31-66641a15bfc9-ovsdbserver-nb\") pod \"dnsmasq-dns-7c456cb699-8ngvp\" (UID: \"ac58cb73-6174-4cdd-ab31-66641a15bfc9\") " pod="openstack/dnsmasq-dns-7c456cb699-8ngvp" Oct 08 18:33:25 crc kubenswrapper[4859]: I1008 18:33:25.074028 4859 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/ac58cb73-6174-4cdd-ab31-66641a15bfc9-dns-svc\") pod \"dnsmasq-dns-7c456cb699-8ngvp\" (UID: \"ac58cb73-6174-4cdd-ab31-66641a15bfc9\") " pod="openstack/dnsmasq-dns-7c456cb699-8ngvp" Oct 08 18:33:25 crc kubenswrapper[4859]: I1008 18:33:25.074043 4859 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/ac58cb73-6174-4cdd-ab31-66641a15bfc9-ovsdbserver-sb\") pod \"dnsmasq-dns-7c456cb699-8ngvp\" (UID: \"ac58cb73-6174-4cdd-ab31-66641a15bfc9\") " pod="openstack/dnsmasq-dns-7c456cb699-8ngvp" Oct 08 18:33:25 crc kubenswrapper[4859]: I1008 18:33:25.075264 4859 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/ac58cb73-6174-4cdd-ab31-66641a15bfc9-ovsdbserver-sb\") pod \"dnsmasq-dns-7c456cb699-8ngvp\" (UID: \"ac58cb73-6174-4cdd-ab31-66641a15bfc9\") " pod="openstack/dnsmasq-dns-7c456cb699-8ngvp" Oct 08 18:33:25 crc kubenswrapper[4859]: I1008 18:33:25.075281 4859 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/ac58cb73-6174-4cdd-ab31-66641a15bfc9-ovsdbserver-nb\") pod \"dnsmasq-dns-7c456cb699-8ngvp\" (UID: \"ac58cb73-6174-4cdd-ab31-66641a15bfc9\") " pod="openstack/dnsmasq-dns-7c456cb699-8ngvp" Oct 08 18:33:25 crc kubenswrapper[4859]: I1008 18:33:25.075438 4859 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/ac58cb73-6174-4cdd-ab31-66641a15bfc9-dns-svc\") pod \"dnsmasq-dns-7c456cb699-8ngvp\" (UID: \"ac58cb73-6174-4cdd-ab31-66641a15bfc9\") " pod="openstack/dnsmasq-dns-7c456cb699-8ngvp" Oct 08 18:33:25 crc kubenswrapper[4859]: I1008 18:33:25.076055 4859 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/ac58cb73-6174-4cdd-ab31-66641a15bfc9-config\") pod \"dnsmasq-dns-7c456cb699-8ngvp\" (UID: \"ac58cb73-6174-4cdd-ab31-66641a15bfc9\") " pod="openstack/dnsmasq-dns-7c456cb699-8ngvp" Oct 08 18:33:25 crc kubenswrapper[4859]: I1008 18:33:25.093708 4859 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-gk46v\" (UniqueName: \"kubernetes.io/projected/ac58cb73-6174-4cdd-ab31-66641a15bfc9-kube-api-access-gk46v\") pod \"dnsmasq-dns-7c456cb699-8ngvp\" (UID: \"ac58cb73-6174-4cdd-ab31-66641a15bfc9\") " pod="openstack/dnsmasq-dns-7c456cb699-8ngvp" Oct 08 18:33:25 crc kubenswrapper[4859]: I1008 18:33:25.126752 4859 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-7c456cb699-8ngvp" Oct 08 18:33:25 crc kubenswrapper[4859]: I1008 18:33:25.410093 4859 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/cinder-158c-account-create-vvkvz"] Oct 08 18:33:25 crc kubenswrapper[4859]: I1008 18:33:25.411807 4859 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-158c-account-create-vvkvz" Oct 08 18:33:25 crc kubenswrapper[4859]: I1008 18:33:25.413963 4859 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-db-secret" Oct 08 18:33:25 crc kubenswrapper[4859]: I1008 18:33:25.417276 4859 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-158c-account-create-vvkvz"] Oct 08 18:33:25 crc kubenswrapper[4859]: I1008 18:33:25.421291 4859 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"c860866f-d309-4824-a31c-e88c5d820712","Type":"ContainerStarted","Data":"793c7e2a1506af023bd2e961dd8f40f4a1bb3251fafc08f4e1e2465f78de53d7"} Oct 08 18:33:25 crc kubenswrapper[4859]: I1008 18:33:25.421331 4859 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"c860866f-d309-4824-a31c-e88c5d820712","Type":"ContainerStarted","Data":"3cbd02af6e8473138404c19a2b889551ae1a22ee9b742396c8166c3ac50ed154"} Oct 08 18:33:25 crc kubenswrapper[4859]: I1008 18:33:25.461110 4859 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/swift-storage-0" podStartSLOduration=36.560244491 podStartE2EDuration="45.461094069s" podCreationTimestamp="2025-10-08 18:32:40 +0000 UTC" firstStartedPulling="2025-10-08 18:33:14.34075106 +0000 UTC m=+964.587590439" lastFinishedPulling="2025-10-08 18:33:23.241600638 +0000 UTC m=+973.488440017" observedRunningTime="2025-10-08 18:33:25.46078303 +0000 UTC m=+975.707622419" watchObservedRunningTime="2025-10-08 18:33:25.461094069 +0000 UTC m=+975.707933448" Oct 08 18:33:25 crc kubenswrapper[4859]: I1008 18:33:25.584963 4859 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-6xf29\" (UniqueName: \"kubernetes.io/projected/ff748840-6d03-4e9a-8284-a4ca9d753278-kube-api-access-6xf29\") pod \"cinder-158c-account-create-vvkvz\" (UID: \"ff748840-6d03-4e9a-8284-a4ca9d753278\") " pod="openstack/cinder-158c-account-create-vvkvz" Oct 08 18:33:25 crc kubenswrapper[4859]: I1008 18:33:25.621572 4859 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/barbican-c89b-account-create-bd78m"] Oct 08 18:33:25 crc kubenswrapper[4859]: I1008 18:33:25.623231 4859 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-c89b-account-create-bd78m" Oct 08 18:33:25 crc kubenswrapper[4859]: I1008 18:33:25.625297 4859 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"barbican-db-secret" Oct 08 18:33:25 crc kubenswrapper[4859]: I1008 18:33:25.639556 4859 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-c89b-account-create-bd78m"] Oct 08 18:33:25 crc kubenswrapper[4859]: I1008 18:33:25.651379 4859 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-7c456cb699-8ngvp"] Oct 08 18:33:25 crc kubenswrapper[4859]: W1008 18:33:25.659794 4859 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podac58cb73_6174_4cdd_ab31_66641a15bfc9.slice/crio-2250c06a70511880ba7bdde51622ac969c8a91744fb5c23caa6c9aec4d43a771 WatchSource:0}: Error finding container 2250c06a70511880ba7bdde51622ac969c8a91744fb5c23caa6c9aec4d43a771: Status 404 returned error can't find the container with id 2250c06a70511880ba7bdde51622ac969c8a91744fb5c23caa6c9aec4d43a771 Oct 08 18:33:25 crc kubenswrapper[4859]: I1008 18:33:25.686697 4859 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-6xf29\" (UniqueName: \"kubernetes.io/projected/ff748840-6d03-4e9a-8284-a4ca9d753278-kube-api-access-6xf29\") pod \"cinder-158c-account-create-vvkvz\" (UID: \"ff748840-6d03-4e9a-8284-a4ca9d753278\") " pod="openstack/cinder-158c-account-create-vvkvz" Oct 08 18:33:25 crc kubenswrapper[4859]: I1008 18:33:25.712780 4859 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-6xf29\" (UniqueName: \"kubernetes.io/projected/ff748840-6d03-4e9a-8284-a4ca9d753278-kube-api-access-6xf29\") pod \"cinder-158c-account-create-vvkvz\" (UID: \"ff748840-6d03-4e9a-8284-a4ca9d753278\") " pod="openstack/cinder-158c-account-create-vvkvz" Oct 08 18:33:25 crc kubenswrapper[4859]: I1008 18:33:25.735224 4859 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-158c-account-create-vvkvz" Oct 08 18:33:25 crc kubenswrapper[4859]: I1008 18:33:25.790223 4859 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-9h2f5\" (UniqueName: \"kubernetes.io/projected/4a64adc4-3689-4d56-b514-f86372be9938-kube-api-access-9h2f5\") pod \"barbican-c89b-account-create-bd78m\" (UID: \"4a64adc4-3689-4d56-b514-f86372be9938\") " pod="openstack/barbican-c89b-account-create-bd78m" Oct 08 18:33:25 crc kubenswrapper[4859]: I1008 18:33:25.815396 4859 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/neutron-d728-account-create-prt4m"] Oct 08 18:33:25 crc kubenswrapper[4859]: I1008 18:33:25.817752 4859 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-d728-account-create-prt4m" Oct 08 18:33:25 crc kubenswrapper[4859]: I1008 18:33:25.822380 4859 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"neutron-db-secret" Oct 08 18:33:25 crc kubenswrapper[4859]: I1008 18:33:25.836812 4859 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-d728-account-create-prt4m"] Oct 08 18:33:25 crc kubenswrapper[4859]: I1008 18:33:25.891676 4859 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-9h2f5\" (UniqueName: \"kubernetes.io/projected/4a64adc4-3689-4d56-b514-f86372be9938-kube-api-access-9h2f5\") pod \"barbican-c89b-account-create-bd78m\" (UID: \"4a64adc4-3689-4d56-b514-f86372be9938\") " pod="openstack/barbican-c89b-account-create-bd78m" Oct 08 18:33:25 crc kubenswrapper[4859]: I1008 18:33:25.891800 4859 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-9xd5w\" (UniqueName: \"kubernetes.io/projected/8503bae7-a295-4935-bcc9-a19971c4f417-kube-api-access-9xd5w\") pod \"neutron-d728-account-create-prt4m\" (UID: \"8503bae7-a295-4935-bcc9-a19971c4f417\") " pod="openstack/neutron-d728-account-create-prt4m" Oct 08 18:33:25 crc kubenswrapper[4859]: I1008 18:33:25.916590 4859 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-9h2f5\" (UniqueName: \"kubernetes.io/projected/4a64adc4-3689-4d56-b514-f86372be9938-kube-api-access-9h2f5\") pod \"barbican-c89b-account-create-bd78m\" (UID: \"4a64adc4-3689-4d56-b514-f86372be9938\") " pod="openstack/barbican-c89b-account-create-bd78m" Oct 08 18:33:25 crc kubenswrapper[4859]: I1008 18:33:25.925254 4859 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-7c456cb699-8ngvp"] Oct 08 18:33:25 crc kubenswrapper[4859]: I1008 18:33:25.952442 4859 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-c89b-account-create-bd78m" Oct 08 18:33:25 crc kubenswrapper[4859]: I1008 18:33:25.961650 4859 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-568bf4d7b5-ljcdc"] Oct 08 18:33:25 crc kubenswrapper[4859]: I1008 18:33:25.963399 4859 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-568bf4d7b5-ljcdc" Oct 08 18:33:25 crc kubenswrapper[4859]: I1008 18:33:25.975512 4859 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-568bf4d7b5-ljcdc"] Oct 08 18:33:25 crc kubenswrapper[4859]: I1008 18:33:25.978752 4859 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"dns-swift-storage-0" Oct 08 18:33:25 crc kubenswrapper[4859]: I1008 18:33:25.997120 4859 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-9xd5w\" (UniqueName: \"kubernetes.io/projected/8503bae7-a295-4935-bcc9-a19971c4f417-kube-api-access-9xd5w\") pod \"neutron-d728-account-create-prt4m\" (UID: \"8503bae7-a295-4935-bcc9-a19971c4f417\") " pod="openstack/neutron-d728-account-create-prt4m" Oct 08 18:33:26 crc kubenswrapper[4859]: I1008 18:33:26.017348 4859 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-9xd5w\" (UniqueName: \"kubernetes.io/projected/8503bae7-a295-4935-bcc9-a19971c4f417-kube-api-access-9xd5w\") pod \"neutron-d728-account-create-prt4m\" (UID: \"8503bae7-a295-4935-bcc9-a19971c4f417\") " pod="openstack/neutron-d728-account-create-prt4m" Oct 08 18:33:26 crc kubenswrapper[4859]: I1008 18:33:26.091497 4859 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-d728-account-create-prt4m" Oct 08 18:33:26 crc kubenswrapper[4859]: I1008 18:33:26.098678 4859 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-s92r8\" (UniqueName: \"kubernetes.io/projected/bc0261a4-dfa9-4356-9a9d-8a46a187a74c-kube-api-access-s92r8\") pod \"dnsmasq-dns-568bf4d7b5-ljcdc\" (UID: \"bc0261a4-dfa9-4356-9a9d-8a46a187a74c\") " pod="openstack/dnsmasq-dns-568bf4d7b5-ljcdc" Oct 08 18:33:26 crc kubenswrapper[4859]: I1008 18:33:26.098787 4859 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/bc0261a4-dfa9-4356-9a9d-8a46a187a74c-dns-svc\") pod \"dnsmasq-dns-568bf4d7b5-ljcdc\" (UID: \"bc0261a4-dfa9-4356-9a9d-8a46a187a74c\") " pod="openstack/dnsmasq-dns-568bf4d7b5-ljcdc" Oct 08 18:33:26 crc kubenswrapper[4859]: I1008 18:33:26.098815 4859 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/bc0261a4-dfa9-4356-9a9d-8a46a187a74c-ovsdbserver-nb\") pod \"dnsmasq-dns-568bf4d7b5-ljcdc\" (UID: \"bc0261a4-dfa9-4356-9a9d-8a46a187a74c\") " pod="openstack/dnsmasq-dns-568bf4d7b5-ljcdc" Oct 08 18:33:26 crc kubenswrapper[4859]: I1008 18:33:26.098840 4859 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/bc0261a4-dfa9-4356-9a9d-8a46a187a74c-dns-swift-storage-0\") pod \"dnsmasq-dns-568bf4d7b5-ljcdc\" (UID: \"bc0261a4-dfa9-4356-9a9d-8a46a187a74c\") " pod="openstack/dnsmasq-dns-568bf4d7b5-ljcdc" Oct 08 18:33:26 crc kubenswrapper[4859]: I1008 18:33:26.098867 4859 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/bc0261a4-dfa9-4356-9a9d-8a46a187a74c-config\") pod \"dnsmasq-dns-568bf4d7b5-ljcdc\" (UID: \"bc0261a4-dfa9-4356-9a9d-8a46a187a74c\") " pod="openstack/dnsmasq-dns-568bf4d7b5-ljcdc" Oct 08 18:33:26 crc kubenswrapper[4859]: I1008 18:33:26.098902 4859 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/bc0261a4-dfa9-4356-9a9d-8a46a187a74c-ovsdbserver-sb\") pod \"dnsmasq-dns-568bf4d7b5-ljcdc\" (UID: \"bc0261a4-dfa9-4356-9a9d-8a46a187a74c\") " pod="openstack/dnsmasq-dns-568bf4d7b5-ljcdc" Oct 08 18:33:26 crc kubenswrapper[4859]: I1008 18:33:26.200863 4859 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/bc0261a4-dfa9-4356-9a9d-8a46a187a74c-dns-svc\") pod \"dnsmasq-dns-568bf4d7b5-ljcdc\" (UID: \"bc0261a4-dfa9-4356-9a9d-8a46a187a74c\") " pod="openstack/dnsmasq-dns-568bf4d7b5-ljcdc" Oct 08 18:33:26 crc kubenswrapper[4859]: I1008 18:33:26.200914 4859 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/bc0261a4-dfa9-4356-9a9d-8a46a187a74c-ovsdbserver-nb\") pod \"dnsmasq-dns-568bf4d7b5-ljcdc\" (UID: \"bc0261a4-dfa9-4356-9a9d-8a46a187a74c\") " pod="openstack/dnsmasq-dns-568bf4d7b5-ljcdc" Oct 08 18:33:26 crc kubenswrapper[4859]: I1008 18:33:26.200938 4859 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/bc0261a4-dfa9-4356-9a9d-8a46a187a74c-dns-swift-storage-0\") pod \"dnsmasq-dns-568bf4d7b5-ljcdc\" (UID: \"bc0261a4-dfa9-4356-9a9d-8a46a187a74c\") " pod="openstack/dnsmasq-dns-568bf4d7b5-ljcdc" Oct 08 18:33:26 crc kubenswrapper[4859]: I1008 18:33:26.200962 4859 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/bc0261a4-dfa9-4356-9a9d-8a46a187a74c-config\") pod \"dnsmasq-dns-568bf4d7b5-ljcdc\" (UID: \"bc0261a4-dfa9-4356-9a9d-8a46a187a74c\") " pod="openstack/dnsmasq-dns-568bf4d7b5-ljcdc" Oct 08 18:33:26 crc kubenswrapper[4859]: I1008 18:33:26.200993 4859 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/bc0261a4-dfa9-4356-9a9d-8a46a187a74c-ovsdbserver-sb\") pod \"dnsmasq-dns-568bf4d7b5-ljcdc\" (UID: \"bc0261a4-dfa9-4356-9a9d-8a46a187a74c\") " pod="openstack/dnsmasq-dns-568bf4d7b5-ljcdc" Oct 08 18:33:26 crc kubenswrapper[4859]: I1008 18:33:26.201056 4859 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s92r8\" (UniqueName: \"kubernetes.io/projected/bc0261a4-dfa9-4356-9a9d-8a46a187a74c-kube-api-access-s92r8\") pod \"dnsmasq-dns-568bf4d7b5-ljcdc\" (UID: \"bc0261a4-dfa9-4356-9a9d-8a46a187a74c\") " pod="openstack/dnsmasq-dns-568bf4d7b5-ljcdc" Oct 08 18:33:26 crc kubenswrapper[4859]: I1008 18:33:26.205495 4859 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/bc0261a4-dfa9-4356-9a9d-8a46a187a74c-dns-svc\") pod \"dnsmasq-dns-568bf4d7b5-ljcdc\" (UID: \"bc0261a4-dfa9-4356-9a9d-8a46a187a74c\") " pod="openstack/dnsmasq-dns-568bf4d7b5-ljcdc" Oct 08 18:33:26 crc kubenswrapper[4859]: I1008 18:33:26.206041 4859 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/bc0261a4-dfa9-4356-9a9d-8a46a187a74c-ovsdbserver-nb\") pod \"dnsmasq-dns-568bf4d7b5-ljcdc\" (UID: \"bc0261a4-dfa9-4356-9a9d-8a46a187a74c\") " pod="openstack/dnsmasq-dns-568bf4d7b5-ljcdc" Oct 08 18:33:26 crc kubenswrapper[4859]: I1008 18:33:26.206527 4859 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/bc0261a4-dfa9-4356-9a9d-8a46a187a74c-config\") pod \"dnsmasq-dns-568bf4d7b5-ljcdc\" (UID: \"bc0261a4-dfa9-4356-9a9d-8a46a187a74c\") " pod="openstack/dnsmasq-dns-568bf4d7b5-ljcdc" Oct 08 18:33:26 crc kubenswrapper[4859]: I1008 18:33:26.206609 4859 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/bc0261a4-dfa9-4356-9a9d-8a46a187a74c-ovsdbserver-sb\") pod \"dnsmasq-dns-568bf4d7b5-ljcdc\" (UID: \"bc0261a4-dfa9-4356-9a9d-8a46a187a74c\") " pod="openstack/dnsmasq-dns-568bf4d7b5-ljcdc" Oct 08 18:33:26 crc kubenswrapper[4859]: I1008 18:33:26.207057 4859 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/bc0261a4-dfa9-4356-9a9d-8a46a187a74c-dns-swift-storage-0\") pod \"dnsmasq-dns-568bf4d7b5-ljcdc\" (UID: \"bc0261a4-dfa9-4356-9a9d-8a46a187a74c\") " pod="openstack/dnsmasq-dns-568bf4d7b5-ljcdc" Oct 08 18:33:26 crc kubenswrapper[4859]: I1008 18:33:26.221494 4859 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-s92r8\" (UniqueName: \"kubernetes.io/projected/bc0261a4-dfa9-4356-9a9d-8a46a187a74c-kube-api-access-s92r8\") pod \"dnsmasq-dns-568bf4d7b5-ljcdc\" (UID: \"bc0261a4-dfa9-4356-9a9d-8a46a187a74c\") " pod="openstack/dnsmasq-dns-568bf4d7b5-ljcdc" Oct 08 18:33:26 crc kubenswrapper[4859]: I1008 18:33:26.330315 4859 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-158c-account-create-vvkvz"] Oct 08 18:33:26 crc kubenswrapper[4859]: W1008 18:33:26.337275 4859 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podff748840_6d03_4e9a_8284_a4ca9d753278.slice/crio-683d6beb079f21e33e46a1beed704f18ced8f288f065c0c57296caa88cdc3131 WatchSource:0}: Error finding container 683d6beb079f21e33e46a1beed704f18ced8f288f065c0c57296caa88cdc3131: Status 404 returned error can't find the container with id 683d6beb079f21e33e46a1beed704f18ced8f288f065c0c57296caa88cdc3131 Oct 08 18:33:26 crc kubenswrapper[4859]: I1008 18:33:26.425211 4859 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-568bf4d7b5-ljcdc" Oct 08 18:33:26 crc kubenswrapper[4859]: I1008 18:33:26.431852 4859 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-158c-account-create-vvkvz" event={"ID":"ff748840-6d03-4e9a-8284-a4ca9d753278","Type":"ContainerStarted","Data":"683d6beb079f21e33e46a1beed704f18ced8f288f065c0c57296caa88cdc3131"} Oct 08 18:33:26 crc kubenswrapper[4859]: I1008 18:33:26.435626 4859 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-7c456cb699-8ngvp" event={"ID":"ac58cb73-6174-4cdd-ab31-66641a15bfc9","Type":"ContainerStarted","Data":"2250c06a70511880ba7bdde51622ac969c8a91744fb5c23caa6c9aec4d43a771"} Oct 08 18:33:26 crc kubenswrapper[4859]: I1008 18:33:26.455767 4859 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-c89b-account-create-bd78m"] Oct 08 18:33:26 crc kubenswrapper[4859]: I1008 18:33:26.591173 4859 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-d728-account-create-prt4m"] Oct 08 18:33:26 crc kubenswrapper[4859]: I1008 18:33:26.970555 4859 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-568bf4d7b5-ljcdc"] Oct 08 18:33:27 crc kubenswrapper[4859]: I1008 18:33:27.443182 4859 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-c89b-account-create-bd78m" event={"ID":"4a64adc4-3689-4d56-b514-f86372be9938","Type":"ContainerStarted","Data":"752de51991f367023c1433d111ae85780d623c32e960fcb5f4b8b3784a11fd80"} Oct 08 18:33:27 crc kubenswrapper[4859]: I1008 18:33:27.444482 4859 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-568bf4d7b5-ljcdc" event={"ID":"bc0261a4-dfa9-4356-9a9d-8a46a187a74c","Type":"ContainerStarted","Data":"3ed08058125117480591800f7330fffb24051967b2ab5875ffcf15602430e2e6"} Oct 08 18:33:27 crc kubenswrapper[4859]: I1008 18:33:27.445659 4859 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-d728-account-create-prt4m" event={"ID":"8503bae7-a295-4935-bcc9-a19971c4f417","Type":"ContainerStarted","Data":"27202892116bb983ec9f6b62b34ea3890e33afe333085ee32b68d1afb383875a"} Oct 08 18:33:29 crc kubenswrapper[4859]: I1008 18:33:29.464341 4859 generic.go:334] "Generic (PLEG): container finished" podID="bc0261a4-dfa9-4356-9a9d-8a46a187a74c" containerID="0dd53ffc93d496397b8e5cca4560b706d2969d4371b2e906faed9663a825e98d" exitCode=0 Oct 08 18:33:29 crc kubenswrapper[4859]: I1008 18:33:29.464421 4859 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-568bf4d7b5-ljcdc" event={"ID":"bc0261a4-dfa9-4356-9a9d-8a46a187a74c","Type":"ContainerDied","Data":"0dd53ffc93d496397b8e5cca4560b706d2969d4371b2e906faed9663a825e98d"} Oct 08 18:33:29 crc kubenswrapper[4859]: I1008 18:33:29.467906 4859 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-158c-account-create-vvkvz" event={"ID":"ff748840-6d03-4e9a-8284-a4ca9d753278","Type":"ContainerStarted","Data":"9a099402c7e32c8080d9dd5a77c08917d15daa6b42628d4aa8b9826531ac0060"} Oct 08 18:33:29 crc kubenswrapper[4859]: I1008 18:33:29.470589 4859 generic.go:334] "Generic (PLEG): container finished" podID="ac58cb73-6174-4cdd-ab31-66641a15bfc9" containerID="9cd5fb4e738297e6f60582dbcd57dce8132a985d29ad73e28fed44b880b5d393" exitCode=0 Oct 08 18:33:29 crc kubenswrapper[4859]: I1008 18:33:29.470639 4859 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-7c456cb699-8ngvp" event={"ID":"ac58cb73-6174-4cdd-ab31-66641a15bfc9","Type":"ContainerDied","Data":"9cd5fb4e738297e6f60582dbcd57dce8132a985d29ad73e28fed44b880b5d393"} Oct 08 18:33:29 crc kubenswrapper[4859]: I1008 18:33:29.478951 4859 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-d728-account-create-prt4m" event={"ID":"8503bae7-a295-4935-bcc9-a19971c4f417","Type":"ContainerStarted","Data":"3817135e05ff0cedaf55bc7e268e2b05f26fca26eb2f615d768d6ce0b81c2cd2"} Oct 08 18:33:29 crc kubenswrapper[4859]: I1008 18:33:29.494350 4859 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-c89b-account-create-bd78m" event={"ID":"4a64adc4-3689-4d56-b514-f86372be9938","Type":"ContainerStarted","Data":"5e20ee9937e1ddfebd11edf873c8c454923508a0da0ee8406c208dbf0b60306f"} Oct 08 18:33:29 crc kubenswrapper[4859]: I1008 18:33:29.536839 4859 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/neutron-d728-account-create-prt4m" podStartSLOduration=4.536820084 podStartE2EDuration="4.536820084s" podCreationTimestamp="2025-10-08 18:33:25 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-08 18:33:29.530117637 +0000 UTC m=+979.776957026" watchObservedRunningTime="2025-10-08 18:33:29.536820084 +0000 UTC m=+979.783659453" Oct 08 18:33:29 crc kubenswrapper[4859]: I1008 18:33:29.554247 4859 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/cinder-158c-account-create-vvkvz" podStartSLOduration=4.554228306 podStartE2EDuration="4.554228306s" podCreationTimestamp="2025-10-08 18:33:25 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-08 18:33:29.552890847 +0000 UTC m=+979.799730226" watchObservedRunningTime="2025-10-08 18:33:29.554228306 +0000 UTC m=+979.801067695" Oct 08 18:33:29 crc kubenswrapper[4859]: I1008 18:33:29.572281 4859 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/barbican-c89b-account-create-bd78m" podStartSLOduration=4.572265226 podStartE2EDuration="4.572265226s" podCreationTimestamp="2025-10-08 18:33:25 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-08 18:33:29.568510956 +0000 UTC m=+979.815350335" watchObservedRunningTime="2025-10-08 18:33:29.572265226 +0000 UTC m=+979.819104605" Oct 08 18:33:29 crc kubenswrapper[4859]: I1008 18:33:29.903947 4859 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-7c456cb699-8ngvp" Oct 08 18:33:29 crc kubenswrapper[4859]: I1008 18:33:29.988622 4859 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/ac58cb73-6174-4cdd-ab31-66641a15bfc9-ovsdbserver-sb\") pod \"ac58cb73-6174-4cdd-ab31-66641a15bfc9\" (UID: \"ac58cb73-6174-4cdd-ab31-66641a15bfc9\") " Oct 08 18:33:29 crc kubenswrapper[4859]: I1008 18:33:29.988726 4859 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/ac58cb73-6174-4cdd-ab31-66641a15bfc9-ovsdbserver-nb\") pod \"ac58cb73-6174-4cdd-ab31-66641a15bfc9\" (UID: \"ac58cb73-6174-4cdd-ab31-66641a15bfc9\") " Oct 08 18:33:29 crc kubenswrapper[4859]: I1008 18:33:29.988766 4859 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/ac58cb73-6174-4cdd-ab31-66641a15bfc9-config\") pod \"ac58cb73-6174-4cdd-ab31-66641a15bfc9\" (UID: \"ac58cb73-6174-4cdd-ab31-66641a15bfc9\") " Oct 08 18:33:29 crc kubenswrapper[4859]: I1008 18:33:29.988815 4859 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-gk46v\" (UniqueName: \"kubernetes.io/projected/ac58cb73-6174-4cdd-ab31-66641a15bfc9-kube-api-access-gk46v\") pod \"ac58cb73-6174-4cdd-ab31-66641a15bfc9\" (UID: \"ac58cb73-6174-4cdd-ab31-66641a15bfc9\") " Oct 08 18:33:29 crc kubenswrapper[4859]: I1008 18:33:29.988854 4859 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/ac58cb73-6174-4cdd-ab31-66641a15bfc9-dns-svc\") pod \"ac58cb73-6174-4cdd-ab31-66641a15bfc9\" (UID: \"ac58cb73-6174-4cdd-ab31-66641a15bfc9\") " Oct 08 18:33:29 crc kubenswrapper[4859]: I1008 18:33:29.992750 4859 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/ac58cb73-6174-4cdd-ab31-66641a15bfc9-kube-api-access-gk46v" (OuterVolumeSpecName: "kube-api-access-gk46v") pod "ac58cb73-6174-4cdd-ab31-66641a15bfc9" (UID: "ac58cb73-6174-4cdd-ab31-66641a15bfc9"). InnerVolumeSpecName "kube-api-access-gk46v". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 08 18:33:30 crc kubenswrapper[4859]: I1008 18:33:30.009490 4859 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/ac58cb73-6174-4cdd-ab31-66641a15bfc9-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "ac58cb73-6174-4cdd-ab31-66641a15bfc9" (UID: "ac58cb73-6174-4cdd-ab31-66641a15bfc9"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 08 18:33:30 crc kubenswrapper[4859]: I1008 18:33:30.014796 4859 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/ac58cb73-6174-4cdd-ab31-66641a15bfc9-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "ac58cb73-6174-4cdd-ab31-66641a15bfc9" (UID: "ac58cb73-6174-4cdd-ab31-66641a15bfc9"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 08 18:33:30 crc kubenswrapper[4859]: I1008 18:33:30.018287 4859 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/ac58cb73-6174-4cdd-ab31-66641a15bfc9-config" (OuterVolumeSpecName: "config") pod "ac58cb73-6174-4cdd-ab31-66641a15bfc9" (UID: "ac58cb73-6174-4cdd-ab31-66641a15bfc9"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 08 18:33:30 crc kubenswrapper[4859]: I1008 18:33:30.019943 4859 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/ac58cb73-6174-4cdd-ab31-66641a15bfc9-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "ac58cb73-6174-4cdd-ab31-66641a15bfc9" (UID: "ac58cb73-6174-4cdd-ab31-66641a15bfc9"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 08 18:33:30 crc kubenswrapper[4859]: I1008 18:33:30.091576 4859 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/ac58cb73-6174-4cdd-ab31-66641a15bfc9-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Oct 08 18:33:30 crc kubenswrapper[4859]: I1008 18:33:30.091611 4859 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/ac58cb73-6174-4cdd-ab31-66641a15bfc9-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Oct 08 18:33:30 crc kubenswrapper[4859]: I1008 18:33:30.091636 4859 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/ac58cb73-6174-4cdd-ab31-66641a15bfc9-config\") on node \"crc\" DevicePath \"\"" Oct 08 18:33:30 crc kubenswrapper[4859]: I1008 18:33:30.091647 4859 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-gk46v\" (UniqueName: \"kubernetes.io/projected/ac58cb73-6174-4cdd-ab31-66641a15bfc9-kube-api-access-gk46v\") on node \"crc\" DevicePath \"\"" Oct 08 18:33:30 crc kubenswrapper[4859]: I1008 18:33:30.091659 4859 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/ac58cb73-6174-4cdd-ab31-66641a15bfc9-dns-svc\") on node \"crc\" DevicePath \"\"" Oct 08 18:33:30 crc kubenswrapper[4859]: I1008 18:33:30.523191 4859 generic.go:334] "Generic (PLEG): container finished" podID="8503bae7-a295-4935-bcc9-a19971c4f417" containerID="3817135e05ff0cedaf55bc7e268e2b05f26fca26eb2f615d768d6ce0b81c2cd2" exitCode=0 Oct 08 18:33:30 crc kubenswrapper[4859]: I1008 18:33:30.523547 4859 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-d728-account-create-prt4m" event={"ID":"8503bae7-a295-4935-bcc9-a19971c4f417","Type":"ContainerDied","Data":"3817135e05ff0cedaf55bc7e268e2b05f26fca26eb2f615d768d6ce0b81c2cd2"} Oct 08 18:33:30 crc kubenswrapper[4859]: I1008 18:33:30.528314 4859 generic.go:334] "Generic (PLEG): container finished" podID="4a64adc4-3689-4d56-b514-f86372be9938" containerID="5e20ee9937e1ddfebd11edf873c8c454923508a0da0ee8406c208dbf0b60306f" exitCode=0 Oct 08 18:33:30 crc kubenswrapper[4859]: I1008 18:33:30.528564 4859 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-c89b-account-create-bd78m" event={"ID":"4a64adc4-3689-4d56-b514-f86372be9938","Type":"ContainerDied","Data":"5e20ee9937e1ddfebd11edf873c8c454923508a0da0ee8406c208dbf0b60306f"} Oct 08 18:33:30 crc kubenswrapper[4859]: I1008 18:33:30.536933 4859 generic.go:334] "Generic (PLEG): container finished" podID="498d1fd9-4fd1-4dce-a122-6f79d4274882" containerID="0188f80e29f014fdf763c155e5438e9a2bebbf2b89f5e7918a56107b4fe05af2" exitCode=0 Oct 08 18:33:30 crc kubenswrapper[4859]: I1008 18:33:30.537036 4859 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-db-sync-gsq5w" event={"ID":"498d1fd9-4fd1-4dce-a122-6f79d4274882","Type":"ContainerDied","Data":"0188f80e29f014fdf763c155e5438e9a2bebbf2b89f5e7918a56107b4fe05af2"} Oct 08 18:33:30 crc kubenswrapper[4859]: I1008 18:33:30.539812 4859 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-568bf4d7b5-ljcdc" event={"ID":"bc0261a4-dfa9-4356-9a9d-8a46a187a74c","Type":"ContainerStarted","Data":"d725695d953ff6d725f64d0a36703d79675001b7ae193612af150d9372e72f20"} Oct 08 18:33:30 crc kubenswrapper[4859]: I1008 18:33:30.540131 4859 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-568bf4d7b5-ljcdc" Oct 08 18:33:30 crc kubenswrapper[4859]: I1008 18:33:30.556881 4859 generic.go:334] "Generic (PLEG): container finished" podID="ff748840-6d03-4e9a-8284-a4ca9d753278" containerID="9a099402c7e32c8080d9dd5a77c08917d15daa6b42628d4aa8b9826531ac0060" exitCode=0 Oct 08 18:33:30 crc kubenswrapper[4859]: I1008 18:33:30.557204 4859 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-158c-account-create-vvkvz" event={"ID":"ff748840-6d03-4e9a-8284-a4ca9d753278","Type":"ContainerDied","Data":"9a099402c7e32c8080d9dd5a77c08917d15daa6b42628d4aa8b9826531ac0060"} Oct 08 18:33:30 crc kubenswrapper[4859]: I1008 18:33:30.563139 4859 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-7c456cb699-8ngvp" event={"ID":"ac58cb73-6174-4cdd-ab31-66641a15bfc9","Type":"ContainerDied","Data":"2250c06a70511880ba7bdde51622ac969c8a91744fb5c23caa6c9aec4d43a771"} Oct 08 18:33:30 crc kubenswrapper[4859]: I1008 18:33:30.563233 4859 scope.go:117] "RemoveContainer" containerID="9cd5fb4e738297e6f60582dbcd57dce8132a985d29ad73e28fed44b880b5d393" Oct 08 18:33:30 crc kubenswrapper[4859]: I1008 18:33:30.563371 4859 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-7c456cb699-8ngvp" Oct 08 18:33:30 crc kubenswrapper[4859]: I1008 18:33:30.618503 4859 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-568bf4d7b5-ljcdc" podStartSLOduration=5.618484819 podStartE2EDuration="5.618484819s" podCreationTimestamp="2025-10-08 18:33:25 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-08 18:33:30.617219572 +0000 UTC m=+980.864058951" watchObservedRunningTime="2025-10-08 18:33:30.618484819 +0000 UTC m=+980.865324198" Oct 08 18:33:30 crc kubenswrapper[4859]: I1008 18:33:30.661515 4859 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-7c456cb699-8ngvp"] Oct 08 18:33:30 crc kubenswrapper[4859]: I1008 18:33:30.668000 4859 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-7c456cb699-8ngvp"] Oct 08 18:33:32 crc kubenswrapper[4859]: I1008 18:33:32.084352 4859 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-db-sync-gsq5w" Oct 08 18:33:32 crc kubenswrapper[4859]: I1008 18:33:32.091861 4859 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-c89b-account-create-bd78m" Oct 08 18:33:32 crc kubenswrapper[4859]: I1008 18:33:32.106531 4859 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-d728-account-create-prt4m" Oct 08 18:33:32 crc kubenswrapper[4859]: I1008 18:33:32.114924 4859 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-158c-account-create-vvkvz" Oct 08 18:33:32 crc kubenswrapper[4859]: I1008 18:33:32.232864 4859 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-9h2f5\" (UniqueName: \"kubernetes.io/projected/4a64adc4-3689-4d56-b514-f86372be9938-kube-api-access-9h2f5\") pod \"4a64adc4-3689-4d56-b514-f86372be9938\" (UID: \"4a64adc4-3689-4d56-b514-f86372be9938\") " Oct 08 18:33:32 crc kubenswrapper[4859]: I1008 18:33:32.232909 4859 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-6xf29\" (UniqueName: \"kubernetes.io/projected/ff748840-6d03-4e9a-8284-a4ca9d753278-kube-api-access-6xf29\") pod \"ff748840-6d03-4e9a-8284-a4ca9d753278\" (UID: \"ff748840-6d03-4e9a-8284-a4ca9d753278\") " Oct 08 18:33:32 crc kubenswrapper[4859]: I1008 18:33:32.233211 4859 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/498d1fd9-4fd1-4dce-a122-6f79d4274882-config-data\") pod \"498d1fd9-4fd1-4dce-a122-6f79d4274882\" (UID: \"498d1fd9-4fd1-4dce-a122-6f79d4274882\") " Oct 08 18:33:32 crc kubenswrapper[4859]: I1008 18:33:32.233266 4859 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/498d1fd9-4fd1-4dce-a122-6f79d4274882-combined-ca-bundle\") pod \"498d1fd9-4fd1-4dce-a122-6f79d4274882\" (UID: \"498d1fd9-4fd1-4dce-a122-6f79d4274882\") " Oct 08 18:33:32 crc kubenswrapper[4859]: I1008 18:33:32.233302 4859 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-4x9vr\" (UniqueName: \"kubernetes.io/projected/498d1fd9-4fd1-4dce-a122-6f79d4274882-kube-api-access-4x9vr\") pod \"498d1fd9-4fd1-4dce-a122-6f79d4274882\" (UID: \"498d1fd9-4fd1-4dce-a122-6f79d4274882\") " Oct 08 18:33:32 crc kubenswrapper[4859]: I1008 18:33:32.233364 4859 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-9xd5w\" (UniqueName: \"kubernetes.io/projected/8503bae7-a295-4935-bcc9-a19971c4f417-kube-api-access-9xd5w\") pod \"8503bae7-a295-4935-bcc9-a19971c4f417\" (UID: \"8503bae7-a295-4935-bcc9-a19971c4f417\") " Oct 08 18:33:32 crc kubenswrapper[4859]: I1008 18:33:32.239879 4859 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8503bae7-a295-4935-bcc9-a19971c4f417-kube-api-access-9xd5w" (OuterVolumeSpecName: "kube-api-access-9xd5w") pod "8503bae7-a295-4935-bcc9-a19971c4f417" (UID: "8503bae7-a295-4935-bcc9-a19971c4f417"). InnerVolumeSpecName "kube-api-access-9xd5w". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 08 18:33:32 crc kubenswrapper[4859]: I1008 18:33:32.239923 4859 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/ff748840-6d03-4e9a-8284-a4ca9d753278-kube-api-access-6xf29" (OuterVolumeSpecName: "kube-api-access-6xf29") pod "ff748840-6d03-4e9a-8284-a4ca9d753278" (UID: "ff748840-6d03-4e9a-8284-a4ca9d753278"). InnerVolumeSpecName "kube-api-access-6xf29". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 08 18:33:32 crc kubenswrapper[4859]: I1008 18:33:32.240416 4859 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/498d1fd9-4fd1-4dce-a122-6f79d4274882-kube-api-access-4x9vr" (OuterVolumeSpecName: "kube-api-access-4x9vr") pod "498d1fd9-4fd1-4dce-a122-6f79d4274882" (UID: "498d1fd9-4fd1-4dce-a122-6f79d4274882"). InnerVolumeSpecName "kube-api-access-4x9vr". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 08 18:33:32 crc kubenswrapper[4859]: I1008 18:33:32.242123 4859 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/4a64adc4-3689-4d56-b514-f86372be9938-kube-api-access-9h2f5" (OuterVolumeSpecName: "kube-api-access-9h2f5") pod "4a64adc4-3689-4d56-b514-f86372be9938" (UID: "4a64adc4-3689-4d56-b514-f86372be9938"). InnerVolumeSpecName "kube-api-access-9h2f5". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 08 18:33:32 crc kubenswrapper[4859]: I1008 18:33:32.260208 4859 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/498d1fd9-4fd1-4dce-a122-6f79d4274882-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "498d1fd9-4fd1-4dce-a122-6f79d4274882" (UID: "498d1fd9-4fd1-4dce-a122-6f79d4274882"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 08 18:33:32 crc kubenswrapper[4859]: I1008 18:33:32.279639 4859 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/498d1fd9-4fd1-4dce-a122-6f79d4274882-config-data" (OuterVolumeSpecName: "config-data") pod "498d1fd9-4fd1-4dce-a122-6f79d4274882" (UID: "498d1fd9-4fd1-4dce-a122-6f79d4274882"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 08 18:33:32 crc kubenswrapper[4859]: I1008 18:33:32.335027 4859 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-9xd5w\" (UniqueName: \"kubernetes.io/projected/8503bae7-a295-4935-bcc9-a19971c4f417-kube-api-access-9xd5w\") on node \"crc\" DevicePath \"\"" Oct 08 18:33:32 crc kubenswrapper[4859]: I1008 18:33:32.335084 4859 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-9h2f5\" (UniqueName: \"kubernetes.io/projected/4a64adc4-3689-4d56-b514-f86372be9938-kube-api-access-9h2f5\") on node \"crc\" DevicePath \"\"" Oct 08 18:33:32 crc kubenswrapper[4859]: I1008 18:33:32.335098 4859 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-6xf29\" (UniqueName: \"kubernetes.io/projected/ff748840-6d03-4e9a-8284-a4ca9d753278-kube-api-access-6xf29\") on node \"crc\" DevicePath \"\"" Oct 08 18:33:32 crc kubenswrapper[4859]: I1008 18:33:32.335112 4859 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/498d1fd9-4fd1-4dce-a122-6f79d4274882-config-data\") on node \"crc\" DevicePath \"\"" Oct 08 18:33:32 crc kubenswrapper[4859]: I1008 18:33:32.335128 4859 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/498d1fd9-4fd1-4dce-a122-6f79d4274882-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 08 18:33:32 crc kubenswrapper[4859]: I1008 18:33:32.335139 4859 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-4x9vr\" (UniqueName: \"kubernetes.io/projected/498d1fd9-4fd1-4dce-a122-6f79d4274882-kube-api-access-4x9vr\") on node \"crc\" DevicePath \"\"" Oct 08 18:33:32 crc kubenswrapper[4859]: I1008 18:33:32.480013 4859 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="ac58cb73-6174-4cdd-ab31-66641a15bfc9" path="/var/lib/kubelet/pods/ac58cb73-6174-4cdd-ab31-66641a15bfc9/volumes" Oct 08 18:33:32 crc kubenswrapper[4859]: I1008 18:33:32.618576 4859 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-d728-account-create-prt4m" event={"ID":"8503bae7-a295-4935-bcc9-a19971c4f417","Type":"ContainerDied","Data":"27202892116bb983ec9f6b62b34ea3890e33afe333085ee32b68d1afb383875a"} Oct 08 18:33:32 crc kubenswrapper[4859]: I1008 18:33:32.618926 4859 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-d728-account-create-prt4m" Oct 08 18:33:32 crc kubenswrapper[4859]: I1008 18:33:32.618943 4859 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="27202892116bb983ec9f6b62b34ea3890e33afe333085ee32b68d1afb383875a" Oct 08 18:33:32 crc kubenswrapper[4859]: I1008 18:33:32.620633 4859 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-c89b-account-create-bd78m" event={"ID":"4a64adc4-3689-4d56-b514-f86372be9938","Type":"ContainerDied","Data":"752de51991f367023c1433d111ae85780d623c32e960fcb5f4b8b3784a11fd80"} Oct 08 18:33:32 crc kubenswrapper[4859]: I1008 18:33:32.620658 4859 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="752de51991f367023c1433d111ae85780d623c32e960fcb5f4b8b3784a11fd80" Oct 08 18:33:32 crc kubenswrapper[4859]: I1008 18:33:32.620800 4859 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-c89b-account-create-bd78m" Oct 08 18:33:32 crc kubenswrapper[4859]: I1008 18:33:32.622118 4859 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-db-sync-gsq5w" event={"ID":"498d1fd9-4fd1-4dce-a122-6f79d4274882","Type":"ContainerDied","Data":"56183b00ad1b371d8528e2f952cf250c9f5f256f000435c1a78b61ee2fad014d"} Oct 08 18:33:32 crc kubenswrapper[4859]: I1008 18:33:32.622148 4859 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="56183b00ad1b371d8528e2f952cf250c9f5f256f000435c1a78b61ee2fad014d" Oct 08 18:33:32 crc kubenswrapper[4859]: I1008 18:33:32.622156 4859 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-db-sync-gsq5w" Oct 08 18:33:32 crc kubenswrapper[4859]: I1008 18:33:32.623352 4859 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-158c-account-create-vvkvz" event={"ID":"ff748840-6d03-4e9a-8284-a4ca9d753278","Type":"ContainerDied","Data":"683d6beb079f21e33e46a1beed704f18ced8f288f065c0c57296caa88cdc3131"} Oct 08 18:33:32 crc kubenswrapper[4859]: I1008 18:33:32.623372 4859 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="683d6beb079f21e33e46a1beed704f18ced8f288f065c0c57296caa88cdc3131" Oct 08 18:33:32 crc kubenswrapper[4859]: I1008 18:33:32.623404 4859 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-158c-account-create-vvkvz" Oct 08 18:33:32 crc kubenswrapper[4859]: I1008 18:33:32.809634 4859 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-568bf4d7b5-ljcdc"] Oct 08 18:33:32 crc kubenswrapper[4859]: I1008 18:33:32.809861 4859 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-568bf4d7b5-ljcdc" podUID="bc0261a4-dfa9-4356-9a9d-8a46a187a74c" containerName="dnsmasq-dns" containerID="cri-o://d725695d953ff6d725f64d0a36703d79675001b7ae193612af150d9372e72f20" gracePeriod=10 Oct 08 18:33:32 crc kubenswrapper[4859]: I1008 18:33:32.850098 4859 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-59df9f9d49-446ss"] Oct 08 18:33:32 crc kubenswrapper[4859]: E1008 18:33:32.850575 4859 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ff748840-6d03-4e9a-8284-a4ca9d753278" containerName="mariadb-account-create" Oct 08 18:33:32 crc kubenswrapper[4859]: I1008 18:33:32.850594 4859 state_mem.go:107] "Deleted CPUSet assignment" podUID="ff748840-6d03-4e9a-8284-a4ca9d753278" containerName="mariadb-account-create" Oct 08 18:33:32 crc kubenswrapper[4859]: E1008 18:33:32.850606 4859 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ac58cb73-6174-4cdd-ab31-66641a15bfc9" containerName="init" Oct 08 18:33:32 crc kubenswrapper[4859]: I1008 18:33:32.850612 4859 state_mem.go:107] "Deleted CPUSet assignment" podUID="ac58cb73-6174-4cdd-ab31-66641a15bfc9" containerName="init" Oct 08 18:33:32 crc kubenswrapper[4859]: E1008 18:33:32.850638 4859 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="4a64adc4-3689-4d56-b514-f86372be9938" containerName="mariadb-account-create" Oct 08 18:33:32 crc kubenswrapper[4859]: I1008 18:33:32.850645 4859 state_mem.go:107] "Deleted CPUSet assignment" podUID="4a64adc4-3689-4d56-b514-f86372be9938" containerName="mariadb-account-create" Oct 08 18:33:32 crc kubenswrapper[4859]: E1008 18:33:32.850656 4859 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8503bae7-a295-4935-bcc9-a19971c4f417" containerName="mariadb-account-create" Oct 08 18:33:32 crc kubenswrapper[4859]: I1008 18:33:32.850663 4859 state_mem.go:107] "Deleted CPUSet assignment" podUID="8503bae7-a295-4935-bcc9-a19971c4f417" containerName="mariadb-account-create" Oct 08 18:33:32 crc kubenswrapper[4859]: E1008 18:33:32.850682 4859 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="498d1fd9-4fd1-4dce-a122-6f79d4274882" containerName="keystone-db-sync" Oct 08 18:33:32 crc kubenswrapper[4859]: I1008 18:33:32.850710 4859 state_mem.go:107] "Deleted CPUSet assignment" podUID="498d1fd9-4fd1-4dce-a122-6f79d4274882" containerName="keystone-db-sync" Oct 08 18:33:32 crc kubenswrapper[4859]: I1008 18:33:32.851374 4859 memory_manager.go:354] "RemoveStaleState removing state" podUID="ac58cb73-6174-4cdd-ab31-66641a15bfc9" containerName="init" Oct 08 18:33:32 crc kubenswrapper[4859]: I1008 18:33:32.851429 4859 memory_manager.go:354] "RemoveStaleState removing state" podUID="8503bae7-a295-4935-bcc9-a19971c4f417" containerName="mariadb-account-create" Oct 08 18:33:32 crc kubenswrapper[4859]: I1008 18:33:32.851440 4859 memory_manager.go:354] "RemoveStaleState removing state" podUID="498d1fd9-4fd1-4dce-a122-6f79d4274882" containerName="keystone-db-sync" Oct 08 18:33:32 crc kubenswrapper[4859]: I1008 18:33:32.851447 4859 memory_manager.go:354] "RemoveStaleState removing state" podUID="4a64adc4-3689-4d56-b514-f86372be9938" containerName="mariadb-account-create" Oct 08 18:33:32 crc kubenswrapper[4859]: I1008 18:33:32.851462 4859 memory_manager.go:354] "RemoveStaleState removing state" podUID="ff748840-6d03-4e9a-8284-a4ca9d753278" containerName="mariadb-account-create" Oct 08 18:33:32 crc kubenswrapper[4859]: I1008 18:33:32.857165 4859 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-59df9f9d49-446ss" Oct 08 18:33:32 crc kubenswrapper[4859]: I1008 18:33:32.893483 4859 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-59df9f9d49-446ss"] Oct 08 18:33:32 crc kubenswrapper[4859]: I1008 18:33:32.933803 4859 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/keystone-bootstrap-mqwmv"] Oct 08 18:33:32 crc kubenswrapper[4859]: I1008 18:33:32.938269 4859 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-bootstrap-mqwmv" Oct 08 18:33:32 crc kubenswrapper[4859]: I1008 18:33:32.949109 4859 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone" Oct 08 18:33:32 crc kubenswrapper[4859]: I1008 18:33:32.949364 4859 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-config-data" Oct 08 18:33:32 crc kubenswrapper[4859]: I1008 18:33:32.949422 4859 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/59231da1-64f0-4cac-a821-20fcbdcd5e30-ovsdbserver-sb\") pod \"dnsmasq-dns-59df9f9d49-446ss\" (UID: \"59231da1-64f0-4cac-a821-20fcbdcd5e30\") " pod="openstack/dnsmasq-dns-59df9f9d49-446ss" Oct 08 18:33:32 crc kubenswrapper[4859]: I1008 18:33:32.949480 4859 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-scripts" Oct 08 18:33:32 crc kubenswrapper[4859]: I1008 18:33:32.949492 4859 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/59231da1-64f0-4cac-a821-20fcbdcd5e30-config\") pod \"dnsmasq-dns-59df9f9d49-446ss\" (UID: \"59231da1-64f0-4cac-a821-20fcbdcd5e30\") " pod="openstack/dnsmasq-dns-59df9f9d49-446ss" Oct 08 18:33:32 crc kubenswrapper[4859]: I1008 18:33:32.949521 4859 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/59231da1-64f0-4cac-a821-20fcbdcd5e30-dns-svc\") pod \"dnsmasq-dns-59df9f9d49-446ss\" (UID: \"59231da1-64f0-4cac-a821-20fcbdcd5e30\") " pod="openstack/dnsmasq-dns-59df9f9d49-446ss" Oct 08 18:33:32 crc kubenswrapper[4859]: I1008 18:33:32.949558 4859 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/59231da1-64f0-4cac-a821-20fcbdcd5e30-dns-swift-storage-0\") pod \"dnsmasq-dns-59df9f9d49-446ss\" (UID: \"59231da1-64f0-4cac-a821-20fcbdcd5e30\") " pod="openstack/dnsmasq-dns-59df9f9d49-446ss" Oct 08 18:33:32 crc kubenswrapper[4859]: I1008 18:33:32.949628 4859 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-lbgjh\" (UniqueName: \"kubernetes.io/projected/59231da1-64f0-4cac-a821-20fcbdcd5e30-kube-api-access-lbgjh\") pod \"dnsmasq-dns-59df9f9d49-446ss\" (UID: \"59231da1-64f0-4cac-a821-20fcbdcd5e30\") " pod="openstack/dnsmasq-dns-59df9f9d49-446ss" Oct 08 18:33:32 crc kubenswrapper[4859]: I1008 18:33:32.949664 4859 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/59231da1-64f0-4cac-a821-20fcbdcd5e30-ovsdbserver-nb\") pod \"dnsmasq-dns-59df9f9d49-446ss\" (UID: \"59231da1-64f0-4cac-a821-20fcbdcd5e30\") " pod="openstack/dnsmasq-dns-59df9f9d49-446ss" Oct 08 18:33:32 crc kubenswrapper[4859]: I1008 18:33:32.957361 4859 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-keystone-dockercfg-jrkml" Oct 08 18:33:32 crc kubenswrapper[4859]: I1008 18:33:32.995757 4859 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-bootstrap-mqwmv"] Oct 08 18:33:33 crc kubenswrapper[4859]: I1008 18:33:33.051133 4859 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/59231da1-64f0-4cac-a821-20fcbdcd5e30-ovsdbserver-sb\") pod \"dnsmasq-dns-59df9f9d49-446ss\" (UID: \"59231da1-64f0-4cac-a821-20fcbdcd5e30\") " pod="openstack/dnsmasq-dns-59df9f9d49-446ss" Oct 08 18:33:33 crc kubenswrapper[4859]: I1008 18:33:33.051225 4859 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/59231da1-64f0-4cac-a821-20fcbdcd5e30-config\") pod \"dnsmasq-dns-59df9f9d49-446ss\" (UID: \"59231da1-64f0-4cac-a821-20fcbdcd5e30\") " pod="openstack/dnsmasq-dns-59df9f9d49-446ss" Oct 08 18:33:33 crc kubenswrapper[4859]: I1008 18:33:33.051248 4859 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/59231da1-64f0-4cac-a821-20fcbdcd5e30-dns-svc\") pod \"dnsmasq-dns-59df9f9d49-446ss\" (UID: \"59231da1-64f0-4cac-a821-20fcbdcd5e30\") " pod="openstack/dnsmasq-dns-59df9f9d49-446ss" Oct 08 18:33:33 crc kubenswrapper[4859]: I1008 18:33:33.051272 4859 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/9d1b6fef-16a5-4c1b-a3bb-95f46f1d92f7-fernet-keys\") pod \"keystone-bootstrap-mqwmv\" (UID: \"9d1b6fef-16a5-4c1b-a3bb-95f46f1d92f7\") " pod="openstack/keystone-bootstrap-mqwmv" Oct 08 18:33:33 crc kubenswrapper[4859]: I1008 18:33:33.051307 4859 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/59231da1-64f0-4cac-a821-20fcbdcd5e30-dns-swift-storage-0\") pod \"dnsmasq-dns-59df9f9d49-446ss\" (UID: \"59231da1-64f0-4cac-a821-20fcbdcd5e30\") " pod="openstack/dnsmasq-dns-59df9f9d49-446ss" Oct 08 18:33:33 crc kubenswrapper[4859]: I1008 18:33:33.051352 4859 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-txptt\" (UniqueName: \"kubernetes.io/projected/9d1b6fef-16a5-4c1b-a3bb-95f46f1d92f7-kube-api-access-txptt\") pod \"keystone-bootstrap-mqwmv\" (UID: \"9d1b6fef-16a5-4c1b-a3bb-95f46f1d92f7\") " pod="openstack/keystone-bootstrap-mqwmv" Oct 08 18:33:33 crc kubenswrapper[4859]: I1008 18:33:33.051386 4859 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/9d1b6fef-16a5-4c1b-a3bb-95f46f1d92f7-config-data\") pod \"keystone-bootstrap-mqwmv\" (UID: \"9d1b6fef-16a5-4c1b-a3bb-95f46f1d92f7\") " pod="openstack/keystone-bootstrap-mqwmv" Oct 08 18:33:33 crc kubenswrapper[4859]: I1008 18:33:33.051406 4859 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/9d1b6fef-16a5-4c1b-a3bb-95f46f1d92f7-credential-keys\") pod \"keystone-bootstrap-mqwmv\" (UID: \"9d1b6fef-16a5-4c1b-a3bb-95f46f1d92f7\") " pod="openstack/keystone-bootstrap-mqwmv" Oct 08 18:33:33 crc kubenswrapper[4859]: I1008 18:33:33.051459 4859 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-lbgjh\" (UniqueName: \"kubernetes.io/projected/59231da1-64f0-4cac-a821-20fcbdcd5e30-kube-api-access-lbgjh\") pod \"dnsmasq-dns-59df9f9d49-446ss\" (UID: \"59231da1-64f0-4cac-a821-20fcbdcd5e30\") " pod="openstack/dnsmasq-dns-59df9f9d49-446ss" Oct 08 18:33:33 crc kubenswrapper[4859]: I1008 18:33:33.051502 4859 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/59231da1-64f0-4cac-a821-20fcbdcd5e30-ovsdbserver-nb\") pod \"dnsmasq-dns-59df9f9d49-446ss\" (UID: \"59231da1-64f0-4cac-a821-20fcbdcd5e30\") " pod="openstack/dnsmasq-dns-59df9f9d49-446ss" Oct 08 18:33:33 crc kubenswrapper[4859]: I1008 18:33:33.051529 4859 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/9d1b6fef-16a5-4c1b-a3bb-95f46f1d92f7-scripts\") pod \"keystone-bootstrap-mqwmv\" (UID: \"9d1b6fef-16a5-4c1b-a3bb-95f46f1d92f7\") " pod="openstack/keystone-bootstrap-mqwmv" Oct 08 18:33:33 crc kubenswrapper[4859]: I1008 18:33:33.051588 4859 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9d1b6fef-16a5-4c1b-a3bb-95f46f1d92f7-combined-ca-bundle\") pod \"keystone-bootstrap-mqwmv\" (UID: \"9d1b6fef-16a5-4c1b-a3bb-95f46f1d92f7\") " pod="openstack/keystone-bootstrap-mqwmv" Oct 08 18:33:33 crc kubenswrapper[4859]: I1008 18:33:33.052703 4859 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/59231da1-64f0-4cac-a821-20fcbdcd5e30-config\") pod \"dnsmasq-dns-59df9f9d49-446ss\" (UID: \"59231da1-64f0-4cac-a821-20fcbdcd5e30\") " pod="openstack/dnsmasq-dns-59df9f9d49-446ss" Oct 08 18:33:33 crc kubenswrapper[4859]: I1008 18:33:33.053154 4859 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/59231da1-64f0-4cac-a821-20fcbdcd5e30-ovsdbserver-nb\") pod \"dnsmasq-dns-59df9f9d49-446ss\" (UID: \"59231da1-64f0-4cac-a821-20fcbdcd5e30\") " pod="openstack/dnsmasq-dns-59df9f9d49-446ss" Oct 08 18:33:33 crc kubenswrapper[4859]: I1008 18:33:33.053184 4859 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/59231da1-64f0-4cac-a821-20fcbdcd5e30-dns-svc\") pod \"dnsmasq-dns-59df9f9d49-446ss\" (UID: \"59231da1-64f0-4cac-a821-20fcbdcd5e30\") " pod="openstack/dnsmasq-dns-59df9f9d49-446ss" Oct 08 18:33:33 crc kubenswrapper[4859]: I1008 18:33:33.053288 4859 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/59231da1-64f0-4cac-a821-20fcbdcd5e30-dns-swift-storage-0\") pod \"dnsmasq-dns-59df9f9d49-446ss\" (UID: \"59231da1-64f0-4cac-a821-20fcbdcd5e30\") " pod="openstack/dnsmasq-dns-59df9f9d49-446ss" Oct 08 18:33:33 crc kubenswrapper[4859]: I1008 18:33:33.053506 4859 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/59231da1-64f0-4cac-a821-20fcbdcd5e30-ovsdbserver-sb\") pod \"dnsmasq-dns-59df9f9d49-446ss\" (UID: \"59231da1-64f0-4cac-a821-20fcbdcd5e30\") " pod="openstack/dnsmasq-dns-59df9f9d49-446ss" Oct 08 18:33:33 crc kubenswrapper[4859]: I1008 18:33:33.085353 4859 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/horizon-7444cc4f69-5xs5x"] Oct 08 18:33:33 crc kubenswrapper[4859]: I1008 18:33:33.089786 4859 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-7444cc4f69-5xs5x" Oct 08 18:33:33 crc kubenswrapper[4859]: I1008 18:33:33.099626 4859 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"horizon-config-data" Oct 08 18:33:33 crc kubenswrapper[4859]: I1008 18:33:33.099992 4859 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"horizon-scripts" Oct 08 18:33:33 crc kubenswrapper[4859]: I1008 18:33:33.100134 4859 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"horizon-horizon-dockercfg-gdtfc" Oct 08 18:33:33 crc kubenswrapper[4859]: I1008 18:33:33.100746 4859 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"horizon" Oct 08 18:33:33 crc kubenswrapper[4859]: I1008 18:33:33.141917 4859 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-lbgjh\" (UniqueName: \"kubernetes.io/projected/59231da1-64f0-4cac-a821-20fcbdcd5e30-kube-api-access-lbgjh\") pod \"dnsmasq-dns-59df9f9d49-446ss\" (UID: \"59231da1-64f0-4cac-a821-20fcbdcd5e30\") " pod="openstack/dnsmasq-dns-59df9f9d49-446ss" Oct 08 18:33:33 crc kubenswrapper[4859]: I1008 18:33:33.155513 4859 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9d1b6fef-16a5-4c1b-a3bb-95f46f1d92f7-combined-ca-bundle\") pod \"keystone-bootstrap-mqwmv\" (UID: \"9d1b6fef-16a5-4c1b-a3bb-95f46f1d92f7\") " pod="openstack/keystone-bootstrap-mqwmv" Oct 08 18:33:33 crc kubenswrapper[4859]: I1008 18:33:33.155652 4859 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/9d1b6fef-16a5-4c1b-a3bb-95f46f1d92f7-fernet-keys\") pod \"keystone-bootstrap-mqwmv\" (UID: \"9d1b6fef-16a5-4c1b-a3bb-95f46f1d92f7\") " pod="openstack/keystone-bootstrap-mqwmv" Oct 08 18:33:33 crc kubenswrapper[4859]: I1008 18:33:33.155732 4859 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-txptt\" (UniqueName: \"kubernetes.io/projected/9d1b6fef-16a5-4c1b-a3bb-95f46f1d92f7-kube-api-access-txptt\") pod \"keystone-bootstrap-mqwmv\" (UID: \"9d1b6fef-16a5-4c1b-a3bb-95f46f1d92f7\") " pod="openstack/keystone-bootstrap-mqwmv" Oct 08 18:33:33 crc kubenswrapper[4859]: I1008 18:33:33.155766 4859 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/9d1b6fef-16a5-4c1b-a3bb-95f46f1d92f7-config-data\") pod \"keystone-bootstrap-mqwmv\" (UID: \"9d1b6fef-16a5-4c1b-a3bb-95f46f1d92f7\") " pod="openstack/keystone-bootstrap-mqwmv" Oct 08 18:33:33 crc kubenswrapper[4859]: I1008 18:33:33.155786 4859 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/9d1b6fef-16a5-4c1b-a3bb-95f46f1d92f7-credential-keys\") pod \"keystone-bootstrap-mqwmv\" (UID: \"9d1b6fef-16a5-4c1b-a3bb-95f46f1d92f7\") " pod="openstack/keystone-bootstrap-mqwmv" Oct 08 18:33:33 crc kubenswrapper[4859]: I1008 18:33:33.155860 4859 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/9d1b6fef-16a5-4c1b-a3bb-95f46f1d92f7-scripts\") pod \"keystone-bootstrap-mqwmv\" (UID: \"9d1b6fef-16a5-4c1b-a3bb-95f46f1d92f7\") " pod="openstack/keystone-bootstrap-mqwmv" Oct 08 18:33:33 crc kubenswrapper[4859]: I1008 18:33:33.167912 4859 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/9d1b6fef-16a5-4c1b-a3bb-95f46f1d92f7-scripts\") pod \"keystone-bootstrap-mqwmv\" (UID: \"9d1b6fef-16a5-4c1b-a3bb-95f46f1d92f7\") " pod="openstack/keystone-bootstrap-mqwmv" Oct 08 18:33:33 crc kubenswrapper[4859]: I1008 18:33:33.173768 4859 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/9d1b6fef-16a5-4c1b-a3bb-95f46f1d92f7-config-data\") pod \"keystone-bootstrap-mqwmv\" (UID: \"9d1b6fef-16a5-4c1b-a3bb-95f46f1d92f7\") " pod="openstack/keystone-bootstrap-mqwmv" Oct 08 18:33:33 crc kubenswrapper[4859]: I1008 18:33:33.177335 4859 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9d1b6fef-16a5-4c1b-a3bb-95f46f1d92f7-combined-ca-bundle\") pod \"keystone-bootstrap-mqwmv\" (UID: \"9d1b6fef-16a5-4c1b-a3bb-95f46f1d92f7\") " pod="openstack/keystone-bootstrap-mqwmv" Oct 08 18:33:33 crc kubenswrapper[4859]: I1008 18:33:33.193323 4859 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/9d1b6fef-16a5-4c1b-a3bb-95f46f1d92f7-credential-keys\") pod \"keystone-bootstrap-mqwmv\" (UID: \"9d1b6fef-16a5-4c1b-a3bb-95f46f1d92f7\") " pod="openstack/keystone-bootstrap-mqwmv" Oct 08 18:33:33 crc kubenswrapper[4859]: I1008 18:33:33.212130 4859 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/horizon-7444cc4f69-5xs5x"] Oct 08 18:33:33 crc kubenswrapper[4859]: I1008 18:33:33.213176 4859 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/9d1b6fef-16a5-4c1b-a3bb-95f46f1d92f7-fernet-keys\") pod \"keystone-bootstrap-mqwmv\" (UID: \"9d1b6fef-16a5-4c1b-a3bb-95f46f1d92f7\") " pod="openstack/keystone-bootstrap-mqwmv" Oct 08 18:33:33 crc kubenswrapper[4859]: I1008 18:33:33.238158 4859 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-txptt\" (UniqueName: \"kubernetes.io/projected/9d1b6fef-16a5-4c1b-a3bb-95f46f1d92f7-kube-api-access-txptt\") pod \"keystone-bootstrap-mqwmv\" (UID: \"9d1b6fef-16a5-4c1b-a3bb-95f46f1d92f7\") " pod="openstack/keystone-bootstrap-mqwmv" Oct 08 18:33:33 crc kubenswrapper[4859]: I1008 18:33:33.251137 4859 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-59df9f9d49-446ss"] Oct 08 18:33:33 crc kubenswrapper[4859]: I1008 18:33:33.251913 4859 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-59df9f9d49-446ss" Oct 08 18:33:33 crc kubenswrapper[4859]: I1008 18:33:33.257596 4859 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/ac9272b3-433e-4afb-afca-0d647a9d9be5-horizon-secret-key\") pod \"horizon-7444cc4f69-5xs5x\" (UID: \"ac9272b3-433e-4afb-afca-0d647a9d9be5\") " pod="openstack/horizon-7444cc4f69-5xs5x" Oct 08 18:33:33 crc kubenswrapper[4859]: I1008 18:33:33.257635 4859 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/ac9272b3-433e-4afb-afca-0d647a9d9be5-config-data\") pod \"horizon-7444cc4f69-5xs5x\" (UID: \"ac9272b3-433e-4afb-afca-0d647a9d9be5\") " pod="openstack/horizon-7444cc4f69-5xs5x" Oct 08 18:33:33 crc kubenswrapper[4859]: I1008 18:33:33.257708 4859 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/ac9272b3-433e-4afb-afca-0d647a9d9be5-logs\") pod \"horizon-7444cc4f69-5xs5x\" (UID: \"ac9272b3-433e-4afb-afca-0d647a9d9be5\") " pod="openstack/horizon-7444cc4f69-5xs5x" Oct 08 18:33:33 crc kubenswrapper[4859]: I1008 18:33:33.257788 4859 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/ac9272b3-433e-4afb-afca-0d647a9d9be5-scripts\") pod \"horizon-7444cc4f69-5xs5x\" (UID: \"ac9272b3-433e-4afb-afca-0d647a9d9be5\") " pod="openstack/horizon-7444cc4f69-5xs5x" Oct 08 18:33:33 crc kubenswrapper[4859]: I1008 18:33:33.257873 4859 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-2mf2r\" (UniqueName: \"kubernetes.io/projected/ac9272b3-433e-4afb-afca-0d647a9d9be5-kube-api-access-2mf2r\") pod \"horizon-7444cc4f69-5xs5x\" (UID: \"ac9272b3-433e-4afb-afca-0d647a9d9be5\") " pod="openstack/horizon-7444cc4f69-5xs5x" Oct 08 18:33:33 crc kubenswrapper[4859]: I1008 18:33:33.297678 4859 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-74776f5dd7-89sbw"] Oct 08 18:33:33 crc kubenswrapper[4859]: I1008 18:33:33.299307 4859 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-74776f5dd7-89sbw" Oct 08 18:33:33 crc kubenswrapper[4859]: I1008 18:33:33.315352 4859 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/placement-db-sync-lsbrj"] Oct 08 18:33:33 crc kubenswrapper[4859]: I1008 18:33:33.317609 4859 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/placement-db-sync-lsbrj" Oct 08 18:33:33 crc kubenswrapper[4859]: I1008 18:33:33.321268 4859 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"placement-scripts" Oct 08 18:33:33 crc kubenswrapper[4859]: I1008 18:33:33.321534 4859 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"placement-placement-dockercfg-gwsrg" Oct 08 18:33:33 crc kubenswrapper[4859]: I1008 18:33:33.331630 4859 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-74776f5dd7-89sbw"] Oct 08 18:33:33 crc kubenswrapper[4859]: I1008 18:33:33.332492 4859 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"placement-config-data" Oct 08 18:33:33 crc kubenswrapper[4859]: I1008 18:33:33.349961 4859 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/horizon-86b9b87697-gmwqv"] Oct 08 18:33:33 crc kubenswrapper[4859]: I1008 18:33:33.363571 4859 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/placement-db-sync-lsbrj"] Oct 08 18:33:33 crc kubenswrapper[4859]: I1008 18:33:33.363723 4859 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-86b9b87697-gmwqv" Oct 08 18:33:33 crc kubenswrapper[4859]: I1008 18:33:33.368915 4859 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-2mf2r\" (UniqueName: \"kubernetes.io/projected/ac9272b3-433e-4afb-afca-0d647a9d9be5-kube-api-access-2mf2r\") pod \"horizon-7444cc4f69-5xs5x\" (UID: \"ac9272b3-433e-4afb-afca-0d647a9d9be5\") " pod="openstack/horizon-7444cc4f69-5xs5x" Oct 08 18:33:33 crc kubenswrapper[4859]: I1008 18:33:33.369190 4859 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/ac9272b3-433e-4afb-afca-0d647a9d9be5-horizon-secret-key\") pod \"horizon-7444cc4f69-5xs5x\" (UID: \"ac9272b3-433e-4afb-afca-0d647a9d9be5\") " pod="openstack/horizon-7444cc4f69-5xs5x" Oct 08 18:33:33 crc kubenswrapper[4859]: I1008 18:33:33.369309 4859 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/ac9272b3-433e-4afb-afca-0d647a9d9be5-config-data\") pod \"horizon-7444cc4f69-5xs5x\" (UID: \"ac9272b3-433e-4afb-afca-0d647a9d9be5\") " pod="openstack/horizon-7444cc4f69-5xs5x" Oct 08 18:33:33 crc kubenswrapper[4859]: I1008 18:33:33.369414 4859 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/ac9272b3-433e-4afb-afca-0d647a9d9be5-logs\") pod \"horizon-7444cc4f69-5xs5x\" (UID: \"ac9272b3-433e-4afb-afca-0d647a9d9be5\") " pod="openstack/horizon-7444cc4f69-5xs5x" Oct 08 18:33:33 crc kubenswrapper[4859]: I1008 18:33:33.369538 4859 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/ac9272b3-433e-4afb-afca-0d647a9d9be5-scripts\") pod \"horizon-7444cc4f69-5xs5x\" (UID: \"ac9272b3-433e-4afb-afca-0d647a9d9be5\") " pod="openstack/horizon-7444cc4f69-5xs5x" Oct 08 18:33:33 crc kubenswrapper[4859]: I1008 18:33:33.377637 4859 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/horizon-86b9b87697-gmwqv"] Oct 08 18:33:33 crc kubenswrapper[4859]: I1008 18:33:33.381946 4859 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-bootstrap-mqwmv" Oct 08 18:33:33 crc kubenswrapper[4859]: I1008 18:33:33.383527 4859 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/ac9272b3-433e-4afb-afca-0d647a9d9be5-scripts\") pod \"horizon-7444cc4f69-5xs5x\" (UID: \"ac9272b3-433e-4afb-afca-0d647a9d9be5\") " pod="openstack/horizon-7444cc4f69-5xs5x" Oct 08 18:33:33 crc kubenswrapper[4859]: I1008 18:33:33.391531 4859 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/ac9272b3-433e-4afb-afca-0d647a9d9be5-config-data\") pod \"horizon-7444cc4f69-5xs5x\" (UID: \"ac9272b3-433e-4afb-afca-0d647a9d9be5\") " pod="openstack/horizon-7444cc4f69-5xs5x" Oct 08 18:33:33 crc kubenswrapper[4859]: I1008 18:33:33.402013 4859 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/ac9272b3-433e-4afb-afca-0d647a9d9be5-logs\") pod \"horizon-7444cc4f69-5xs5x\" (UID: \"ac9272b3-433e-4afb-afca-0d647a9d9be5\") " pod="openstack/horizon-7444cc4f69-5xs5x" Oct 08 18:33:33 crc kubenswrapper[4859]: I1008 18:33:33.410639 4859 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/ac9272b3-433e-4afb-afca-0d647a9d9be5-horizon-secret-key\") pod \"horizon-7444cc4f69-5xs5x\" (UID: \"ac9272b3-433e-4afb-afca-0d647a9d9be5\") " pod="openstack/horizon-7444cc4f69-5xs5x" Oct 08 18:33:33 crc kubenswrapper[4859]: I1008 18:33:33.425644 4859 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-2mf2r\" (UniqueName: \"kubernetes.io/projected/ac9272b3-433e-4afb-afca-0d647a9d9be5-kube-api-access-2mf2r\") pod \"horizon-7444cc4f69-5xs5x\" (UID: \"ac9272b3-433e-4afb-afca-0d647a9d9be5\") " pod="openstack/horizon-7444cc4f69-5xs5x" Oct 08 18:33:33 crc kubenswrapper[4859]: I1008 18:33:33.435766 4859 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/glance-default-internal-api-0"] Oct 08 18:33:33 crc kubenswrapper[4859]: I1008 18:33:33.444743 4859 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-internal-api-0" Oct 08 18:33:33 crc kubenswrapper[4859]: I1008 18:33:33.451923 4859 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-scripts" Oct 08 18:33:33 crc kubenswrapper[4859]: I1008 18:33:33.453296 4859 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-default-internal-config-data" Oct 08 18:33:33 crc kubenswrapper[4859]: I1008 18:33:33.453428 4859 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-glance-default-internal-svc" Oct 08 18:33:33 crc kubenswrapper[4859]: I1008 18:33:33.453605 4859 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-glance-dockercfg-2fnvk" Oct 08 18:33:33 crc kubenswrapper[4859]: I1008 18:33:33.463924 4859 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-internal-api-0"] Oct 08 18:33:33 crc kubenswrapper[4859]: I1008 18:33:33.471610 4859 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-fjb7b\" (UniqueName: \"kubernetes.io/projected/ed95b639-b2c8-4879-bedb-16681e0166cf-kube-api-access-fjb7b\") pod \"placement-db-sync-lsbrj\" (UID: \"ed95b639-b2c8-4879-bedb-16681e0166cf\") " pod="openstack/placement-db-sync-lsbrj" Oct 08 18:33:33 crc kubenswrapper[4859]: I1008 18:33:33.471654 4859 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/f08c21e6-a74c-4ad9-a3fa-4e56ff2c6432-ovsdbserver-nb\") pod \"dnsmasq-dns-74776f5dd7-89sbw\" (UID: \"f08c21e6-a74c-4ad9-a3fa-4e56ff2c6432\") " pod="openstack/dnsmasq-dns-74776f5dd7-89sbw" Oct 08 18:33:33 crc kubenswrapper[4859]: I1008 18:33:33.471676 4859 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/ed95b639-b2c8-4879-bedb-16681e0166cf-logs\") pod \"placement-db-sync-lsbrj\" (UID: \"ed95b639-b2c8-4879-bedb-16681e0166cf\") " pod="openstack/placement-db-sync-lsbrj" Oct 08 18:33:33 crc kubenswrapper[4859]: I1008 18:33:33.471730 4859 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/f08c21e6-a74c-4ad9-a3fa-4e56ff2c6432-dns-swift-storage-0\") pod \"dnsmasq-dns-74776f5dd7-89sbw\" (UID: \"f08c21e6-a74c-4ad9-a3fa-4e56ff2c6432\") " pod="openstack/dnsmasq-dns-74776f5dd7-89sbw" Oct 08 18:33:33 crc kubenswrapper[4859]: I1008 18:33:33.471751 4859 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/f08c21e6-a74c-4ad9-a3fa-4e56ff2c6432-config\") pod \"dnsmasq-dns-74776f5dd7-89sbw\" (UID: \"f08c21e6-a74c-4ad9-a3fa-4e56ff2c6432\") " pod="openstack/dnsmasq-dns-74776f5dd7-89sbw" Oct 08 18:33:33 crc kubenswrapper[4859]: I1008 18:33:33.471772 4859 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/1acee243-a101-4dc6-a599-aaee4821dbfa-scripts\") pod \"horizon-86b9b87697-gmwqv\" (UID: \"1acee243-a101-4dc6-a599-aaee4821dbfa\") " pod="openstack/horizon-86b9b87697-gmwqv" Oct 08 18:33:33 crc kubenswrapper[4859]: I1008 18:33:33.471796 4859 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ed95b639-b2c8-4879-bedb-16681e0166cf-combined-ca-bundle\") pod \"placement-db-sync-lsbrj\" (UID: \"ed95b639-b2c8-4879-bedb-16681e0166cf\") " pod="openstack/placement-db-sync-lsbrj" Oct 08 18:33:33 crc kubenswrapper[4859]: I1008 18:33:33.471813 4859 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ed95b639-b2c8-4879-bedb-16681e0166cf-config-data\") pod \"placement-db-sync-lsbrj\" (UID: \"ed95b639-b2c8-4879-bedb-16681e0166cf\") " pod="openstack/placement-db-sync-lsbrj" Oct 08 18:33:33 crc kubenswrapper[4859]: I1008 18:33:33.471829 4859 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-4czsn\" (UniqueName: \"kubernetes.io/projected/f08c21e6-a74c-4ad9-a3fa-4e56ff2c6432-kube-api-access-4czsn\") pod \"dnsmasq-dns-74776f5dd7-89sbw\" (UID: \"f08c21e6-a74c-4ad9-a3fa-4e56ff2c6432\") " pod="openstack/dnsmasq-dns-74776f5dd7-89sbw" Oct 08 18:33:33 crc kubenswrapper[4859]: I1008 18:33:33.471853 4859 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/f08c21e6-a74c-4ad9-a3fa-4e56ff2c6432-ovsdbserver-sb\") pod \"dnsmasq-dns-74776f5dd7-89sbw\" (UID: \"f08c21e6-a74c-4ad9-a3fa-4e56ff2c6432\") " pod="openstack/dnsmasq-dns-74776f5dd7-89sbw" Oct 08 18:33:33 crc kubenswrapper[4859]: I1008 18:33:33.471872 4859 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/ed95b639-b2c8-4879-bedb-16681e0166cf-scripts\") pod \"placement-db-sync-lsbrj\" (UID: \"ed95b639-b2c8-4879-bedb-16681e0166cf\") " pod="openstack/placement-db-sync-lsbrj" Oct 08 18:33:33 crc kubenswrapper[4859]: I1008 18:33:33.471893 4859 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/1acee243-a101-4dc6-a599-aaee4821dbfa-config-data\") pod \"horizon-86b9b87697-gmwqv\" (UID: \"1acee243-a101-4dc6-a599-aaee4821dbfa\") " pod="openstack/horizon-86b9b87697-gmwqv" Oct 08 18:33:33 crc kubenswrapper[4859]: I1008 18:33:33.471913 4859 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/1acee243-a101-4dc6-a599-aaee4821dbfa-horizon-secret-key\") pod \"horizon-86b9b87697-gmwqv\" (UID: \"1acee243-a101-4dc6-a599-aaee4821dbfa\") " pod="openstack/horizon-86b9b87697-gmwqv" Oct 08 18:33:33 crc kubenswrapper[4859]: I1008 18:33:33.471931 4859 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-d9csb\" (UniqueName: \"kubernetes.io/projected/1acee243-a101-4dc6-a599-aaee4821dbfa-kube-api-access-d9csb\") pod \"horizon-86b9b87697-gmwqv\" (UID: \"1acee243-a101-4dc6-a599-aaee4821dbfa\") " pod="openstack/horizon-86b9b87697-gmwqv" Oct 08 18:33:33 crc kubenswrapper[4859]: I1008 18:33:33.471981 4859 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/f08c21e6-a74c-4ad9-a3fa-4e56ff2c6432-dns-svc\") pod \"dnsmasq-dns-74776f5dd7-89sbw\" (UID: \"f08c21e6-a74c-4ad9-a3fa-4e56ff2c6432\") " pod="openstack/dnsmasq-dns-74776f5dd7-89sbw" Oct 08 18:33:33 crc kubenswrapper[4859]: I1008 18:33:33.471998 4859 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/1acee243-a101-4dc6-a599-aaee4821dbfa-logs\") pod \"horizon-86b9b87697-gmwqv\" (UID: \"1acee243-a101-4dc6-a599-aaee4821dbfa\") " pod="openstack/horizon-86b9b87697-gmwqv" Oct 08 18:33:33 crc kubenswrapper[4859]: I1008 18:33:33.541788 4859 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/glance-default-external-api-0"] Oct 08 18:33:33 crc kubenswrapper[4859]: I1008 18:33:33.543550 4859 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-external-api-0" Oct 08 18:33:33 crc kubenswrapper[4859]: I1008 18:33:33.547375 4859 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-568bf4d7b5-ljcdc" Oct 08 18:33:33 crc kubenswrapper[4859]: I1008 18:33:33.554189 4859 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-external-api-0"] Oct 08 18:33:33 crc kubenswrapper[4859]: I1008 18:33:33.559222 4859 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ceilometer-0"] Oct 08 18:33:33 crc kubenswrapper[4859]: E1008 18:33:33.559577 4859 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="bc0261a4-dfa9-4356-9a9d-8a46a187a74c" containerName="dnsmasq-dns" Oct 08 18:33:33 crc kubenswrapper[4859]: I1008 18:33:33.559589 4859 state_mem.go:107] "Deleted CPUSet assignment" podUID="bc0261a4-dfa9-4356-9a9d-8a46a187a74c" containerName="dnsmasq-dns" Oct 08 18:33:33 crc kubenswrapper[4859]: E1008 18:33:33.559610 4859 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="bc0261a4-dfa9-4356-9a9d-8a46a187a74c" containerName="init" Oct 08 18:33:33 crc kubenswrapper[4859]: I1008 18:33:33.559616 4859 state_mem.go:107] "Deleted CPUSet assignment" podUID="bc0261a4-dfa9-4356-9a9d-8a46a187a74c" containerName="init" Oct 08 18:33:33 crc kubenswrapper[4859]: I1008 18:33:33.559906 4859 memory_manager.go:354] "RemoveStaleState removing state" podUID="bc0261a4-dfa9-4356-9a9d-8a46a187a74c" containerName="dnsmasq-dns" Oct 08 18:33:33 crc kubenswrapper[4859]: I1008 18:33:33.560368 4859 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-glance-default-public-svc" Oct 08 18:33:33 crc kubenswrapper[4859]: I1008 18:33:33.560590 4859 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-default-external-config-data" Oct 08 18:33:33 crc kubenswrapper[4859]: I1008 18:33:33.561292 4859 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Oct 08 18:33:33 crc kubenswrapper[4859]: I1008 18:33:33.573741 4859 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-scripts" Oct 08 18:33:33 crc kubenswrapper[4859]: I1008 18:33:33.573885 4859 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-config-data" Oct 08 18:33:33 crc kubenswrapper[4859]: I1008 18:33:33.581603 4859 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/f08c21e6-a74c-4ad9-a3fa-4e56ff2c6432-ovsdbserver-sb\") pod \"dnsmasq-dns-74776f5dd7-89sbw\" (UID: \"f08c21e6-a74c-4ad9-a3fa-4e56ff2c6432\") " pod="openstack/dnsmasq-dns-74776f5dd7-89sbw" Oct 08 18:33:33 crc kubenswrapper[4859]: I1008 18:33:33.581641 4859 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/ed95b639-b2c8-4879-bedb-16681e0166cf-scripts\") pod \"placement-db-sync-lsbrj\" (UID: \"ed95b639-b2c8-4879-bedb-16681e0166cf\") " pod="openstack/placement-db-sync-lsbrj" Oct 08 18:33:33 crc kubenswrapper[4859]: I1008 18:33:33.581666 4859 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/1acee243-a101-4dc6-a599-aaee4821dbfa-config-data\") pod \"horizon-86b9b87697-gmwqv\" (UID: \"1acee243-a101-4dc6-a599-aaee4821dbfa\") " pod="openstack/horizon-86b9b87697-gmwqv" Oct 08 18:33:33 crc kubenswrapper[4859]: I1008 18:33:33.581713 4859 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/21a3071c-7cf9-4008-a37b-ae90f125ddbc-logs\") pod \"glance-default-internal-api-0\" (UID: \"21a3071c-7cf9-4008-a37b-ae90f125ddbc\") " pod="openstack/glance-default-internal-api-0" Oct 08 18:33:33 crc kubenswrapper[4859]: I1008 18:33:33.581734 4859 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/1acee243-a101-4dc6-a599-aaee4821dbfa-horizon-secret-key\") pod \"horizon-86b9b87697-gmwqv\" (UID: \"1acee243-a101-4dc6-a599-aaee4821dbfa\") " pod="openstack/horizon-86b9b87697-gmwqv" Oct 08 18:33:33 crc kubenswrapper[4859]: I1008 18:33:33.581750 4859 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-d9csb\" (UniqueName: \"kubernetes.io/projected/1acee243-a101-4dc6-a599-aaee4821dbfa-kube-api-access-d9csb\") pod \"horizon-86b9b87697-gmwqv\" (UID: \"1acee243-a101-4dc6-a599-aaee4821dbfa\") " pod="openstack/horizon-86b9b87697-gmwqv" Oct 08 18:33:33 crc kubenswrapper[4859]: I1008 18:33:33.581795 4859 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/21a3071c-7cf9-4008-a37b-ae90f125ddbc-combined-ca-bundle\") pod \"glance-default-internal-api-0\" (UID: \"21a3071c-7cf9-4008-a37b-ae90f125ddbc\") " pod="openstack/glance-default-internal-api-0" Oct 08 18:33:33 crc kubenswrapper[4859]: I1008 18:33:33.581858 4859 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/f08c21e6-a74c-4ad9-a3fa-4e56ff2c6432-dns-svc\") pod \"dnsmasq-dns-74776f5dd7-89sbw\" (UID: \"f08c21e6-a74c-4ad9-a3fa-4e56ff2c6432\") " pod="openstack/dnsmasq-dns-74776f5dd7-89sbw" Oct 08 18:33:33 crc kubenswrapper[4859]: I1008 18:33:33.581873 4859 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/1acee243-a101-4dc6-a599-aaee4821dbfa-logs\") pod \"horizon-86b9b87697-gmwqv\" (UID: \"1acee243-a101-4dc6-a599-aaee4821dbfa\") " pod="openstack/horizon-86b9b87697-gmwqv" Oct 08 18:33:33 crc kubenswrapper[4859]: I1008 18:33:33.581903 4859 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-fjb7b\" (UniqueName: \"kubernetes.io/projected/ed95b639-b2c8-4879-bedb-16681e0166cf-kube-api-access-fjb7b\") pod \"placement-db-sync-lsbrj\" (UID: \"ed95b639-b2c8-4879-bedb-16681e0166cf\") " pod="openstack/placement-db-sync-lsbrj" Oct 08 18:33:33 crc kubenswrapper[4859]: I1008 18:33:33.581924 4859 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/21a3071c-7cf9-4008-a37b-ae90f125ddbc-internal-tls-certs\") pod \"glance-default-internal-api-0\" (UID: \"21a3071c-7cf9-4008-a37b-ae90f125ddbc\") " pod="openstack/glance-default-internal-api-0" Oct 08 18:33:33 crc kubenswrapper[4859]: I1008 18:33:33.581940 4859 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/f08c21e6-a74c-4ad9-a3fa-4e56ff2c6432-ovsdbserver-nb\") pod \"dnsmasq-dns-74776f5dd7-89sbw\" (UID: \"f08c21e6-a74c-4ad9-a3fa-4e56ff2c6432\") " pod="openstack/dnsmasq-dns-74776f5dd7-89sbw" Oct 08 18:33:33 crc kubenswrapper[4859]: I1008 18:33:33.581960 4859 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/ed95b639-b2c8-4879-bedb-16681e0166cf-logs\") pod \"placement-db-sync-lsbrj\" (UID: \"ed95b639-b2c8-4879-bedb-16681e0166cf\") " pod="openstack/placement-db-sync-lsbrj" Oct 08 18:33:33 crc kubenswrapper[4859]: I1008 18:33:33.581980 4859 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-9mgdb\" (UniqueName: \"kubernetes.io/projected/21a3071c-7cf9-4008-a37b-ae90f125ddbc-kube-api-access-9mgdb\") pod \"glance-default-internal-api-0\" (UID: \"21a3071c-7cf9-4008-a37b-ae90f125ddbc\") " pod="openstack/glance-default-internal-api-0" Oct 08 18:33:33 crc kubenswrapper[4859]: I1008 18:33:33.582003 4859 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/21a3071c-7cf9-4008-a37b-ae90f125ddbc-scripts\") pod \"glance-default-internal-api-0\" (UID: \"21a3071c-7cf9-4008-a37b-ae90f125ddbc\") " pod="openstack/glance-default-internal-api-0" Oct 08 18:33:33 crc kubenswrapper[4859]: I1008 18:33:33.582019 4859 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/f08c21e6-a74c-4ad9-a3fa-4e56ff2c6432-dns-swift-storage-0\") pod \"dnsmasq-dns-74776f5dd7-89sbw\" (UID: \"f08c21e6-a74c-4ad9-a3fa-4e56ff2c6432\") " pod="openstack/dnsmasq-dns-74776f5dd7-89sbw" Oct 08 18:33:33 crc kubenswrapper[4859]: I1008 18:33:33.582048 4859 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/f08c21e6-a74c-4ad9-a3fa-4e56ff2c6432-config\") pod \"dnsmasq-dns-74776f5dd7-89sbw\" (UID: \"f08c21e6-a74c-4ad9-a3fa-4e56ff2c6432\") " pod="openstack/dnsmasq-dns-74776f5dd7-89sbw" Oct 08 18:33:33 crc kubenswrapper[4859]: I1008 18:33:33.582069 4859 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/1acee243-a101-4dc6-a599-aaee4821dbfa-scripts\") pod \"horizon-86b9b87697-gmwqv\" (UID: \"1acee243-a101-4dc6-a599-aaee4821dbfa\") " pod="openstack/horizon-86b9b87697-gmwqv" Oct 08 18:33:33 crc kubenswrapper[4859]: I1008 18:33:33.582088 4859 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/21a3071c-7cf9-4008-a37b-ae90f125ddbc-httpd-run\") pod \"glance-default-internal-api-0\" (UID: \"21a3071c-7cf9-4008-a37b-ae90f125ddbc\") " pod="openstack/glance-default-internal-api-0" Oct 08 18:33:33 crc kubenswrapper[4859]: I1008 18:33:33.582105 4859 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/21a3071c-7cf9-4008-a37b-ae90f125ddbc-config-data\") pod \"glance-default-internal-api-0\" (UID: \"21a3071c-7cf9-4008-a37b-ae90f125ddbc\") " pod="openstack/glance-default-internal-api-0" Oct 08 18:33:33 crc kubenswrapper[4859]: I1008 18:33:33.582122 4859 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ed95b639-b2c8-4879-bedb-16681e0166cf-combined-ca-bundle\") pod \"placement-db-sync-lsbrj\" (UID: \"ed95b639-b2c8-4879-bedb-16681e0166cf\") " pod="openstack/placement-db-sync-lsbrj" Oct 08 18:33:33 crc kubenswrapper[4859]: I1008 18:33:33.582139 4859 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ed95b639-b2c8-4879-bedb-16681e0166cf-config-data\") pod \"placement-db-sync-lsbrj\" (UID: \"ed95b639-b2c8-4879-bedb-16681e0166cf\") " pod="openstack/placement-db-sync-lsbrj" Oct 08 18:33:33 crc kubenswrapper[4859]: I1008 18:33:33.582159 4859 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage10-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage10-crc\") pod \"glance-default-internal-api-0\" (UID: \"21a3071c-7cf9-4008-a37b-ae90f125ddbc\") " pod="openstack/glance-default-internal-api-0" Oct 08 18:33:33 crc kubenswrapper[4859]: I1008 18:33:33.582181 4859 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-4czsn\" (UniqueName: \"kubernetes.io/projected/f08c21e6-a74c-4ad9-a3fa-4e56ff2c6432-kube-api-access-4czsn\") pod \"dnsmasq-dns-74776f5dd7-89sbw\" (UID: \"f08c21e6-a74c-4ad9-a3fa-4e56ff2c6432\") " pod="openstack/dnsmasq-dns-74776f5dd7-89sbw" Oct 08 18:33:33 crc kubenswrapper[4859]: I1008 18:33:33.583515 4859 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/f08c21e6-a74c-4ad9-a3fa-4e56ff2c6432-ovsdbserver-sb\") pod \"dnsmasq-dns-74776f5dd7-89sbw\" (UID: \"f08c21e6-a74c-4ad9-a3fa-4e56ff2c6432\") " pod="openstack/dnsmasq-dns-74776f5dd7-89sbw" Oct 08 18:33:33 crc kubenswrapper[4859]: I1008 18:33:33.584183 4859 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-7444cc4f69-5xs5x" Oct 08 18:33:33 crc kubenswrapper[4859]: I1008 18:33:33.585496 4859 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/ed95b639-b2c8-4879-bedb-16681e0166cf-logs\") pod \"placement-db-sync-lsbrj\" (UID: \"ed95b639-b2c8-4879-bedb-16681e0166cf\") " pod="openstack/placement-db-sync-lsbrj" Oct 08 18:33:33 crc kubenswrapper[4859]: I1008 18:33:33.586755 4859 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/1acee243-a101-4dc6-a599-aaee4821dbfa-logs\") pod \"horizon-86b9b87697-gmwqv\" (UID: \"1acee243-a101-4dc6-a599-aaee4821dbfa\") " pod="openstack/horizon-86b9b87697-gmwqv" Oct 08 18:33:33 crc kubenswrapper[4859]: I1008 18:33:33.586768 4859 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/1acee243-a101-4dc6-a599-aaee4821dbfa-config-data\") pod \"horizon-86b9b87697-gmwqv\" (UID: \"1acee243-a101-4dc6-a599-aaee4821dbfa\") " pod="openstack/horizon-86b9b87697-gmwqv" Oct 08 18:33:33 crc kubenswrapper[4859]: I1008 18:33:33.587231 4859 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/f08c21e6-a74c-4ad9-a3fa-4e56ff2c6432-dns-swift-storage-0\") pod \"dnsmasq-dns-74776f5dd7-89sbw\" (UID: \"f08c21e6-a74c-4ad9-a3fa-4e56ff2c6432\") " pod="openstack/dnsmasq-dns-74776f5dd7-89sbw" Oct 08 18:33:33 crc kubenswrapper[4859]: I1008 18:33:33.587416 4859 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/f08c21e6-a74c-4ad9-a3fa-4e56ff2c6432-config\") pod \"dnsmasq-dns-74776f5dd7-89sbw\" (UID: \"f08c21e6-a74c-4ad9-a3fa-4e56ff2c6432\") " pod="openstack/dnsmasq-dns-74776f5dd7-89sbw" Oct 08 18:33:33 crc kubenswrapper[4859]: I1008 18:33:33.587421 4859 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/f08c21e6-a74c-4ad9-a3fa-4e56ff2c6432-ovsdbserver-nb\") pod \"dnsmasq-dns-74776f5dd7-89sbw\" (UID: \"f08c21e6-a74c-4ad9-a3fa-4e56ff2c6432\") " pod="openstack/dnsmasq-dns-74776f5dd7-89sbw" Oct 08 18:33:33 crc kubenswrapper[4859]: I1008 18:33:33.587827 4859 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/1acee243-a101-4dc6-a599-aaee4821dbfa-scripts\") pod \"horizon-86b9b87697-gmwqv\" (UID: \"1acee243-a101-4dc6-a599-aaee4821dbfa\") " pod="openstack/horizon-86b9b87697-gmwqv" Oct 08 18:33:33 crc kubenswrapper[4859]: I1008 18:33:33.588186 4859 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/f08c21e6-a74c-4ad9-a3fa-4e56ff2c6432-dns-svc\") pod \"dnsmasq-dns-74776f5dd7-89sbw\" (UID: \"f08c21e6-a74c-4ad9-a3fa-4e56ff2c6432\") " pod="openstack/dnsmasq-dns-74776f5dd7-89sbw" Oct 08 18:33:33 crc kubenswrapper[4859]: I1008 18:33:33.599032 4859 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ed95b639-b2c8-4879-bedb-16681e0166cf-config-data\") pod \"placement-db-sync-lsbrj\" (UID: \"ed95b639-b2c8-4879-bedb-16681e0166cf\") " pod="openstack/placement-db-sync-lsbrj" Oct 08 18:33:33 crc kubenswrapper[4859]: I1008 18:33:33.612993 4859 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Oct 08 18:33:33 crc kubenswrapper[4859]: I1008 18:33:33.615130 4859 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ed95b639-b2c8-4879-bedb-16681e0166cf-combined-ca-bundle\") pod \"placement-db-sync-lsbrj\" (UID: \"ed95b639-b2c8-4879-bedb-16681e0166cf\") " pod="openstack/placement-db-sync-lsbrj" Oct 08 18:33:33 crc kubenswrapper[4859]: I1008 18:33:33.621172 4859 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-fjb7b\" (UniqueName: \"kubernetes.io/projected/ed95b639-b2c8-4879-bedb-16681e0166cf-kube-api-access-fjb7b\") pod \"placement-db-sync-lsbrj\" (UID: \"ed95b639-b2c8-4879-bedb-16681e0166cf\") " pod="openstack/placement-db-sync-lsbrj" Oct 08 18:33:33 crc kubenswrapper[4859]: I1008 18:33:33.621584 4859 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/ed95b639-b2c8-4879-bedb-16681e0166cf-scripts\") pod \"placement-db-sync-lsbrj\" (UID: \"ed95b639-b2c8-4879-bedb-16681e0166cf\") " pod="openstack/placement-db-sync-lsbrj" Oct 08 18:33:33 crc kubenswrapper[4859]: I1008 18:33:33.622073 4859 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/1acee243-a101-4dc6-a599-aaee4821dbfa-horizon-secret-key\") pod \"horizon-86b9b87697-gmwqv\" (UID: \"1acee243-a101-4dc6-a599-aaee4821dbfa\") " pod="openstack/horizon-86b9b87697-gmwqv" Oct 08 18:33:33 crc kubenswrapper[4859]: I1008 18:33:33.633261 4859 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-d9csb\" (UniqueName: \"kubernetes.io/projected/1acee243-a101-4dc6-a599-aaee4821dbfa-kube-api-access-d9csb\") pod \"horizon-86b9b87697-gmwqv\" (UID: \"1acee243-a101-4dc6-a599-aaee4821dbfa\") " pod="openstack/horizon-86b9b87697-gmwqv" Oct 08 18:33:33 crc kubenswrapper[4859]: I1008 18:33:33.634820 4859 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-4czsn\" (UniqueName: \"kubernetes.io/projected/f08c21e6-a74c-4ad9-a3fa-4e56ff2c6432-kube-api-access-4czsn\") pod \"dnsmasq-dns-74776f5dd7-89sbw\" (UID: \"f08c21e6-a74c-4ad9-a3fa-4e56ff2c6432\") " pod="openstack/dnsmasq-dns-74776f5dd7-89sbw" Oct 08 18:33:33 crc kubenswrapper[4859]: I1008 18:33:33.673752 4859 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-74776f5dd7-89sbw" Oct 08 18:33:33 crc kubenswrapper[4859]: I1008 18:33:33.684027 4859 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/bc0261a4-dfa9-4356-9a9d-8a46a187a74c-ovsdbserver-sb\") pod \"bc0261a4-dfa9-4356-9a9d-8a46a187a74c\" (UID: \"bc0261a4-dfa9-4356-9a9d-8a46a187a74c\") " Oct 08 18:33:33 crc kubenswrapper[4859]: I1008 18:33:33.684089 4859 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/bc0261a4-dfa9-4356-9a9d-8a46a187a74c-config\") pod \"bc0261a4-dfa9-4356-9a9d-8a46a187a74c\" (UID: \"bc0261a4-dfa9-4356-9a9d-8a46a187a74c\") " Oct 08 18:33:33 crc kubenswrapper[4859]: I1008 18:33:33.684148 4859 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-s92r8\" (UniqueName: \"kubernetes.io/projected/bc0261a4-dfa9-4356-9a9d-8a46a187a74c-kube-api-access-s92r8\") pod \"bc0261a4-dfa9-4356-9a9d-8a46a187a74c\" (UID: \"bc0261a4-dfa9-4356-9a9d-8a46a187a74c\") " Oct 08 18:33:33 crc kubenswrapper[4859]: I1008 18:33:33.684288 4859 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/bc0261a4-dfa9-4356-9a9d-8a46a187a74c-ovsdbserver-nb\") pod \"bc0261a4-dfa9-4356-9a9d-8a46a187a74c\" (UID: \"bc0261a4-dfa9-4356-9a9d-8a46a187a74c\") " Oct 08 18:33:33 crc kubenswrapper[4859]: I1008 18:33:33.684319 4859 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/bc0261a4-dfa9-4356-9a9d-8a46a187a74c-dns-svc\") pod \"bc0261a4-dfa9-4356-9a9d-8a46a187a74c\" (UID: \"bc0261a4-dfa9-4356-9a9d-8a46a187a74c\") " Oct 08 18:33:33 crc kubenswrapper[4859]: I1008 18:33:33.684370 4859 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/bc0261a4-dfa9-4356-9a9d-8a46a187a74c-dns-swift-storage-0\") pod \"bc0261a4-dfa9-4356-9a9d-8a46a187a74c\" (UID: \"bc0261a4-dfa9-4356-9a9d-8a46a187a74c\") " Oct 08 18:33:33 crc kubenswrapper[4859]: I1008 18:33:33.684559 4859 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/21a3071c-7cf9-4008-a37b-ae90f125ddbc-httpd-run\") pod \"glance-default-internal-api-0\" (UID: \"21a3071c-7cf9-4008-a37b-ae90f125ddbc\") " pod="openstack/glance-default-internal-api-0" Oct 08 18:33:33 crc kubenswrapper[4859]: I1008 18:33:33.684579 4859 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/21a3071c-7cf9-4008-a37b-ae90f125ddbc-config-data\") pod \"glance-default-internal-api-0\" (UID: \"21a3071c-7cf9-4008-a37b-ae90f125ddbc\") " pod="openstack/glance-default-internal-api-0" Oct 08 18:33:33 crc kubenswrapper[4859]: I1008 18:33:33.684600 4859 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage10-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage10-crc\") pod \"glance-default-internal-api-0\" (UID: \"21a3071c-7cf9-4008-a37b-ae90f125ddbc\") " pod="openstack/glance-default-internal-api-0" Oct 08 18:33:33 crc kubenswrapper[4859]: I1008 18:33:33.684621 4859 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/87f0eecb-178a-438a-8643-e29c19390884-log-httpd\") pod \"ceilometer-0\" (UID: \"87f0eecb-178a-438a-8643-e29c19390884\") " pod="openstack/ceilometer-0" Oct 08 18:33:33 crc kubenswrapper[4859]: I1008 18:33:33.684645 4859 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/bfc391b0-6e05-41b9-8c78-7cf870f5cfe4-logs\") pod \"glance-default-external-api-0\" (UID: \"bfc391b0-6e05-41b9-8c78-7cf870f5cfe4\") " pod="openstack/glance-default-external-api-0" Oct 08 18:33:33 crc kubenswrapper[4859]: I1008 18:33:33.684662 4859 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/bfc391b0-6e05-41b9-8c78-7cf870f5cfe4-public-tls-certs\") pod \"glance-default-external-api-0\" (UID: \"bfc391b0-6e05-41b9-8c78-7cf870f5cfe4\") " pod="openstack/glance-default-external-api-0" Oct 08 18:33:33 crc kubenswrapper[4859]: I1008 18:33:33.684705 4859 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/21a3071c-7cf9-4008-a37b-ae90f125ddbc-logs\") pod \"glance-default-internal-api-0\" (UID: \"21a3071c-7cf9-4008-a37b-ae90f125ddbc\") " pod="openstack/glance-default-internal-api-0" Oct 08 18:33:33 crc kubenswrapper[4859]: I1008 18:33:33.684721 4859 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/87f0eecb-178a-438a-8643-e29c19390884-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"87f0eecb-178a-438a-8643-e29c19390884\") " pod="openstack/ceilometer-0" Oct 08 18:33:33 crc kubenswrapper[4859]: I1008 18:33:33.684745 4859 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/bfc391b0-6e05-41b9-8c78-7cf870f5cfe4-config-data\") pod \"glance-default-external-api-0\" (UID: \"bfc391b0-6e05-41b9-8c78-7cf870f5cfe4\") " pod="openstack/glance-default-external-api-0" Oct 08 18:33:33 crc kubenswrapper[4859]: I1008 18:33:33.684784 4859 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/21a3071c-7cf9-4008-a37b-ae90f125ddbc-combined-ca-bundle\") pod \"glance-default-internal-api-0\" (UID: \"21a3071c-7cf9-4008-a37b-ae90f125ddbc\") " pod="openstack/glance-default-internal-api-0" Oct 08 18:33:33 crc kubenswrapper[4859]: I1008 18:33:33.684807 4859 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-4kfnf\" (UniqueName: \"kubernetes.io/projected/bfc391b0-6e05-41b9-8c78-7cf870f5cfe4-kube-api-access-4kfnf\") pod \"glance-default-external-api-0\" (UID: \"bfc391b0-6e05-41b9-8c78-7cf870f5cfe4\") " pod="openstack/glance-default-external-api-0" Oct 08 18:33:33 crc kubenswrapper[4859]: I1008 18:33:33.684833 4859 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/87f0eecb-178a-438a-8643-e29c19390884-run-httpd\") pod \"ceilometer-0\" (UID: \"87f0eecb-178a-438a-8643-e29c19390884\") " pod="openstack/ceilometer-0" Oct 08 18:33:33 crc kubenswrapper[4859]: I1008 18:33:33.684851 4859 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/87f0eecb-178a-438a-8643-e29c19390884-scripts\") pod \"ceilometer-0\" (UID: \"87f0eecb-178a-438a-8643-e29c19390884\") " pod="openstack/ceilometer-0" Oct 08 18:33:33 crc kubenswrapper[4859]: I1008 18:33:33.684867 4859 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/bfc391b0-6e05-41b9-8c78-7cf870f5cfe4-httpd-run\") pod \"glance-default-external-api-0\" (UID: \"bfc391b0-6e05-41b9-8c78-7cf870f5cfe4\") " pod="openstack/glance-default-external-api-0" Oct 08 18:33:33 crc kubenswrapper[4859]: I1008 18:33:33.684890 4859 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/21a3071c-7cf9-4008-a37b-ae90f125ddbc-internal-tls-certs\") pod \"glance-default-internal-api-0\" (UID: \"21a3071c-7cf9-4008-a37b-ae90f125ddbc\") " pod="openstack/glance-default-internal-api-0" Oct 08 18:33:33 crc kubenswrapper[4859]: I1008 18:33:33.684909 4859 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/87f0eecb-178a-438a-8643-e29c19390884-config-data\") pod \"ceilometer-0\" (UID: \"87f0eecb-178a-438a-8643-e29c19390884\") " pod="openstack/ceilometer-0" Oct 08 18:33:33 crc kubenswrapper[4859]: I1008 18:33:33.684928 4859 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-nx5wn\" (UniqueName: \"kubernetes.io/projected/87f0eecb-178a-438a-8643-e29c19390884-kube-api-access-nx5wn\") pod \"ceilometer-0\" (UID: \"87f0eecb-178a-438a-8643-e29c19390884\") " pod="openstack/ceilometer-0" Oct 08 18:33:33 crc kubenswrapper[4859]: I1008 18:33:33.684945 4859 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-9mgdb\" (UniqueName: \"kubernetes.io/projected/21a3071c-7cf9-4008-a37b-ae90f125ddbc-kube-api-access-9mgdb\") pod \"glance-default-internal-api-0\" (UID: \"21a3071c-7cf9-4008-a37b-ae90f125ddbc\") " pod="openstack/glance-default-internal-api-0" Oct 08 18:33:33 crc kubenswrapper[4859]: I1008 18:33:33.684982 4859 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/bfc391b0-6e05-41b9-8c78-7cf870f5cfe4-scripts\") pod \"glance-default-external-api-0\" (UID: \"bfc391b0-6e05-41b9-8c78-7cf870f5cfe4\") " pod="openstack/glance-default-external-api-0" Oct 08 18:33:33 crc kubenswrapper[4859]: I1008 18:33:33.684999 4859 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/21a3071c-7cf9-4008-a37b-ae90f125ddbc-scripts\") pod \"glance-default-internal-api-0\" (UID: \"21a3071c-7cf9-4008-a37b-ae90f125ddbc\") " pod="openstack/glance-default-internal-api-0" Oct 08 18:33:33 crc kubenswrapper[4859]: I1008 18:33:33.685020 4859 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage12-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage12-crc\") pod \"glance-default-external-api-0\" (UID: \"bfc391b0-6e05-41b9-8c78-7cf870f5cfe4\") " pod="openstack/glance-default-external-api-0" Oct 08 18:33:33 crc kubenswrapper[4859]: I1008 18:33:33.685037 4859 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/87f0eecb-178a-438a-8643-e29c19390884-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"87f0eecb-178a-438a-8643-e29c19390884\") " pod="openstack/ceilometer-0" Oct 08 18:33:33 crc kubenswrapper[4859]: I1008 18:33:33.685052 4859 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/bfc391b0-6e05-41b9-8c78-7cf870f5cfe4-combined-ca-bundle\") pod \"glance-default-external-api-0\" (UID: \"bfc391b0-6e05-41b9-8c78-7cf870f5cfe4\") " pod="openstack/glance-default-external-api-0" Oct 08 18:33:33 crc kubenswrapper[4859]: I1008 18:33:33.697129 4859 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage10-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage10-crc\") pod \"glance-default-internal-api-0\" (UID: \"21a3071c-7cf9-4008-a37b-ae90f125ddbc\") device mount path \"/mnt/openstack/pv10\"" pod="openstack/glance-default-internal-api-0" Oct 08 18:33:33 crc kubenswrapper[4859]: I1008 18:33:33.700216 4859 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/21a3071c-7cf9-4008-a37b-ae90f125ddbc-logs\") pod \"glance-default-internal-api-0\" (UID: \"21a3071c-7cf9-4008-a37b-ae90f125ddbc\") " pod="openstack/glance-default-internal-api-0" Oct 08 18:33:33 crc kubenswrapper[4859]: I1008 18:33:33.709709 4859 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/21a3071c-7cf9-4008-a37b-ae90f125ddbc-httpd-run\") pod \"glance-default-internal-api-0\" (UID: \"21a3071c-7cf9-4008-a37b-ae90f125ddbc\") " pod="openstack/glance-default-internal-api-0" Oct 08 18:33:33 crc kubenswrapper[4859]: I1008 18:33:33.712284 4859 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/placement-db-sync-lsbrj" Oct 08 18:33:33 crc kubenswrapper[4859]: I1008 18:33:33.713923 4859 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/21a3071c-7cf9-4008-a37b-ae90f125ddbc-config-data\") pod \"glance-default-internal-api-0\" (UID: \"21a3071c-7cf9-4008-a37b-ae90f125ddbc\") " pod="openstack/glance-default-internal-api-0" Oct 08 18:33:33 crc kubenswrapper[4859]: I1008 18:33:33.746125 4859 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/bc0261a4-dfa9-4356-9a9d-8a46a187a74c-kube-api-access-s92r8" (OuterVolumeSpecName: "kube-api-access-s92r8") pod "bc0261a4-dfa9-4356-9a9d-8a46a187a74c" (UID: "bc0261a4-dfa9-4356-9a9d-8a46a187a74c"). InnerVolumeSpecName "kube-api-access-s92r8". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 08 18:33:33 crc kubenswrapper[4859]: I1008 18:33:33.746542 4859 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-86b9b87697-gmwqv" Oct 08 18:33:33 crc kubenswrapper[4859]: I1008 18:33:33.762140 4859 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/21a3071c-7cf9-4008-a37b-ae90f125ddbc-combined-ca-bundle\") pod \"glance-default-internal-api-0\" (UID: \"21a3071c-7cf9-4008-a37b-ae90f125ddbc\") " pod="openstack/glance-default-internal-api-0" Oct 08 18:33:33 crc kubenswrapper[4859]: I1008 18:33:33.765258 4859 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/21a3071c-7cf9-4008-a37b-ae90f125ddbc-internal-tls-certs\") pod \"glance-default-internal-api-0\" (UID: \"21a3071c-7cf9-4008-a37b-ae90f125ddbc\") " pod="openstack/glance-default-internal-api-0" Oct 08 18:33:33 crc kubenswrapper[4859]: I1008 18:33:33.766951 4859 generic.go:334] "Generic (PLEG): container finished" podID="bc0261a4-dfa9-4356-9a9d-8a46a187a74c" containerID="d725695d953ff6d725f64d0a36703d79675001b7ae193612af150d9372e72f20" exitCode=0 Oct 08 18:33:33 crc kubenswrapper[4859]: I1008 18:33:33.766987 4859 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-568bf4d7b5-ljcdc" event={"ID":"bc0261a4-dfa9-4356-9a9d-8a46a187a74c","Type":"ContainerDied","Data":"d725695d953ff6d725f64d0a36703d79675001b7ae193612af150d9372e72f20"} Oct 08 18:33:33 crc kubenswrapper[4859]: I1008 18:33:33.767014 4859 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-568bf4d7b5-ljcdc" event={"ID":"bc0261a4-dfa9-4356-9a9d-8a46a187a74c","Type":"ContainerDied","Data":"3ed08058125117480591800f7330fffb24051967b2ab5875ffcf15602430e2e6"} Oct 08 18:33:33 crc kubenswrapper[4859]: I1008 18:33:33.767029 4859 scope.go:117] "RemoveContainer" containerID="d725695d953ff6d725f64d0a36703d79675001b7ae193612af150d9372e72f20" Oct 08 18:33:33 crc kubenswrapper[4859]: I1008 18:33:33.767156 4859 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-568bf4d7b5-ljcdc" Oct 08 18:33:33 crc kubenswrapper[4859]: I1008 18:33:33.782621 4859 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-9mgdb\" (UniqueName: \"kubernetes.io/projected/21a3071c-7cf9-4008-a37b-ae90f125ddbc-kube-api-access-9mgdb\") pod \"glance-default-internal-api-0\" (UID: \"21a3071c-7cf9-4008-a37b-ae90f125ddbc\") " pod="openstack/glance-default-internal-api-0" Oct 08 18:33:33 crc kubenswrapper[4859]: I1008 18:33:33.787322 4859 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/87f0eecb-178a-438a-8643-e29c19390884-log-httpd\") pod \"ceilometer-0\" (UID: \"87f0eecb-178a-438a-8643-e29c19390884\") " pod="openstack/ceilometer-0" Oct 08 18:33:33 crc kubenswrapper[4859]: I1008 18:33:33.787374 4859 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/bfc391b0-6e05-41b9-8c78-7cf870f5cfe4-logs\") pod \"glance-default-external-api-0\" (UID: \"bfc391b0-6e05-41b9-8c78-7cf870f5cfe4\") " pod="openstack/glance-default-external-api-0" Oct 08 18:33:33 crc kubenswrapper[4859]: I1008 18:33:33.787392 4859 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/bfc391b0-6e05-41b9-8c78-7cf870f5cfe4-public-tls-certs\") pod \"glance-default-external-api-0\" (UID: \"bfc391b0-6e05-41b9-8c78-7cf870f5cfe4\") " pod="openstack/glance-default-external-api-0" Oct 08 18:33:33 crc kubenswrapper[4859]: I1008 18:33:33.787437 4859 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/87f0eecb-178a-438a-8643-e29c19390884-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"87f0eecb-178a-438a-8643-e29c19390884\") " pod="openstack/ceilometer-0" Oct 08 18:33:33 crc kubenswrapper[4859]: I1008 18:33:33.787466 4859 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/bfc391b0-6e05-41b9-8c78-7cf870f5cfe4-config-data\") pod \"glance-default-external-api-0\" (UID: \"bfc391b0-6e05-41b9-8c78-7cf870f5cfe4\") " pod="openstack/glance-default-external-api-0" Oct 08 18:33:33 crc kubenswrapper[4859]: I1008 18:33:33.787510 4859 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-4kfnf\" (UniqueName: \"kubernetes.io/projected/bfc391b0-6e05-41b9-8c78-7cf870f5cfe4-kube-api-access-4kfnf\") pod \"glance-default-external-api-0\" (UID: \"bfc391b0-6e05-41b9-8c78-7cf870f5cfe4\") " pod="openstack/glance-default-external-api-0" Oct 08 18:33:33 crc kubenswrapper[4859]: I1008 18:33:33.787535 4859 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/87f0eecb-178a-438a-8643-e29c19390884-run-httpd\") pod \"ceilometer-0\" (UID: \"87f0eecb-178a-438a-8643-e29c19390884\") " pod="openstack/ceilometer-0" Oct 08 18:33:33 crc kubenswrapper[4859]: I1008 18:33:33.787555 4859 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/87f0eecb-178a-438a-8643-e29c19390884-scripts\") pod \"ceilometer-0\" (UID: \"87f0eecb-178a-438a-8643-e29c19390884\") " pod="openstack/ceilometer-0" Oct 08 18:33:33 crc kubenswrapper[4859]: I1008 18:33:33.787574 4859 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/bfc391b0-6e05-41b9-8c78-7cf870f5cfe4-httpd-run\") pod \"glance-default-external-api-0\" (UID: \"bfc391b0-6e05-41b9-8c78-7cf870f5cfe4\") " pod="openstack/glance-default-external-api-0" Oct 08 18:33:33 crc kubenswrapper[4859]: I1008 18:33:33.787596 4859 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/87f0eecb-178a-438a-8643-e29c19390884-config-data\") pod \"ceilometer-0\" (UID: \"87f0eecb-178a-438a-8643-e29c19390884\") " pod="openstack/ceilometer-0" Oct 08 18:33:33 crc kubenswrapper[4859]: I1008 18:33:33.787613 4859 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-nx5wn\" (UniqueName: \"kubernetes.io/projected/87f0eecb-178a-438a-8643-e29c19390884-kube-api-access-nx5wn\") pod \"ceilometer-0\" (UID: \"87f0eecb-178a-438a-8643-e29c19390884\") " pod="openstack/ceilometer-0" Oct 08 18:33:33 crc kubenswrapper[4859]: I1008 18:33:33.787634 4859 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/bfc391b0-6e05-41b9-8c78-7cf870f5cfe4-scripts\") pod \"glance-default-external-api-0\" (UID: \"bfc391b0-6e05-41b9-8c78-7cf870f5cfe4\") " pod="openstack/glance-default-external-api-0" Oct 08 18:33:33 crc kubenswrapper[4859]: I1008 18:33:33.787661 4859 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage12-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage12-crc\") pod \"glance-default-external-api-0\" (UID: \"bfc391b0-6e05-41b9-8c78-7cf870f5cfe4\") " pod="openstack/glance-default-external-api-0" Oct 08 18:33:33 crc kubenswrapper[4859]: I1008 18:33:33.787678 4859 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/87f0eecb-178a-438a-8643-e29c19390884-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"87f0eecb-178a-438a-8643-e29c19390884\") " pod="openstack/ceilometer-0" Oct 08 18:33:33 crc kubenswrapper[4859]: I1008 18:33:33.787707 4859 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/bfc391b0-6e05-41b9-8c78-7cf870f5cfe4-combined-ca-bundle\") pod \"glance-default-external-api-0\" (UID: \"bfc391b0-6e05-41b9-8c78-7cf870f5cfe4\") " pod="openstack/glance-default-external-api-0" Oct 08 18:33:33 crc kubenswrapper[4859]: I1008 18:33:33.787760 4859 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-s92r8\" (UniqueName: \"kubernetes.io/projected/bc0261a4-dfa9-4356-9a9d-8a46a187a74c-kube-api-access-s92r8\") on node \"crc\" DevicePath \"\"" Oct 08 18:33:33 crc kubenswrapper[4859]: I1008 18:33:33.789364 4859 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/87f0eecb-178a-438a-8643-e29c19390884-run-httpd\") pod \"ceilometer-0\" (UID: \"87f0eecb-178a-438a-8643-e29c19390884\") " pod="openstack/ceilometer-0" Oct 08 18:33:33 crc kubenswrapper[4859]: I1008 18:33:33.789411 4859 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/87f0eecb-178a-438a-8643-e29c19390884-log-httpd\") pod \"ceilometer-0\" (UID: \"87f0eecb-178a-438a-8643-e29c19390884\") " pod="openstack/ceilometer-0" Oct 08 18:33:33 crc kubenswrapper[4859]: I1008 18:33:33.790842 4859 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage12-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage12-crc\") pod \"glance-default-external-api-0\" (UID: \"bfc391b0-6e05-41b9-8c78-7cf870f5cfe4\") device mount path \"/mnt/openstack/pv12\"" pod="openstack/glance-default-external-api-0" Oct 08 18:33:33 crc kubenswrapper[4859]: I1008 18:33:33.792262 4859 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/bfc391b0-6e05-41b9-8c78-7cf870f5cfe4-httpd-run\") pod \"glance-default-external-api-0\" (UID: \"bfc391b0-6e05-41b9-8c78-7cf870f5cfe4\") " pod="openstack/glance-default-external-api-0" Oct 08 18:33:33 crc kubenswrapper[4859]: I1008 18:33:33.802227 4859 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/bfc391b0-6e05-41b9-8c78-7cf870f5cfe4-logs\") pod \"glance-default-external-api-0\" (UID: \"bfc391b0-6e05-41b9-8c78-7cf870f5cfe4\") " pod="openstack/glance-default-external-api-0" Oct 08 18:33:33 crc kubenswrapper[4859]: I1008 18:33:33.841735 4859 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/87f0eecb-178a-438a-8643-e29c19390884-scripts\") pod \"ceilometer-0\" (UID: \"87f0eecb-178a-438a-8643-e29c19390884\") " pod="openstack/ceilometer-0" Oct 08 18:33:33 crc kubenswrapper[4859]: I1008 18:33:33.842633 4859 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/21a3071c-7cf9-4008-a37b-ae90f125ddbc-scripts\") pod \"glance-default-internal-api-0\" (UID: \"21a3071c-7cf9-4008-a37b-ae90f125ddbc\") " pod="openstack/glance-default-internal-api-0" Oct 08 18:33:33 crc kubenswrapper[4859]: I1008 18:33:33.848359 4859 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/87f0eecb-178a-438a-8643-e29c19390884-config-data\") pod \"ceilometer-0\" (UID: \"87f0eecb-178a-438a-8643-e29c19390884\") " pod="openstack/ceilometer-0" Oct 08 18:33:33 crc kubenswrapper[4859]: I1008 18:33:33.851929 4859 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/bfc391b0-6e05-41b9-8c78-7cf870f5cfe4-combined-ca-bundle\") pod \"glance-default-external-api-0\" (UID: \"bfc391b0-6e05-41b9-8c78-7cf870f5cfe4\") " pod="openstack/glance-default-external-api-0" Oct 08 18:33:33 crc kubenswrapper[4859]: I1008 18:33:33.869410 4859 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/87f0eecb-178a-438a-8643-e29c19390884-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"87f0eecb-178a-438a-8643-e29c19390884\") " pod="openstack/ceilometer-0" Oct 08 18:33:33 crc kubenswrapper[4859]: I1008 18:33:33.876994 4859 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/bfc391b0-6e05-41b9-8c78-7cf870f5cfe4-scripts\") pod \"glance-default-external-api-0\" (UID: \"bfc391b0-6e05-41b9-8c78-7cf870f5cfe4\") " pod="openstack/glance-default-external-api-0" Oct 08 18:33:33 crc kubenswrapper[4859]: I1008 18:33:33.885039 4859 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/87f0eecb-178a-438a-8643-e29c19390884-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"87f0eecb-178a-438a-8643-e29c19390884\") " pod="openstack/ceilometer-0" Oct 08 18:33:33 crc kubenswrapper[4859]: I1008 18:33:33.888542 4859 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-nx5wn\" (UniqueName: \"kubernetes.io/projected/87f0eecb-178a-438a-8643-e29c19390884-kube-api-access-nx5wn\") pod \"ceilometer-0\" (UID: \"87f0eecb-178a-438a-8643-e29c19390884\") " pod="openstack/ceilometer-0" Oct 08 18:33:33 crc kubenswrapper[4859]: I1008 18:33:33.889453 4859 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-4kfnf\" (UniqueName: \"kubernetes.io/projected/bfc391b0-6e05-41b9-8c78-7cf870f5cfe4-kube-api-access-4kfnf\") pod \"glance-default-external-api-0\" (UID: \"bfc391b0-6e05-41b9-8c78-7cf870f5cfe4\") " pod="openstack/glance-default-external-api-0" Oct 08 18:33:33 crc kubenswrapper[4859]: I1008 18:33:33.898112 4859 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/bfc391b0-6e05-41b9-8c78-7cf870f5cfe4-public-tls-certs\") pod \"glance-default-external-api-0\" (UID: \"bfc391b0-6e05-41b9-8c78-7cf870f5cfe4\") " pod="openstack/glance-default-external-api-0" Oct 08 18:33:33 crc kubenswrapper[4859]: I1008 18:33:33.898804 4859 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/bfc391b0-6e05-41b9-8c78-7cf870f5cfe4-config-data\") pod \"glance-default-external-api-0\" (UID: \"bfc391b0-6e05-41b9-8c78-7cf870f5cfe4\") " pod="openstack/glance-default-external-api-0" Oct 08 18:33:33 crc kubenswrapper[4859]: I1008 18:33:33.909804 4859 scope.go:117] "RemoveContainer" containerID="0dd53ffc93d496397b8e5cca4560b706d2969d4371b2e906faed9663a825e98d" Oct 08 18:33:33 crc kubenswrapper[4859]: I1008 18:33:33.921887 4859 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Oct 08 18:33:34 crc kubenswrapper[4859]: I1008 18:33:34.008463 4859 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/bc0261a4-dfa9-4356-9a9d-8a46a187a74c-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "bc0261a4-dfa9-4356-9a9d-8a46a187a74c" (UID: "bc0261a4-dfa9-4356-9a9d-8a46a187a74c"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 08 18:33:34 crc kubenswrapper[4859]: I1008 18:33:34.021483 4859 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage12-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage12-crc\") pod \"glance-default-external-api-0\" (UID: \"bfc391b0-6e05-41b9-8c78-7cf870f5cfe4\") " pod="openstack/glance-default-external-api-0" Oct 08 18:33:34 crc kubenswrapper[4859]: I1008 18:33:34.039156 4859 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage10-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage10-crc\") pod \"glance-default-internal-api-0\" (UID: \"21a3071c-7cf9-4008-a37b-ae90f125ddbc\") " pod="openstack/glance-default-internal-api-0" Oct 08 18:33:34 crc kubenswrapper[4859]: I1008 18:33:34.040697 4859 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/bc0261a4-dfa9-4356-9a9d-8a46a187a74c-dns-swift-storage-0" (OuterVolumeSpecName: "dns-swift-storage-0") pod "bc0261a4-dfa9-4356-9a9d-8a46a187a74c" (UID: "bc0261a4-dfa9-4356-9a9d-8a46a187a74c"). InnerVolumeSpecName "dns-swift-storage-0". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 08 18:33:34 crc kubenswrapper[4859]: I1008 18:33:34.044439 4859 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/bc0261a4-dfa9-4356-9a9d-8a46a187a74c-config" (OuterVolumeSpecName: "config") pod "bc0261a4-dfa9-4356-9a9d-8a46a187a74c" (UID: "bc0261a4-dfa9-4356-9a9d-8a46a187a74c"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 08 18:33:34 crc kubenswrapper[4859]: I1008 18:33:34.066785 4859 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-internal-api-0" Oct 08 18:33:34 crc kubenswrapper[4859]: I1008 18:33:34.068523 4859 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/bc0261a4-dfa9-4356-9a9d-8a46a187a74c-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "bc0261a4-dfa9-4356-9a9d-8a46a187a74c" (UID: "bc0261a4-dfa9-4356-9a9d-8a46a187a74c"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 08 18:33:34 crc kubenswrapper[4859]: I1008 18:33:34.070987 4859 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/bc0261a4-dfa9-4356-9a9d-8a46a187a74c-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "bc0261a4-dfa9-4356-9a9d-8a46a187a74c" (UID: "bc0261a4-dfa9-4356-9a9d-8a46a187a74c"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 08 18:33:34 crc kubenswrapper[4859]: I1008 18:33:34.100681 4859 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/bc0261a4-dfa9-4356-9a9d-8a46a187a74c-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Oct 08 18:33:34 crc kubenswrapper[4859]: I1008 18:33:34.100738 4859 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/bc0261a4-dfa9-4356-9a9d-8a46a187a74c-dns-svc\") on node \"crc\" DevicePath \"\"" Oct 08 18:33:34 crc kubenswrapper[4859]: I1008 18:33:34.100750 4859 reconciler_common.go:293] "Volume detached for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/bc0261a4-dfa9-4356-9a9d-8a46a187a74c-dns-swift-storage-0\") on node \"crc\" DevicePath \"\"" Oct 08 18:33:34 crc kubenswrapper[4859]: I1008 18:33:34.100763 4859 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/bc0261a4-dfa9-4356-9a9d-8a46a187a74c-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Oct 08 18:33:34 crc kubenswrapper[4859]: I1008 18:33:34.100771 4859 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/bc0261a4-dfa9-4356-9a9d-8a46a187a74c-config\") on node \"crc\" DevicePath \"\"" Oct 08 18:33:34 crc kubenswrapper[4859]: I1008 18:33:34.113508 4859 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-59df9f9d49-446ss"] Oct 08 18:33:34 crc kubenswrapper[4859]: I1008 18:33:34.181797 4859 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-external-api-0" Oct 08 18:33:34 crc kubenswrapper[4859]: I1008 18:33:34.283896 4859 scope.go:117] "RemoveContainer" containerID="d725695d953ff6d725f64d0a36703d79675001b7ae193612af150d9372e72f20" Oct 08 18:33:34 crc kubenswrapper[4859]: E1008 18:33:34.286359 4859 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"d725695d953ff6d725f64d0a36703d79675001b7ae193612af150d9372e72f20\": container with ID starting with d725695d953ff6d725f64d0a36703d79675001b7ae193612af150d9372e72f20 not found: ID does not exist" containerID="d725695d953ff6d725f64d0a36703d79675001b7ae193612af150d9372e72f20" Oct 08 18:33:34 crc kubenswrapper[4859]: I1008 18:33:34.286399 4859 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"d725695d953ff6d725f64d0a36703d79675001b7ae193612af150d9372e72f20"} err="failed to get container status \"d725695d953ff6d725f64d0a36703d79675001b7ae193612af150d9372e72f20\": rpc error: code = NotFound desc = could not find container \"d725695d953ff6d725f64d0a36703d79675001b7ae193612af150d9372e72f20\": container with ID starting with d725695d953ff6d725f64d0a36703d79675001b7ae193612af150d9372e72f20 not found: ID does not exist" Oct 08 18:33:34 crc kubenswrapper[4859]: I1008 18:33:34.286419 4859 scope.go:117] "RemoveContainer" containerID="0dd53ffc93d496397b8e5cca4560b706d2969d4371b2e906faed9663a825e98d" Oct 08 18:33:34 crc kubenswrapper[4859]: E1008 18:33:34.287141 4859 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"0dd53ffc93d496397b8e5cca4560b706d2969d4371b2e906faed9663a825e98d\": container with ID starting with 0dd53ffc93d496397b8e5cca4560b706d2969d4371b2e906faed9663a825e98d not found: ID does not exist" containerID="0dd53ffc93d496397b8e5cca4560b706d2969d4371b2e906faed9663a825e98d" Oct 08 18:33:34 crc kubenswrapper[4859]: I1008 18:33:34.287169 4859 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"0dd53ffc93d496397b8e5cca4560b706d2969d4371b2e906faed9663a825e98d"} err="failed to get container status \"0dd53ffc93d496397b8e5cca4560b706d2969d4371b2e906faed9663a825e98d\": rpc error: code = NotFound desc = could not find container \"0dd53ffc93d496397b8e5cca4560b706d2969d4371b2e906faed9663a825e98d\": container with ID starting with 0dd53ffc93d496397b8e5cca4560b706d2969d4371b2e906faed9663a825e98d not found: ID does not exist" Oct 08 18:33:34 crc kubenswrapper[4859]: I1008 18:33:34.305820 4859 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-bootstrap-mqwmv"] Oct 08 18:33:34 crc kubenswrapper[4859]: W1008 18:33:34.318306 4859 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod9d1b6fef_16a5_4c1b_a3bb_95f46f1d92f7.slice/crio-082ec68b2336d53fa12edeb6e2e6896b2dfcb72c1887ba0fb688c69dc1e894f3 WatchSource:0}: Error finding container 082ec68b2336d53fa12edeb6e2e6896b2dfcb72c1887ba0fb688c69dc1e894f3: Status 404 returned error can't find the container with id 082ec68b2336d53fa12edeb6e2e6896b2dfcb72c1887ba0fb688c69dc1e894f3 Oct 08 18:33:34 crc kubenswrapper[4859]: I1008 18:33:34.420186 4859 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-568bf4d7b5-ljcdc"] Oct 08 18:33:34 crc kubenswrapper[4859]: I1008 18:33:34.441310 4859 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-568bf4d7b5-ljcdc"] Oct 08 18:33:34 crc kubenswrapper[4859]: I1008 18:33:34.491178 4859 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="bc0261a4-dfa9-4356-9a9d-8a46a187a74c" path="/var/lib/kubelet/pods/bc0261a4-dfa9-4356-9a9d-8a46a187a74c/volumes" Oct 08 18:33:34 crc kubenswrapper[4859]: I1008 18:33:34.570763 4859 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/horizon-7444cc4f69-5xs5x"] Oct 08 18:33:34 crc kubenswrapper[4859]: W1008 18:33:34.719508 4859 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-poded95b639_b2c8_4879_bedb_16681e0166cf.slice/crio-df7e476e592572851bdab7449f59a5936c9d9eb10ebb99616a2965aeba03e560 WatchSource:0}: Error finding container df7e476e592572851bdab7449f59a5936c9d9eb10ebb99616a2965aeba03e560: Status 404 returned error can't find the container with id df7e476e592572851bdab7449f59a5936c9d9eb10ebb99616a2965aeba03e560 Oct 08 18:33:34 crc kubenswrapper[4859]: I1008 18:33:34.720416 4859 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/placement-db-sync-lsbrj"] Oct 08 18:33:34 crc kubenswrapper[4859]: I1008 18:33:34.732842 4859 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-74776f5dd7-89sbw"] Oct 08 18:33:34 crc kubenswrapper[4859]: I1008 18:33:34.738988 4859 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/horizon-86b9b87697-gmwqv"] Oct 08 18:33:34 crc kubenswrapper[4859]: W1008 18:33:34.742369 4859 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod1acee243_a101_4dc6_a599_aaee4821dbfa.slice/crio-e34d4711b5a7eed7ee113c9435cd61a30a628c24c4ea192e6eecf5c18f11bcad WatchSource:0}: Error finding container e34d4711b5a7eed7ee113c9435cd61a30a628c24c4ea192e6eecf5c18f11bcad: Status 404 returned error can't find the container with id e34d4711b5a7eed7ee113c9435cd61a30a628c24c4ea192e6eecf5c18f11bcad Oct 08 18:33:34 crc kubenswrapper[4859]: I1008 18:33:34.782748 4859 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-db-sync-lsbrj" event={"ID":"ed95b639-b2c8-4879-bedb-16681e0166cf","Type":"ContainerStarted","Data":"df7e476e592572851bdab7449f59a5936c9d9eb10ebb99616a2965aeba03e560"} Oct 08 18:33:34 crc kubenswrapper[4859]: I1008 18:33:34.790240 4859 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-59df9f9d49-446ss" event={"ID":"59231da1-64f0-4cac-a821-20fcbdcd5e30","Type":"ContainerDied","Data":"b7a3ebb4c349e8595427019f6888747ec30f80e5647bd49f80b85043adb827ba"} Oct 08 18:33:34 crc kubenswrapper[4859]: I1008 18:33:34.790265 4859 generic.go:334] "Generic (PLEG): container finished" podID="59231da1-64f0-4cac-a821-20fcbdcd5e30" containerID="b7a3ebb4c349e8595427019f6888747ec30f80e5647bd49f80b85043adb827ba" exitCode=0 Oct 08 18:33:34 crc kubenswrapper[4859]: I1008 18:33:34.790303 4859 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-59df9f9d49-446ss" event={"ID":"59231da1-64f0-4cac-a821-20fcbdcd5e30","Type":"ContainerStarted","Data":"6e3f208043bc7132b17baaa6c02478587f5d58307c3346d288f5143187a1c333"} Oct 08 18:33:34 crc kubenswrapper[4859]: I1008 18:33:34.793433 4859 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-7444cc4f69-5xs5x" event={"ID":"ac9272b3-433e-4afb-afca-0d647a9d9be5","Type":"ContainerStarted","Data":"36c78cc9e0cd813d0049f15532a403a7bafccd0c34aa89a7ce96d714b689cb2a"} Oct 08 18:33:34 crc kubenswrapper[4859]: I1008 18:33:34.801244 4859 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-74776f5dd7-89sbw" event={"ID":"f08c21e6-a74c-4ad9-a3fa-4e56ff2c6432","Type":"ContainerStarted","Data":"4a13491c7c60974dd907317f0881466a73694103f0829dc45b85c65b1f0e8bdb"} Oct 08 18:33:34 crc kubenswrapper[4859]: I1008 18:33:34.808337 4859 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-86b9b87697-gmwqv" event={"ID":"1acee243-a101-4dc6-a599-aaee4821dbfa","Type":"ContainerStarted","Data":"e34d4711b5a7eed7ee113c9435cd61a30a628c24c4ea192e6eecf5c18f11bcad"} Oct 08 18:33:34 crc kubenswrapper[4859]: I1008 18:33:34.826899 4859 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-bootstrap-mqwmv" event={"ID":"9d1b6fef-16a5-4c1b-a3bb-95f46f1d92f7","Type":"ContainerStarted","Data":"e48d82fde98292fc3ac341cff6cc00d37119be8b99b5e5c33a5f112eb3f15c4a"} Oct 08 18:33:34 crc kubenswrapper[4859]: I1008 18:33:34.826954 4859 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-bootstrap-mqwmv" event={"ID":"9d1b6fef-16a5-4c1b-a3bb-95f46f1d92f7","Type":"ContainerStarted","Data":"082ec68b2336d53fa12edeb6e2e6896b2dfcb72c1887ba0fb688c69dc1e894f3"} Oct 08 18:33:34 crc kubenswrapper[4859]: I1008 18:33:34.848596 4859 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/keystone-bootstrap-mqwmv" podStartSLOduration=2.848577832 podStartE2EDuration="2.848577832s" podCreationTimestamp="2025-10-08 18:33:32 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-08 18:33:34.842443741 +0000 UTC m=+985.089283140" watchObservedRunningTime="2025-10-08 18:33:34.848577832 +0000 UTC m=+985.095417211" Oct 08 18:33:34 crc kubenswrapper[4859]: I1008 18:33:34.874811 4859 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Oct 08 18:33:34 crc kubenswrapper[4859]: I1008 18:33:34.975372 4859 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-internal-api-0"] Oct 08 18:33:36 crc kubenswrapper[4859]: I1008 18:33:35.094613 4859 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-external-api-0"] Oct 08 18:33:36 crc kubenswrapper[4859]: I1008 18:33:35.251482 4859 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-59df9f9d49-446ss" Oct 08 18:33:36 crc kubenswrapper[4859]: I1008 18:33:35.296496 4859 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-default-external-api-0"] Oct 08 18:33:36 crc kubenswrapper[4859]: I1008 18:33:35.337432 4859 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/horizon-7444cc4f69-5xs5x"] Oct 08 18:33:36 crc kubenswrapper[4859]: I1008 18:33:35.426769 4859 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/horizon-85b86687bf-6gp89"] Oct 08 18:33:36 crc kubenswrapper[4859]: E1008 18:33:35.441362 4859 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="59231da1-64f0-4cac-a821-20fcbdcd5e30" containerName="init" Oct 08 18:33:36 crc kubenswrapper[4859]: I1008 18:33:35.441399 4859 state_mem.go:107] "Deleted CPUSet assignment" podUID="59231da1-64f0-4cac-a821-20fcbdcd5e30" containerName="init" Oct 08 18:33:36 crc kubenswrapper[4859]: I1008 18:33:35.442450 4859 memory_manager.go:354] "RemoveStaleState removing state" podUID="59231da1-64f0-4cac-a821-20fcbdcd5e30" containerName="init" Oct 08 18:33:36 crc kubenswrapper[4859]: I1008 18:33:35.447083 4859 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-lbgjh\" (UniqueName: \"kubernetes.io/projected/59231da1-64f0-4cac-a821-20fcbdcd5e30-kube-api-access-lbgjh\") pod \"59231da1-64f0-4cac-a821-20fcbdcd5e30\" (UID: \"59231da1-64f0-4cac-a821-20fcbdcd5e30\") " Oct 08 18:33:36 crc kubenswrapper[4859]: I1008 18:33:35.447168 4859 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/59231da1-64f0-4cac-a821-20fcbdcd5e30-ovsdbserver-sb\") pod \"59231da1-64f0-4cac-a821-20fcbdcd5e30\" (UID: \"59231da1-64f0-4cac-a821-20fcbdcd5e30\") " Oct 08 18:33:36 crc kubenswrapper[4859]: I1008 18:33:35.447258 4859 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/59231da1-64f0-4cac-a821-20fcbdcd5e30-ovsdbserver-nb\") pod \"59231da1-64f0-4cac-a821-20fcbdcd5e30\" (UID: \"59231da1-64f0-4cac-a821-20fcbdcd5e30\") " Oct 08 18:33:36 crc kubenswrapper[4859]: I1008 18:33:35.447342 4859 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/59231da1-64f0-4cac-a821-20fcbdcd5e30-config\") pod \"59231da1-64f0-4cac-a821-20fcbdcd5e30\" (UID: \"59231da1-64f0-4cac-a821-20fcbdcd5e30\") " Oct 08 18:33:36 crc kubenswrapper[4859]: I1008 18:33:35.447416 4859 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/59231da1-64f0-4cac-a821-20fcbdcd5e30-dns-svc\") pod \"59231da1-64f0-4cac-a821-20fcbdcd5e30\" (UID: \"59231da1-64f0-4cac-a821-20fcbdcd5e30\") " Oct 08 18:33:36 crc kubenswrapper[4859]: I1008 18:33:35.447464 4859 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/59231da1-64f0-4cac-a821-20fcbdcd5e30-dns-swift-storage-0\") pod \"59231da1-64f0-4cac-a821-20fcbdcd5e30\" (UID: \"59231da1-64f0-4cac-a821-20fcbdcd5e30\") " Oct 08 18:33:36 crc kubenswrapper[4859]: I1008 18:33:35.447847 4859 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/horizon-85b86687bf-6gp89"] Oct 08 18:33:36 crc kubenswrapper[4859]: I1008 18:33:35.447936 4859 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-85b86687bf-6gp89" Oct 08 18:33:36 crc kubenswrapper[4859]: I1008 18:33:35.453383 4859 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-default-internal-api-0"] Oct 08 18:33:36 crc kubenswrapper[4859]: I1008 18:33:35.460540 4859 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/59231da1-64f0-4cac-a821-20fcbdcd5e30-kube-api-access-lbgjh" (OuterVolumeSpecName: "kube-api-access-lbgjh") pod "59231da1-64f0-4cac-a821-20fcbdcd5e30" (UID: "59231da1-64f0-4cac-a821-20fcbdcd5e30"). InnerVolumeSpecName "kube-api-access-lbgjh". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 08 18:33:36 crc kubenswrapper[4859]: I1008 18:33:35.462014 4859 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Oct 08 18:33:36 crc kubenswrapper[4859]: I1008 18:33:35.496023 4859 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/59231da1-64f0-4cac-a821-20fcbdcd5e30-config" (OuterVolumeSpecName: "config") pod "59231da1-64f0-4cac-a821-20fcbdcd5e30" (UID: "59231da1-64f0-4cac-a821-20fcbdcd5e30"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 08 18:33:36 crc kubenswrapper[4859]: I1008 18:33:35.501501 4859 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/59231da1-64f0-4cac-a821-20fcbdcd5e30-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "59231da1-64f0-4cac-a821-20fcbdcd5e30" (UID: "59231da1-64f0-4cac-a821-20fcbdcd5e30"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 08 18:33:36 crc kubenswrapper[4859]: I1008 18:33:35.501588 4859 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/59231da1-64f0-4cac-a821-20fcbdcd5e30-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "59231da1-64f0-4cac-a821-20fcbdcd5e30" (UID: "59231da1-64f0-4cac-a821-20fcbdcd5e30"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 08 18:33:36 crc kubenswrapper[4859]: I1008 18:33:35.515047 4859 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/59231da1-64f0-4cac-a821-20fcbdcd5e30-dns-swift-storage-0" (OuterVolumeSpecName: "dns-swift-storage-0") pod "59231da1-64f0-4cac-a821-20fcbdcd5e30" (UID: "59231da1-64f0-4cac-a821-20fcbdcd5e30"). InnerVolumeSpecName "dns-swift-storage-0". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 08 18:33:36 crc kubenswrapper[4859]: I1008 18:33:35.541315 4859 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/59231da1-64f0-4cac-a821-20fcbdcd5e30-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "59231da1-64f0-4cac-a821-20fcbdcd5e30" (UID: "59231da1-64f0-4cac-a821-20fcbdcd5e30"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 08 18:33:36 crc kubenswrapper[4859]: I1008 18:33:35.551385 4859 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/9ed85886-9589-4d26-ad59-7a6021df659e-config-data\") pod \"horizon-85b86687bf-6gp89\" (UID: \"9ed85886-9589-4d26-ad59-7a6021df659e\") " pod="openstack/horizon-85b86687bf-6gp89" Oct 08 18:33:36 crc kubenswrapper[4859]: I1008 18:33:35.551485 4859 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/9ed85886-9589-4d26-ad59-7a6021df659e-scripts\") pod \"horizon-85b86687bf-6gp89\" (UID: \"9ed85886-9589-4d26-ad59-7a6021df659e\") " pod="openstack/horizon-85b86687bf-6gp89" Oct 08 18:33:36 crc kubenswrapper[4859]: I1008 18:33:35.552037 4859 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/9ed85886-9589-4d26-ad59-7a6021df659e-logs\") pod \"horizon-85b86687bf-6gp89\" (UID: \"9ed85886-9589-4d26-ad59-7a6021df659e\") " pod="openstack/horizon-85b86687bf-6gp89" Oct 08 18:33:36 crc kubenswrapper[4859]: I1008 18:33:35.552543 4859 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/9ed85886-9589-4d26-ad59-7a6021df659e-horizon-secret-key\") pod \"horizon-85b86687bf-6gp89\" (UID: \"9ed85886-9589-4d26-ad59-7a6021df659e\") " pod="openstack/horizon-85b86687bf-6gp89" Oct 08 18:33:36 crc kubenswrapper[4859]: I1008 18:33:35.552633 4859 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-vvdq9\" (UniqueName: \"kubernetes.io/projected/9ed85886-9589-4d26-ad59-7a6021df659e-kube-api-access-vvdq9\") pod \"horizon-85b86687bf-6gp89\" (UID: \"9ed85886-9589-4d26-ad59-7a6021df659e\") " pod="openstack/horizon-85b86687bf-6gp89" Oct 08 18:33:36 crc kubenswrapper[4859]: I1008 18:33:35.553734 4859 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/59231da1-64f0-4cac-a821-20fcbdcd5e30-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Oct 08 18:33:36 crc kubenswrapper[4859]: I1008 18:33:35.554456 4859 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/59231da1-64f0-4cac-a821-20fcbdcd5e30-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Oct 08 18:33:36 crc kubenswrapper[4859]: I1008 18:33:35.554480 4859 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/59231da1-64f0-4cac-a821-20fcbdcd5e30-config\") on node \"crc\" DevicePath \"\"" Oct 08 18:33:36 crc kubenswrapper[4859]: I1008 18:33:35.554492 4859 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/59231da1-64f0-4cac-a821-20fcbdcd5e30-dns-svc\") on node \"crc\" DevicePath \"\"" Oct 08 18:33:36 crc kubenswrapper[4859]: I1008 18:33:35.554528 4859 reconciler_common.go:293] "Volume detached for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/59231da1-64f0-4cac-a821-20fcbdcd5e30-dns-swift-storage-0\") on node \"crc\" DevicePath \"\"" Oct 08 18:33:36 crc kubenswrapper[4859]: I1008 18:33:35.554541 4859 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-lbgjh\" (UniqueName: \"kubernetes.io/projected/59231da1-64f0-4cac-a821-20fcbdcd5e30-kube-api-access-lbgjh\") on node \"crc\" DevicePath \"\"" Oct 08 18:33:36 crc kubenswrapper[4859]: I1008 18:33:35.655920 4859 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/9ed85886-9589-4d26-ad59-7a6021df659e-logs\") pod \"horizon-85b86687bf-6gp89\" (UID: \"9ed85886-9589-4d26-ad59-7a6021df659e\") " pod="openstack/horizon-85b86687bf-6gp89" Oct 08 18:33:36 crc kubenswrapper[4859]: I1008 18:33:35.655987 4859 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/9ed85886-9589-4d26-ad59-7a6021df659e-horizon-secret-key\") pod \"horizon-85b86687bf-6gp89\" (UID: \"9ed85886-9589-4d26-ad59-7a6021df659e\") " pod="openstack/horizon-85b86687bf-6gp89" Oct 08 18:33:36 crc kubenswrapper[4859]: I1008 18:33:35.656162 4859 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-vvdq9\" (UniqueName: \"kubernetes.io/projected/9ed85886-9589-4d26-ad59-7a6021df659e-kube-api-access-vvdq9\") pod \"horizon-85b86687bf-6gp89\" (UID: \"9ed85886-9589-4d26-ad59-7a6021df659e\") " pod="openstack/horizon-85b86687bf-6gp89" Oct 08 18:33:36 crc kubenswrapper[4859]: I1008 18:33:35.656305 4859 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/9ed85886-9589-4d26-ad59-7a6021df659e-logs\") pod \"horizon-85b86687bf-6gp89\" (UID: \"9ed85886-9589-4d26-ad59-7a6021df659e\") " pod="openstack/horizon-85b86687bf-6gp89" Oct 08 18:33:36 crc kubenswrapper[4859]: I1008 18:33:35.656380 4859 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/9ed85886-9589-4d26-ad59-7a6021df659e-config-data\") pod \"horizon-85b86687bf-6gp89\" (UID: \"9ed85886-9589-4d26-ad59-7a6021df659e\") " pod="openstack/horizon-85b86687bf-6gp89" Oct 08 18:33:36 crc kubenswrapper[4859]: I1008 18:33:35.656441 4859 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/9ed85886-9589-4d26-ad59-7a6021df659e-scripts\") pod \"horizon-85b86687bf-6gp89\" (UID: \"9ed85886-9589-4d26-ad59-7a6021df659e\") " pod="openstack/horizon-85b86687bf-6gp89" Oct 08 18:33:36 crc kubenswrapper[4859]: I1008 18:33:35.657169 4859 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/9ed85886-9589-4d26-ad59-7a6021df659e-scripts\") pod \"horizon-85b86687bf-6gp89\" (UID: \"9ed85886-9589-4d26-ad59-7a6021df659e\") " pod="openstack/horizon-85b86687bf-6gp89" Oct 08 18:33:36 crc kubenswrapper[4859]: I1008 18:33:35.658233 4859 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/9ed85886-9589-4d26-ad59-7a6021df659e-config-data\") pod \"horizon-85b86687bf-6gp89\" (UID: \"9ed85886-9589-4d26-ad59-7a6021df659e\") " pod="openstack/horizon-85b86687bf-6gp89" Oct 08 18:33:36 crc kubenswrapper[4859]: I1008 18:33:35.659904 4859 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/9ed85886-9589-4d26-ad59-7a6021df659e-horizon-secret-key\") pod \"horizon-85b86687bf-6gp89\" (UID: \"9ed85886-9589-4d26-ad59-7a6021df659e\") " pod="openstack/horizon-85b86687bf-6gp89" Oct 08 18:33:36 crc kubenswrapper[4859]: I1008 18:33:35.686596 4859 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-vvdq9\" (UniqueName: \"kubernetes.io/projected/9ed85886-9589-4d26-ad59-7a6021df659e-kube-api-access-vvdq9\") pod \"horizon-85b86687bf-6gp89\" (UID: \"9ed85886-9589-4d26-ad59-7a6021df659e\") " pod="openstack/horizon-85b86687bf-6gp89" Oct 08 18:33:36 crc kubenswrapper[4859]: I1008 18:33:35.809511 4859 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/cinder-db-sync-jqwrh"] Oct 08 18:33:36 crc kubenswrapper[4859]: I1008 18:33:35.810846 4859 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-db-sync-jqwrh" Oct 08 18:33:36 crc kubenswrapper[4859]: I1008 18:33:35.860132 4859 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-config-data" Oct 08 18:33:36 crc kubenswrapper[4859]: I1008 18:33:35.860644 4859 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-cinder-dockercfg-gldt6" Oct 08 18:33:36 crc kubenswrapper[4859]: I1008 18:33:35.860768 4859 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-scripts" Oct 08 18:33:36 crc kubenswrapper[4859]: I1008 18:33:35.875855 4859 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-85b86687bf-6gp89" Oct 08 18:33:36 crc kubenswrapper[4859]: I1008 18:33:35.897160 4859 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"bfc391b0-6e05-41b9-8c78-7cf870f5cfe4","Type":"ContainerStarted","Data":"2ad247c3a483c8e2e59d1bc0aa8d4df493b21a6998eb9b1922b243053a801353"} Oct 08 18:33:36 crc kubenswrapper[4859]: I1008 18:33:35.899917 4859 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"21a3071c-7cf9-4008-a37b-ae90f125ddbc","Type":"ContainerStarted","Data":"f6b7101c206d49941deb736f964625f6291e2b703f2978608c79d5f676a49527"} Oct 08 18:33:36 crc kubenswrapper[4859]: I1008 18:33:35.904924 4859 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-59df9f9d49-446ss" Oct 08 18:33:36 crc kubenswrapper[4859]: I1008 18:33:35.905321 4859 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-59df9f9d49-446ss" event={"ID":"59231da1-64f0-4cac-a821-20fcbdcd5e30","Type":"ContainerDied","Data":"6e3f208043bc7132b17baaa6c02478587f5d58307c3346d288f5143187a1c333"} Oct 08 18:33:36 crc kubenswrapper[4859]: I1008 18:33:35.905373 4859 scope.go:117] "RemoveContainer" containerID="b7a3ebb4c349e8595427019f6888747ec30f80e5647bd49f80b85043adb827ba" Oct 08 18:33:36 crc kubenswrapper[4859]: I1008 18:33:35.911159 4859 generic.go:334] "Generic (PLEG): container finished" podID="f08c21e6-a74c-4ad9-a3fa-4e56ff2c6432" containerID="f0557242a2e57c4c6dfb47d6597787e847f098ef8c3c670c0d077683329bae49" exitCode=0 Oct 08 18:33:36 crc kubenswrapper[4859]: I1008 18:33:35.911219 4859 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-74776f5dd7-89sbw" event={"ID":"f08c21e6-a74c-4ad9-a3fa-4e56ff2c6432","Type":"ContainerDied","Data":"f0557242a2e57c4c6dfb47d6597787e847f098ef8c3c670c0d077683329bae49"} Oct 08 18:33:36 crc kubenswrapper[4859]: I1008 18:33:35.914363 4859 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-db-sync-jqwrh"] Oct 08 18:33:36 crc kubenswrapper[4859]: I1008 18:33:35.933921 4859 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"87f0eecb-178a-438a-8643-e29c19390884","Type":"ContainerStarted","Data":"78f9e1079ba0c58f7bdefd70eb6b6101a88872c51f57d7f91919f7b8db0bdcb2"} Oct 08 18:33:36 crc kubenswrapper[4859]: I1008 18:33:35.968607 4859 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-bdbmw\" (UniqueName: \"kubernetes.io/projected/720db6c0-f18f-429b-9589-fa503d212139-kube-api-access-bdbmw\") pod \"cinder-db-sync-jqwrh\" (UID: \"720db6c0-f18f-429b-9589-fa503d212139\") " pod="openstack/cinder-db-sync-jqwrh" Oct 08 18:33:36 crc kubenswrapper[4859]: I1008 18:33:35.968735 4859 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/720db6c0-f18f-429b-9589-fa503d212139-db-sync-config-data\") pod \"cinder-db-sync-jqwrh\" (UID: \"720db6c0-f18f-429b-9589-fa503d212139\") " pod="openstack/cinder-db-sync-jqwrh" Oct 08 18:33:36 crc kubenswrapper[4859]: I1008 18:33:35.968800 4859 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/720db6c0-f18f-429b-9589-fa503d212139-etc-machine-id\") pod \"cinder-db-sync-jqwrh\" (UID: \"720db6c0-f18f-429b-9589-fa503d212139\") " pod="openstack/cinder-db-sync-jqwrh" Oct 08 18:33:36 crc kubenswrapper[4859]: I1008 18:33:35.968962 4859 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/720db6c0-f18f-429b-9589-fa503d212139-scripts\") pod \"cinder-db-sync-jqwrh\" (UID: \"720db6c0-f18f-429b-9589-fa503d212139\") " pod="openstack/cinder-db-sync-jqwrh" Oct 08 18:33:36 crc kubenswrapper[4859]: I1008 18:33:35.968996 4859 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/720db6c0-f18f-429b-9589-fa503d212139-combined-ca-bundle\") pod \"cinder-db-sync-jqwrh\" (UID: \"720db6c0-f18f-429b-9589-fa503d212139\") " pod="openstack/cinder-db-sync-jqwrh" Oct 08 18:33:36 crc kubenswrapper[4859]: I1008 18:33:35.969020 4859 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/720db6c0-f18f-429b-9589-fa503d212139-config-data\") pod \"cinder-db-sync-jqwrh\" (UID: \"720db6c0-f18f-429b-9589-fa503d212139\") " pod="openstack/cinder-db-sync-jqwrh" Oct 08 18:33:36 crc kubenswrapper[4859]: I1008 18:33:36.035942 4859 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/barbican-db-sync-x99jc"] Oct 08 18:33:36 crc kubenswrapper[4859]: I1008 18:33:36.037400 4859 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-db-sync-x99jc" Oct 08 18:33:36 crc kubenswrapper[4859]: I1008 18:33:36.039956 4859 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"barbican-barbican-dockercfg-tkhlt" Oct 08 18:33:36 crc kubenswrapper[4859]: I1008 18:33:36.040164 4859 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"barbican-config-data" Oct 08 18:33:36 crc kubenswrapper[4859]: I1008 18:33:36.048430 4859 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-db-sync-x99jc"] Oct 08 18:33:36 crc kubenswrapper[4859]: I1008 18:33:36.076099 4859 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/720db6c0-f18f-429b-9589-fa503d212139-scripts\") pod \"cinder-db-sync-jqwrh\" (UID: \"720db6c0-f18f-429b-9589-fa503d212139\") " pod="openstack/cinder-db-sync-jqwrh" Oct 08 18:33:36 crc kubenswrapper[4859]: I1008 18:33:36.076157 4859 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/720db6c0-f18f-429b-9589-fa503d212139-combined-ca-bundle\") pod \"cinder-db-sync-jqwrh\" (UID: \"720db6c0-f18f-429b-9589-fa503d212139\") " pod="openstack/cinder-db-sync-jqwrh" Oct 08 18:33:36 crc kubenswrapper[4859]: I1008 18:33:36.076550 4859 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/720db6c0-f18f-429b-9589-fa503d212139-config-data\") pod \"cinder-db-sync-jqwrh\" (UID: \"720db6c0-f18f-429b-9589-fa503d212139\") " pod="openstack/cinder-db-sync-jqwrh" Oct 08 18:33:36 crc kubenswrapper[4859]: I1008 18:33:36.076631 4859 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-bdbmw\" (UniqueName: \"kubernetes.io/projected/720db6c0-f18f-429b-9589-fa503d212139-kube-api-access-bdbmw\") pod \"cinder-db-sync-jqwrh\" (UID: \"720db6c0-f18f-429b-9589-fa503d212139\") " pod="openstack/cinder-db-sync-jqwrh" Oct 08 18:33:36 crc kubenswrapper[4859]: I1008 18:33:36.076679 4859 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/720db6c0-f18f-429b-9589-fa503d212139-db-sync-config-data\") pod \"cinder-db-sync-jqwrh\" (UID: \"720db6c0-f18f-429b-9589-fa503d212139\") " pod="openstack/cinder-db-sync-jqwrh" Oct 08 18:33:36 crc kubenswrapper[4859]: I1008 18:33:36.076766 4859 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/720db6c0-f18f-429b-9589-fa503d212139-etc-machine-id\") pod \"cinder-db-sync-jqwrh\" (UID: \"720db6c0-f18f-429b-9589-fa503d212139\") " pod="openstack/cinder-db-sync-jqwrh" Oct 08 18:33:36 crc kubenswrapper[4859]: I1008 18:33:36.076896 4859 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/720db6c0-f18f-429b-9589-fa503d212139-etc-machine-id\") pod \"cinder-db-sync-jqwrh\" (UID: \"720db6c0-f18f-429b-9589-fa503d212139\") " pod="openstack/cinder-db-sync-jqwrh" Oct 08 18:33:36 crc kubenswrapper[4859]: I1008 18:33:36.092507 4859 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/720db6c0-f18f-429b-9589-fa503d212139-config-data\") pod \"cinder-db-sync-jqwrh\" (UID: \"720db6c0-f18f-429b-9589-fa503d212139\") " pod="openstack/cinder-db-sync-jqwrh" Oct 08 18:33:36 crc kubenswrapper[4859]: I1008 18:33:36.093032 4859 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/720db6c0-f18f-429b-9589-fa503d212139-db-sync-config-data\") pod \"cinder-db-sync-jqwrh\" (UID: \"720db6c0-f18f-429b-9589-fa503d212139\") " pod="openstack/cinder-db-sync-jqwrh" Oct 08 18:33:36 crc kubenswrapper[4859]: I1008 18:33:36.101827 4859 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/720db6c0-f18f-429b-9589-fa503d212139-combined-ca-bundle\") pod \"cinder-db-sync-jqwrh\" (UID: \"720db6c0-f18f-429b-9589-fa503d212139\") " pod="openstack/cinder-db-sync-jqwrh" Oct 08 18:33:36 crc kubenswrapper[4859]: I1008 18:33:36.104821 4859 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-bdbmw\" (UniqueName: \"kubernetes.io/projected/720db6c0-f18f-429b-9589-fa503d212139-kube-api-access-bdbmw\") pod \"cinder-db-sync-jqwrh\" (UID: \"720db6c0-f18f-429b-9589-fa503d212139\") " pod="openstack/cinder-db-sync-jqwrh" Oct 08 18:33:36 crc kubenswrapper[4859]: I1008 18:33:36.107700 4859 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/720db6c0-f18f-429b-9589-fa503d212139-scripts\") pod \"cinder-db-sync-jqwrh\" (UID: \"720db6c0-f18f-429b-9589-fa503d212139\") " pod="openstack/cinder-db-sync-jqwrh" Oct 08 18:33:36 crc kubenswrapper[4859]: I1008 18:33:36.179214 4859 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/3b2953d1-3ebe-4d7d-b60c-6ad07d02af5d-db-sync-config-data\") pod \"barbican-db-sync-x99jc\" (UID: \"3b2953d1-3ebe-4d7d-b60c-6ad07d02af5d\") " pod="openstack/barbican-db-sync-x99jc" Oct 08 18:33:36 crc kubenswrapper[4859]: I1008 18:33:36.179291 4859 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-llvcl\" (UniqueName: \"kubernetes.io/projected/3b2953d1-3ebe-4d7d-b60c-6ad07d02af5d-kube-api-access-llvcl\") pod \"barbican-db-sync-x99jc\" (UID: \"3b2953d1-3ebe-4d7d-b60c-6ad07d02af5d\") " pod="openstack/barbican-db-sync-x99jc" Oct 08 18:33:36 crc kubenswrapper[4859]: I1008 18:33:36.179325 4859 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3b2953d1-3ebe-4d7d-b60c-6ad07d02af5d-combined-ca-bundle\") pod \"barbican-db-sync-x99jc\" (UID: \"3b2953d1-3ebe-4d7d-b60c-6ad07d02af5d\") " pod="openstack/barbican-db-sync-x99jc" Oct 08 18:33:36 crc kubenswrapper[4859]: I1008 18:33:36.191377 4859 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-db-sync-jqwrh" Oct 08 18:33:36 crc kubenswrapper[4859]: I1008 18:33:36.226495 4859 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/neutron-db-sync-cw6k2"] Oct 08 18:33:36 crc kubenswrapper[4859]: I1008 18:33:36.227952 4859 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-db-sync-cw6k2" Oct 08 18:33:36 crc kubenswrapper[4859]: I1008 18:33:36.230031 4859 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"neutron-config" Oct 08 18:33:36 crc kubenswrapper[4859]: I1008 18:33:36.230280 4859 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"neutron-httpd-config" Oct 08 18:33:36 crc kubenswrapper[4859]: I1008 18:33:36.230424 4859 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"neutron-neutron-dockercfg-t6c4d" Oct 08 18:33:36 crc kubenswrapper[4859]: I1008 18:33:36.236186 4859 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-db-sync-cw6k2"] Oct 08 18:33:36 crc kubenswrapper[4859]: I1008 18:33:36.284612 4859 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/3b2953d1-3ebe-4d7d-b60c-6ad07d02af5d-db-sync-config-data\") pod \"barbican-db-sync-x99jc\" (UID: \"3b2953d1-3ebe-4d7d-b60c-6ad07d02af5d\") " pod="openstack/barbican-db-sync-x99jc" Oct 08 18:33:36 crc kubenswrapper[4859]: I1008 18:33:36.284714 4859 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-llvcl\" (UniqueName: \"kubernetes.io/projected/3b2953d1-3ebe-4d7d-b60c-6ad07d02af5d-kube-api-access-llvcl\") pod \"barbican-db-sync-x99jc\" (UID: \"3b2953d1-3ebe-4d7d-b60c-6ad07d02af5d\") " pod="openstack/barbican-db-sync-x99jc" Oct 08 18:33:36 crc kubenswrapper[4859]: I1008 18:33:36.284763 4859 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3b2953d1-3ebe-4d7d-b60c-6ad07d02af5d-combined-ca-bundle\") pod \"barbican-db-sync-x99jc\" (UID: \"3b2953d1-3ebe-4d7d-b60c-6ad07d02af5d\") " pod="openstack/barbican-db-sync-x99jc" Oct 08 18:33:36 crc kubenswrapper[4859]: I1008 18:33:36.294230 4859 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3b2953d1-3ebe-4d7d-b60c-6ad07d02af5d-combined-ca-bundle\") pod \"barbican-db-sync-x99jc\" (UID: \"3b2953d1-3ebe-4d7d-b60c-6ad07d02af5d\") " pod="openstack/barbican-db-sync-x99jc" Oct 08 18:33:36 crc kubenswrapper[4859]: I1008 18:33:36.300007 4859 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/3b2953d1-3ebe-4d7d-b60c-6ad07d02af5d-db-sync-config-data\") pod \"barbican-db-sync-x99jc\" (UID: \"3b2953d1-3ebe-4d7d-b60c-6ad07d02af5d\") " pod="openstack/barbican-db-sync-x99jc" Oct 08 18:33:36 crc kubenswrapper[4859]: I1008 18:33:36.300070 4859 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-59df9f9d49-446ss"] Oct 08 18:33:36 crc kubenswrapper[4859]: I1008 18:33:36.302537 4859 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-llvcl\" (UniqueName: \"kubernetes.io/projected/3b2953d1-3ebe-4d7d-b60c-6ad07d02af5d-kube-api-access-llvcl\") pod \"barbican-db-sync-x99jc\" (UID: \"3b2953d1-3ebe-4d7d-b60c-6ad07d02af5d\") " pod="openstack/barbican-db-sync-x99jc" Oct 08 18:33:36 crc kubenswrapper[4859]: I1008 18:33:36.312108 4859 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-59df9f9d49-446ss"] Oct 08 18:33:36 crc kubenswrapper[4859]: I1008 18:33:36.391394 4859 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/56c33135-38c1-4e29-9804-098312529f0f-combined-ca-bundle\") pod \"neutron-db-sync-cw6k2\" (UID: \"56c33135-38c1-4e29-9804-098312529f0f\") " pod="openstack/neutron-db-sync-cw6k2" Oct 08 18:33:36 crc kubenswrapper[4859]: I1008 18:33:36.391648 4859 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-6pcvs\" (UniqueName: \"kubernetes.io/projected/56c33135-38c1-4e29-9804-098312529f0f-kube-api-access-6pcvs\") pod \"neutron-db-sync-cw6k2\" (UID: \"56c33135-38c1-4e29-9804-098312529f0f\") " pod="openstack/neutron-db-sync-cw6k2" Oct 08 18:33:36 crc kubenswrapper[4859]: I1008 18:33:36.391682 4859 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/56c33135-38c1-4e29-9804-098312529f0f-config\") pod \"neutron-db-sync-cw6k2\" (UID: \"56c33135-38c1-4e29-9804-098312529f0f\") " pod="openstack/neutron-db-sync-cw6k2" Oct 08 18:33:36 crc kubenswrapper[4859]: I1008 18:33:36.493040 4859 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-6pcvs\" (UniqueName: \"kubernetes.io/projected/56c33135-38c1-4e29-9804-098312529f0f-kube-api-access-6pcvs\") pod \"neutron-db-sync-cw6k2\" (UID: \"56c33135-38c1-4e29-9804-098312529f0f\") " pod="openstack/neutron-db-sync-cw6k2" Oct 08 18:33:36 crc kubenswrapper[4859]: I1008 18:33:36.493111 4859 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/56c33135-38c1-4e29-9804-098312529f0f-config\") pod \"neutron-db-sync-cw6k2\" (UID: \"56c33135-38c1-4e29-9804-098312529f0f\") " pod="openstack/neutron-db-sync-cw6k2" Oct 08 18:33:36 crc kubenswrapper[4859]: I1008 18:33:36.493322 4859 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/56c33135-38c1-4e29-9804-098312529f0f-combined-ca-bundle\") pod \"neutron-db-sync-cw6k2\" (UID: \"56c33135-38c1-4e29-9804-098312529f0f\") " pod="openstack/neutron-db-sync-cw6k2" Oct 08 18:33:36 crc kubenswrapper[4859]: I1008 18:33:36.497349 4859 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/56c33135-38c1-4e29-9804-098312529f0f-combined-ca-bundle\") pod \"neutron-db-sync-cw6k2\" (UID: \"56c33135-38c1-4e29-9804-098312529f0f\") " pod="openstack/neutron-db-sync-cw6k2" Oct 08 18:33:36 crc kubenswrapper[4859]: I1008 18:33:36.507201 4859 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/secret/56c33135-38c1-4e29-9804-098312529f0f-config\") pod \"neutron-db-sync-cw6k2\" (UID: \"56c33135-38c1-4e29-9804-098312529f0f\") " pod="openstack/neutron-db-sync-cw6k2" Oct 08 18:33:36 crc kubenswrapper[4859]: I1008 18:33:36.508341 4859 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-6pcvs\" (UniqueName: \"kubernetes.io/projected/56c33135-38c1-4e29-9804-098312529f0f-kube-api-access-6pcvs\") pod \"neutron-db-sync-cw6k2\" (UID: \"56c33135-38c1-4e29-9804-098312529f0f\") " pod="openstack/neutron-db-sync-cw6k2" Oct 08 18:33:36 crc kubenswrapper[4859]: I1008 18:33:36.509226 4859 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="59231da1-64f0-4cac-a821-20fcbdcd5e30" path="/var/lib/kubelet/pods/59231da1-64f0-4cac-a821-20fcbdcd5e30/volumes" Oct 08 18:33:36 crc kubenswrapper[4859]: I1008 18:33:36.567288 4859 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-db-sync-x99jc" Oct 08 18:33:36 crc kubenswrapper[4859]: I1008 18:33:36.587138 4859 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-db-sync-cw6k2" Oct 08 18:33:36 crc kubenswrapper[4859]: I1008 18:33:36.766927 4859 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/horizon-85b86687bf-6gp89"] Oct 08 18:33:36 crc kubenswrapper[4859]: W1008 18:33:36.784741 4859 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod9ed85886_9589_4d26_ad59_7a6021df659e.slice/crio-1a0b89180efa9e1f99c0c69d71c18251e60c4bad6befa9c2c5ebb38d4f8be499 WatchSource:0}: Error finding container 1a0b89180efa9e1f99c0c69d71c18251e60c4bad6befa9c2c5ebb38d4f8be499: Status 404 returned error can't find the container with id 1a0b89180efa9e1f99c0c69d71c18251e60c4bad6befa9c2c5ebb38d4f8be499 Oct 08 18:33:36 crc kubenswrapper[4859]: W1008 18:33:36.857295 4859 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod720db6c0_f18f_429b_9589_fa503d212139.slice/crio-6158396a7a7a7bb9ea51437d23c0b6b2e9b2ef400482d5e49cc2cee88629328c WatchSource:0}: Error finding container 6158396a7a7a7bb9ea51437d23c0b6b2e9b2ef400482d5e49cc2cee88629328c: Status 404 returned error can't find the container with id 6158396a7a7a7bb9ea51437d23c0b6b2e9b2ef400482d5e49cc2cee88629328c Oct 08 18:33:36 crc kubenswrapper[4859]: I1008 18:33:36.860590 4859 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-db-sync-jqwrh"] Oct 08 18:33:36 crc kubenswrapper[4859]: I1008 18:33:36.946050 4859 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-db-sync-jqwrh" event={"ID":"720db6c0-f18f-429b-9589-fa503d212139","Type":"ContainerStarted","Data":"6158396a7a7a7bb9ea51437d23c0b6b2e9b2ef400482d5e49cc2cee88629328c"} Oct 08 18:33:36 crc kubenswrapper[4859]: I1008 18:33:36.951056 4859 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"bfc391b0-6e05-41b9-8c78-7cf870f5cfe4","Type":"ContainerStarted","Data":"6413f62ae3e9c179cba99f16d4f32c63326b5a3dc4e809c333d5364e3aac6c77"} Oct 08 18:33:36 crc kubenswrapper[4859]: I1008 18:33:36.956501 4859 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-85b86687bf-6gp89" event={"ID":"9ed85886-9589-4d26-ad59-7a6021df659e","Type":"ContainerStarted","Data":"1a0b89180efa9e1f99c0c69d71c18251e60c4bad6befa9c2c5ebb38d4f8be499"} Oct 08 18:33:36 crc kubenswrapper[4859]: I1008 18:33:36.961800 4859 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"21a3071c-7cf9-4008-a37b-ae90f125ddbc","Type":"ContainerStarted","Data":"72c30502678312848e2a6186412b97211a55d0089e7d406b78c0cc1efb9c2f27"} Oct 08 18:33:36 crc kubenswrapper[4859]: I1008 18:33:36.970657 4859 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-74776f5dd7-89sbw" event={"ID":"f08c21e6-a74c-4ad9-a3fa-4e56ff2c6432","Type":"ContainerStarted","Data":"f4cadb69d4dbae11278e1c5580f08f34fd3b6a5866e898ae3481fabd25c41b30"} Oct 08 18:33:36 crc kubenswrapper[4859]: I1008 18:33:36.971018 4859 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-74776f5dd7-89sbw" Oct 08 18:33:37 crc kubenswrapper[4859]: I1008 18:33:37.009057 4859 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-74776f5dd7-89sbw" podStartSLOduration=4.009029097 podStartE2EDuration="4.009029097s" podCreationTimestamp="2025-10-08 18:33:33 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-08 18:33:36.995372686 +0000 UTC m=+987.242212075" watchObservedRunningTime="2025-10-08 18:33:37.009029097 +0000 UTC m=+987.255868476" Oct 08 18:33:37 crc kubenswrapper[4859]: I1008 18:33:37.077065 4859 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-db-sync-x99jc"] Oct 08 18:33:37 crc kubenswrapper[4859]: I1008 18:33:37.278837 4859 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-db-sync-cw6k2"] Oct 08 18:33:37 crc kubenswrapper[4859]: W1008 18:33:37.294062 4859 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod56c33135_38c1_4e29_9804_098312529f0f.slice/crio-7f4b57ab0611ccc6e613abee019be0e27cd21ccc590e873186cba81c6373cd7b WatchSource:0}: Error finding container 7f4b57ab0611ccc6e613abee019be0e27cd21ccc590e873186cba81c6373cd7b: Status 404 returned error can't find the container with id 7f4b57ab0611ccc6e613abee019be0e27cd21ccc590e873186cba81c6373cd7b Oct 08 18:33:37 crc kubenswrapper[4859]: I1008 18:33:37.983462 4859 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-db-sync-cw6k2" event={"ID":"56c33135-38c1-4e29-9804-098312529f0f","Type":"ContainerStarted","Data":"954a12574e8a4ca3fce7cfa866b30b0090d26d8e434f0ca540812c65bf12435d"} Oct 08 18:33:37 crc kubenswrapper[4859]: I1008 18:33:37.983735 4859 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-db-sync-cw6k2" event={"ID":"56c33135-38c1-4e29-9804-098312529f0f","Type":"ContainerStarted","Data":"7f4b57ab0611ccc6e613abee019be0e27cd21ccc590e873186cba81c6373cd7b"} Oct 08 18:33:37 crc kubenswrapper[4859]: I1008 18:33:37.994376 4859 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"21a3071c-7cf9-4008-a37b-ae90f125ddbc","Type":"ContainerStarted","Data":"a44772a4cd0d4ee3d17393a0bc729acb973b7a3199f86de123b075f6994eae7c"} Oct 08 18:33:37 crc kubenswrapper[4859]: I1008 18:33:37.994510 4859 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/glance-default-internal-api-0" podUID="21a3071c-7cf9-4008-a37b-ae90f125ddbc" containerName="glance-log" containerID="cri-o://72c30502678312848e2a6186412b97211a55d0089e7d406b78c0cc1efb9c2f27" gracePeriod=30 Oct 08 18:33:37 crc kubenswrapper[4859]: I1008 18:33:37.994755 4859 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/glance-default-internal-api-0" podUID="21a3071c-7cf9-4008-a37b-ae90f125ddbc" containerName="glance-httpd" containerID="cri-o://a44772a4cd0d4ee3d17393a0bc729acb973b7a3199f86de123b075f6994eae7c" gracePeriod=30 Oct 08 18:33:37 crc kubenswrapper[4859]: I1008 18:33:37.999128 4859 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-db-sync-x99jc" event={"ID":"3b2953d1-3ebe-4d7d-b60c-6ad07d02af5d","Type":"ContainerStarted","Data":"5326bbd772abed3f9cc2690e3fb5c558ee843f4383d3dd7134f5126d43cf5d02"} Oct 08 18:33:38 crc kubenswrapper[4859]: I1008 18:33:38.032796 4859 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/glance-default-external-api-0" podUID="bfc391b0-6e05-41b9-8c78-7cf870f5cfe4" containerName="glance-log" containerID="cri-o://6413f62ae3e9c179cba99f16d4f32c63326b5a3dc4e809c333d5364e3aac6c77" gracePeriod=30 Oct 08 18:33:38 crc kubenswrapper[4859]: I1008 18:33:38.033028 4859 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"bfc391b0-6e05-41b9-8c78-7cf870f5cfe4","Type":"ContainerStarted","Data":"092813e771436337b8a8182b40e005937742352405076340ee1eb8c6aa53c4ca"} Oct 08 18:33:38 crc kubenswrapper[4859]: I1008 18:33:38.033522 4859 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/glance-default-external-api-0" podUID="bfc391b0-6e05-41b9-8c78-7cf870f5cfe4" containerName="glance-httpd" containerID="cri-o://092813e771436337b8a8182b40e005937742352405076340ee1eb8c6aa53c4ca" gracePeriod=30 Oct 08 18:33:38 crc kubenswrapper[4859]: I1008 18:33:38.045732 4859 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/neutron-db-sync-cw6k2" podStartSLOduration=2.04570748 podStartE2EDuration="2.04570748s" podCreationTimestamp="2025-10-08 18:33:36 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-08 18:33:38.007961561 +0000 UTC m=+988.254800940" watchObservedRunningTime="2025-10-08 18:33:38.04570748 +0000 UTC m=+988.292546859" Oct 08 18:33:38 crc kubenswrapper[4859]: I1008 18:33:38.048727 4859 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/glance-default-internal-api-0" podStartSLOduration=5.048707128 podStartE2EDuration="5.048707128s" podCreationTimestamp="2025-10-08 18:33:33 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-08 18:33:38.02903888 +0000 UTC m=+988.275878259" watchObservedRunningTime="2025-10-08 18:33:38.048707128 +0000 UTC m=+988.295546517" Oct 08 18:33:38 crc kubenswrapper[4859]: I1008 18:33:38.076843 4859 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/glance-default-external-api-0" podStartSLOduration=5.076824615 podStartE2EDuration="5.076824615s" podCreationTimestamp="2025-10-08 18:33:33 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-08 18:33:38.062950797 +0000 UTC m=+988.309790176" watchObservedRunningTime="2025-10-08 18:33:38.076824615 +0000 UTC m=+988.323663994" Oct 08 18:33:39 crc kubenswrapper[4859]: I1008 18:33:39.061161 4859 generic.go:334] "Generic (PLEG): container finished" podID="9d1b6fef-16a5-4c1b-a3bb-95f46f1d92f7" containerID="e48d82fde98292fc3ac341cff6cc00d37119be8b99b5e5c33a5f112eb3f15c4a" exitCode=0 Oct 08 18:33:39 crc kubenswrapper[4859]: I1008 18:33:39.061199 4859 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-bootstrap-mqwmv" event={"ID":"9d1b6fef-16a5-4c1b-a3bb-95f46f1d92f7","Type":"ContainerDied","Data":"e48d82fde98292fc3ac341cff6cc00d37119be8b99b5e5c33a5f112eb3f15c4a"} Oct 08 18:33:39 crc kubenswrapper[4859]: I1008 18:33:39.074022 4859 generic.go:334] "Generic (PLEG): container finished" podID="21a3071c-7cf9-4008-a37b-ae90f125ddbc" containerID="a44772a4cd0d4ee3d17393a0bc729acb973b7a3199f86de123b075f6994eae7c" exitCode=0 Oct 08 18:33:39 crc kubenswrapper[4859]: I1008 18:33:39.074050 4859 generic.go:334] "Generic (PLEG): container finished" podID="21a3071c-7cf9-4008-a37b-ae90f125ddbc" containerID="72c30502678312848e2a6186412b97211a55d0089e7d406b78c0cc1efb9c2f27" exitCode=143 Oct 08 18:33:39 crc kubenswrapper[4859]: I1008 18:33:39.074085 4859 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"21a3071c-7cf9-4008-a37b-ae90f125ddbc","Type":"ContainerDied","Data":"a44772a4cd0d4ee3d17393a0bc729acb973b7a3199f86de123b075f6994eae7c"} Oct 08 18:33:39 crc kubenswrapper[4859]: I1008 18:33:39.074105 4859 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"21a3071c-7cf9-4008-a37b-ae90f125ddbc","Type":"ContainerDied","Data":"72c30502678312848e2a6186412b97211a55d0089e7d406b78c0cc1efb9c2f27"} Oct 08 18:33:39 crc kubenswrapper[4859]: I1008 18:33:39.078480 4859 generic.go:334] "Generic (PLEG): container finished" podID="bfc391b0-6e05-41b9-8c78-7cf870f5cfe4" containerID="092813e771436337b8a8182b40e005937742352405076340ee1eb8c6aa53c4ca" exitCode=0 Oct 08 18:33:39 crc kubenswrapper[4859]: I1008 18:33:39.078499 4859 generic.go:334] "Generic (PLEG): container finished" podID="bfc391b0-6e05-41b9-8c78-7cf870f5cfe4" containerID="6413f62ae3e9c179cba99f16d4f32c63326b5a3dc4e809c333d5364e3aac6c77" exitCode=143 Oct 08 18:33:39 crc kubenswrapper[4859]: I1008 18:33:39.079005 4859 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"bfc391b0-6e05-41b9-8c78-7cf870f5cfe4","Type":"ContainerDied","Data":"092813e771436337b8a8182b40e005937742352405076340ee1eb8c6aa53c4ca"} Oct 08 18:33:39 crc kubenswrapper[4859]: I1008 18:33:39.079085 4859 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"bfc391b0-6e05-41b9-8c78-7cf870f5cfe4","Type":"ContainerDied","Data":"6413f62ae3e9c179cba99f16d4f32c63326b5a3dc4e809c333d5364e3aac6c77"} Oct 08 18:33:41 crc kubenswrapper[4859]: I1008 18:33:41.942399 4859 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/horizon-86b9b87697-gmwqv"] Oct 08 18:33:41 crc kubenswrapper[4859]: I1008 18:33:41.971545 4859 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/horizon-7d58687c8-fhctx"] Oct 08 18:33:41 crc kubenswrapper[4859]: I1008 18:33:41.973973 4859 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-7d58687c8-fhctx" Oct 08 18:33:41 crc kubenswrapper[4859]: I1008 18:33:41.975804 4859 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-horizon-svc" Oct 08 18:33:41 crc kubenswrapper[4859]: I1008 18:33:41.986994 4859 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/horizon-7d58687c8-fhctx"] Oct 08 18:33:42 crc kubenswrapper[4859]: I1008 18:33:42.039203 4859 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/1a92487e-3309-41c4-8c82-cd57f517ab4b-logs\") pod \"horizon-7d58687c8-fhctx\" (UID: \"1a92487e-3309-41c4-8c82-cd57f517ab4b\") " pod="openstack/horizon-7d58687c8-fhctx" Oct 08 18:33:42 crc kubenswrapper[4859]: I1008 18:33:42.039265 4859 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/1a92487e-3309-41c4-8c82-cd57f517ab4b-scripts\") pod \"horizon-7d58687c8-fhctx\" (UID: \"1a92487e-3309-41c4-8c82-cd57f517ab4b\") " pod="openstack/horizon-7d58687c8-fhctx" Oct 08 18:33:42 crc kubenswrapper[4859]: I1008 18:33:42.039293 4859 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-7q7tg\" (UniqueName: \"kubernetes.io/projected/1a92487e-3309-41c4-8c82-cd57f517ab4b-kube-api-access-7q7tg\") pod \"horizon-7d58687c8-fhctx\" (UID: \"1a92487e-3309-41c4-8c82-cd57f517ab4b\") " pod="openstack/horizon-7d58687c8-fhctx" Oct 08 18:33:42 crc kubenswrapper[4859]: I1008 18:33:42.039584 4859 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/1a92487e-3309-41c4-8c82-cd57f517ab4b-horizon-secret-key\") pod \"horizon-7d58687c8-fhctx\" (UID: \"1a92487e-3309-41c4-8c82-cd57f517ab4b\") " pod="openstack/horizon-7d58687c8-fhctx" Oct 08 18:33:42 crc kubenswrapper[4859]: I1008 18:33:42.039759 4859 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"horizon-tls-certs\" (UniqueName: \"kubernetes.io/secret/1a92487e-3309-41c4-8c82-cd57f517ab4b-horizon-tls-certs\") pod \"horizon-7d58687c8-fhctx\" (UID: \"1a92487e-3309-41c4-8c82-cd57f517ab4b\") " pod="openstack/horizon-7d58687c8-fhctx" Oct 08 18:33:42 crc kubenswrapper[4859]: I1008 18:33:42.039811 4859 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1a92487e-3309-41c4-8c82-cd57f517ab4b-combined-ca-bundle\") pod \"horizon-7d58687c8-fhctx\" (UID: \"1a92487e-3309-41c4-8c82-cd57f517ab4b\") " pod="openstack/horizon-7d58687c8-fhctx" Oct 08 18:33:42 crc kubenswrapper[4859]: I1008 18:33:42.039868 4859 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/1a92487e-3309-41c4-8c82-cd57f517ab4b-config-data\") pod \"horizon-7d58687c8-fhctx\" (UID: \"1a92487e-3309-41c4-8c82-cd57f517ab4b\") " pod="openstack/horizon-7d58687c8-fhctx" Oct 08 18:33:42 crc kubenswrapper[4859]: I1008 18:33:42.062184 4859 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/horizon-85b86687bf-6gp89"] Oct 08 18:33:42 crc kubenswrapper[4859]: I1008 18:33:42.120141 4859 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/horizon-6f569d5894-mx8v5"] Oct 08 18:33:42 crc kubenswrapper[4859]: I1008 18:33:42.121701 4859 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-6f569d5894-mx8v5" Oct 08 18:33:42 crc kubenswrapper[4859]: I1008 18:33:42.137327 4859 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/horizon-6f569d5894-mx8v5"] Oct 08 18:33:42 crc kubenswrapper[4859]: I1008 18:33:42.142563 4859 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/50b5de31-90f6-43cf-9e16-7b89b09f1e57-horizon-secret-key\") pod \"horizon-6f569d5894-mx8v5\" (UID: \"50b5de31-90f6-43cf-9e16-7b89b09f1e57\") " pod="openstack/horizon-6f569d5894-mx8v5" Oct 08 18:33:42 crc kubenswrapper[4859]: I1008 18:33:42.142625 4859 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/1a92487e-3309-41c4-8c82-cd57f517ab4b-scripts\") pod \"horizon-7d58687c8-fhctx\" (UID: \"1a92487e-3309-41c4-8c82-cd57f517ab4b\") " pod="openstack/horizon-7d58687c8-fhctx" Oct 08 18:33:42 crc kubenswrapper[4859]: I1008 18:33:42.142664 4859 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-7q7tg\" (UniqueName: \"kubernetes.io/projected/1a92487e-3309-41c4-8c82-cd57f517ab4b-kube-api-access-7q7tg\") pod \"horizon-7d58687c8-fhctx\" (UID: \"1a92487e-3309-41c4-8c82-cd57f517ab4b\") " pod="openstack/horizon-7d58687c8-fhctx" Oct 08 18:33:42 crc kubenswrapper[4859]: I1008 18:33:42.142818 4859 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/50b5de31-90f6-43cf-9e16-7b89b09f1e57-config-data\") pod \"horizon-6f569d5894-mx8v5\" (UID: \"50b5de31-90f6-43cf-9e16-7b89b09f1e57\") " pod="openstack/horizon-6f569d5894-mx8v5" Oct 08 18:33:42 crc kubenswrapper[4859]: I1008 18:33:42.143031 4859 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/1a92487e-3309-41c4-8c82-cd57f517ab4b-horizon-secret-key\") pod \"horizon-7d58687c8-fhctx\" (UID: \"1a92487e-3309-41c4-8c82-cd57f517ab4b\") " pod="openstack/horizon-7d58687c8-fhctx" Oct 08 18:33:42 crc kubenswrapper[4859]: I1008 18:33:42.143085 4859 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-vcnq5\" (UniqueName: \"kubernetes.io/projected/50b5de31-90f6-43cf-9e16-7b89b09f1e57-kube-api-access-vcnq5\") pod \"horizon-6f569d5894-mx8v5\" (UID: \"50b5de31-90f6-43cf-9e16-7b89b09f1e57\") " pod="openstack/horizon-6f569d5894-mx8v5" Oct 08 18:33:42 crc kubenswrapper[4859]: I1008 18:33:42.143127 4859 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/50b5de31-90f6-43cf-9e16-7b89b09f1e57-scripts\") pod \"horizon-6f569d5894-mx8v5\" (UID: \"50b5de31-90f6-43cf-9e16-7b89b09f1e57\") " pod="openstack/horizon-6f569d5894-mx8v5" Oct 08 18:33:42 crc kubenswrapper[4859]: I1008 18:33:42.143165 4859 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/50b5de31-90f6-43cf-9e16-7b89b09f1e57-logs\") pod \"horizon-6f569d5894-mx8v5\" (UID: \"50b5de31-90f6-43cf-9e16-7b89b09f1e57\") " pod="openstack/horizon-6f569d5894-mx8v5" Oct 08 18:33:42 crc kubenswrapper[4859]: I1008 18:33:42.143197 4859 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"horizon-tls-certs\" (UniqueName: \"kubernetes.io/secret/1a92487e-3309-41c4-8c82-cd57f517ab4b-horizon-tls-certs\") pod \"horizon-7d58687c8-fhctx\" (UID: \"1a92487e-3309-41c4-8c82-cd57f517ab4b\") " pod="openstack/horizon-7d58687c8-fhctx" Oct 08 18:33:42 crc kubenswrapper[4859]: I1008 18:33:42.143212 4859 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"horizon-tls-certs\" (UniqueName: \"kubernetes.io/secret/50b5de31-90f6-43cf-9e16-7b89b09f1e57-horizon-tls-certs\") pod \"horizon-6f569d5894-mx8v5\" (UID: \"50b5de31-90f6-43cf-9e16-7b89b09f1e57\") " pod="openstack/horizon-6f569d5894-mx8v5" Oct 08 18:33:42 crc kubenswrapper[4859]: I1008 18:33:42.143236 4859 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/50b5de31-90f6-43cf-9e16-7b89b09f1e57-combined-ca-bundle\") pod \"horizon-6f569d5894-mx8v5\" (UID: \"50b5de31-90f6-43cf-9e16-7b89b09f1e57\") " pod="openstack/horizon-6f569d5894-mx8v5" Oct 08 18:33:42 crc kubenswrapper[4859]: I1008 18:33:42.143261 4859 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1a92487e-3309-41c4-8c82-cd57f517ab4b-combined-ca-bundle\") pod \"horizon-7d58687c8-fhctx\" (UID: \"1a92487e-3309-41c4-8c82-cd57f517ab4b\") " pod="openstack/horizon-7d58687c8-fhctx" Oct 08 18:33:42 crc kubenswrapper[4859]: I1008 18:33:42.143306 4859 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/1a92487e-3309-41c4-8c82-cd57f517ab4b-config-data\") pod \"horizon-7d58687c8-fhctx\" (UID: \"1a92487e-3309-41c4-8c82-cd57f517ab4b\") " pod="openstack/horizon-7d58687c8-fhctx" Oct 08 18:33:42 crc kubenswrapper[4859]: I1008 18:33:42.143418 4859 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/1a92487e-3309-41c4-8c82-cd57f517ab4b-logs\") pod \"horizon-7d58687c8-fhctx\" (UID: \"1a92487e-3309-41c4-8c82-cd57f517ab4b\") " pod="openstack/horizon-7d58687c8-fhctx" Oct 08 18:33:42 crc kubenswrapper[4859]: I1008 18:33:42.143674 4859 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/1a92487e-3309-41c4-8c82-cd57f517ab4b-scripts\") pod \"horizon-7d58687c8-fhctx\" (UID: \"1a92487e-3309-41c4-8c82-cd57f517ab4b\") " pod="openstack/horizon-7d58687c8-fhctx" Oct 08 18:33:42 crc kubenswrapper[4859]: I1008 18:33:42.143854 4859 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/1a92487e-3309-41c4-8c82-cd57f517ab4b-logs\") pod \"horizon-7d58687c8-fhctx\" (UID: \"1a92487e-3309-41c4-8c82-cd57f517ab4b\") " pod="openstack/horizon-7d58687c8-fhctx" Oct 08 18:33:42 crc kubenswrapper[4859]: I1008 18:33:42.145416 4859 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/1a92487e-3309-41c4-8c82-cd57f517ab4b-config-data\") pod \"horizon-7d58687c8-fhctx\" (UID: \"1a92487e-3309-41c4-8c82-cd57f517ab4b\") " pod="openstack/horizon-7d58687c8-fhctx" Oct 08 18:33:42 crc kubenswrapper[4859]: I1008 18:33:42.150871 4859 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"horizon-tls-certs\" (UniqueName: \"kubernetes.io/secret/1a92487e-3309-41c4-8c82-cd57f517ab4b-horizon-tls-certs\") pod \"horizon-7d58687c8-fhctx\" (UID: \"1a92487e-3309-41c4-8c82-cd57f517ab4b\") " pod="openstack/horizon-7d58687c8-fhctx" Oct 08 18:33:42 crc kubenswrapper[4859]: I1008 18:33:42.152965 4859 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/1a92487e-3309-41c4-8c82-cd57f517ab4b-horizon-secret-key\") pod \"horizon-7d58687c8-fhctx\" (UID: \"1a92487e-3309-41c4-8c82-cd57f517ab4b\") " pod="openstack/horizon-7d58687c8-fhctx" Oct 08 18:33:42 crc kubenswrapper[4859]: I1008 18:33:42.161189 4859 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1a92487e-3309-41c4-8c82-cd57f517ab4b-combined-ca-bundle\") pod \"horizon-7d58687c8-fhctx\" (UID: \"1a92487e-3309-41c4-8c82-cd57f517ab4b\") " pod="openstack/horizon-7d58687c8-fhctx" Oct 08 18:33:42 crc kubenswrapper[4859]: I1008 18:33:42.174179 4859 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-7q7tg\" (UniqueName: \"kubernetes.io/projected/1a92487e-3309-41c4-8c82-cd57f517ab4b-kube-api-access-7q7tg\") pod \"horizon-7d58687c8-fhctx\" (UID: \"1a92487e-3309-41c4-8c82-cd57f517ab4b\") " pod="openstack/horizon-7d58687c8-fhctx" Oct 08 18:33:42 crc kubenswrapper[4859]: I1008 18:33:42.246486 4859 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/50b5de31-90f6-43cf-9e16-7b89b09f1e57-horizon-secret-key\") pod \"horizon-6f569d5894-mx8v5\" (UID: \"50b5de31-90f6-43cf-9e16-7b89b09f1e57\") " pod="openstack/horizon-6f569d5894-mx8v5" Oct 08 18:33:42 crc kubenswrapper[4859]: I1008 18:33:42.247048 4859 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/50b5de31-90f6-43cf-9e16-7b89b09f1e57-config-data\") pod \"horizon-6f569d5894-mx8v5\" (UID: \"50b5de31-90f6-43cf-9e16-7b89b09f1e57\") " pod="openstack/horizon-6f569d5894-mx8v5" Oct 08 18:33:42 crc kubenswrapper[4859]: I1008 18:33:42.247289 4859 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-vcnq5\" (UniqueName: \"kubernetes.io/projected/50b5de31-90f6-43cf-9e16-7b89b09f1e57-kube-api-access-vcnq5\") pod \"horizon-6f569d5894-mx8v5\" (UID: \"50b5de31-90f6-43cf-9e16-7b89b09f1e57\") " pod="openstack/horizon-6f569d5894-mx8v5" Oct 08 18:33:42 crc kubenswrapper[4859]: I1008 18:33:42.247343 4859 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/50b5de31-90f6-43cf-9e16-7b89b09f1e57-scripts\") pod \"horizon-6f569d5894-mx8v5\" (UID: \"50b5de31-90f6-43cf-9e16-7b89b09f1e57\") " pod="openstack/horizon-6f569d5894-mx8v5" Oct 08 18:33:42 crc kubenswrapper[4859]: I1008 18:33:42.247383 4859 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/50b5de31-90f6-43cf-9e16-7b89b09f1e57-logs\") pod \"horizon-6f569d5894-mx8v5\" (UID: \"50b5de31-90f6-43cf-9e16-7b89b09f1e57\") " pod="openstack/horizon-6f569d5894-mx8v5" Oct 08 18:33:42 crc kubenswrapper[4859]: I1008 18:33:42.247412 4859 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"horizon-tls-certs\" (UniqueName: \"kubernetes.io/secret/50b5de31-90f6-43cf-9e16-7b89b09f1e57-horizon-tls-certs\") pod \"horizon-6f569d5894-mx8v5\" (UID: \"50b5de31-90f6-43cf-9e16-7b89b09f1e57\") " pod="openstack/horizon-6f569d5894-mx8v5" Oct 08 18:33:42 crc kubenswrapper[4859]: I1008 18:33:42.247440 4859 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/50b5de31-90f6-43cf-9e16-7b89b09f1e57-combined-ca-bundle\") pod \"horizon-6f569d5894-mx8v5\" (UID: \"50b5de31-90f6-43cf-9e16-7b89b09f1e57\") " pod="openstack/horizon-6f569d5894-mx8v5" Oct 08 18:33:42 crc kubenswrapper[4859]: I1008 18:33:42.248208 4859 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/50b5de31-90f6-43cf-9e16-7b89b09f1e57-scripts\") pod \"horizon-6f569d5894-mx8v5\" (UID: \"50b5de31-90f6-43cf-9e16-7b89b09f1e57\") " pod="openstack/horizon-6f569d5894-mx8v5" Oct 08 18:33:42 crc kubenswrapper[4859]: I1008 18:33:42.248851 4859 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/50b5de31-90f6-43cf-9e16-7b89b09f1e57-config-data\") pod \"horizon-6f569d5894-mx8v5\" (UID: \"50b5de31-90f6-43cf-9e16-7b89b09f1e57\") " pod="openstack/horizon-6f569d5894-mx8v5" Oct 08 18:33:42 crc kubenswrapper[4859]: I1008 18:33:42.249951 4859 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/50b5de31-90f6-43cf-9e16-7b89b09f1e57-logs\") pod \"horizon-6f569d5894-mx8v5\" (UID: \"50b5de31-90f6-43cf-9e16-7b89b09f1e57\") " pod="openstack/horizon-6f569d5894-mx8v5" Oct 08 18:33:42 crc kubenswrapper[4859]: I1008 18:33:42.261381 4859 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/50b5de31-90f6-43cf-9e16-7b89b09f1e57-horizon-secret-key\") pod \"horizon-6f569d5894-mx8v5\" (UID: \"50b5de31-90f6-43cf-9e16-7b89b09f1e57\") " pod="openstack/horizon-6f569d5894-mx8v5" Oct 08 18:33:42 crc kubenswrapper[4859]: I1008 18:33:42.261951 4859 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"horizon-tls-certs\" (UniqueName: \"kubernetes.io/secret/50b5de31-90f6-43cf-9e16-7b89b09f1e57-horizon-tls-certs\") pod \"horizon-6f569d5894-mx8v5\" (UID: \"50b5de31-90f6-43cf-9e16-7b89b09f1e57\") " pod="openstack/horizon-6f569d5894-mx8v5" Oct 08 18:33:42 crc kubenswrapper[4859]: I1008 18:33:42.264924 4859 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/50b5de31-90f6-43cf-9e16-7b89b09f1e57-combined-ca-bundle\") pod \"horizon-6f569d5894-mx8v5\" (UID: \"50b5de31-90f6-43cf-9e16-7b89b09f1e57\") " pod="openstack/horizon-6f569d5894-mx8v5" Oct 08 18:33:42 crc kubenswrapper[4859]: I1008 18:33:42.272918 4859 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-vcnq5\" (UniqueName: \"kubernetes.io/projected/50b5de31-90f6-43cf-9e16-7b89b09f1e57-kube-api-access-vcnq5\") pod \"horizon-6f569d5894-mx8v5\" (UID: \"50b5de31-90f6-43cf-9e16-7b89b09f1e57\") " pod="openstack/horizon-6f569d5894-mx8v5" Oct 08 18:33:42 crc kubenswrapper[4859]: I1008 18:33:42.306459 4859 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-7d58687c8-fhctx" Oct 08 18:33:42 crc kubenswrapper[4859]: I1008 18:33:42.440114 4859 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-6f569d5894-mx8v5" Oct 08 18:33:43 crc kubenswrapper[4859]: I1008 18:33:43.676300 4859 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-74776f5dd7-89sbw" Oct 08 18:33:43 crc kubenswrapper[4859]: I1008 18:33:43.783597 4859 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-7b587f8db7-lm5t5"] Oct 08 18:33:43 crc kubenswrapper[4859]: I1008 18:33:43.784200 4859 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-7b587f8db7-lm5t5" podUID="528e6076-58d3-4fcd-ac6e-4fc5d32dd85f" containerName="dnsmasq-dns" containerID="cri-o://4277ff351d5cc7e7b03ffd72a0f5d7bb37123fc379df3099fb15b5d20798ff85" gracePeriod=10 Oct 08 18:33:45 crc kubenswrapper[4859]: I1008 18:33:45.143020 4859 generic.go:334] "Generic (PLEG): container finished" podID="528e6076-58d3-4fcd-ac6e-4fc5d32dd85f" containerID="4277ff351d5cc7e7b03ffd72a0f5d7bb37123fc379df3099fb15b5d20798ff85" exitCode=0 Oct 08 18:33:45 crc kubenswrapper[4859]: I1008 18:33:45.143074 4859 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-7b587f8db7-lm5t5" event={"ID":"528e6076-58d3-4fcd-ac6e-4fc5d32dd85f","Type":"ContainerDied","Data":"4277ff351d5cc7e7b03ffd72a0f5d7bb37123fc379df3099fb15b5d20798ff85"} Oct 08 18:33:45 crc kubenswrapper[4859]: I1008 18:33:45.451128 4859 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/dnsmasq-dns-7b587f8db7-lm5t5" podUID="528e6076-58d3-4fcd-ac6e-4fc5d32dd85f" containerName="dnsmasq-dns" probeResult="failure" output="dial tcp 10.217.0.116:5353: connect: connection refused" Oct 08 18:33:47 crc kubenswrapper[4859]: I1008 18:33:47.463130 4859 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-external-api-0" Oct 08 18:33:47 crc kubenswrapper[4859]: I1008 18:33:47.471589 4859 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-internal-api-0" Oct 08 18:33:47 crc kubenswrapper[4859]: I1008 18:33:47.501168 4859 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-bootstrap-mqwmv" Oct 08 18:33:47 crc kubenswrapper[4859]: I1008 18:33:47.574824 4859 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/bfc391b0-6e05-41b9-8c78-7cf870f5cfe4-scripts\") pod \"bfc391b0-6e05-41b9-8c78-7cf870f5cfe4\" (UID: \"bfc391b0-6e05-41b9-8c78-7cf870f5cfe4\") " Oct 08 18:33:47 crc kubenswrapper[4859]: I1008 18:33:47.574871 4859 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-4kfnf\" (UniqueName: \"kubernetes.io/projected/bfc391b0-6e05-41b9-8c78-7cf870f5cfe4-kube-api-access-4kfnf\") pod \"bfc391b0-6e05-41b9-8c78-7cf870f5cfe4\" (UID: \"bfc391b0-6e05-41b9-8c78-7cf870f5cfe4\") " Oct 08 18:33:47 crc kubenswrapper[4859]: I1008 18:33:47.574930 4859 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/bfc391b0-6e05-41b9-8c78-7cf870f5cfe4-config-data\") pod \"bfc391b0-6e05-41b9-8c78-7cf870f5cfe4\" (UID: \"bfc391b0-6e05-41b9-8c78-7cf870f5cfe4\") " Oct 08 18:33:47 crc kubenswrapper[4859]: I1008 18:33:47.574952 4859 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/21a3071c-7cf9-4008-a37b-ae90f125ddbc-httpd-run\") pod \"21a3071c-7cf9-4008-a37b-ae90f125ddbc\" (UID: \"21a3071c-7cf9-4008-a37b-ae90f125ddbc\") " Oct 08 18:33:47 crc kubenswrapper[4859]: I1008 18:33:47.574982 4859 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/21a3071c-7cf9-4008-a37b-ae90f125ddbc-logs\") pod \"21a3071c-7cf9-4008-a37b-ae90f125ddbc\" (UID: \"21a3071c-7cf9-4008-a37b-ae90f125ddbc\") " Oct 08 18:33:47 crc kubenswrapper[4859]: I1008 18:33:47.575039 4859 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/21a3071c-7cf9-4008-a37b-ae90f125ddbc-scripts\") pod \"21a3071c-7cf9-4008-a37b-ae90f125ddbc\" (UID: \"21a3071c-7cf9-4008-a37b-ae90f125ddbc\") " Oct 08 18:33:47 crc kubenswrapper[4859]: I1008 18:33:47.575096 4859 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/9d1b6fef-16a5-4c1b-a3bb-95f46f1d92f7-credential-keys\") pod \"9d1b6fef-16a5-4c1b-a3bb-95f46f1d92f7\" (UID: \"9d1b6fef-16a5-4c1b-a3bb-95f46f1d92f7\") " Oct 08 18:33:47 crc kubenswrapper[4859]: I1008 18:33:47.575119 4859 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"glance\" (UniqueName: \"kubernetes.io/local-volume/local-storage12-crc\") pod \"bfc391b0-6e05-41b9-8c78-7cf870f5cfe4\" (UID: \"bfc391b0-6e05-41b9-8c78-7cf870f5cfe4\") " Oct 08 18:33:47 crc kubenswrapper[4859]: I1008 18:33:47.575152 4859 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/9d1b6fef-16a5-4c1b-a3bb-95f46f1d92f7-scripts\") pod \"9d1b6fef-16a5-4c1b-a3bb-95f46f1d92f7\" (UID: \"9d1b6fef-16a5-4c1b-a3bb-95f46f1d92f7\") " Oct 08 18:33:47 crc kubenswrapper[4859]: I1008 18:33:47.575181 4859 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-9mgdb\" (UniqueName: \"kubernetes.io/projected/21a3071c-7cf9-4008-a37b-ae90f125ddbc-kube-api-access-9mgdb\") pod \"21a3071c-7cf9-4008-a37b-ae90f125ddbc\" (UID: \"21a3071c-7cf9-4008-a37b-ae90f125ddbc\") " Oct 08 18:33:47 crc kubenswrapper[4859]: I1008 18:33:47.575203 4859 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/bfc391b0-6e05-41b9-8c78-7cf870f5cfe4-httpd-run\") pod \"bfc391b0-6e05-41b9-8c78-7cf870f5cfe4\" (UID: \"bfc391b0-6e05-41b9-8c78-7cf870f5cfe4\") " Oct 08 18:33:47 crc kubenswrapper[4859]: I1008 18:33:47.575222 4859 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-txptt\" (UniqueName: \"kubernetes.io/projected/9d1b6fef-16a5-4c1b-a3bb-95f46f1d92f7-kube-api-access-txptt\") pod \"9d1b6fef-16a5-4c1b-a3bb-95f46f1d92f7\" (UID: \"9d1b6fef-16a5-4c1b-a3bb-95f46f1d92f7\") " Oct 08 18:33:47 crc kubenswrapper[4859]: I1008 18:33:47.575238 4859 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/bfc391b0-6e05-41b9-8c78-7cf870f5cfe4-public-tls-certs\") pod \"bfc391b0-6e05-41b9-8c78-7cf870f5cfe4\" (UID: \"bfc391b0-6e05-41b9-8c78-7cf870f5cfe4\") " Oct 08 18:33:47 crc kubenswrapper[4859]: I1008 18:33:47.575259 4859 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/21a3071c-7cf9-4008-a37b-ae90f125ddbc-config-data\") pod \"21a3071c-7cf9-4008-a37b-ae90f125ddbc\" (UID: \"21a3071c-7cf9-4008-a37b-ae90f125ddbc\") " Oct 08 18:33:47 crc kubenswrapper[4859]: I1008 18:33:47.575278 4859 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/21a3071c-7cf9-4008-a37b-ae90f125ddbc-combined-ca-bundle\") pod \"21a3071c-7cf9-4008-a37b-ae90f125ddbc\" (UID: \"21a3071c-7cf9-4008-a37b-ae90f125ddbc\") " Oct 08 18:33:47 crc kubenswrapper[4859]: I1008 18:33:47.575297 4859 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/21a3071c-7cf9-4008-a37b-ae90f125ddbc-internal-tls-certs\") pod \"21a3071c-7cf9-4008-a37b-ae90f125ddbc\" (UID: \"21a3071c-7cf9-4008-a37b-ae90f125ddbc\") " Oct 08 18:33:47 crc kubenswrapper[4859]: I1008 18:33:47.575317 4859 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/9d1b6fef-16a5-4c1b-a3bb-95f46f1d92f7-fernet-keys\") pod \"9d1b6fef-16a5-4c1b-a3bb-95f46f1d92f7\" (UID: \"9d1b6fef-16a5-4c1b-a3bb-95f46f1d92f7\") " Oct 08 18:33:47 crc kubenswrapper[4859]: I1008 18:33:47.575336 4859 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/9d1b6fef-16a5-4c1b-a3bb-95f46f1d92f7-config-data\") pod \"9d1b6fef-16a5-4c1b-a3bb-95f46f1d92f7\" (UID: \"9d1b6fef-16a5-4c1b-a3bb-95f46f1d92f7\") " Oct 08 18:33:47 crc kubenswrapper[4859]: I1008 18:33:47.575356 4859 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"glance\" (UniqueName: \"kubernetes.io/local-volume/local-storage10-crc\") pod \"21a3071c-7cf9-4008-a37b-ae90f125ddbc\" (UID: \"21a3071c-7cf9-4008-a37b-ae90f125ddbc\") " Oct 08 18:33:47 crc kubenswrapper[4859]: I1008 18:33:47.575384 4859 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/bfc391b0-6e05-41b9-8c78-7cf870f5cfe4-logs\") pod \"bfc391b0-6e05-41b9-8c78-7cf870f5cfe4\" (UID: \"bfc391b0-6e05-41b9-8c78-7cf870f5cfe4\") " Oct 08 18:33:47 crc kubenswrapper[4859]: I1008 18:33:47.575402 4859 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/bfc391b0-6e05-41b9-8c78-7cf870f5cfe4-combined-ca-bundle\") pod \"bfc391b0-6e05-41b9-8c78-7cf870f5cfe4\" (UID: \"bfc391b0-6e05-41b9-8c78-7cf870f5cfe4\") " Oct 08 18:33:47 crc kubenswrapper[4859]: I1008 18:33:47.575435 4859 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9d1b6fef-16a5-4c1b-a3bb-95f46f1d92f7-combined-ca-bundle\") pod \"9d1b6fef-16a5-4c1b-a3bb-95f46f1d92f7\" (UID: \"9d1b6fef-16a5-4c1b-a3bb-95f46f1d92f7\") " Oct 08 18:33:47 crc kubenswrapper[4859]: I1008 18:33:47.575473 4859 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/21a3071c-7cf9-4008-a37b-ae90f125ddbc-httpd-run" (OuterVolumeSpecName: "httpd-run") pod "21a3071c-7cf9-4008-a37b-ae90f125ddbc" (UID: "21a3071c-7cf9-4008-a37b-ae90f125ddbc"). InnerVolumeSpecName "httpd-run". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 08 18:33:47 crc kubenswrapper[4859]: I1008 18:33:47.575613 4859 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/21a3071c-7cf9-4008-a37b-ae90f125ddbc-logs" (OuterVolumeSpecName: "logs") pod "21a3071c-7cf9-4008-a37b-ae90f125ddbc" (UID: "21a3071c-7cf9-4008-a37b-ae90f125ddbc"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 08 18:33:47 crc kubenswrapper[4859]: I1008 18:33:47.576284 4859 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/bfc391b0-6e05-41b9-8c78-7cf870f5cfe4-httpd-run" (OuterVolumeSpecName: "httpd-run") pod "bfc391b0-6e05-41b9-8c78-7cf870f5cfe4" (UID: "bfc391b0-6e05-41b9-8c78-7cf870f5cfe4"). InnerVolumeSpecName "httpd-run". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 08 18:33:47 crc kubenswrapper[4859]: I1008 18:33:47.576880 4859 reconciler_common.go:293] "Volume detached for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/21a3071c-7cf9-4008-a37b-ae90f125ddbc-httpd-run\") on node \"crc\" DevicePath \"\"" Oct 08 18:33:47 crc kubenswrapper[4859]: I1008 18:33:47.576916 4859 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/21a3071c-7cf9-4008-a37b-ae90f125ddbc-logs\") on node \"crc\" DevicePath \"\"" Oct 08 18:33:47 crc kubenswrapper[4859]: I1008 18:33:47.576928 4859 reconciler_common.go:293] "Volume detached for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/bfc391b0-6e05-41b9-8c78-7cf870f5cfe4-httpd-run\") on node \"crc\" DevicePath \"\"" Oct 08 18:33:47 crc kubenswrapper[4859]: I1008 18:33:47.582400 4859 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/bfc391b0-6e05-41b9-8c78-7cf870f5cfe4-logs" (OuterVolumeSpecName: "logs") pod "bfc391b0-6e05-41b9-8c78-7cf870f5cfe4" (UID: "bfc391b0-6e05-41b9-8c78-7cf870f5cfe4"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 08 18:33:47 crc kubenswrapper[4859]: I1008 18:33:47.582408 4859 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/bfc391b0-6e05-41b9-8c78-7cf870f5cfe4-scripts" (OuterVolumeSpecName: "scripts") pod "bfc391b0-6e05-41b9-8c78-7cf870f5cfe4" (UID: "bfc391b0-6e05-41b9-8c78-7cf870f5cfe4"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 08 18:33:47 crc kubenswrapper[4859]: I1008 18:33:47.583285 4859 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/local-volume/local-storage12-crc" (OuterVolumeSpecName: "glance") pod "bfc391b0-6e05-41b9-8c78-7cf870f5cfe4" (UID: "bfc391b0-6e05-41b9-8c78-7cf870f5cfe4"). InnerVolumeSpecName "local-storage12-crc". PluginName "kubernetes.io/local-volume", VolumeGidValue "" Oct 08 18:33:47 crc kubenswrapper[4859]: I1008 18:33:47.583704 4859 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/9d1b6fef-16a5-4c1b-a3bb-95f46f1d92f7-credential-keys" (OuterVolumeSpecName: "credential-keys") pod "9d1b6fef-16a5-4c1b-a3bb-95f46f1d92f7" (UID: "9d1b6fef-16a5-4c1b-a3bb-95f46f1d92f7"). InnerVolumeSpecName "credential-keys". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 08 18:33:47 crc kubenswrapper[4859]: I1008 18:33:47.584002 4859 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/21a3071c-7cf9-4008-a37b-ae90f125ddbc-kube-api-access-9mgdb" (OuterVolumeSpecName: "kube-api-access-9mgdb") pod "21a3071c-7cf9-4008-a37b-ae90f125ddbc" (UID: "21a3071c-7cf9-4008-a37b-ae90f125ddbc"). InnerVolumeSpecName "kube-api-access-9mgdb". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 08 18:33:47 crc kubenswrapper[4859]: I1008 18:33:47.584064 4859 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/local-volume/local-storage10-crc" (OuterVolumeSpecName: "glance") pod "21a3071c-7cf9-4008-a37b-ae90f125ddbc" (UID: "21a3071c-7cf9-4008-a37b-ae90f125ddbc"). InnerVolumeSpecName "local-storage10-crc". PluginName "kubernetes.io/local-volume", VolumeGidValue "" Oct 08 18:33:47 crc kubenswrapper[4859]: I1008 18:33:47.584456 4859 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/bfc391b0-6e05-41b9-8c78-7cf870f5cfe4-kube-api-access-4kfnf" (OuterVolumeSpecName: "kube-api-access-4kfnf") pod "bfc391b0-6e05-41b9-8c78-7cf870f5cfe4" (UID: "bfc391b0-6e05-41b9-8c78-7cf870f5cfe4"). InnerVolumeSpecName "kube-api-access-4kfnf". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 08 18:33:47 crc kubenswrapper[4859]: I1008 18:33:47.585105 4859 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/9d1b6fef-16a5-4c1b-a3bb-95f46f1d92f7-scripts" (OuterVolumeSpecName: "scripts") pod "9d1b6fef-16a5-4c1b-a3bb-95f46f1d92f7" (UID: "9d1b6fef-16a5-4c1b-a3bb-95f46f1d92f7"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 08 18:33:47 crc kubenswrapper[4859]: I1008 18:33:47.587812 4859 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/9d1b6fef-16a5-4c1b-a3bb-95f46f1d92f7-fernet-keys" (OuterVolumeSpecName: "fernet-keys") pod "9d1b6fef-16a5-4c1b-a3bb-95f46f1d92f7" (UID: "9d1b6fef-16a5-4c1b-a3bb-95f46f1d92f7"). InnerVolumeSpecName "fernet-keys". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 08 18:33:47 crc kubenswrapper[4859]: I1008 18:33:47.597828 4859 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/21a3071c-7cf9-4008-a37b-ae90f125ddbc-scripts" (OuterVolumeSpecName: "scripts") pod "21a3071c-7cf9-4008-a37b-ae90f125ddbc" (UID: "21a3071c-7cf9-4008-a37b-ae90f125ddbc"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 08 18:33:47 crc kubenswrapper[4859]: I1008 18:33:47.633325 4859 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/9d1b6fef-16a5-4c1b-a3bb-95f46f1d92f7-kube-api-access-txptt" (OuterVolumeSpecName: "kube-api-access-txptt") pod "9d1b6fef-16a5-4c1b-a3bb-95f46f1d92f7" (UID: "9d1b6fef-16a5-4c1b-a3bb-95f46f1d92f7"). InnerVolumeSpecName "kube-api-access-txptt". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 08 18:33:47 crc kubenswrapper[4859]: I1008 18:33:47.637243 4859 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/9d1b6fef-16a5-4c1b-a3bb-95f46f1d92f7-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "9d1b6fef-16a5-4c1b-a3bb-95f46f1d92f7" (UID: "9d1b6fef-16a5-4c1b-a3bb-95f46f1d92f7"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 08 18:33:47 crc kubenswrapper[4859]: I1008 18:33:47.639002 4859 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/bfc391b0-6e05-41b9-8c78-7cf870f5cfe4-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "bfc391b0-6e05-41b9-8c78-7cf870f5cfe4" (UID: "bfc391b0-6e05-41b9-8c78-7cf870f5cfe4"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 08 18:33:47 crc kubenswrapper[4859]: I1008 18:33:47.650490 4859 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/9d1b6fef-16a5-4c1b-a3bb-95f46f1d92f7-config-data" (OuterVolumeSpecName: "config-data") pod "9d1b6fef-16a5-4c1b-a3bb-95f46f1d92f7" (UID: "9d1b6fef-16a5-4c1b-a3bb-95f46f1d92f7"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 08 18:33:47 crc kubenswrapper[4859]: I1008 18:33:47.655174 4859 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/21a3071c-7cf9-4008-a37b-ae90f125ddbc-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "21a3071c-7cf9-4008-a37b-ae90f125ddbc" (UID: "21a3071c-7cf9-4008-a37b-ae90f125ddbc"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 08 18:33:47 crc kubenswrapper[4859]: I1008 18:33:47.675410 4859 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/bfc391b0-6e05-41b9-8c78-7cf870f5cfe4-public-tls-certs" (OuterVolumeSpecName: "public-tls-certs") pod "bfc391b0-6e05-41b9-8c78-7cf870f5cfe4" (UID: "bfc391b0-6e05-41b9-8c78-7cf870f5cfe4"). InnerVolumeSpecName "public-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 08 18:33:47 crc kubenswrapper[4859]: I1008 18:33:47.678234 4859 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/21a3071c-7cf9-4008-a37b-ae90f125ddbc-internal-tls-certs" (OuterVolumeSpecName: "internal-tls-certs") pod "21a3071c-7cf9-4008-a37b-ae90f125ddbc" (UID: "21a3071c-7cf9-4008-a37b-ae90f125ddbc"). InnerVolumeSpecName "internal-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 08 18:33:47 crc kubenswrapper[4859]: I1008 18:33:47.678866 4859 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/21a3071c-7cf9-4008-a37b-ae90f125ddbc-scripts\") on node \"crc\" DevicePath \"\"" Oct 08 18:33:47 crc kubenswrapper[4859]: I1008 18:33:47.678907 4859 reconciler_common.go:293] "Volume detached for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/9d1b6fef-16a5-4c1b-a3bb-95f46f1d92f7-credential-keys\") on node \"crc\" DevicePath \"\"" Oct 08 18:33:47 crc kubenswrapper[4859]: I1008 18:33:47.678943 4859 reconciler_common.go:286] "operationExecutor.UnmountDevice started for volume \"local-storage12-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage12-crc\") on node \"crc\" " Oct 08 18:33:47 crc kubenswrapper[4859]: I1008 18:33:47.678959 4859 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/9d1b6fef-16a5-4c1b-a3bb-95f46f1d92f7-scripts\") on node \"crc\" DevicePath \"\"" Oct 08 18:33:47 crc kubenswrapper[4859]: I1008 18:33:47.678974 4859 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-9mgdb\" (UniqueName: \"kubernetes.io/projected/21a3071c-7cf9-4008-a37b-ae90f125ddbc-kube-api-access-9mgdb\") on node \"crc\" DevicePath \"\"" Oct 08 18:33:47 crc kubenswrapper[4859]: I1008 18:33:47.678986 4859 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-txptt\" (UniqueName: \"kubernetes.io/projected/9d1b6fef-16a5-4c1b-a3bb-95f46f1d92f7-kube-api-access-txptt\") on node \"crc\" DevicePath \"\"" Oct 08 18:33:47 crc kubenswrapper[4859]: I1008 18:33:47.678998 4859 reconciler_common.go:293] "Volume detached for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/bfc391b0-6e05-41b9-8c78-7cf870f5cfe4-public-tls-certs\") on node \"crc\" DevicePath \"\"" Oct 08 18:33:47 crc kubenswrapper[4859]: I1008 18:33:47.679009 4859 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/21a3071c-7cf9-4008-a37b-ae90f125ddbc-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 08 18:33:47 crc kubenswrapper[4859]: I1008 18:33:47.679022 4859 reconciler_common.go:293] "Volume detached for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/21a3071c-7cf9-4008-a37b-ae90f125ddbc-internal-tls-certs\") on node \"crc\" DevicePath \"\"" Oct 08 18:33:47 crc kubenswrapper[4859]: I1008 18:33:47.679033 4859 reconciler_common.go:293] "Volume detached for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/9d1b6fef-16a5-4c1b-a3bb-95f46f1d92f7-fernet-keys\") on node \"crc\" DevicePath \"\"" Oct 08 18:33:47 crc kubenswrapper[4859]: I1008 18:33:47.679044 4859 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/9d1b6fef-16a5-4c1b-a3bb-95f46f1d92f7-config-data\") on node \"crc\" DevicePath \"\"" Oct 08 18:33:47 crc kubenswrapper[4859]: I1008 18:33:47.679063 4859 reconciler_common.go:286] "operationExecutor.UnmountDevice started for volume \"local-storage10-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage10-crc\") on node \"crc\" " Oct 08 18:33:47 crc kubenswrapper[4859]: I1008 18:33:47.679076 4859 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/bfc391b0-6e05-41b9-8c78-7cf870f5cfe4-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 08 18:33:47 crc kubenswrapper[4859]: I1008 18:33:47.679087 4859 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/bfc391b0-6e05-41b9-8c78-7cf870f5cfe4-logs\") on node \"crc\" DevicePath \"\"" Oct 08 18:33:47 crc kubenswrapper[4859]: I1008 18:33:47.679098 4859 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9d1b6fef-16a5-4c1b-a3bb-95f46f1d92f7-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 08 18:33:47 crc kubenswrapper[4859]: I1008 18:33:47.679110 4859 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/bfc391b0-6e05-41b9-8c78-7cf870f5cfe4-scripts\") on node \"crc\" DevicePath \"\"" Oct 08 18:33:47 crc kubenswrapper[4859]: I1008 18:33:47.679121 4859 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-4kfnf\" (UniqueName: \"kubernetes.io/projected/bfc391b0-6e05-41b9-8c78-7cf870f5cfe4-kube-api-access-4kfnf\") on node \"crc\" DevicePath \"\"" Oct 08 18:33:47 crc kubenswrapper[4859]: I1008 18:33:47.694683 4859 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/bfc391b0-6e05-41b9-8c78-7cf870f5cfe4-config-data" (OuterVolumeSpecName: "config-data") pod "bfc391b0-6e05-41b9-8c78-7cf870f5cfe4" (UID: "bfc391b0-6e05-41b9-8c78-7cf870f5cfe4"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 08 18:33:47 crc kubenswrapper[4859]: I1008 18:33:47.696857 4859 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/21a3071c-7cf9-4008-a37b-ae90f125ddbc-config-data" (OuterVolumeSpecName: "config-data") pod "21a3071c-7cf9-4008-a37b-ae90f125ddbc" (UID: "21a3071c-7cf9-4008-a37b-ae90f125ddbc"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 08 18:33:47 crc kubenswrapper[4859]: I1008 18:33:47.713661 4859 operation_generator.go:917] UnmountDevice succeeded for volume "local-storage10-crc" (UniqueName: "kubernetes.io/local-volume/local-storage10-crc") on node "crc" Oct 08 18:33:47 crc kubenswrapper[4859]: I1008 18:33:47.714822 4859 operation_generator.go:917] UnmountDevice succeeded for volume "local-storage12-crc" (UniqueName: "kubernetes.io/local-volume/local-storage12-crc") on node "crc" Oct 08 18:33:47 crc kubenswrapper[4859]: I1008 18:33:47.780471 4859 reconciler_common.go:293] "Volume detached for volume \"local-storage10-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage10-crc\") on node \"crc\" DevicePath \"\"" Oct 08 18:33:47 crc kubenswrapper[4859]: I1008 18:33:47.780507 4859 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/bfc391b0-6e05-41b9-8c78-7cf870f5cfe4-config-data\") on node \"crc\" DevicePath \"\"" Oct 08 18:33:47 crc kubenswrapper[4859]: I1008 18:33:47.780519 4859 reconciler_common.go:293] "Volume detached for volume \"local-storage12-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage12-crc\") on node \"crc\" DevicePath \"\"" Oct 08 18:33:47 crc kubenswrapper[4859]: I1008 18:33:47.780527 4859 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/21a3071c-7cf9-4008-a37b-ae90f125ddbc-config-data\") on node \"crc\" DevicePath \"\"" Oct 08 18:33:47 crc kubenswrapper[4859]: I1008 18:33:47.924946 4859 patch_prober.go:28] interesting pod/machine-config-daemon-82s52 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 08 18:33:47 crc kubenswrapper[4859]: I1008 18:33:47.925318 4859 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-82s52" podUID="b23a6a6c-9d92-4e7b-840e-55cfda873a2d" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 08 18:33:48 crc kubenswrapper[4859]: I1008 18:33:48.170941 4859 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"bfc391b0-6e05-41b9-8c78-7cf870f5cfe4","Type":"ContainerDied","Data":"2ad247c3a483c8e2e59d1bc0aa8d4df493b21a6998eb9b1922b243053a801353"} Oct 08 18:33:48 crc kubenswrapper[4859]: I1008 18:33:48.170992 4859 scope.go:117] "RemoveContainer" containerID="092813e771436337b8a8182b40e005937742352405076340ee1eb8c6aa53c4ca" Oct 08 18:33:48 crc kubenswrapper[4859]: I1008 18:33:48.171128 4859 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-external-api-0" Oct 08 18:33:48 crc kubenswrapper[4859]: I1008 18:33:48.174450 4859 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-bootstrap-mqwmv" event={"ID":"9d1b6fef-16a5-4c1b-a3bb-95f46f1d92f7","Type":"ContainerDied","Data":"082ec68b2336d53fa12edeb6e2e6896b2dfcb72c1887ba0fb688c69dc1e894f3"} Oct 08 18:33:48 crc kubenswrapper[4859]: I1008 18:33:48.174501 4859 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="082ec68b2336d53fa12edeb6e2e6896b2dfcb72c1887ba0fb688c69dc1e894f3" Oct 08 18:33:48 crc kubenswrapper[4859]: I1008 18:33:48.174472 4859 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-bootstrap-mqwmv" Oct 08 18:33:48 crc kubenswrapper[4859]: I1008 18:33:48.176906 4859 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"21a3071c-7cf9-4008-a37b-ae90f125ddbc","Type":"ContainerDied","Data":"f6b7101c206d49941deb736f964625f6291e2b703f2978608c79d5f676a49527"} Oct 08 18:33:48 crc kubenswrapper[4859]: I1008 18:33:48.176972 4859 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-internal-api-0" Oct 08 18:33:48 crc kubenswrapper[4859]: I1008 18:33:48.227867 4859 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-default-external-api-0"] Oct 08 18:33:48 crc kubenswrapper[4859]: I1008 18:33:48.239601 4859 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/glance-default-external-api-0"] Oct 08 18:33:48 crc kubenswrapper[4859]: I1008 18:33:48.250493 4859 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-default-internal-api-0"] Oct 08 18:33:48 crc kubenswrapper[4859]: I1008 18:33:48.269198 4859 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/glance-default-external-api-0"] Oct 08 18:33:48 crc kubenswrapper[4859]: E1008 18:33:48.269604 4859 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9d1b6fef-16a5-4c1b-a3bb-95f46f1d92f7" containerName="keystone-bootstrap" Oct 08 18:33:48 crc kubenswrapper[4859]: I1008 18:33:48.269621 4859 state_mem.go:107] "Deleted CPUSet assignment" podUID="9d1b6fef-16a5-4c1b-a3bb-95f46f1d92f7" containerName="keystone-bootstrap" Oct 08 18:33:48 crc kubenswrapper[4859]: E1008 18:33:48.269635 4859 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="bfc391b0-6e05-41b9-8c78-7cf870f5cfe4" containerName="glance-httpd" Oct 08 18:33:48 crc kubenswrapper[4859]: I1008 18:33:48.269641 4859 state_mem.go:107] "Deleted CPUSet assignment" podUID="bfc391b0-6e05-41b9-8c78-7cf870f5cfe4" containerName="glance-httpd" Oct 08 18:33:48 crc kubenswrapper[4859]: E1008 18:33:48.269649 4859 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="bfc391b0-6e05-41b9-8c78-7cf870f5cfe4" containerName="glance-log" Oct 08 18:33:48 crc kubenswrapper[4859]: I1008 18:33:48.269656 4859 state_mem.go:107] "Deleted CPUSet assignment" podUID="bfc391b0-6e05-41b9-8c78-7cf870f5cfe4" containerName="glance-log" Oct 08 18:33:48 crc kubenswrapper[4859]: E1008 18:33:48.269675 4859 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="21a3071c-7cf9-4008-a37b-ae90f125ddbc" containerName="glance-log" Oct 08 18:33:48 crc kubenswrapper[4859]: I1008 18:33:48.269702 4859 state_mem.go:107] "Deleted CPUSet assignment" podUID="21a3071c-7cf9-4008-a37b-ae90f125ddbc" containerName="glance-log" Oct 08 18:33:48 crc kubenswrapper[4859]: E1008 18:33:48.269729 4859 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="21a3071c-7cf9-4008-a37b-ae90f125ddbc" containerName="glance-httpd" Oct 08 18:33:48 crc kubenswrapper[4859]: I1008 18:33:48.269735 4859 state_mem.go:107] "Deleted CPUSet assignment" podUID="21a3071c-7cf9-4008-a37b-ae90f125ddbc" containerName="glance-httpd" Oct 08 18:33:48 crc kubenswrapper[4859]: I1008 18:33:48.269884 4859 memory_manager.go:354] "RemoveStaleState removing state" podUID="bfc391b0-6e05-41b9-8c78-7cf870f5cfe4" containerName="glance-httpd" Oct 08 18:33:48 crc kubenswrapper[4859]: I1008 18:33:48.269897 4859 memory_manager.go:354] "RemoveStaleState removing state" podUID="21a3071c-7cf9-4008-a37b-ae90f125ddbc" containerName="glance-httpd" Oct 08 18:33:48 crc kubenswrapper[4859]: I1008 18:33:48.269913 4859 memory_manager.go:354] "RemoveStaleState removing state" podUID="bfc391b0-6e05-41b9-8c78-7cf870f5cfe4" containerName="glance-log" Oct 08 18:33:48 crc kubenswrapper[4859]: I1008 18:33:48.269920 4859 memory_manager.go:354] "RemoveStaleState removing state" podUID="9d1b6fef-16a5-4c1b-a3bb-95f46f1d92f7" containerName="keystone-bootstrap" Oct 08 18:33:48 crc kubenswrapper[4859]: I1008 18:33:48.269930 4859 memory_manager.go:354] "RemoveStaleState removing state" podUID="21a3071c-7cf9-4008-a37b-ae90f125ddbc" containerName="glance-log" Oct 08 18:33:48 crc kubenswrapper[4859]: I1008 18:33:48.270838 4859 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-external-api-0" Oct 08 18:33:48 crc kubenswrapper[4859]: I1008 18:33:48.275193 4859 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-default-external-config-data" Oct 08 18:33:48 crc kubenswrapper[4859]: I1008 18:33:48.275410 4859 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-glance-dockercfg-2fnvk" Oct 08 18:33:48 crc kubenswrapper[4859]: I1008 18:33:48.275506 4859 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-scripts" Oct 08 18:33:48 crc kubenswrapper[4859]: I1008 18:33:48.276776 4859 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-glance-default-public-svc" Oct 08 18:33:48 crc kubenswrapper[4859]: I1008 18:33:48.277322 4859 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/glance-default-internal-api-0"] Oct 08 18:33:48 crc kubenswrapper[4859]: I1008 18:33:48.286921 4859 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-external-api-0"] Oct 08 18:33:48 crc kubenswrapper[4859]: I1008 18:33:48.287524 4859 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/b8433969-f079-49a2-95f3-ce4783d6a62c-config-data\") pod \"glance-default-external-api-0\" (UID: \"b8433969-f079-49a2-95f3-ce4783d6a62c\") " pod="openstack/glance-default-external-api-0" Oct 08 18:33:48 crc kubenswrapper[4859]: I1008 18:33:48.287585 4859 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/b8433969-f079-49a2-95f3-ce4783d6a62c-public-tls-certs\") pod \"glance-default-external-api-0\" (UID: \"b8433969-f079-49a2-95f3-ce4783d6a62c\") " pod="openstack/glance-default-external-api-0" Oct 08 18:33:48 crc kubenswrapper[4859]: I1008 18:33:48.287626 4859 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/b8433969-f079-49a2-95f3-ce4783d6a62c-scripts\") pod \"glance-default-external-api-0\" (UID: \"b8433969-f079-49a2-95f3-ce4783d6a62c\") " pod="openstack/glance-default-external-api-0" Oct 08 18:33:48 crc kubenswrapper[4859]: I1008 18:33:48.287660 4859 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/b8433969-f079-49a2-95f3-ce4783d6a62c-httpd-run\") pod \"glance-default-external-api-0\" (UID: \"b8433969-f079-49a2-95f3-ce4783d6a62c\") " pod="openstack/glance-default-external-api-0" Oct 08 18:33:48 crc kubenswrapper[4859]: I1008 18:33:48.287705 4859 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage12-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage12-crc\") pod \"glance-default-external-api-0\" (UID: \"b8433969-f079-49a2-95f3-ce4783d6a62c\") " pod="openstack/glance-default-external-api-0" Oct 08 18:33:48 crc kubenswrapper[4859]: I1008 18:33:48.287723 4859 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/b8433969-f079-49a2-95f3-ce4783d6a62c-logs\") pod \"glance-default-external-api-0\" (UID: \"b8433969-f079-49a2-95f3-ce4783d6a62c\") " pod="openstack/glance-default-external-api-0" Oct 08 18:33:48 crc kubenswrapper[4859]: I1008 18:33:48.287762 4859 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b8433969-f079-49a2-95f3-ce4783d6a62c-combined-ca-bundle\") pod \"glance-default-external-api-0\" (UID: \"b8433969-f079-49a2-95f3-ce4783d6a62c\") " pod="openstack/glance-default-external-api-0" Oct 08 18:33:48 crc kubenswrapper[4859]: I1008 18:33:48.287780 4859 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-xvctl\" (UniqueName: \"kubernetes.io/projected/b8433969-f079-49a2-95f3-ce4783d6a62c-kube-api-access-xvctl\") pod \"glance-default-external-api-0\" (UID: \"b8433969-f079-49a2-95f3-ce4783d6a62c\") " pod="openstack/glance-default-external-api-0" Oct 08 18:33:48 crc kubenswrapper[4859]: I1008 18:33:48.310527 4859 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/glance-default-internal-api-0"] Oct 08 18:33:48 crc kubenswrapper[4859]: I1008 18:33:48.312372 4859 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-internal-api-0" Oct 08 18:33:48 crc kubenswrapper[4859]: I1008 18:33:48.315555 4859 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-glance-default-internal-svc" Oct 08 18:33:48 crc kubenswrapper[4859]: I1008 18:33:48.315742 4859 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-default-internal-config-data" Oct 08 18:33:48 crc kubenswrapper[4859]: I1008 18:33:48.322884 4859 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-internal-api-0"] Oct 08 18:33:48 crc kubenswrapper[4859]: I1008 18:33:48.389598 4859 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/b8433969-f079-49a2-95f3-ce4783d6a62c-config-data\") pod \"glance-default-external-api-0\" (UID: \"b8433969-f079-49a2-95f3-ce4783d6a62c\") " pod="openstack/glance-default-external-api-0" Oct 08 18:33:48 crc kubenswrapper[4859]: I1008 18:33:48.389760 4859 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/b8433969-f079-49a2-95f3-ce4783d6a62c-public-tls-certs\") pod \"glance-default-external-api-0\" (UID: \"b8433969-f079-49a2-95f3-ce4783d6a62c\") " pod="openstack/glance-default-external-api-0" Oct 08 18:33:48 crc kubenswrapper[4859]: I1008 18:33:48.389847 4859 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/b8433969-f079-49a2-95f3-ce4783d6a62c-scripts\") pod \"glance-default-external-api-0\" (UID: \"b8433969-f079-49a2-95f3-ce4783d6a62c\") " pod="openstack/glance-default-external-api-0" Oct 08 18:33:48 crc kubenswrapper[4859]: I1008 18:33:48.389893 4859 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/b8433969-f079-49a2-95f3-ce4783d6a62c-httpd-run\") pod \"glance-default-external-api-0\" (UID: \"b8433969-f079-49a2-95f3-ce4783d6a62c\") " pod="openstack/glance-default-external-api-0" Oct 08 18:33:48 crc kubenswrapper[4859]: I1008 18:33:48.389926 4859 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage12-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage12-crc\") pod \"glance-default-external-api-0\" (UID: \"b8433969-f079-49a2-95f3-ce4783d6a62c\") " pod="openstack/glance-default-external-api-0" Oct 08 18:33:48 crc kubenswrapper[4859]: I1008 18:33:48.389956 4859 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/b8433969-f079-49a2-95f3-ce4783d6a62c-logs\") pod \"glance-default-external-api-0\" (UID: \"b8433969-f079-49a2-95f3-ce4783d6a62c\") " pod="openstack/glance-default-external-api-0" Oct 08 18:33:48 crc kubenswrapper[4859]: I1008 18:33:48.390009 4859 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b8433969-f079-49a2-95f3-ce4783d6a62c-combined-ca-bundle\") pod \"glance-default-external-api-0\" (UID: \"b8433969-f079-49a2-95f3-ce4783d6a62c\") " pod="openstack/glance-default-external-api-0" Oct 08 18:33:48 crc kubenswrapper[4859]: I1008 18:33:48.390031 4859 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-xvctl\" (UniqueName: \"kubernetes.io/projected/b8433969-f079-49a2-95f3-ce4783d6a62c-kube-api-access-xvctl\") pod \"glance-default-external-api-0\" (UID: \"b8433969-f079-49a2-95f3-ce4783d6a62c\") " pod="openstack/glance-default-external-api-0" Oct 08 18:33:48 crc kubenswrapper[4859]: I1008 18:33:48.390915 4859 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage12-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage12-crc\") pod \"glance-default-external-api-0\" (UID: \"b8433969-f079-49a2-95f3-ce4783d6a62c\") device mount path \"/mnt/openstack/pv12\"" pod="openstack/glance-default-external-api-0" Oct 08 18:33:48 crc kubenswrapper[4859]: I1008 18:33:48.391953 4859 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/b8433969-f079-49a2-95f3-ce4783d6a62c-httpd-run\") pod \"glance-default-external-api-0\" (UID: \"b8433969-f079-49a2-95f3-ce4783d6a62c\") " pod="openstack/glance-default-external-api-0" Oct 08 18:33:48 crc kubenswrapper[4859]: I1008 18:33:48.395575 4859 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/b8433969-f079-49a2-95f3-ce4783d6a62c-config-data\") pod \"glance-default-external-api-0\" (UID: \"b8433969-f079-49a2-95f3-ce4783d6a62c\") " pod="openstack/glance-default-external-api-0" Oct 08 18:33:48 crc kubenswrapper[4859]: I1008 18:33:48.396102 4859 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/b8433969-f079-49a2-95f3-ce4783d6a62c-logs\") pod \"glance-default-external-api-0\" (UID: \"b8433969-f079-49a2-95f3-ce4783d6a62c\") " pod="openstack/glance-default-external-api-0" Oct 08 18:33:48 crc kubenswrapper[4859]: I1008 18:33:48.398945 4859 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/b8433969-f079-49a2-95f3-ce4783d6a62c-public-tls-certs\") pod \"glance-default-external-api-0\" (UID: \"b8433969-f079-49a2-95f3-ce4783d6a62c\") " pod="openstack/glance-default-external-api-0" Oct 08 18:33:48 crc kubenswrapper[4859]: I1008 18:33:48.399359 4859 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b8433969-f079-49a2-95f3-ce4783d6a62c-combined-ca-bundle\") pod \"glance-default-external-api-0\" (UID: \"b8433969-f079-49a2-95f3-ce4783d6a62c\") " pod="openstack/glance-default-external-api-0" Oct 08 18:33:48 crc kubenswrapper[4859]: I1008 18:33:48.401321 4859 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/b8433969-f079-49a2-95f3-ce4783d6a62c-scripts\") pod \"glance-default-external-api-0\" (UID: \"b8433969-f079-49a2-95f3-ce4783d6a62c\") " pod="openstack/glance-default-external-api-0" Oct 08 18:33:48 crc kubenswrapper[4859]: I1008 18:33:48.409314 4859 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-xvctl\" (UniqueName: \"kubernetes.io/projected/b8433969-f079-49a2-95f3-ce4783d6a62c-kube-api-access-xvctl\") pod \"glance-default-external-api-0\" (UID: \"b8433969-f079-49a2-95f3-ce4783d6a62c\") " pod="openstack/glance-default-external-api-0" Oct 08 18:33:48 crc kubenswrapper[4859]: I1008 18:33:48.444167 4859 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage12-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage12-crc\") pod \"glance-default-external-api-0\" (UID: \"b8433969-f079-49a2-95f3-ce4783d6a62c\") " pod="openstack/glance-default-external-api-0" Oct 08 18:33:48 crc kubenswrapper[4859]: I1008 18:33:48.493038 4859 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/7bf3f855-41ae-4de9-ba6c-0fb4f51df141-config-data\") pod \"glance-default-internal-api-0\" (UID: \"7bf3f855-41ae-4de9-ba6c-0fb4f51df141\") " pod="openstack/glance-default-internal-api-0" Oct 08 18:33:48 crc kubenswrapper[4859]: I1008 18:33:48.493122 4859 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7bf3f855-41ae-4de9-ba6c-0fb4f51df141-combined-ca-bundle\") pod \"glance-default-internal-api-0\" (UID: \"7bf3f855-41ae-4de9-ba6c-0fb4f51df141\") " pod="openstack/glance-default-internal-api-0" Oct 08 18:33:48 crc kubenswrapper[4859]: I1008 18:33:48.493161 4859 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/7bf3f855-41ae-4de9-ba6c-0fb4f51df141-internal-tls-certs\") pod \"glance-default-internal-api-0\" (UID: \"7bf3f855-41ae-4de9-ba6c-0fb4f51df141\") " pod="openstack/glance-default-internal-api-0" Oct 08 18:33:48 crc kubenswrapper[4859]: I1008 18:33:48.493188 4859 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/7bf3f855-41ae-4de9-ba6c-0fb4f51df141-httpd-run\") pod \"glance-default-internal-api-0\" (UID: \"7bf3f855-41ae-4de9-ba6c-0fb4f51df141\") " pod="openstack/glance-default-internal-api-0" Oct 08 18:33:48 crc kubenswrapper[4859]: I1008 18:33:48.493249 4859 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage10-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage10-crc\") pod \"glance-default-internal-api-0\" (UID: \"7bf3f855-41ae-4de9-ba6c-0fb4f51df141\") " pod="openstack/glance-default-internal-api-0" Oct 08 18:33:48 crc kubenswrapper[4859]: I1008 18:33:48.493279 4859 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/7bf3f855-41ae-4de9-ba6c-0fb4f51df141-logs\") pod \"glance-default-internal-api-0\" (UID: \"7bf3f855-41ae-4de9-ba6c-0fb4f51df141\") " pod="openstack/glance-default-internal-api-0" Oct 08 18:33:48 crc kubenswrapper[4859]: I1008 18:33:48.493395 4859 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-4ms4c\" (UniqueName: \"kubernetes.io/projected/7bf3f855-41ae-4de9-ba6c-0fb4f51df141-kube-api-access-4ms4c\") pod \"glance-default-internal-api-0\" (UID: \"7bf3f855-41ae-4de9-ba6c-0fb4f51df141\") " pod="openstack/glance-default-internal-api-0" Oct 08 18:33:48 crc kubenswrapper[4859]: I1008 18:33:48.493424 4859 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/7bf3f855-41ae-4de9-ba6c-0fb4f51df141-scripts\") pod \"glance-default-internal-api-0\" (UID: \"7bf3f855-41ae-4de9-ba6c-0fb4f51df141\") " pod="openstack/glance-default-internal-api-0" Oct 08 18:33:48 crc kubenswrapper[4859]: I1008 18:33:48.512803 4859 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="21a3071c-7cf9-4008-a37b-ae90f125ddbc" path="/var/lib/kubelet/pods/21a3071c-7cf9-4008-a37b-ae90f125ddbc/volumes" Oct 08 18:33:48 crc kubenswrapper[4859]: I1008 18:33:48.514480 4859 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="bfc391b0-6e05-41b9-8c78-7cf870f5cfe4" path="/var/lib/kubelet/pods/bfc391b0-6e05-41b9-8c78-7cf870f5cfe4/volumes" Oct 08 18:33:48 crc kubenswrapper[4859]: I1008 18:33:48.594520 4859 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/7bf3f855-41ae-4de9-ba6c-0fb4f51df141-scripts\") pod \"glance-default-internal-api-0\" (UID: \"7bf3f855-41ae-4de9-ba6c-0fb4f51df141\") " pod="openstack/glance-default-internal-api-0" Oct 08 18:33:48 crc kubenswrapper[4859]: I1008 18:33:48.594605 4859 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/7bf3f855-41ae-4de9-ba6c-0fb4f51df141-config-data\") pod \"glance-default-internal-api-0\" (UID: \"7bf3f855-41ae-4de9-ba6c-0fb4f51df141\") " pod="openstack/glance-default-internal-api-0" Oct 08 18:33:48 crc kubenswrapper[4859]: I1008 18:33:48.594642 4859 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7bf3f855-41ae-4de9-ba6c-0fb4f51df141-combined-ca-bundle\") pod \"glance-default-internal-api-0\" (UID: \"7bf3f855-41ae-4de9-ba6c-0fb4f51df141\") " pod="openstack/glance-default-internal-api-0" Oct 08 18:33:48 crc kubenswrapper[4859]: I1008 18:33:48.594675 4859 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/7bf3f855-41ae-4de9-ba6c-0fb4f51df141-internal-tls-certs\") pod \"glance-default-internal-api-0\" (UID: \"7bf3f855-41ae-4de9-ba6c-0fb4f51df141\") " pod="openstack/glance-default-internal-api-0" Oct 08 18:33:48 crc kubenswrapper[4859]: I1008 18:33:48.594716 4859 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/7bf3f855-41ae-4de9-ba6c-0fb4f51df141-httpd-run\") pod \"glance-default-internal-api-0\" (UID: \"7bf3f855-41ae-4de9-ba6c-0fb4f51df141\") " pod="openstack/glance-default-internal-api-0" Oct 08 18:33:48 crc kubenswrapper[4859]: I1008 18:33:48.594763 4859 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage10-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage10-crc\") pod \"glance-default-internal-api-0\" (UID: \"7bf3f855-41ae-4de9-ba6c-0fb4f51df141\") " pod="openstack/glance-default-internal-api-0" Oct 08 18:33:48 crc kubenswrapper[4859]: I1008 18:33:48.594785 4859 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/7bf3f855-41ae-4de9-ba6c-0fb4f51df141-logs\") pod \"glance-default-internal-api-0\" (UID: \"7bf3f855-41ae-4de9-ba6c-0fb4f51df141\") " pod="openstack/glance-default-internal-api-0" Oct 08 18:33:48 crc kubenswrapper[4859]: I1008 18:33:48.594917 4859 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-4ms4c\" (UniqueName: \"kubernetes.io/projected/7bf3f855-41ae-4de9-ba6c-0fb4f51df141-kube-api-access-4ms4c\") pod \"glance-default-internal-api-0\" (UID: \"7bf3f855-41ae-4de9-ba6c-0fb4f51df141\") " pod="openstack/glance-default-internal-api-0" Oct 08 18:33:48 crc kubenswrapper[4859]: I1008 18:33:48.595352 4859 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage10-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage10-crc\") pod \"glance-default-internal-api-0\" (UID: \"7bf3f855-41ae-4de9-ba6c-0fb4f51df141\") device mount path \"/mnt/openstack/pv10\"" pod="openstack/glance-default-internal-api-0" Oct 08 18:33:48 crc kubenswrapper[4859]: I1008 18:33:48.596538 4859 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/7bf3f855-41ae-4de9-ba6c-0fb4f51df141-logs\") pod \"glance-default-internal-api-0\" (UID: \"7bf3f855-41ae-4de9-ba6c-0fb4f51df141\") " pod="openstack/glance-default-internal-api-0" Oct 08 18:33:48 crc kubenswrapper[4859]: I1008 18:33:48.598574 4859 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/7bf3f855-41ae-4de9-ba6c-0fb4f51df141-httpd-run\") pod \"glance-default-internal-api-0\" (UID: \"7bf3f855-41ae-4de9-ba6c-0fb4f51df141\") " pod="openstack/glance-default-internal-api-0" Oct 08 18:33:48 crc kubenswrapper[4859]: I1008 18:33:48.599440 4859 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/7bf3f855-41ae-4de9-ba6c-0fb4f51df141-config-data\") pod \"glance-default-internal-api-0\" (UID: \"7bf3f855-41ae-4de9-ba6c-0fb4f51df141\") " pod="openstack/glance-default-internal-api-0" Oct 08 18:33:48 crc kubenswrapper[4859]: I1008 18:33:48.602537 4859 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7bf3f855-41ae-4de9-ba6c-0fb4f51df141-combined-ca-bundle\") pod \"glance-default-internal-api-0\" (UID: \"7bf3f855-41ae-4de9-ba6c-0fb4f51df141\") " pod="openstack/glance-default-internal-api-0" Oct 08 18:33:48 crc kubenswrapper[4859]: I1008 18:33:48.603539 4859 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/7bf3f855-41ae-4de9-ba6c-0fb4f51df141-scripts\") pod \"glance-default-internal-api-0\" (UID: \"7bf3f855-41ae-4de9-ba6c-0fb4f51df141\") " pod="openstack/glance-default-internal-api-0" Oct 08 18:33:48 crc kubenswrapper[4859]: I1008 18:33:48.609331 4859 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/7bf3f855-41ae-4de9-ba6c-0fb4f51df141-internal-tls-certs\") pod \"glance-default-internal-api-0\" (UID: \"7bf3f855-41ae-4de9-ba6c-0fb4f51df141\") " pod="openstack/glance-default-internal-api-0" Oct 08 18:33:48 crc kubenswrapper[4859]: I1008 18:33:48.623297 4859 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/keystone-bootstrap-mqwmv"] Oct 08 18:33:48 crc kubenswrapper[4859]: I1008 18:33:48.627508 4859 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-4ms4c\" (UniqueName: \"kubernetes.io/projected/7bf3f855-41ae-4de9-ba6c-0fb4f51df141-kube-api-access-4ms4c\") pod \"glance-default-internal-api-0\" (UID: \"7bf3f855-41ae-4de9-ba6c-0fb4f51df141\") " pod="openstack/glance-default-internal-api-0" Oct 08 18:33:48 crc kubenswrapper[4859]: I1008 18:33:48.627868 4859 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/keystone-bootstrap-mqwmv"] Oct 08 18:33:48 crc kubenswrapper[4859]: I1008 18:33:48.629952 4859 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-external-api-0" Oct 08 18:33:48 crc kubenswrapper[4859]: I1008 18:33:48.639994 4859 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage10-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage10-crc\") pod \"glance-default-internal-api-0\" (UID: \"7bf3f855-41ae-4de9-ba6c-0fb4f51df141\") " pod="openstack/glance-default-internal-api-0" Oct 08 18:33:48 crc kubenswrapper[4859]: I1008 18:33:48.647185 4859 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-internal-api-0" Oct 08 18:33:48 crc kubenswrapper[4859]: I1008 18:33:48.713972 4859 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/keystone-bootstrap-75br8"] Oct 08 18:33:48 crc kubenswrapper[4859]: I1008 18:33:48.715380 4859 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-bootstrap-75br8" Oct 08 18:33:48 crc kubenswrapper[4859]: I1008 18:33:48.718305 4859 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-scripts" Oct 08 18:33:48 crc kubenswrapper[4859]: I1008 18:33:48.718515 4859 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone" Oct 08 18:33:48 crc kubenswrapper[4859]: I1008 18:33:48.718787 4859 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-config-data" Oct 08 18:33:48 crc kubenswrapper[4859]: I1008 18:33:48.719094 4859 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-keystone-dockercfg-jrkml" Oct 08 18:33:48 crc kubenswrapper[4859]: I1008 18:33:48.722295 4859 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-bootstrap-75br8"] Oct 08 18:33:48 crc kubenswrapper[4859]: I1008 18:33:48.798730 4859 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/91a7b430-b4d6-49f6-8951-0d62d51d6685-combined-ca-bundle\") pod \"keystone-bootstrap-75br8\" (UID: \"91a7b430-b4d6-49f6-8951-0d62d51d6685\") " pod="openstack/keystone-bootstrap-75br8" Oct 08 18:33:48 crc kubenswrapper[4859]: I1008 18:33:48.798878 4859 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-bhq6l\" (UniqueName: \"kubernetes.io/projected/91a7b430-b4d6-49f6-8951-0d62d51d6685-kube-api-access-bhq6l\") pod \"keystone-bootstrap-75br8\" (UID: \"91a7b430-b4d6-49f6-8951-0d62d51d6685\") " pod="openstack/keystone-bootstrap-75br8" Oct 08 18:33:48 crc kubenswrapper[4859]: I1008 18:33:48.799020 4859 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/91a7b430-b4d6-49f6-8951-0d62d51d6685-credential-keys\") pod \"keystone-bootstrap-75br8\" (UID: \"91a7b430-b4d6-49f6-8951-0d62d51d6685\") " pod="openstack/keystone-bootstrap-75br8" Oct 08 18:33:48 crc kubenswrapper[4859]: I1008 18:33:48.799133 4859 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/91a7b430-b4d6-49f6-8951-0d62d51d6685-config-data\") pod \"keystone-bootstrap-75br8\" (UID: \"91a7b430-b4d6-49f6-8951-0d62d51d6685\") " pod="openstack/keystone-bootstrap-75br8" Oct 08 18:33:48 crc kubenswrapper[4859]: I1008 18:33:48.799183 4859 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/91a7b430-b4d6-49f6-8951-0d62d51d6685-fernet-keys\") pod \"keystone-bootstrap-75br8\" (UID: \"91a7b430-b4d6-49f6-8951-0d62d51d6685\") " pod="openstack/keystone-bootstrap-75br8" Oct 08 18:33:48 crc kubenswrapper[4859]: I1008 18:33:48.799235 4859 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/91a7b430-b4d6-49f6-8951-0d62d51d6685-scripts\") pod \"keystone-bootstrap-75br8\" (UID: \"91a7b430-b4d6-49f6-8951-0d62d51d6685\") " pod="openstack/keystone-bootstrap-75br8" Oct 08 18:33:48 crc kubenswrapper[4859]: I1008 18:33:48.901372 4859 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/91a7b430-b4d6-49f6-8951-0d62d51d6685-combined-ca-bundle\") pod \"keystone-bootstrap-75br8\" (UID: \"91a7b430-b4d6-49f6-8951-0d62d51d6685\") " pod="openstack/keystone-bootstrap-75br8" Oct 08 18:33:48 crc kubenswrapper[4859]: I1008 18:33:48.901477 4859 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-bhq6l\" (UniqueName: \"kubernetes.io/projected/91a7b430-b4d6-49f6-8951-0d62d51d6685-kube-api-access-bhq6l\") pod \"keystone-bootstrap-75br8\" (UID: \"91a7b430-b4d6-49f6-8951-0d62d51d6685\") " pod="openstack/keystone-bootstrap-75br8" Oct 08 18:33:48 crc kubenswrapper[4859]: I1008 18:33:48.901513 4859 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/91a7b430-b4d6-49f6-8951-0d62d51d6685-credential-keys\") pod \"keystone-bootstrap-75br8\" (UID: \"91a7b430-b4d6-49f6-8951-0d62d51d6685\") " pod="openstack/keystone-bootstrap-75br8" Oct 08 18:33:48 crc kubenswrapper[4859]: I1008 18:33:48.901551 4859 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/91a7b430-b4d6-49f6-8951-0d62d51d6685-config-data\") pod \"keystone-bootstrap-75br8\" (UID: \"91a7b430-b4d6-49f6-8951-0d62d51d6685\") " pod="openstack/keystone-bootstrap-75br8" Oct 08 18:33:48 crc kubenswrapper[4859]: I1008 18:33:48.901581 4859 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/91a7b430-b4d6-49f6-8951-0d62d51d6685-fernet-keys\") pod \"keystone-bootstrap-75br8\" (UID: \"91a7b430-b4d6-49f6-8951-0d62d51d6685\") " pod="openstack/keystone-bootstrap-75br8" Oct 08 18:33:48 crc kubenswrapper[4859]: I1008 18:33:48.901625 4859 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/91a7b430-b4d6-49f6-8951-0d62d51d6685-scripts\") pod \"keystone-bootstrap-75br8\" (UID: \"91a7b430-b4d6-49f6-8951-0d62d51d6685\") " pod="openstack/keystone-bootstrap-75br8" Oct 08 18:33:48 crc kubenswrapper[4859]: I1008 18:33:48.905660 4859 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/91a7b430-b4d6-49f6-8951-0d62d51d6685-config-data\") pod \"keystone-bootstrap-75br8\" (UID: \"91a7b430-b4d6-49f6-8951-0d62d51d6685\") " pod="openstack/keystone-bootstrap-75br8" Oct 08 18:33:48 crc kubenswrapper[4859]: I1008 18:33:48.906078 4859 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/91a7b430-b4d6-49f6-8951-0d62d51d6685-combined-ca-bundle\") pod \"keystone-bootstrap-75br8\" (UID: \"91a7b430-b4d6-49f6-8951-0d62d51d6685\") " pod="openstack/keystone-bootstrap-75br8" Oct 08 18:33:48 crc kubenswrapper[4859]: I1008 18:33:48.906267 4859 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/91a7b430-b4d6-49f6-8951-0d62d51d6685-fernet-keys\") pod \"keystone-bootstrap-75br8\" (UID: \"91a7b430-b4d6-49f6-8951-0d62d51d6685\") " pod="openstack/keystone-bootstrap-75br8" Oct 08 18:33:48 crc kubenswrapper[4859]: I1008 18:33:48.906603 4859 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/91a7b430-b4d6-49f6-8951-0d62d51d6685-credential-keys\") pod \"keystone-bootstrap-75br8\" (UID: \"91a7b430-b4d6-49f6-8951-0d62d51d6685\") " pod="openstack/keystone-bootstrap-75br8" Oct 08 18:33:48 crc kubenswrapper[4859]: I1008 18:33:48.908105 4859 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/91a7b430-b4d6-49f6-8951-0d62d51d6685-scripts\") pod \"keystone-bootstrap-75br8\" (UID: \"91a7b430-b4d6-49f6-8951-0d62d51d6685\") " pod="openstack/keystone-bootstrap-75br8" Oct 08 18:33:48 crc kubenswrapper[4859]: I1008 18:33:48.921312 4859 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-bhq6l\" (UniqueName: \"kubernetes.io/projected/91a7b430-b4d6-49f6-8951-0d62d51d6685-kube-api-access-bhq6l\") pod \"keystone-bootstrap-75br8\" (UID: \"91a7b430-b4d6-49f6-8951-0d62d51d6685\") " pod="openstack/keystone-bootstrap-75br8" Oct 08 18:33:49 crc kubenswrapper[4859]: I1008 18:33:49.040867 4859 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-bootstrap-75br8" Oct 08 18:33:50 crc kubenswrapper[4859]: I1008 18:33:50.451681 4859 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/dnsmasq-dns-7b587f8db7-lm5t5" podUID="528e6076-58d3-4fcd-ac6e-4fc5d32dd85f" containerName="dnsmasq-dns" probeResult="failure" output="dial tcp 10.217.0.116:5353: connect: connection refused" Oct 08 18:33:50 crc kubenswrapper[4859]: I1008 18:33:50.487964 4859 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="9d1b6fef-16a5-4c1b-a3bb-95f46f1d92f7" path="/var/lib/kubelet/pods/9d1b6fef-16a5-4c1b-a3bb-95f46f1d92f7/volumes" Oct 08 18:33:53 crc kubenswrapper[4859]: E1008 18:33:53.179648 4859 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/podified-antelope-centos9/openstack-horizon@sha256:03b4bb79b71d5ca7792d19c4c0ee08a5e5a407ad844c087305c42dd909ee7490" Oct 08 18:33:53 crc kubenswrapper[4859]: E1008 18:33:53.180296 4859 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:horizon-log,Image:quay.io/podified-antelope-centos9/openstack-horizon@sha256:03b4bb79b71d5ca7792d19c4c0ee08a5e5a407ad844c087305c42dd909ee7490,Command:[/bin/bash],Args:[-c tail -n+1 -F /var/log/horizon/horizon.log],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:CONFIG_HASH,Value:ndfh5dfhdbh9fh58bh674hb4h557h76h5ffh57fh645h594h587h58hb8h7h695h556h5dfhcbh544h654h5dch5ffh5fbh5c6h58fh64ch549hcfh79q,ValueFrom:nil,},EnvVar{Name:ENABLE_DESIGNATE,Value:yes,ValueFrom:nil,},EnvVar{Name:ENABLE_HEAT,Value:yes,ValueFrom:nil,},EnvVar{Name:ENABLE_IRONIC,Value:yes,ValueFrom:nil,},EnvVar{Name:ENABLE_MANILA,Value:yes,ValueFrom:nil,},EnvVar{Name:ENABLE_OCTAVIA,Value:yes,ValueFrom:nil,},EnvVar{Name:ENABLE_WATCHER,Value:no,ValueFrom:nil,},EnvVar{Name:KOLLA_CONFIG_STRATEGY,Value:COPY_ALWAYS,ValueFrom:nil,},EnvVar{Name:UNPACK_THEME,Value:true,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:logs,ReadOnly:false,MountPath:/var/log/horizon,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-2mf2r,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*48,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*true,RunAsGroup:*42400,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod horizon-7444cc4f69-5xs5x_openstack(ac9272b3-433e-4afb-afca-0d647a9d9be5): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Oct 08 18:33:53 crc kubenswrapper[4859]: E1008 18:33:53.184250 4859 pod_workers.go:1301] "Error syncing pod, skipping" err="[failed to \"StartContainer\" for \"horizon-log\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\", failed to \"StartContainer\" for \"horizon\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/podified-antelope-centos9/openstack-horizon@sha256:03b4bb79b71d5ca7792d19c4c0ee08a5e5a407ad844c087305c42dd909ee7490\\\"\"]" pod="openstack/horizon-7444cc4f69-5xs5x" podUID="ac9272b3-433e-4afb-afca-0d647a9d9be5" Oct 08 18:33:55 crc kubenswrapper[4859]: E1008 18:33:55.006333 4859 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/podified-antelope-centos9/openstack-horizon@sha256:03b4bb79b71d5ca7792d19c4c0ee08a5e5a407ad844c087305c42dd909ee7490" Oct 08 18:33:55 crc kubenswrapper[4859]: E1008 18:33:55.006882 4859 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:horizon-log,Image:quay.io/podified-antelope-centos9/openstack-horizon@sha256:03b4bb79b71d5ca7792d19c4c0ee08a5e5a407ad844c087305c42dd909ee7490,Command:[/bin/bash],Args:[-c tail -n+1 -F /var/log/horizon/horizon.log],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:CONFIG_HASH,Value:n664hd8h675hd4hcfh557h689h57h7bh675h667h5c6h5b8h4h5cbh8dh7hcfh58fh57fh5c7hc7hdbh67chb4h66h5bch89h67fh574h8fh5bfq,ValueFrom:nil,},EnvVar{Name:ENABLE_DESIGNATE,Value:yes,ValueFrom:nil,},EnvVar{Name:ENABLE_HEAT,Value:yes,ValueFrom:nil,},EnvVar{Name:ENABLE_IRONIC,Value:yes,ValueFrom:nil,},EnvVar{Name:ENABLE_MANILA,Value:yes,ValueFrom:nil,},EnvVar{Name:ENABLE_OCTAVIA,Value:yes,ValueFrom:nil,},EnvVar{Name:ENABLE_WATCHER,Value:no,ValueFrom:nil,},EnvVar{Name:KOLLA_CONFIG_STRATEGY,Value:COPY_ALWAYS,ValueFrom:nil,},EnvVar{Name:UNPACK_THEME,Value:true,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:logs,ReadOnly:false,MountPath:/var/log/horizon,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-vvdq9,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*48,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*true,RunAsGroup:*42400,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod horizon-85b86687bf-6gp89_openstack(9ed85886-9589-4d26-ad59-7a6021df659e): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Oct 08 18:33:55 crc kubenswrapper[4859]: E1008 18:33:55.009164 4859 pod_workers.go:1301] "Error syncing pod, skipping" err="[failed to \"StartContainer\" for \"horizon-log\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\", failed to \"StartContainer\" for \"horizon\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/podified-antelope-centos9/openstack-horizon@sha256:03b4bb79b71d5ca7792d19c4c0ee08a5e5a407ad844c087305c42dd909ee7490\\\"\"]" pod="openstack/horizon-85b86687bf-6gp89" podUID="9ed85886-9589-4d26-ad59-7a6021df659e" Oct 08 18:33:55 crc kubenswrapper[4859]: E1008 18:33:55.023961 4859 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/podified-antelope-centos9/openstack-horizon@sha256:03b4bb79b71d5ca7792d19c4c0ee08a5e5a407ad844c087305c42dd909ee7490" Oct 08 18:33:55 crc kubenswrapper[4859]: E1008 18:33:55.024139 4859 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:horizon-log,Image:quay.io/podified-antelope-centos9/openstack-horizon@sha256:03b4bb79b71d5ca7792d19c4c0ee08a5e5a407ad844c087305c42dd909ee7490,Command:[/bin/bash],Args:[-c tail -n+1 -F /var/log/horizon/horizon.log],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:CONFIG_HASH,Value:n65ch578h8dh5d7h5c5h5bdh5cbh98h5h68dh4hcfhf8h658h589h549h599h99hdchc7h5f9h5cfhd7hb8h79h698h66ch578h5fbh576h54fh54q,ValueFrom:nil,},EnvVar{Name:ENABLE_DESIGNATE,Value:yes,ValueFrom:nil,},EnvVar{Name:ENABLE_HEAT,Value:yes,ValueFrom:nil,},EnvVar{Name:ENABLE_IRONIC,Value:yes,ValueFrom:nil,},EnvVar{Name:ENABLE_MANILA,Value:yes,ValueFrom:nil,},EnvVar{Name:ENABLE_OCTAVIA,Value:yes,ValueFrom:nil,},EnvVar{Name:ENABLE_WATCHER,Value:no,ValueFrom:nil,},EnvVar{Name:KOLLA_CONFIG_STRATEGY,Value:COPY_ALWAYS,ValueFrom:nil,},EnvVar{Name:UNPACK_THEME,Value:true,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:logs,ReadOnly:false,MountPath:/var/log/horizon,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-d9csb,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*48,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*true,RunAsGroup:*42400,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod horizon-86b9b87697-gmwqv_openstack(1acee243-a101-4dc6-a599-aaee4821dbfa): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Oct 08 18:33:55 crc kubenswrapper[4859]: E1008 18:33:55.029405 4859 pod_workers.go:1301] "Error syncing pod, skipping" err="[failed to \"StartContainer\" for \"horizon-log\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\", failed to \"StartContainer\" for \"horizon\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/podified-antelope-centos9/openstack-horizon@sha256:03b4bb79b71d5ca7792d19c4c0ee08a5e5a407ad844c087305c42dd909ee7490\\\"\"]" pod="openstack/horizon-86b9b87697-gmwqv" podUID="1acee243-a101-4dc6-a599-aaee4821dbfa" Oct 08 18:34:00 crc kubenswrapper[4859]: I1008 18:34:00.451098 4859 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/dnsmasq-dns-7b587f8db7-lm5t5" podUID="528e6076-58d3-4fcd-ac6e-4fc5d32dd85f" containerName="dnsmasq-dns" probeResult="failure" output="dial tcp 10.217.0.116:5353: i/o timeout" Oct 08 18:34:00 crc kubenswrapper[4859]: I1008 18:34:00.451888 4859 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-7b587f8db7-lm5t5" Oct 08 18:34:01 crc kubenswrapper[4859]: I1008 18:34:01.289766 4859 generic.go:334] "Generic (PLEG): container finished" podID="56c33135-38c1-4e29-9804-098312529f0f" containerID="954a12574e8a4ca3fce7cfa866b30b0090d26d8e434f0ca540812c65bf12435d" exitCode=0 Oct 08 18:34:01 crc kubenswrapper[4859]: I1008 18:34:01.289808 4859 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-db-sync-cw6k2" event={"ID":"56c33135-38c1-4e29-9804-098312529f0f","Type":"ContainerDied","Data":"954a12574e8a4ca3fce7cfa866b30b0090d26d8e434f0ca540812c65bf12435d"} Oct 08 18:34:03 crc kubenswrapper[4859]: I1008 18:34:03.250024 4859 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-7444cc4f69-5xs5x" Oct 08 18:34:03 crc kubenswrapper[4859]: I1008 18:34:03.305497 4859 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-7444cc4f69-5xs5x" event={"ID":"ac9272b3-433e-4afb-afca-0d647a9d9be5","Type":"ContainerDied","Data":"36c78cc9e0cd813d0049f15532a403a7bafccd0c34aa89a7ce96d714b689cb2a"} Oct 08 18:34:03 crc kubenswrapper[4859]: I1008 18:34:03.305538 4859 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-7444cc4f69-5xs5x" Oct 08 18:34:03 crc kubenswrapper[4859]: I1008 18:34:03.445775 4859 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/ac9272b3-433e-4afb-afca-0d647a9d9be5-config-data\") pod \"ac9272b3-433e-4afb-afca-0d647a9d9be5\" (UID: \"ac9272b3-433e-4afb-afca-0d647a9d9be5\") " Oct 08 18:34:03 crc kubenswrapper[4859]: I1008 18:34:03.445853 4859 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-2mf2r\" (UniqueName: \"kubernetes.io/projected/ac9272b3-433e-4afb-afca-0d647a9d9be5-kube-api-access-2mf2r\") pod \"ac9272b3-433e-4afb-afca-0d647a9d9be5\" (UID: \"ac9272b3-433e-4afb-afca-0d647a9d9be5\") " Oct 08 18:34:03 crc kubenswrapper[4859]: I1008 18:34:03.445952 4859 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/ac9272b3-433e-4afb-afca-0d647a9d9be5-scripts\") pod \"ac9272b3-433e-4afb-afca-0d647a9d9be5\" (UID: \"ac9272b3-433e-4afb-afca-0d647a9d9be5\") " Oct 08 18:34:03 crc kubenswrapper[4859]: I1008 18:34:03.445976 4859 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/ac9272b3-433e-4afb-afca-0d647a9d9be5-horizon-secret-key\") pod \"ac9272b3-433e-4afb-afca-0d647a9d9be5\" (UID: \"ac9272b3-433e-4afb-afca-0d647a9d9be5\") " Oct 08 18:34:03 crc kubenswrapper[4859]: I1008 18:34:03.446049 4859 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/ac9272b3-433e-4afb-afca-0d647a9d9be5-logs\") pod \"ac9272b3-433e-4afb-afca-0d647a9d9be5\" (UID: \"ac9272b3-433e-4afb-afca-0d647a9d9be5\") " Oct 08 18:34:03 crc kubenswrapper[4859]: I1008 18:34:03.446759 4859 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/ac9272b3-433e-4afb-afca-0d647a9d9be5-logs" (OuterVolumeSpecName: "logs") pod "ac9272b3-433e-4afb-afca-0d647a9d9be5" (UID: "ac9272b3-433e-4afb-afca-0d647a9d9be5"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 08 18:34:03 crc kubenswrapper[4859]: I1008 18:34:03.447179 4859 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/ac9272b3-433e-4afb-afca-0d647a9d9be5-scripts" (OuterVolumeSpecName: "scripts") pod "ac9272b3-433e-4afb-afca-0d647a9d9be5" (UID: "ac9272b3-433e-4afb-afca-0d647a9d9be5"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 08 18:34:03 crc kubenswrapper[4859]: I1008 18:34:03.447239 4859 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/ac9272b3-433e-4afb-afca-0d647a9d9be5-config-data" (OuterVolumeSpecName: "config-data") pod "ac9272b3-433e-4afb-afca-0d647a9d9be5" (UID: "ac9272b3-433e-4afb-afca-0d647a9d9be5"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 08 18:34:03 crc kubenswrapper[4859]: I1008 18:34:03.452047 4859 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/ac9272b3-433e-4afb-afca-0d647a9d9be5-horizon-secret-key" (OuterVolumeSpecName: "horizon-secret-key") pod "ac9272b3-433e-4afb-afca-0d647a9d9be5" (UID: "ac9272b3-433e-4afb-afca-0d647a9d9be5"). InnerVolumeSpecName "horizon-secret-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 08 18:34:03 crc kubenswrapper[4859]: I1008 18:34:03.466672 4859 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/ac9272b3-433e-4afb-afca-0d647a9d9be5-kube-api-access-2mf2r" (OuterVolumeSpecName: "kube-api-access-2mf2r") pod "ac9272b3-433e-4afb-afca-0d647a9d9be5" (UID: "ac9272b3-433e-4afb-afca-0d647a9d9be5"). InnerVolumeSpecName "kube-api-access-2mf2r". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 08 18:34:03 crc kubenswrapper[4859]: I1008 18:34:03.548946 4859 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/ac9272b3-433e-4afb-afca-0d647a9d9be5-logs\") on node \"crc\" DevicePath \"\"" Oct 08 18:34:03 crc kubenswrapper[4859]: I1008 18:34:03.549027 4859 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/ac9272b3-433e-4afb-afca-0d647a9d9be5-config-data\") on node \"crc\" DevicePath \"\"" Oct 08 18:34:03 crc kubenswrapper[4859]: I1008 18:34:03.549040 4859 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-2mf2r\" (UniqueName: \"kubernetes.io/projected/ac9272b3-433e-4afb-afca-0d647a9d9be5-kube-api-access-2mf2r\") on node \"crc\" DevicePath \"\"" Oct 08 18:34:03 crc kubenswrapper[4859]: I1008 18:34:03.549050 4859 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/ac9272b3-433e-4afb-afca-0d647a9d9be5-scripts\") on node \"crc\" DevicePath \"\"" Oct 08 18:34:03 crc kubenswrapper[4859]: I1008 18:34:03.549059 4859 reconciler_common.go:293] "Volume detached for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/ac9272b3-433e-4afb-afca-0d647a9d9be5-horizon-secret-key\") on node \"crc\" DevicePath \"\"" Oct 08 18:34:03 crc kubenswrapper[4859]: I1008 18:34:03.671534 4859 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/horizon-7444cc4f69-5xs5x"] Oct 08 18:34:03 crc kubenswrapper[4859]: I1008 18:34:03.678165 4859 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/horizon-7444cc4f69-5xs5x"] Oct 08 18:34:03 crc kubenswrapper[4859]: E1008 18:34:03.726648 4859 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/podified-antelope-centos9/openstack-barbican-api@sha256:cbe345acb37e57986ecf6685d28c72d0e639bdb493a18e9d3ba947d6c3a16384" Oct 08 18:34:03 crc kubenswrapper[4859]: E1008 18:34:03.726914 4859 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:barbican-db-sync,Image:quay.io/podified-antelope-centos9/openstack-barbican-api@sha256:cbe345acb37e57986ecf6685d28c72d0e639bdb493a18e9d3ba947d6c3a16384,Command:[/bin/bash],Args:[-c barbican-manage db upgrade],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:KOLLA_BOOTSTRAP,Value:TRUE,ValueFrom:nil,},EnvVar{Name:KOLLA_CONFIG_STRATEGY,Value:COPY_ALWAYS,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:db-sync-config-data,ReadOnly:true,MountPath:/etc/barbican/barbican.conf.d,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:combined-ca-bundle,ReadOnly:true,MountPath:/etc/pki/ca-trust/extracted/pem/tls-ca-bundle.pem,SubPath:tls-ca-bundle.pem,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-llvcl,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*42403,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:nil,RunAsGroup:*42403,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod barbican-db-sync-x99jc_openstack(3b2953d1-3ebe-4d7d-b60c-6ad07d02af5d): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Oct 08 18:34:03 crc kubenswrapper[4859]: E1008 18:34:03.728699 4859 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"barbican-db-sync\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack/barbican-db-sync-x99jc" podUID="3b2953d1-3ebe-4d7d-b60c-6ad07d02af5d" Oct 08 18:34:03 crc kubenswrapper[4859]: I1008 18:34:03.771581 4859 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-7b587f8db7-lm5t5" Oct 08 18:34:03 crc kubenswrapper[4859]: I1008 18:34:03.780739 4859 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-db-sync-cw6k2" Oct 08 18:34:03 crc kubenswrapper[4859]: I1008 18:34:03.787267 4859 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-86b9b87697-gmwqv" Oct 08 18:34:03 crc kubenswrapper[4859]: I1008 18:34:03.799454 4859 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-85b86687bf-6gp89" Oct 08 18:34:03 crc kubenswrapper[4859]: I1008 18:34:03.955982 4859 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/9ed85886-9589-4d26-ad59-7a6021df659e-horizon-secret-key\") pod \"9ed85886-9589-4d26-ad59-7a6021df659e\" (UID: \"9ed85886-9589-4d26-ad59-7a6021df659e\") " Oct 08 18:34:03 crc kubenswrapper[4859]: I1008 18:34:03.956049 4859 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/1acee243-a101-4dc6-a599-aaee4821dbfa-config-data\") pod \"1acee243-a101-4dc6-a599-aaee4821dbfa\" (UID: \"1acee243-a101-4dc6-a599-aaee4821dbfa\") " Oct 08 18:34:03 crc kubenswrapper[4859]: I1008 18:34:03.956082 4859 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-mxlls\" (UniqueName: \"kubernetes.io/projected/528e6076-58d3-4fcd-ac6e-4fc5d32dd85f-kube-api-access-mxlls\") pod \"528e6076-58d3-4fcd-ac6e-4fc5d32dd85f\" (UID: \"528e6076-58d3-4fcd-ac6e-4fc5d32dd85f\") " Oct 08 18:34:03 crc kubenswrapper[4859]: I1008 18:34:03.956136 4859 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/56c33135-38c1-4e29-9804-098312529f0f-combined-ca-bundle\") pod \"56c33135-38c1-4e29-9804-098312529f0f\" (UID: \"56c33135-38c1-4e29-9804-098312529f0f\") " Oct 08 18:34:03 crc kubenswrapper[4859]: I1008 18:34:03.956198 4859 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/1acee243-a101-4dc6-a599-aaee4821dbfa-horizon-secret-key\") pod \"1acee243-a101-4dc6-a599-aaee4821dbfa\" (UID: \"1acee243-a101-4dc6-a599-aaee4821dbfa\") " Oct 08 18:34:03 crc kubenswrapper[4859]: I1008 18:34:03.956230 4859 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/9ed85886-9589-4d26-ad59-7a6021df659e-config-data\") pod \"9ed85886-9589-4d26-ad59-7a6021df659e\" (UID: \"9ed85886-9589-4d26-ad59-7a6021df659e\") " Oct 08 18:34:03 crc kubenswrapper[4859]: I1008 18:34:03.956284 4859 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/528e6076-58d3-4fcd-ac6e-4fc5d32dd85f-ovsdbserver-nb\") pod \"528e6076-58d3-4fcd-ac6e-4fc5d32dd85f\" (UID: \"528e6076-58d3-4fcd-ac6e-4fc5d32dd85f\") " Oct 08 18:34:03 crc kubenswrapper[4859]: I1008 18:34:03.956317 4859 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/528e6076-58d3-4fcd-ac6e-4fc5d32dd85f-dns-svc\") pod \"528e6076-58d3-4fcd-ac6e-4fc5d32dd85f\" (UID: \"528e6076-58d3-4fcd-ac6e-4fc5d32dd85f\") " Oct 08 18:34:03 crc kubenswrapper[4859]: I1008 18:34:03.956385 4859 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/56c33135-38c1-4e29-9804-098312529f0f-config\") pod \"56c33135-38c1-4e29-9804-098312529f0f\" (UID: \"56c33135-38c1-4e29-9804-098312529f0f\") " Oct 08 18:34:03 crc kubenswrapper[4859]: I1008 18:34:03.956428 4859 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/9ed85886-9589-4d26-ad59-7a6021df659e-scripts\") pod \"9ed85886-9589-4d26-ad59-7a6021df659e\" (UID: \"9ed85886-9589-4d26-ad59-7a6021df659e\") " Oct 08 18:34:03 crc kubenswrapper[4859]: I1008 18:34:03.956448 4859 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-6pcvs\" (UniqueName: \"kubernetes.io/projected/56c33135-38c1-4e29-9804-098312529f0f-kube-api-access-6pcvs\") pod \"56c33135-38c1-4e29-9804-098312529f0f\" (UID: \"56c33135-38c1-4e29-9804-098312529f0f\") " Oct 08 18:34:03 crc kubenswrapper[4859]: I1008 18:34:03.956517 4859 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/1acee243-a101-4dc6-a599-aaee4821dbfa-scripts\") pod \"1acee243-a101-4dc6-a599-aaee4821dbfa\" (UID: \"1acee243-a101-4dc6-a599-aaee4821dbfa\") " Oct 08 18:34:03 crc kubenswrapper[4859]: I1008 18:34:03.956537 4859 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/528e6076-58d3-4fcd-ac6e-4fc5d32dd85f-config\") pod \"528e6076-58d3-4fcd-ac6e-4fc5d32dd85f\" (UID: \"528e6076-58d3-4fcd-ac6e-4fc5d32dd85f\") " Oct 08 18:34:03 crc kubenswrapper[4859]: I1008 18:34:03.956603 4859 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/1acee243-a101-4dc6-a599-aaee4821dbfa-logs\") pod \"1acee243-a101-4dc6-a599-aaee4821dbfa\" (UID: \"1acee243-a101-4dc6-a599-aaee4821dbfa\") " Oct 08 18:34:03 crc kubenswrapper[4859]: I1008 18:34:03.956621 4859 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-vvdq9\" (UniqueName: \"kubernetes.io/projected/9ed85886-9589-4d26-ad59-7a6021df659e-kube-api-access-vvdq9\") pod \"9ed85886-9589-4d26-ad59-7a6021df659e\" (UID: \"9ed85886-9589-4d26-ad59-7a6021df659e\") " Oct 08 18:34:03 crc kubenswrapper[4859]: I1008 18:34:03.956733 4859 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/9ed85886-9589-4d26-ad59-7a6021df659e-logs\") pod \"9ed85886-9589-4d26-ad59-7a6021df659e\" (UID: \"9ed85886-9589-4d26-ad59-7a6021df659e\") " Oct 08 18:34:03 crc kubenswrapper[4859]: I1008 18:34:03.956761 4859 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-d9csb\" (UniqueName: \"kubernetes.io/projected/1acee243-a101-4dc6-a599-aaee4821dbfa-kube-api-access-d9csb\") pod \"1acee243-a101-4dc6-a599-aaee4821dbfa\" (UID: \"1acee243-a101-4dc6-a599-aaee4821dbfa\") " Oct 08 18:34:03 crc kubenswrapper[4859]: I1008 18:34:03.956820 4859 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/528e6076-58d3-4fcd-ac6e-4fc5d32dd85f-ovsdbserver-sb\") pod \"528e6076-58d3-4fcd-ac6e-4fc5d32dd85f\" (UID: \"528e6076-58d3-4fcd-ac6e-4fc5d32dd85f\") " Oct 08 18:34:03 crc kubenswrapper[4859]: I1008 18:34:03.957029 4859 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1acee243-a101-4dc6-a599-aaee4821dbfa-config-data" (OuterVolumeSpecName: "config-data") pod "1acee243-a101-4dc6-a599-aaee4821dbfa" (UID: "1acee243-a101-4dc6-a599-aaee4821dbfa"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 08 18:34:03 crc kubenswrapper[4859]: I1008 18:34:03.957058 4859 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/9ed85886-9589-4d26-ad59-7a6021df659e-config-data" (OuterVolumeSpecName: "config-data") pod "9ed85886-9589-4d26-ad59-7a6021df659e" (UID: "9ed85886-9589-4d26-ad59-7a6021df659e"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 08 18:34:03 crc kubenswrapper[4859]: I1008 18:34:03.957865 4859 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/1acee243-a101-4dc6-a599-aaee4821dbfa-config-data\") on node \"crc\" DevicePath \"\"" Oct 08 18:34:03 crc kubenswrapper[4859]: I1008 18:34:03.957895 4859 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/9ed85886-9589-4d26-ad59-7a6021df659e-config-data\") on node \"crc\" DevicePath \"\"" Oct 08 18:34:03 crc kubenswrapper[4859]: I1008 18:34:03.958722 4859 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1acee243-a101-4dc6-a599-aaee4821dbfa-scripts" (OuterVolumeSpecName: "scripts") pod "1acee243-a101-4dc6-a599-aaee4821dbfa" (UID: "1acee243-a101-4dc6-a599-aaee4821dbfa"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 08 18:34:03 crc kubenswrapper[4859]: I1008 18:34:03.960816 4859 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/9ed85886-9589-4d26-ad59-7a6021df659e-horizon-secret-key" (OuterVolumeSpecName: "horizon-secret-key") pod "9ed85886-9589-4d26-ad59-7a6021df659e" (UID: "9ed85886-9589-4d26-ad59-7a6021df659e"). InnerVolumeSpecName "horizon-secret-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 08 18:34:03 crc kubenswrapper[4859]: I1008 18:34:03.960828 4859 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1acee243-a101-4dc6-a599-aaee4821dbfa-horizon-secret-key" (OuterVolumeSpecName: "horizon-secret-key") pod "1acee243-a101-4dc6-a599-aaee4821dbfa" (UID: "1acee243-a101-4dc6-a599-aaee4821dbfa"). InnerVolumeSpecName "horizon-secret-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 08 18:34:03 crc kubenswrapper[4859]: I1008 18:34:03.962077 4859 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/528e6076-58d3-4fcd-ac6e-4fc5d32dd85f-kube-api-access-mxlls" (OuterVolumeSpecName: "kube-api-access-mxlls") pod "528e6076-58d3-4fcd-ac6e-4fc5d32dd85f" (UID: "528e6076-58d3-4fcd-ac6e-4fc5d32dd85f"). InnerVolumeSpecName "kube-api-access-mxlls". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 08 18:34:03 crc kubenswrapper[4859]: I1008 18:34:03.962315 4859 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/9ed85886-9589-4d26-ad59-7a6021df659e-logs" (OuterVolumeSpecName: "logs") pod "9ed85886-9589-4d26-ad59-7a6021df659e" (UID: "9ed85886-9589-4d26-ad59-7a6021df659e"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 08 18:34:03 crc kubenswrapper[4859]: I1008 18:34:03.965059 4859 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/9ed85886-9589-4d26-ad59-7a6021df659e-kube-api-access-vvdq9" (OuterVolumeSpecName: "kube-api-access-vvdq9") pod "9ed85886-9589-4d26-ad59-7a6021df659e" (UID: "9ed85886-9589-4d26-ad59-7a6021df659e"). InnerVolumeSpecName "kube-api-access-vvdq9". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 08 18:34:03 crc kubenswrapper[4859]: I1008 18:34:03.965453 4859 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/9ed85886-9589-4d26-ad59-7a6021df659e-scripts" (OuterVolumeSpecName: "scripts") pod "9ed85886-9589-4d26-ad59-7a6021df659e" (UID: "9ed85886-9589-4d26-ad59-7a6021df659e"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 08 18:34:03 crc kubenswrapper[4859]: I1008 18:34:03.965618 4859 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/1acee243-a101-4dc6-a599-aaee4821dbfa-logs" (OuterVolumeSpecName: "logs") pod "1acee243-a101-4dc6-a599-aaee4821dbfa" (UID: "1acee243-a101-4dc6-a599-aaee4821dbfa"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 08 18:34:03 crc kubenswrapper[4859]: I1008 18:34:03.966619 4859 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/1acee243-a101-4dc6-a599-aaee4821dbfa-kube-api-access-d9csb" (OuterVolumeSpecName: "kube-api-access-d9csb") pod "1acee243-a101-4dc6-a599-aaee4821dbfa" (UID: "1acee243-a101-4dc6-a599-aaee4821dbfa"). InnerVolumeSpecName "kube-api-access-d9csb". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 08 18:34:03 crc kubenswrapper[4859]: I1008 18:34:03.971940 4859 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/56c33135-38c1-4e29-9804-098312529f0f-kube-api-access-6pcvs" (OuterVolumeSpecName: "kube-api-access-6pcvs") pod "56c33135-38c1-4e29-9804-098312529f0f" (UID: "56c33135-38c1-4e29-9804-098312529f0f"). InnerVolumeSpecName "kube-api-access-6pcvs". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 08 18:34:03 crc kubenswrapper[4859]: I1008 18:34:03.986424 4859 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/56c33135-38c1-4e29-9804-098312529f0f-config" (OuterVolumeSpecName: "config") pod "56c33135-38c1-4e29-9804-098312529f0f" (UID: "56c33135-38c1-4e29-9804-098312529f0f"). InnerVolumeSpecName "config". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 08 18:34:03 crc kubenswrapper[4859]: I1008 18:34:03.988022 4859 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/56c33135-38c1-4e29-9804-098312529f0f-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "56c33135-38c1-4e29-9804-098312529f0f" (UID: "56c33135-38c1-4e29-9804-098312529f0f"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 08 18:34:04 crc kubenswrapper[4859]: I1008 18:34:04.009098 4859 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/528e6076-58d3-4fcd-ac6e-4fc5d32dd85f-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "528e6076-58d3-4fcd-ac6e-4fc5d32dd85f" (UID: "528e6076-58d3-4fcd-ac6e-4fc5d32dd85f"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 08 18:34:04 crc kubenswrapper[4859]: I1008 18:34:04.009367 4859 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/528e6076-58d3-4fcd-ac6e-4fc5d32dd85f-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "528e6076-58d3-4fcd-ac6e-4fc5d32dd85f" (UID: "528e6076-58d3-4fcd-ac6e-4fc5d32dd85f"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 08 18:34:04 crc kubenswrapper[4859]: I1008 18:34:04.010608 4859 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/528e6076-58d3-4fcd-ac6e-4fc5d32dd85f-config" (OuterVolumeSpecName: "config") pod "528e6076-58d3-4fcd-ac6e-4fc5d32dd85f" (UID: "528e6076-58d3-4fcd-ac6e-4fc5d32dd85f"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 08 18:34:04 crc kubenswrapper[4859]: I1008 18:34:04.012069 4859 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/528e6076-58d3-4fcd-ac6e-4fc5d32dd85f-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "528e6076-58d3-4fcd-ac6e-4fc5d32dd85f" (UID: "528e6076-58d3-4fcd-ac6e-4fc5d32dd85f"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 08 18:34:04 crc kubenswrapper[4859]: I1008 18:34:04.059959 4859 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-mxlls\" (UniqueName: \"kubernetes.io/projected/528e6076-58d3-4fcd-ac6e-4fc5d32dd85f-kube-api-access-mxlls\") on node \"crc\" DevicePath \"\"" Oct 08 18:34:04 crc kubenswrapper[4859]: I1008 18:34:04.060002 4859 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/56c33135-38c1-4e29-9804-098312529f0f-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 08 18:34:04 crc kubenswrapper[4859]: I1008 18:34:04.060015 4859 reconciler_common.go:293] "Volume detached for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/1acee243-a101-4dc6-a599-aaee4821dbfa-horizon-secret-key\") on node \"crc\" DevicePath \"\"" Oct 08 18:34:04 crc kubenswrapper[4859]: I1008 18:34:04.060027 4859 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/528e6076-58d3-4fcd-ac6e-4fc5d32dd85f-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Oct 08 18:34:04 crc kubenswrapper[4859]: I1008 18:34:04.060038 4859 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/528e6076-58d3-4fcd-ac6e-4fc5d32dd85f-dns-svc\") on node \"crc\" DevicePath \"\"" Oct 08 18:34:04 crc kubenswrapper[4859]: I1008 18:34:04.060049 4859 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/secret/56c33135-38c1-4e29-9804-098312529f0f-config\") on node \"crc\" DevicePath \"\"" Oct 08 18:34:04 crc kubenswrapper[4859]: I1008 18:34:04.060059 4859 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/9ed85886-9589-4d26-ad59-7a6021df659e-scripts\") on node \"crc\" DevicePath \"\"" Oct 08 18:34:04 crc kubenswrapper[4859]: I1008 18:34:04.060068 4859 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-6pcvs\" (UniqueName: \"kubernetes.io/projected/56c33135-38c1-4e29-9804-098312529f0f-kube-api-access-6pcvs\") on node \"crc\" DevicePath \"\"" Oct 08 18:34:04 crc kubenswrapper[4859]: I1008 18:34:04.060080 4859 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/1acee243-a101-4dc6-a599-aaee4821dbfa-scripts\") on node \"crc\" DevicePath \"\"" Oct 08 18:34:04 crc kubenswrapper[4859]: I1008 18:34:04.060092 4859 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/528e6076-58d3-4fcd-ac6e-4fc5d32dd85f-config\") on node \"crc\" DevicePath \"\"" Oct 08 18:34:04 crc kubenswrapper[4859]: I1008 18:34:04.060103 4859 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/1acee243-a101-4dc6-a599-aaee4821dbfa-logs\") on node \"crc\" DevicePath \"\"" Oct 08 18:34:04 crc kubenswrapper[4859]: I1008 18:34:04.060115 4859 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-vvdq9\" (UniqueName: \"kubernetes.io/projected/9ed85886-9589-4d26-ad59-7a6021df659e-kube-api-access-vvdq9\") on node \"crc\" DevicePath \"\"" Oct 08 18:34:04 crc kubenswrapper[4859]: I1008 18:34:04.060125 4859 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/9ed85886-9589-4d26-ad59-7a6021df659e-logs\") on node \"crc\" DevicePath \"\"" Oct 08 18:34:04 crc kubenswrapper[4859]: I1008 18:34:04.060137 4859 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-d9csb\" (UniqueName: \"kubernetes.io/projected/1acee243-a101-4dc6-a599-aaee4821dbfa-kube-api-access-d9csb\") on node \"crc\" DevicePath \"\"" Oct 08 18:34:04 crc kubenswrapper[4859]: I1008 18:34:04.060146 4859 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/528e6076-58d3-4fcd-ac6e-4fc5d32dd85f-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Oct 08 18:34:04 crc kubenswrapper[4859]: I1008 18:34:04.060154 4859 reconciler_common.go:293] "Volume detached for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/9ed85886-9589-4d26-ad59-7a6021df659e-horizon-secret-key\") on node \"crc\" DevicePath \"\"" Oct 08 18:34:04 crc kubenswrapper[4859]: I1008 18:34:04.340838 4859 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-7b587f8db7-lm5t5" event={"ID":"528e6076-58d3-4fcd-ac6e-4fc5d32dd85f","Type":"ContainerDied","Data":"178b91783a6ecda6b48b4d222dd123d392ae794e7e329ae3c6e12c430eca7ece"} Oct 08 18:34:04 crc kubenswrapper[4859]: I1008 18:34:04.340873 4859 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-7b587f8db7-lm5t5" Oct 08 18:34:04 crc kubenswrapper[4859]: I1008 18:34:04.343436 4859 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-86b9b87697-gmwqv" event={"ID":"1acee243-a101-4dc6-a599-aaee4821dbfa","Type":"ContainerDied","Data":"e34d4711b5a7eed7ee113c9435cd61a30a628c24c4ea192e6eecf5c18f11bcad"} Oct 08 18:34:04 crc kubenswrapper[4859]: I1008 18:34:04.343474 4859 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-86b9b87697-gmwqv" Oct 08 18:34:04 crc kubenswrapper[4859]: I1008 18:34:04.345512 4859 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-db-sync-cw6k2" event={"ID":"56c33135-38c1-4e29-9804-098312529f0f","Type":"ContainerDied","Data":"7f4b57ab0611ccc6e613abee019be0e27cd21ccc590e873186cba81c6373cd7b"} Oct 08 18:34:04 crc kubenswrapper[4859]: I1008 18:34:04.345537 4859 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-db-sync-cw6k2" Oct 08 18:34:04 crc kubenswrapper[4859]: I1008 18:34:04.345543 4859 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="7f4b57ab0611ccc6e613abee019be0e27cd21ccc590e873186cba81c6373cd7b" Oct 08 18:34:04 crc kubenswrapper[4859]: I1008 18:34:04.347244 4859 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-85b86687bf-6gp89" event={"ID":"9ed85886-9589-4d26-ad59-7a6021df659e","Type":"ContainerDied","Data":"1a0b89180efa9e1f99c0c69d71c18251e60c4bad6befa9c2c5ebb38d4f8be499"} Oct 08 18:34:04 crc kubenswrapper[4859]: I1008 18:34:04.347609 4859 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-85b86687bf-6gp89" Oct 08 18:34:04 crc kubenswrapper[4859]: E1008 18:34:04.349765 4859 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"barbican-db-sync\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/podified-antelope-centos9/openstack-barbican-api@sha256:cbe345acb37e57986ecf6685d28c72d0e639bdb493a18e9d3ba947d6c3a16384\\\"\"" pod="openstack/barbican-db-sync-x99jc" podUID="3b2953d1-3ebe-4d7d-b60c-6ad07d02af5d" Oct 08 18:34:04 crc kubenswrapper[4859]: I1008 18:34:04.401885 4859 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-7b587f8db7-lm5t5"] Oct 08 18:34:04 crc kubenswrapper[4859]: I1008 18:34:04.412584 4859 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-7b587f8db7-lm5t5"] Oct 08 18:34:04 crc kubenswrapper[4859]: I1008 18:34:04.455184 4859 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/horizon-86b9b87697-gmwqv"] Oct 08 18:34:04 crc kubenswrapper[4859]: I1008 18:34:04.465309 4859 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/horizon-86b9b87697-gmwqv"] Oct 08 18:34:04 crc kubenswrapper[4859]: I1008 18:34:04.481903 4859 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="1acee243-a101-4dc6-a599-aaee4821dbfa" path="/var/lib/kubelet/pods/1acee243-a101-4dc6-a599-aaee4821dbfa/volumes" Oct 08 18:34:04 crc kubenswrapper[4859]: I1008 18:34:04.482316 4859 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="528e6076-58d3-4fcd-ac6e-4fc5d32dd85f" path="/var/lib/kubelet/pods/528e6076-58d3-4fcd-ac6e-4fc5d32dd85f/volumes" Oct 08 18:34:04 crc kubenswrapper[4859]: I1008 18:34:04.482997 4859 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="ac9272b3-433e-4afb-afca-0d647a9d9be5" path="/var/lib/kubelet/pods/ac9272b3-433e-4afb-afca-0d647a9d9be5/volumes" Oct 08 18:34:04 crc kubenswrapper[4859]: I1008 18:34:04.485256 4859 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/horizon-85b86687bf-6gp89"] Oct 08 18:34:04 crc kubenswrapper[4859]: I1008 18:34:04.492946 4859 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/horizon-85b86687bf-6gp89"] Oct 08 18:34:05 crc kubenswrapper[4859]: I1008 18:34:05.083895 4859 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-58799d59b9-5kp6v"] Oct 08 18:34:05 crc kubenswrapper[4859]: E1008 18:34:05.084573 4859 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="528e6076-58d3-4fcd-ac6e-4fc5d32dd85f" containerName="init" Oct 08 18:34:05 crc kubenswrapper[4859]: I1008 18:34:05.084596 4859 state_mem.go:107] "Deleted CPUSet assignment" podUID="528e6076-58d3-4fcd-ac6e-4fc5d32dd85f" containerName="init" Oct 08 18:34:05 crc kubenswrapper[4859]: E1008 18:34:05.084615 4859 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="528e6076-58d3-4fcd-ac6e-4fc5d32dd85f" containerName="dnsmasq-dns" Oct 08 18:34:05 crc kubenswrapper[4859]: I1008 18:34:05.084624 4859 state_mem.go:107] "Deleted CPUSet assignment" podUID="528e6076-58d3-4fcd-ac6e-4fc5d32dd85f" containerName="dnsmasq-dns" Oct 08 18:34:05 crc kubenswrapper[4859]: E1008 18:34:05.084656 4859 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="56c33135-38c1-4e29-9804-098312529f0f" containerName="neutron-db-sync" Oct 08 18:34:05 crc kubenswrapper[4859]: I1008 18:34:05.084665 4859 state_mem.go:107] "Deleted CPUSet assignment" podUID="56c33135-38c1-4e29-9804-098312529f0f" containerName="neutron-db-sync" Oct 08 18:34:05 crc kubenswrapper[4859]: I1008 18:34:05.084879 4859 memory_manager.go:354] "RemoveStaleState removing state" podUID="528e6076-58d3-4fcd-ac6e-4fc5d32dd85f" containerName="dnsmasq-dns" Oct 08 18:34:05 crc kubenswrapper[4859]: I1008 18:34:05.084902 4859 memory_manager.go:354] "RemoveStaleState removing state" podUID="56c33135-38c1-4e29-9804-098312529f0f" containerName="neutron-db-sync" Oct 08 18:34:05 crc kubenswrapper[4859]: I1008 18:34:05.085874 4859 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-58799d59b9-5kp6v" Oct 08 18:34:05 crc kubenswrapper[4859]: I1008 18:34:05.086795 4859 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/881d266c-6730-404d-8243-ea0df0d3d03f-dns-swift-storage-0\") pod \"dnsmasq-dns-58799d59b9-5kp6v\" (UID: \"881d266c-6730-404d-8243-ea0df0d3d03f\") " pod="openstack/dnsmasq-dns-58799d59b9-5kp6v" Oct 08 18:34:05 crc kubenswrapper[4859]: I1008 18:34:05.086866 4859 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/881d266c-6730-404d-8243-ea0df0d3d03f-ovsdbserver-nb\") pod \"dnsmasq-dns-58799d59b9-5kp6v\" (UID: \"881d266c-6730-404d-8243-ea0df0d3d03f\") " pod="openstack/dnsmasq-dns-58799d59b9-5kp6v" Oct 08 18:34:05 crc kubenswrapper[4859]: I1008 18:34:05.086906 4859 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-2gc55\" (UniqueName: \"kubernetes.io/projected/881d266c-6730-404d-8243-ea0df0d3d03f-kube-api-access-2gc55\") pod \"dnsmasq-dns-58799d59b9-5kp6v\" (UID: \"881d266c-6730-404d-8243-ea0df0d3d03f\") " pod="openstack/dnsmasq-dns-58799d59b9-5kp6v" Oct 08 18:34:05 crc kubenswrapper[4859]: I1008 18:34:05.086925 4859 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/881d266c-6730-404d-8243-ea0df0d3d03f-dns-svc\") pod \"dnsmasq-dns-58799d59b9-5kp6v\" (UID: \"881d266c-6730-404d-8243-ea0df0d3d03f\") " pod="openstack/dnsmasq-dns-58799d59b9-5kp6v" Oct 08 18:34:05 crc kubenswrapper[4859]: I1008 18:34:05.086991 4859 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/881d266c-6730-404d-8243-ea0df0d3d03f-config\") pod \"dnsmasq-dns-58799d59b9-5kp6v\" (UID: \"881d266c-6730-404d-8243-ea0df0d3d03f\") " pod="openstack/dnsmasq-dns-58799d59b9-5kp6v" Oct 08 18:34:05 crc kubenswrapper[4859]: I1008 18:34:05.087014 4859 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/881d266c-6730-404d-8243-ea0df0d3d03f-ovsdbserver-sb\") pod \"dnsmasq-dns-58799d59b9-5kp6v\" (UID: \"881d266c-6730-404d-8243-ea0df0d3d03f\") " pod="openstack/dnsmasq-dns-58799d59b9-5kp6v" Oct 08 18:34:05 crc kubenswrapper[4859]: I1008 18:34:05.109495 4859 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-58799d59b9-5kp6v"] Oct 08 18:34:05 crc kubenswrapper[4859]: I1008 18:34:05.188113 4859 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-2gc55\" (UniqueName: \"kubernetes.io/projected/881d266c-6730-404d-8243-ea0df0d3d03f-kube-api-access-2gc55\") pod \"dnsmasq-dns-58799d59b9-5kp6v\" (UID: \"881d266c-6730-404d-8243-ea0df0d3d03f\") " pod="openstack/dnsmasq-dns-58799d59b9-5kp6v" Oct 08 18:34:05 crc kubenswrapper[4859]: I1008 18:34:05.188153 4859 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/881d266c-6730-404d-8243-ea0df0d3d03f-dns-svc\") pod \"dnsmasq-dns-58799d59b9-5kp6v\" (UID: \"881d266c-6730-404d-8243-ea0df0d3d03f\") " pod="openstack/dnsmasq-dns-58799d59b9-5kp6v" Oct 08 18:34:05 crc kubenswrapper[4859]: I1008 18:34:05.188185 4859 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/881d266c-6730-404d-8243-ea0df0d3d03f-config\") pod \"dnsmasq-dns-58799d59b9-5kp6v\" (UID: \"881d266c-6730-404d-8243-ea0df0d3d03f\") " pod="openstack/dnsmasq-dns-58799d59b9-5kp6v" Oct 08 18:34:05 crc kubenswrapper[4859]: I1008 18:34:05.188210 4859 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/881d266c-6730-404d-8243-ea0df0d3d03f-ovsdbserver-sb\") pod \"dnsmasq-dns-58799d59b9-5kp6v\" (UID: \"881d266c-6730-404d-8243-ea0df0d3d03f\") " pod="openstack/dnsmasq-dns-58799d59b9-5kp6v" Oct 08 18:34:05 crc kubenswrapper[4859]: I1008 18:34:05.188733 4859 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/881d266c-6730-404d-8243-ea0df0d3d03f-dns-swift-storage-0\") pod \"dnsmasq-dns-58799d59b9-5kp6v\" (UID: \"881d266c-6730-404d-8243-ea0df0d3d03f\") " pod="openstack/dnsmasq-dns-58799d59b9-5kp6v" Oct 08 18:34:05 crc kubenswrapper[4859]: I1008 18:34:05.188776 4859 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/881d266c-6730-404d-8243-ea0df0d3d03f-ovsdbserver-nb\") pod \"dnsmasq-dns-58799d59b9-5kp6v\" (UID: \"881d266c-6730-404d-8243-ea0df0d3d03f\") " pod="openstack/dnsmasq-dns-58799d59b9-5kp6v" Oct 08 18:34:05 crc kubenswrapper[4859]: I1008 18:34:05.189521 4859 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/881d266c-6730-404d-8243-ea0df0d3d03f-ovsdbserver-sb\") pod \"dnsmasq-dns-58799d59b9-5kp6v\" (UID: \"881d266c-6730-404d-8243-ea0df0d3d03f\") " pod="openstack/dnsmasq-dns-58799d59b9-5kp6v" Oct 08 18:34:05 crc kubenswrapper[4859]: I1008 18:34:05.189588 4859 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/881d266c-6730-404d-8243-ea0df0d3d03f-ovsdbserver-nb\") pod \"dnsmasq-dns-58799d59b9-5kp6v\" (UID: \"881d266c-6730-404d-8243-ea0df0d3d03f\") " pod="openstack/dnsmasq-dns-58799d59b9-5kp6v" Oct 08 18:34:05 crc kubenswrapper[4859]: I1008 18:34:05.189835 4859 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/881d266c-6730-404d-8243-ea0df0d3d03f-dns-swift-storage-0\") pod \"dnsmasq-dns-58799d59b9-5kp6v\" (UID: \"881d266c-6730-404d-8243-ea0df0d3d03f\") " pod="openstack/dnsmasq-dns-58799d59b9-5kp6v" Oct 08 18:34:05 crc kubenswrapper[4859]: I1008 18:34:05.191021 4859 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/881d266c-6730-404d-8243-ea0df0d3d03f-config\") pod \"dnsmasq-dns-58799d59b9-5kp6v\" (UID: \"881d266c-6730-404d-8243-ea0df0d3d03f\") " pod="openstack/dnsmasq-dns-58799d59b9-5kp6v" Oct 08 18:34:05 crc kubenswrapper[4859]: I1008 18:34:05.191502 4859 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/881d266c-6730-404d-8243-ea0df0d3d03f-dns-svc\") pod \"dnsmasq-dns-58799d59b9-5kp6v\" (UID: \"881d266c-6730-404d-8243-ea0df0d3d03f\") " pod="openstack/dnsmasq-dns-58799d59b9-5kp6v" Oct 08 18:34:05 crc kubenswrapper[4859]: I1008 18:34:05.220314 4859 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/neutron-7755787cf6-5gvlv"] Oct 08 18:34:05 crc kubenswrapper[4859]: I1008 18:34:05.220844 4859 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-2gc55\" (UniqueName: \"kubernetes.io/projected/881d266c-6730-404d-8243-ea0df0d3d03f-kube-api-access-2gc55\") pod \"dnsmasq-dns-58799d59b9-5kp6v\" (UID: \"881d266c-6730-404d-8243-ea0df0d3d03f\") " pod="openstack/dnsmasq-dns-58799d59b9-5kp6v" Oct 08 18:34:05 crc kubenswrapper[4859]: I1008 18:34:05.221916 4859 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-7755787cf6-5gvlv" Oct 08 18:34:05 crc kubenswrapper[4859]: I1008 18:34:05.233458 4859 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"neutron-httpd-config" Oct 08 18:34:05 crc kubenswrapper[4859]: I1008 18:34:05.233634 4859 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-neutron-ovndbs" Oct 08 18:34:05 crc kubenswrapper[4859]: I1008 18:34:05.233750 4859 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"neutron-neutron-dockercfg-t6c4d" Oct 08 18:34:05 crc kubenswrapper[4859]: I1008 18:34:05.233847 4859 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"neutron-config" Oct 08 18:34:05 crc kubenswrapper[4859]: I1008 18:34:05.234336 4859 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-7755787cf6-5gvlv"] Oct 08 18:34:05 crc kubenswrapper[4859]: I1008 18:34:05.392038 4859 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovndb-tls-certs\" (UniqueName: \"kubernetes.io/secret/08efd21e-517e-46dc-bf88-60cfb1943208-ovndb-tls-certs\") pod \"neutron-7755787cf6-5gvlv\" (UID: \"08efd21e-517e-46dc-bf88-60cfb1943208\") " pod="openstack/neutron-7755787cf6-5gvlv" Oct 08 18:34:05 crc kubenswrapper[4859]: I1008 18:34:05.392107 4859 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-wgpl2\" (UniqueName: \"kubernetes.io/projected/08efd21e-517e-46dc-bf88-60cfb1943208-kube-api-access-wgpl2\") pod \"neutron-7755787cf6-5gvlv\" (UID: \"08efd21e-517e-46dc-bf88-60cfb1943208\") " pod="openstack/neutron-7755787cf6-5gvlv" Oct 08 18:34:05 crc kubenswrapper[4859]: I1008 18:34:05.392444 4859 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/08efd21e-517e-46dc-bf88-60cfb1943208-combined-ca-bundle\") pod \"neutron-7755787cf6-5gvlv\" (UID: \"08efd21e-517e-46dc-bf88-60cfb1943208\") " pod="openstack/neutron-7755787cf6-5gvlv" Oct 08 18:34:05 crc kubenswrapper[4859]: I1008 18:34:05.392547 4859 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/08efd21e-517e-46dc-bf88-60cfb1943208-config\") pod \"neutron-7755787cf6-5gvlv\" (UID: \"08efd21e-517e-46dc-bf88-60cfb1943208\") " pod="openstack/neutron-7755787cf6-5gvlv" Oct 08 18:34:05 crc kubenswrapper[4859]: I1008 18:34:05.392623 4859 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"httpd-config\" (UniqueName: \"kubernetes.io/secret/08efd21e-517e-46dc-bf88-60cfb1943208-httpd-config\") pod \"neutron-7755787cf6-5gvlv\" (UID: \"08efd21e-517e-46dc-bf88-60cfb1943208\") " pod="openstack/neutron-7755787cf6-5gvlv" Oct 08 18:34:05 crc kubenswrapper[4859]: I1008 18:34:05.408055 4859 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-58799d59b9-5kp6v" Oct 08 18:34:05 crc kubenswrapper[4859]: I1008 18:34:05.452469 4859 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/dnsmasq-dns-7b587f8db7-lm5t5" podUID="528e6076-58d3-4fcd-ac6e-4fc5d32dd85f" containerName="dnsmasq-dns" probeResult="failure" output="dial tcp 10.217.0.116:5353: i/o timeout" Oct 08 18:34:05 crc kubenswrapper[4859]: I1008 18:34:05.493524 4859 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"httpd-config\" (UniqueName: \"kubernetes.io/secret/08efd21e-517e-46dc-bf88-60cfb1943208-httpd-config\") pod \"neutron-7755787cf6-5gvlv\" (UID: \"08efd21e-517e-46dc-bf88-60cfb1943208\") " pod="openstack/neutron-7755787cf6-5gvlv" Oct 08 18:34:05 crc kubenswrapper[4859]: I1008 18:34:05.493629 4859 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovndb-tls-certs\" (UniqueName: \"kubernetes.io/secret/08efd21e-517e-46dc-bf88-60cfb1943208-ovndb-tls-certs\") pod \"neutron-7755787cf6-5gvlv\" (UID: \"08efd21e-517e-46dc-bf88-60cfb1943208\") " pod="openstack/neutron-7755787cf6-5gvlv" Oct 08 18:34:05 crc kubenswrapper[4859]: I1008 18:34:05.493679 4859 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-wgpl2\" (UniqueName: \"kubernetes.io/projected/08efd21e-517e-46dc-bf88-60cfb1943208-kube-api-access-wgpl2\") pod \"neutron-7755787cf6-5gvlv\" (UID: \"08efd21e-517e-46dc-bf88-60cfb1943208\") " pod="openstack/neutron-7755787cf6-5gvlv" Oct 08 18:34:05 crc kubenswrapper[4859]: I1008 18:34:05.493723 4859 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/08efd21e-517e-46dc-bf88-60cfb1943208-combined-ca-bundle\") pod \"neutron-7755787cf6-5gvlv\" (UID: \"08efd21e-517e-46dc-bf88-60cfb1943208\") " pod="openstack/neutron-7755787cf6-5gvlv" Oct 08 18:34:05 crc kubenswrapper[4859]: I1008 18:34:05.493772 4859 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/08efd21e-517e-46dc-bf88-60cfb1943208-config\") pod \"neutron-7755787cf6-5gvlv\" (UID: \"08efd21e-517e-46dc-bf88-60cfb1943208\") " pod="openstack/neutron-7755787cf6-5gvlv" Oct 08 18:34:05 crc kubenswrapper[4859]: I1008 18:34:05.497249 4859 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/secret/08efd21e-517e-46dc-bf88-60cfb1943208-config\") pod \"neutron-7755787cf6-5gvlv\" (UID: \"08efd21e-517e-46dc-bf88-60cfb1943208\") " pod="openstack/neutron-7755787cf6-5gvlv" Oct 08 18:34:05 crc kubenswrapper[4859]: I1008 18:34:05.497562 4859 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/08efd21e-517e-46dc-bf88-60cfb1943208-combined-ca-bundle\") pod \"neutron-7755787cf6-5gvlv\" (UID: \"08efd21e-517e-46dc-bf88-60cfb1943208\") " pod="openstack/neutron-7755787cf6-5gvlv" Oct 08 18:34:05 crc kubenswrapper[4859]: I1008 18:34:05.497607 4859 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovndb-tls-certs\" (UniqueName: \"kubernetes.io/secret/08efd21e-517e-46dc-bf88-60cfb1943208-ovndb-tls-certs\") pod \"neutron-7755787cf6-5gvlv\" (UID: \"08efd21e-517e-46dc-bf88-60cfb1943208\") " pod="openstack/neutron-7755787cf6-5gvlv" Oct 08 18:34:05 crc kubenswrapper[4859]: I1008 18:34:05.511775 4859 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-wgpl2\" (UniqueName: \"kubernetes.io/projected/08efd21e-517e-46dc-bf88-60cfb1943208-kube-api-access-wgpl2\") pod \"neutron-7755787cf6-5gvlv\" (UID: \"08efd21e-517e-46dc-bf88-60cfb1943208\") " pod="openstack/neutron-7755787cf6-5gvlv" Oct 08 18:34:05 crc kubenswrapper[4859]: I1008 18:34:05.519675 4859 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"httpd-config\" (UniqueName: \"kubernetes.io/secret/08efd21e-517e-46dc-bf88-60cfb1943208-httpd-config\") pod \"neutron-7755787cf6-5gvlv\" (UID: \"08efd21e-517e-46dc-bf88-60cfb1943208\") " pod="openstack/neutron-7755787cf6-5gvlv" Oct 08 18:34:05 crc kubenswrapper[4859]: I1008 18:34:05.582531 4859 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-7755787cf6-5gvlv" Oct 08 18:34:06 crc kubenswrapper[4859]: I1008 18:34:06.483739 4859 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="9ed85886-9589-4d26-ad59-7a6021df659e" path="/var/lib/kubelet/pods/9ed85886-9589-4d26-ad59-7a6021df659e/volumes" Oct 08 18:34:06 crc kubenswrapper[4859]: I1008 18:34:06.602372 4859 scope.go:117] "RemoveContainer" containerID="6413f62ae3e9c179cba99f16d4f32c63326b5a3dc4e809c333d5364e3aac6c77" Oct 08 18:34:06 crc kubenswrapper[4859]: E1008 18:34:06.623398 4859 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/podified-antelope-centos9/openstack-cinder-api@sha256:85c75d60e1bd2f8a9ea0a2bb21a8df64c0a6f7b504cc1a05a355981d4b90e92f" Oct 08 18:34:06 crc kubenswrapper[4859]: E1008 18:34:06.623573 4859 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:cinder-db-sync,Image:quay.io/podified-antelope-centos9/openstack-cinder-api@sha256:85c75d60e1bd2f8a9ea0a2bb21a8df64c0a6f7b504cc1a05a355981d4b90e92f,Command:[/bin/bash],Args:[-c /usr/local/bin/kolla_set_configs && /usr/local/bin/kolla_start],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:KOLLA_BOOTSTRAP,Value:TRUE,ValueFrom:nil,},EnvVar{Name:KOLLA_CONFIG_STRATEGY,Value:COPY_ALWAYS,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:etc-machine-id,ReadOnly:true,MountPath:/etc/machine-id,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:scripts,ReadOnly:true,MountPath:/usr/local/bin/container-scripts,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:config-data,ReadOnly:true,MountPath:/var/lib/config-data/merged,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:config-data,ReadOnly:true,MountPath:/etc/my.cnf,SubPath:my.cnf,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:db-sync-config-data,ReadOnly:true,MountPath:/etc/cinder/cinder.conf.d,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:config-data,ReadOnly:true,MountPath:/var/lib/kolla/config_files/config.json,SubPath:db-sync-config.json,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:combined-ca-bundle,ReadOnly:true,MountPath:/etc/pki/ca-trust/extracted/pem/tls-ca-bundle.pem,SubPath:tls-ca-bundle.pem,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-bdbmw,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:nil,Privileged:nil,SELinuxOptions:nil,RunAsUser:*0,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:nil,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod cinder-db-sync-jqwrh_openstack(720db6c0-f18f-429b-9589-fa503d212139): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Oct 08 18:34:06 crc kubenswrapper[4859]: E1008 18:34:06.624738 4859 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"cinder-db-sync\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack/cinder-db-sync-jqwrh" podUID="720db6c0-f18f-429b-9589-fa503d212139" Oct 08 18:34:06 crc kubenswrapper[4859]: I1008 18:34:06.823947 4859 scope.go:117] "RemoveContainer" containerID="a44772a4cd0d4ee3d17393a0bc729acb973b7a3199f86de123b075f6994eae7c" Oct 08 18:34:06 crc kubenswrapper[4859]: I1008 18:34:06.910642 4859 scope.go:117] "RemoveContainer" containerID="72c30502678312848e2a6186412b97211a55d0089e7d406b78c0cc1efb9c2f27" Oct 08 18:34:06 crc kubenswrapper[4859]: I1008 18:34:06.955982 4859 scope.go:117] "RemoveContainer" containerID="4277ff351d5cc7e7b03ffd72a0f5d7bb37123fc379df3099fb15b5d20798ff85" Oct 08 18:34:07 crc kubenswrapper[4859]: I1008 18:34:07.012246 4859 scope.go:117] "RemoveContainer" containerID="313d9ef521bf8874aff0eb11214820ac2b18d10f20ca005846085c0cb3d3a769" Oct 08 18:34:07 crc kubenswrapper[4859]: I1008 18:34:07.378527 4859 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-db-sync-lsbrj" event={"ID":"ed95b639-b2c8-4879-bedb-16681e0166cf","Type":"ContainerStarted","Data":"e90f0759e2ee9de59ae4c6f1fc51a484554c5850b56b12d4c621266811a5593d"} Oct 08 18:34:07 crc kubenswrapper[4859]: I1008 18:34:07.396900 4859 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/horizon-7d58687c8-fhctx"] Oct 08 18:34:07 crc kubenswrapper[4859]: I1008 18:34:07.409324 4859 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"87f0eecb-178a-438a-8643-e29c19390884","Type":"ContainerStarted","Data":"2a699d4ef1d913cf94815cb3b9d6ef92b6f7bcdf4ce58dbd6b5b6b78663afd86"} Oct 08 18:34:07 crc kubenswrapper[4859]: I1008 18:34:07.409367 4859 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/horizon-6f569d5894-mx8v5"] Oct 08 18:34:07 crc kubenswrapper[4859]: E1008 18:34:07.413939 4859 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"cinder-db-sync\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/podified-antelope-centos9/openstack-cinder-api@sha256:85c75d60e1bd2f8a9ea0a2bb21a8df64c0a6f7b504cc1a05a355981d4b90e92f\\\"\"" pod="openstack/cinder-db-sync-jqwrh" podUID="720db6c0-f18f-429b-9589-fa503d212139" Oct 08 18:34:07 crc kubenswrapper[4859]: I1008 18:34:07.417678 4859 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/placement-db-sync-lsbrj" podStartSLOduration=5.430558213 podStartE2EDuration="34.417654084s" podCreationTimestamp="2025-10-08 18:33:33 +0000 UTC" firstStartedPulling="2025-10-08 18:33:34.725898595 +0000 UTC m=+984.972737974" lastFinishedPulling="2025-10-08 18:34:03.712994466 +0000 UTC m=+1013.959833845" observedRunningTime="2025-10-08 18:34:07.408151335 +0000 UTC m=+1017.654990714" watchObservedRunningTime="2025-10-08 18:34:07.417654084 +0000 UTC m=+1017.664493463" Oct 08 18:34:07 crc kubenswrapper[4859]: I1008 18:34:07.435673 4859 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-58799d59b9-5kp6v"] Oct 08 18:34:07 crc kubenswrapper[4859]: I1008 18:34:07.490834 4859 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-internal-api-0"] Oct 08 18:34:07 crc kubenswrapper[4859]: W1008 18:34:07.495525 4859 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod7bf3f855_41ae_4de9_ba6c_0fb4f51df141.slice/crio-3eb8702bed8879b712d748acd343a4ab86e37d62f33dc654c6a14848f6eaccc5 WatchSource:0}: Error finding container 3eb8702bed8879b712d748acd343a4ab86e37d62f33dc654c6a14848f6eaccc5: Status 404 returned error can't find the container with id 3eb8702bed8879b712d748acd343a4ab86e37d62f33dc654c6a14848f6eaccc5 Oct 08 18:34:07 crc kubenswrapper[4859]: I1008 18:34:07.558586 4859 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-bootstrap-75br8"] Oct 08 18:34:07 crc kubenswrapper[4859]: I1008 18:34:07.566730 4859 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/neutron-5f9b9b6f45-wr5fx"] Oct 08 18:34:07 crc kubenswrapper[4859]: I1008 18:34:07.568301 4859 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-5f9b9b6f45-wr5fx" Oct 08 18:34:07 crc kubenswrapper[4859]: I1008 18:34:07.570386 4859 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-neutron-public-svc" Oct 08 18:34:07 crc kubenswrapper[4859]: I1008 18:34:07.570873 4859 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-neutron-internal-svc" Oct 08 18:34:07 crc kubenswrapper[4859]: I1008 18:34:07.603027 4859 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-5f9b9b6f45-wr5fx"] Oct 08 18:34:07 crc kubenswrapper[4859]: W1008 18:34:07.618823 4859 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod91a7b430_b4d6_49f6_8951_0d62d51d6685.slice/crio-fd001032ca6f5caa4b862ecbbe60f937772c5f0121bd0d6c6a8cf97142847e3f WatchSource:0}: Error finding container fd001032ca6f5caa4b862ecbbe60f937772c5f0121bd0d6c6a8cf97142847e3f: Status 404 returned error can't find the container with id fd001032ca6f5caa4b862ecbbe60f937772c5f0121bd0d6c6a8cf97142847e3f Oct 08 18:34:07 crc kubenswrapper[4859]: I1008 18:34:07.652483 4859 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-external-api-0"] Oct 08 18:34:07 crc kubenswrapper[4859]: I1008 18:34:07.725271 4859 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-7755787cf6-5gvlv"] Oct 08 18:34:07 crc kubenswrapper[4859]: I1008 18:34:07.745030 4859 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/f64840cf-331e-4d8d-b102-49225713df66-internal-tls-certs\") pod \"neutron-5f9b9b6f45-wr5fx\" (UID: \"f64840cf-331e-4d8d-b102-49225713df66\") " pod="openstack/neutron-5f9b9b6f45-wr5fx" Oct 08 18:34:07 crc kubenswrapper[4859]: I1008 18:34:07.745093 4859 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/f64840cf-331e-4d8d-b102-49225713df66-public-tls-certs\") pod \"neutron-5f9b9b6f45-wr5fx\" (UID: \"f64840cf-331e-4d8d-b102-49225713df66\") " pod="openstack/neutron-5f9b9b6f45-wr5fx" Oct 08 18:34:07 crc kubenswrapper[4859]: I1008 18:34:07.745149 4859 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/f64840cf-331e-4d8d-b102-49225713df66-config\") pod \"neutron-5f9b9b6f45-wr5fx\" (UID: \"f64840cf-331e-4d8d-b102-49225713df66\") " pod="openstack/neutron-5f9b9b6f45-wr5fx" Oct 08 18:34:07 crc kubenswrapper[4859]: I1008 18:34:07.745186 4859 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"httpd-config\" (UniqueName: \"kubernetes.io/secret/f64840cf-331e-4d8d-b102-49225713df66-httpd-config\") pod \"neutron-5f9b9b6f45-wr5fx\" (UID: \"f64840cf-331e-4d8d-b102-49225713df66\") " pod="openstack/neutron-5f9b9b6f45-wr5fx" Oct 08 18:34:07 crc kubenswrapper[4859]: I1008 18:34:07.745288 4859 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f64840cf-331e-4d8d-b102-49225713df66-combined-ca-bundle\") pod \"neutron-5f9b9b6f45-wr5fx\" (UID: \"f64840cf-331e-4d8d-b102-49225713df66\") " pod="openstack/neutron-5f9b9b6f45-wr5fx" Oct 08 18:34:07 crc kubenswrapper[4859]: I1008 18:34:07.745315 4859 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-rzjfc\" (UniqueName: \"kubernetes.io/projected/f64840cf-331e-4d8d-b102-49225713df66-kube-api-access-rzjfc\") pod \"neutron-5f9b9b6f45-wr5fx\" (UID: \"f64840cf-331e-4d8d-b102-49225713df66\") " pod="openstack/neutron-5f9b9b6f45-wr5fx" Oct 08 18:34:07 crc kubenswrapper[4859]: I1008 18:34:07.745369 4859 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovndb-tls-certs\" (UniqueName: \"kubernetes.io/secret/f64840cf-331e-4d8d-b102-49225713df66-ovndb-tls-certs\") pod \"neutron-5f9b9b6f45-wr5fx\" (UID: \"f64840cf-331e-4d8d-b102-49225713df66\") " pod="openstack/neutron-5f9b9b6f45-wr5fx" Oct 08 18:34:07 crc kubenswrapper[4859]: I1008 18:34:07.846847 4859 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/f64840cf-331e-4d8d-b102-49225713df66-config\") pod \"neutron-5f9b9b6f45-wr5fx\" (UID: \"f64840cf-331e-4d8d-b102-49225713df66\") " pod="openstack/neutron-5f9b9b6f45-wr5fx" Oct 08 18:34:07 crc kubenswrapper[4859]: I1008 18:34:07.846902 4859 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"httpd-config\" (UniqueName: \"kubernetes.io/secret/f64840cf-331e-4d8d-b102-49225713df66-httpd-config\") pod \"neutron-5f9b9b6f45-wr5fx\" (UID: \"f64840cf-331e-4d8d-b102-49225713df66\") " pod="openstack/neutron-5f9b9b6f45-wr5fx" Oct 08 18:34:07 crc kubenswrapper[4859]: I1008 18:34:07.847001 4859 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f64840cf-331e-4d8d-b102-49225713df66-combined-ca-bundle\") pod \"neutron-5f9b9b6f45-wr5fx\" (UID: \"f64840cf-331e-4d8d-b102-49225713df66\") " pod="openstack/neutron-5f9b9b6f45-wr5fx" Oct 08 18:34:07 crc kubenswrapper[4859]: I1008 18:34:07.847020 4859 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rzjfc\" (UniqueName: \"kubernetes.io/projected/f64840cf-331e-4d8d-b102-49225713df66-kube-api-access-rzjfc\") pod \"neutron-5f9b9b6f45-wr5fx\" (UID: \"f64840cf-331e-4d8d-b102-49225713df66\") " pod="openstack/neutron-5f9b9b6f45-wr5fx" Oct 08 18:34:07 crc kubenswrapper[4859]: I1008 18:34:07.847064 4859 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovndb-tls-certs\" (UniqueName: \"kubernetes.io/secret/f64840cf-331e-4d8d-b102-49225713df66-ovndb-tls-certs\") pod \"neutron-5f9b9b6f45-wr5fx\" (UID: \"f64840cf-331e-4d8d-b102-49225713df66\") " pod="openstack/neutron-5f9b9b6f45-wr5fx" Oct 08 18:34:07 crc kubenswrapper[4859]: I1008 18:34:07.847101 4859 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/f64840cf-331e-4d8d-b102-49225713df66-internal-tls-certs\") pod \"neutron-5f9b9b6f45-wr5fx\" (UID: \"f64840cf-331e-4d8d-b102-49225713df66\") " pod="openstack/neutron-5f9b9b6f45-wr5fx" Oct 08 18:34:07 crc kubenswrapper[4859]: I1008 18:34:07.847117 4859 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/f64840cf-331e-4d8d-b102-49225713df66-public-tls-certs\") pod \"neutron-5f9b9b6f45-wr5fx\" (UID: \"f64840cf-331e-4d8d-b102-49225713df66\") " pod="openstack/neutron-5f9b9b6f45-wr5fx" Oct 08 18:34:07 crc kubenswrapper[4859]: I1008 18:34:07.852118 4859 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/f64840cf-331e-4d8d-b102-49225713df66-public-tls-certs\") pod \"neutron-5f9b9b6f45-wr5fx\" (UID: \"f64840cf-331e-4d8d-b102-49225713df66\") " pod="openstack/neutron-5f9b9b6f45-wr5fx" Oct 08 18:34:07 crc kubenswrapper[4859]: I1008 18:34:07.853561 4859 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/secret/f64840cf-331e-4d8d-b102-49225713df66-config\") pod \"neutron-5f9b9b6f45-wr5fx\" (UID: \"f64840cf-331e-4d8d-b102-49225713df66\") " pod="openstack/neutron-5f9b9b6f45-wr5fx" Oct 08 18:34:07 crc kubenswrapper[4859]: I1008 18:34:07.853734 4859 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/f64840cf-331e-4d8d-b102-49225713df66-internal-tls-certs\") pod \"neutron-5f9b9b6f45-wr5fx\" (UID: \"f64840cf-331e-4d8d-b102-49225713df66\") " pod="openstack/neutron-5f9b9b6f45-wr5fx" Oct 08 18:34:07 crc kubenswrapper[4859]: I1008 18:34:07.854005 4859 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f64840cf-331e-4d8d-b102-49225713df66-combined-ca-bundle\") pod \"neutron-5f9b9b6f45-wr5fx\" (UID: \"f64840cf-331e-4d8d-b102-49225713df66\") " pod="openstack/neutron-5f9b9b6f45-wr5fx" Oct 08 18:34:07 crc kubenswrapper[4859]: I1008 18:34:07.861483 4859 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"httpd-config\" (UniqueName: \"kubernetes.io/secret/f64840cf-331e-4d8d-b102-49225713df66-httpd-config\") pod \"neutron-5f9b9b6f45-wr5fx\" (UID: \"f64840cf-331e-4d8d-b102-49225713df66\") " pod="openstack/neutron-5f9b9b6f45-wr5fx" Oct 08 18:34:07 crc kubenswrapper[4859]: I1008 18:34:07.865250 4859 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovndb-tls-certs\" (UniqueName: \"kubernetes.io/secret/f64840cf-331e-4d8d-b102-49225713df66-ovndb-tls-certs\") pod \"neutron-5f9b9b6f45-wr5fx\" (UID: \"f64840cf-331e-4d8d-b102-49225713df66\") " pod="openstack/neutron-5f9b9b6f45-wr5fx" Oct 08 18:34:07 crc kubenswrapper[4859]: I1008 18:34:07.868237 4859 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rzjfc\" (UniqueName: \"kubernetes.io/projected/f64840cf-331e-4d8d-b102-49225713df66-kube-api-access-rzjfc\") pod \"neutron-5f9b9b6f45-wr5fx\" (UID: \"f64840cf-331e-4d8d-b102-49225713df66\") " pod="openstack/neutron-5f9b9b6f45-wr5fx" Oct 08 18:34:08 crc kubenswrapper[4859]: I1008 18:34:08.120354 4859 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-5f9b9b6f45-wr5fx" Oct 08 18:34:08 crc kubenswrapper[4859]: I1008 18:34:08.463271 4859 generic.go:334] "Generic (PLEG): container finished" podID="881d266c-6730-404d-8243-ea0df0d3d03f" containerID="0079be4d4c960d2bcd1f78238bedd74e2b721bf336a7f9bd857e41e20ed129f6" exitCode=0 Oct 08 18:34:08 crc kubenswrapper[4859]: I1008 18:34:08.463732 4859 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-58799d59b9-5kp6v" event={"ID":"881d266c-6730-404d-8243-ea0df0d3d03f","Type":"ContainerDied","Data":"0079be4d4c960d2bcd1f78238bedd74e2b721bf336a7f9bd857e41e20ed129f6"} Oct 08 18:34:08 crc kubenswrapper[4859]: I1008 18:34:08.463765 4859 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-58799d59b9-5kp6v" event={"ID":"881d266c-6730-404d-8243-ea0df0d3d03f","Type":"ContainerStarted","Data":"5361fd0c4f9aa7d00088e20917cbe941dd6b0518e8bed856f763ca12404f277f"} Oct 08 18:34:08 crc kubenswrapper[4859]: I1008 18:34:08.518635 4859 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-6f569d5894-mx8v5" event={"ID":"50b5de31-90f6-43cf-9e16-7b89b09f1e57","Type":"ContainerStarted","Data":"bf49e814b3bae872b621c2db2effa8146c3370f75eb21768e975a63ec2a44f71"} Oct 08 18:34:08 crc kubenswrapper[4859]: I1008 18:34:08.518977 4859 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-6f569d5894-mx8v5" event={"ID":"50b5de31-90f6-43cf-9e16-7b89b09f1e57","Type":"ContainerStarted","Data":"71918abb02b628c141e31cb218a488f190407b3490118205bad9359808d8efe7"} Oct 08 18:34:08 crc kubenswrapper[4859]: I1008 18:34:08.518987 4859 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"b8433969-f079-49a2-95f3-ce4783d6a62c","Type":"ContainerStarted","Data":"4aa06945ba8e9a81faeb45633d2bc0119700cecd4455590a24af48f3a27a60db"} Oct 08 18:34:08 crc kubenswrapper[4859]: I1008 18:34:08.518998 4859 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-7d58687c8-fhctx" event={"ID":"1a92487e-3309-41c4-8c82-cd57f517ab4b","Type":"ContainerStarted","Data":"fb897bd998eaff60715c21b1cde92b421d7bf9ae9f4afad9a1ed7b20f7cfd7b3"} Oct 08 18:34:08 crc kubenswrapper[4859]: I1008 18:34:08.519008 4859 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-7d58687c8-fhctx" event={"ID":"1a92487e-3309-41c4-8c82-cd57f517ab4b","Type":"ContainerStarted","Data":"c546fc2ab05e42f55afac42efd39b6705f14b4294bf04b9895522878307c7ccf"} Oct 08 18:34:08 crc kubenswrapper[4859]: I1008 18:34:08.519018 4859 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-bootstrap-75br8" event={"ID":"91a7b430-b4d6-49f6-8951-0d62d51d6685","Type":"ContainerStarted","Data":"912ebc52cf289ca586e23593221698f35f08e44164fc786655a89a1fa335f018"} Oct 08 18:34:08 crc kubenswrapper[4859]: I1008 18:34:08.519028 4859 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-bootstrap-75br8" event={"ID":"91a7b430-b4d6-49f6-8951-0d62d51d6685","Type":"ContainerStarted","Data":"fd001032ca6f5caa4b862ecbbe60f937772c5f0121bd0d6c6a8cf97142847e3f"} Oct 08 18:34:08 crc kubenswrapper[4859]: I1008 18:34:08.519037 4859 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-7755787cf6-5gvlv" event={"ID":"08efd21e-517e-46dc-bf88-60cfb1943208","Type":"ContainerStarted","Data":"9c658a235d8e9170379bfc9b21179e87097616e2d09b4f9696d1fc6548e1ac2c"} Oct 08 18:34:08 crc kubenswrapper[4859]: I1008 18:34:08.519045 4859 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-7755787cf6-5gvlv" event={"ID":"08efd21e-517e-46dc-bf88-60cfb1943208","Type":"ContainerStarted","Data":"90e37068995838dc25b4e6fb72406bd39a3a12e4c276b08c48491e71c0dc8d00"} Oct 08 18:34:08 crc kubenswrapper[4859]: I1008 18:34:08.519054 4859 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"7bf3f855-41ae-4de9-ba6c-0fb4f51df141","Type":"ContainerStarted","Data":"3eb8702bed8879b712d748acd343a4ab86e37d62f33dc654c6a14848f6eaccc5"} Oct 08 18:34:08 crc kubenswrapper[4859]: I1008 18:34:08.599348 4859 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/keystone-bootstrap-75br8" podStartSLOduration=20.599328549 podStartE2EDuration="20.599328549s" podCreationTimestamp="2025-10-08 18:33:48 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-08 18:34:08.523434389 +0000 UTC m=+1018.770273768" watchObservedRunningTime="2025-10-08 18:34:08.599328549 +0000 UTC m=+1018.846167928" Oct 08 18:34:08 crc kubenswrapper[4859]: I1008 18:34:08.856667 4859 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-5f9b9b6f45-wr5fx"] Oct 08 18:34:08 crc kubenswrapper[4859]: W1008 18:34:08.890703 4859 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podf64840cf_331e_4d8d_b102_49225713df66.slice/crio-0f054d3e7058bfa49f97db0ac43863573b3aef178768a6a7df370bd534aca07e WatchSource:0}: Error finding container 0f054d3e7058bfa49f97db0ac43863573b3aef178768a6a7df370bd534aca07e: Status 404 returned error can't find the container with id 0f054d3e7058bfa49f97db0ac43863573b3aef178768a6a7df370bd534aca07e Oct 08 18:34:09 crc kubenswrapper[4859]: I1008 18:34:09.521296 4859 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"b8433969-f079-49a2-95f3-ce4783d6a62c","Type":"ContainerStarted","Data":"e58a2a7e6d7cb1671e6f15be2eae9753bf9ba14ba510c49cc6d4d7121ba024ce"} Oct 08 18:34:09 crc kubenswrapper[4859]: I1008 18:34:09.531432 4859 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-7d58687c8-fhctx" event={"ID":"1a92487e-3309-41c4-8c82-cd57f517ab4b","Type":"ContainerStarted","Data":"648c622ce9eb0bdf3d9c43d8313db6636c307abe8d840bde30cf40e0039e99c7"} Oct 08 18:34:09 crc kubenswrapper[4859]: I1008 18:34:09.538058 4859 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-7755787cf6-5gvlv" event={"ID":"08efd21e-517e-46dc-bf88-60cfb1943208","Type":"ContainerStarted","Data":"2822847f53055edb0f62021465c0ea6cda6373ec67373966a4dcf91c5c45ddef"} Oct 08 18:34:09 crc kubenswrapper[4859]: I1008 18:34:09.539187 4859 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/neutron-7755787cf6-5gvlv" Oct 08 18:34:09 crc kubenswrapper[4859]: I1008 18:34:09.549082 4859 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"7bf3f855-41ae-4de9-ba6c-0fb4f51df141","Type":"ContainerStarted","Data":"46ed96c026d937d863774320b18d16cbedb2414f358afc23e4af1c83103bb4e5"} Oct 08 18:34:09 crc kubenswrapper[4859]: I1008 18:34:09.549118 4859 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"7bf3f855-41ae-4de9-ba6c-0fb4f51df141","Type":"ContainerStarted","Data":"28ec01a1442339cbd321103beebc7df267b3a533827492d256693d3f04972c5d"} Oct 08 18:34:09 crc kubenswrapper[4859]: I1008 18:34:09.562240 4859 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/horizon-7d58687c8-fhctx" podStartSLOduration=28.105073754 podStartE2EDuration="28.562222902s" podCreationTimestamp="2025-10-08 18:33:41 +0000 UTC" firstStartedPulling="2025-10-08 18:34:07.420437016 +0000 UTC m=+1017.667276395" lastFinishedPulling="2025-10-08 18:34:07.877586164 +0000 UTC m=+1018.124425543" observedRunningTime="2025-10-08 18:34:09.554856516 +0000 UTC m=+1019.801695925" watchObservedRunningTime="2025-10-08 18:34:09.562222902 +0000 UTC m=+1019.809062271" Oct 08 18:34:09 crc kubenswrapper[4859]: I1008 18:34:09.563775 4859 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-58799d59b9-5kp6v" event={"ID":"881d266c-6730-404d-8243-ea0df0d3d03f","Type":"ContainerStarted","Data":"30a2952f16a49c92b8051f736d0387ff2fd3296b5cd3c22fcca6085b250996db"} Oct 08 18:34:09 crc kubenswrapper[4859]: I1008 18:34:09.563994 4859 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-58799d59b9-5kp6v" Oct 08 18:34:09 crc kubenswrapper[4859]: I1008 18:34:09.570496 4859 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-6f569d5894-mx8v5" event={"ID":"50b5de31-90f6-43cf-9e16-7b89b09f1e57","Type":"ContainerStarted","Data":"6cf92563d64bd83f385b041cec37f00035787026875ed167c9efb3f508b85d6c"} Oct 08 18:34:09 crc kubenswrapper[4859]: I1008 18:34:09.575347 4859 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/neutron-7755787cf6-5gvlv" podStartSLOduration=4.575331268 podStartE2EDuration="4.575331268s" podCreationTimestamp="2025-10-08 18:34:05 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-08 18:34:09.572940497 +0000 UTC m=+1019.819779876" watchObservedRunningTime="2025-10-08 18:34:09.575331268 +0000 UTC m=+1019.822170647" Oct 08 18:34:09 crc kubenswrapper[4859]: I1008 18:34:09.580550 4859 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-5f9b9b6f45-wr5fx" event={"ID":"f64840cf-331e-4d8d-b102-49225713df66","Type":"ContainerStarted","Data":"0f054d3e7058bfa49f97db0ac43863573b3aef178768a6a7df370bd534aca07e"} Oct 08 18:34:09 crc kubenswrapper[4859]: I1008 18:34:09.611177 4859 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/glance-default-internal-api-0" podStartSLOduration=21.611159371 podStartE2EDuration="21.611159371s" podCreationTimestamp="2025-10-08 18:33:48 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-08 18:34:09.594179982 +0000 UTC m=+1019.841019361" watchObservedRunningTime="2025-10-08 18:34:09.611159371 +0000 UTC m=+1019.857998760" Oct 08 18:34:09 crc kubenswrapper[4859]: I1008 18:34:09.628503 4859 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-58799d59b9-5kp6v" podStartSLOduration=4.6284887900000005 podStartE2EDuration="4.62848879s" podCreationTimestamp="2025-10-08 18:34:05 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-08 18:34:09.626041738 +0000 UTC m=+1019.872881107" watchObservedRunningTime="2025-10-08 18:34:09.62848879 +0000 UTC m=+1019.875328169" Oct 08 18:34:09 crc kubenswrapper[4859]: I1008 18:34:09.677289 4859 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/horizon-6f569d5894-mx8v5" podStartSLOduration=27.176512704 podStartE2EDuration="27.677273564s" podCreationTimestamp="2025-10-08 18:33:42 +0000 UTC" firstStartedPulling="2025-10-08 18:34:07.440118244 +0000 UTC m=+1017.686957623" lastFinishedPulling="2025-10-08 18:34:07.940879104 +0000 UTC m=+1018.187718483" observedRunningTime="2025-10-08 18:34:09.657246795 +0000 UTC m=+1019.904086174" watchObservedRunningTime="2025-10-08 18:34:09.677273564 +0000 UTC m=+1019.924112943" Oct 08 18:34:10 crc kubenswrapper[4859]: I1008 18:34:10.617101 4859 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"87f0eecb-178a-438a-8643-e29c19390884","Type":"ContainerStarted","Data":"c76586e5e90236f90858e341d25b5dde536c874c2bca5a632ffc704ce10dd611"} Oct 08 18:34:10 crc kubenswrapper[4859]: I1008 18:34:10.629312 4859 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-5f9b9b6f45-wr5fx" event={"ID":"f64840cf-331e-4d8d-b102-49225713df66","Type":"ContainerStarted","Data":"b1c4ee2dc279e69960c1b0e89bf33994ea2cf4ab948f4fd0236882791cfc0132"} Oct 08 18:34:10 crc kubenswrapper[4859]: I1008 18:34:10.629355 4859 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-5f9b9b6f45-wr5fx" event={"ID":"f64840cf-331e-4d8d-b102-49225713df66","Type":"ContainerStarted","Data":"9c1b90fbc1a269c2376965a6e8bda9e0c2292da6ade00831318d080337c92532"} Oct 08 18:34:10 crc kubenswrapper[4859]: I1008 18:34:10.630198 4859 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/neutron-5f9b9b6f45-wr5fx" Oct 08 18:34:10 crc kubenswrapper[4859]: I1008 18:34:10.636946 4859 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"b8433969-f079-49a2-95f3-ce4783d6a62c","Type":"ContainerStarted","Data":"3d02d784fce2d216c2d98154bf821d302aeba5a184bb06406551e52cbd463ff4"} Oct 08 18:34:10 crc kubenswrapper[4859]: I1008 18:34:10.641940 4859 generic.go:334] "Generic (PLEG): container finished" podID="ed95b639-b2c8-4879-bedb-16681e0166cf" containerID="e90f0759e2ee9de59ae4c6f1fc51a484554c5850b56b12d4c621266811a5593d" exitCode=0 Oct 08 18:34:10 crc kubenswrapper[4859]: I1008 18:34:10.643423 4859 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-db-sync-lsbrj" event={"ID":"ed95b639-b2c8-4879-bedb-16681e0166cf","Type":"ContainerDied","Data":"e90f0759e2ee9de59ae4c6f1fc51a484554c5850b56b12d4c621266811a5593d"} Oct 08 18:34:10 crc kubenswrapper[4859]: I1008 18:34:10.662489 4859 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/neutron-5f9b9b6f45-wr5fx" podStartSLOduration=3.662467344 podStartE2EDuration="3.662467344s" podCreationTimestamp="2025-10-08 18:34:07 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-08 18:34:10.646962568 +0000 UTC m=+1020.893801957" watchObservedRunningTime="2025-10-08 18:34:10.662467344 +0000 UTC m=+1020.909306723" Oct 08 18:34:10 crc kubenswrapper[4859]: I1008 18:34:10.697871 4859 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/glance-default-external-api-0" podStartSLOduration=22.697846904 podStartE2EDuration="22.697846904s" podCreationTimestamp="2025-10-08 18:33:48 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-08 18:34:10.690676753 +0000 UTC m=+1020.937516152" watchObservedRunningTime="2025-10-08 18:34:10.697846904 +0000 UTC m=+1020.944686283" Oct 08 18:34:11 crc kubenswrapper[4859]: I1008 18:34:11.652522 4859 generic.go:334] "Generic (PLEG): container finished" podID="91a7b430-b4d6-49f6-8951-0d62d51d6685" containerID="912ebc52cf289ca586e23593221698f35f08e44164fc786655a89a1fa335f018" exitCode=0 Oct 08 18:34:11 crc kubenswrapper[4859]: I1008 18:34:11.653259 4859 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-bootstrap-75br8" event={"ID":"91a7b430-b4d6-49f6-8951-0d62d51d6685","Type":"ContainerDied","Data":"912ebc52cf289ca586e23593221698f35f08e44164fc786655a89a1fa335f018"} Oct 08 18:34:12 crc kubenswrapper[4859]: I1008 18:34:12.307271 4859 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/horizon-7d58687c8-fhctx" Oct 08 18:34:12 crc kubenswrapper[4859]: I1008 18:34:12.307645 4859 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/horizon-7d58687c8-fhctx" Oct 08 18:34:12 crc kubenswrapper[4859]: I1008 18:34:12.445047 4859 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/horizon-6f569d5894-mx8v5" Oct 08 18:34:12 crc kubenswrapper[4859]: I1008 18:34:12.445110 4859 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/horizon-6f569d5894-mx8v5" Oct 08 18:34:13 crc kubenswrapper[4859]: I1008 18:34:13.203195 4859 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/placement-db-sync-lsbrj" Oct 08 18:34:13 crc kubenswrapper[4859]: I1008 18:34:13.219753 4859 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-bootstrap-75br8" Oct 08 18:34:13 crc kubenswrapper[4859]: I1008 18:34:13.283459 4859 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/91a7b430-b4d6-49f6-8951-0d62d51d6685-config-data\") pod \"91a7b430-b4d6-49f6-8951-0d62d51d6685\" (UID: \"91a7b430-b4d6-49f6-8951-0d62d51d6685\") " Oct 08 18:34:13 crc kubenswrapper[4859]: I1008 18:34:13.283504 4859 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/ed95b639-b2c8-4879-bedb-16681e0166cf-scripts\") pod \"ed95b639-b2c8-4879-bedb-16681e0166cf\" (UID: \"ed95b639-b2c8-4879-bedb-16681e0166cf\") " Oct 08 18:34:13 crc kubenswrapper[4859]: I1008 18:34:13.283536 4859 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/ed95b639-b2c8-4879-bedb-16681e0166cf-logs\") pod \"ed95b639-b2c8-4879-bedb-16681e0166cf\" (UID: \"ed95b639-b2c8-4879-bedb-16681e0166cf\") " Oct 08 18:34:13 crc kubenswrapper[4859]: I1008 18:34:13.283614 4859 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-bhq6l\" (UniqueName: \"kubernetes.io/projected/91a7b430-b4d6-49f6-8951-0d62d51d6685-kube-api-access-bhq6l\") pod \"91a7b430-b4d6-49f6-8951-0d62d51d6685\" (UID: \"91a7b430-b4d6-49f6-8951-0d62d51d6685\") " Oct 08 18:34:13 crc kubenswrapper[4859]: I1008 18:34:13.283671 4859 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/91a7b430-b4d6-49f6-8951-0d62d51d6685-fernet-keys\") pod \"91a7b430-b4d6-49f6-8951-0d62d51d6685\" (UID: \"91a7b430-b4d6-49f6-8951-0d62d51d6685\") " Oct 08 18:34:13 crc kubenswrapper[4859]: I1008 18:34:13.283714 4859 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/91a7b430-b4d6-49f6-8951-0d62d51d6685-scripts\") pod \"91a7b430-b4d6-49f6-8951-0d62d51d6685\" (UID: \"91a7b430-b4d6-49f6-8951-0d62d51d6685\") " Oct 08 18:34:13 crc kubenswrapper[4859]: I1008 18:34:13.283815 4859 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ed95b639-b2c8-4879-bedb-16681e0166cf-config-data\") pod \"ed95b639-b2c8-4879-bedb-16681e0166cf\" (UID: \"ed95b639-b2c8-4879-bedb-16681e0166cf\") " Oct 08 18:34:13 crc kubenswrapper[4859]: I1008 18:34:13.283898 4859 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/91a7b430-b4d6-49f6-8951-0d62d51d6685-combined-ca-bundle\") pod \"91a7b430-b4d6-49f6-8951-0d62d51d6685\" (UID: \"91a7b430-b4d6-49f6-8951-0d62d51d6685\") " Oct 08 18:34:13 crc kubenswrapper[4859]: I1008 18:34:13.283997 4859 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/91a7b430-b4d6-49f6-8951-0d62d51d6685-credential-keys\") pod \"91a7b430-b4d6-49f6-8951-0d62d51d6685\" (UID: \"91a7b430-b4d6-49f6-8951-0d62d51d6685\") " Oct 08 18:34:13 crc kubenswrapper[4859]: I1008 18:34:13.284087 4859 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-fjb7b\" (UniqueName: \"kubernetes.io/projected/ed95b639-b2c8-4879-bedb-16681e0166cf-kube-api-access-fjb7b\") pod \"ed95b639-b2c8-4879-bedb-16681e0166cf\" (UID: \"ed95b639-b2c8-4879-bedb-16681e0166cf\") " Oct 08 18:34:13 crc kubenswrapper[4859]: I1008 18:34:13.284136 4859 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ed95b639-b2c8-4879-bedb-16681e0166cf-combined-ca-bundle\") pod \"ed95b639-b2c8-4879-bedb-16681e0166cf\" (UID: \"ed95b639-b2c8-4879-bedb-16681e0166cf\") " Oct 08 18:34:13 crc kubenswrapper[4859]: I1008 18:34:13.284158 4859 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/ed95b639-b2c8-4879-bedb-16681e0166cf-logs" (OuterVolumeSpecName: "logs") pod "ed95b639-b2c8-4879-bedb-16681e0166cf" (UID: "ed95b639-b2c8-4879-bedb-16681e0166cf"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 08 18:34:13 crc kubenswrapper[4859]: I1008 18:34:13.285347 4859 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/ed95b639-b2c8-4879-bedb-16681e0166cf-logs\") on node \"crc\" DevicePath \"\"" Oct 08 18:34:13 crc kubenswrapper[4859]: I1008 18:34:13.293049 4859 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/ed95b639-b2c8-4879-bedb-16681e0166cf-scripts" (OuterVolumeSpecName: "scripts") pod "ed95b639-b2c8-4879-bedb-16681e0166cf" (UID: "ed95b639-b2c8-4879-bedb-16681e0166cf"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 08 18:34:13 crc kubenswrapper[4859]: I1008 18:34:13.302298 4859 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/91a7b430-b4d6-49f6-8951-0d62d51d6685-scripts" (OuterVolumeSpecName: "scripts") pod "91a7b430-b4d6-49f6-8951-0d62d51d6685" (UID: "91a7b430-b4d6-49f6-8951-0d62d51d6685"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 08 18:34:13 crc kubenswrapper[4859]: I1008 18:34:13.304635 4859 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/91a7b430-b4d6-49f6-8951-0d62d51d6685-credential-keys" (OuterVolumeSpecName: "credential-keys") pod "91a7b430-b4d6-49f6-8951-0d62d51d6685" (UID: "91a7b430-b4d6-49f6-8951-0d62d51d6685"). InnerVolumeSpecName "credential-keys". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 08 18:34:13 crc kubenswrapper[4859]: I1008 18:34:13.305247 4859 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/91a7b430-b4d6-49f6-8951-0d62d51d6685-kube-api-access-bhq6l" (OuterVolumeSpecName: "kube-api-access-bhq6l") pod "91a7b430-b4d6-49f6-8951-0d62d51d6685" (UID: "91a7b430-b4d6-49f6-8951-0d62d51d6685"). InnerVolumeSpecName "kube-api-access-bhq6l". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 08 18:34:13 crc kubenswrapper[4859]: I1008 18:34:13.310569 4859 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/ed95b639-b2c8-4879-bedb-16681e0166cf-kube-api-access-fjb7b" (OuterVolumeSpecName: "kube-api-access-fjb7b") pod "ed95b639-b2c8-4879-bedb-16681e0166cf" (UID: "ed95b639-b2c8-4879-bedb-16681e0166cf"). InnerVolumeSpecName "kube-api-access-fjb7b". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 08 18:34:13 crc kubenswrapper[4859]: I1008 18:34:13.314013 4859 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/91a7b430-b4d6-49f6-8951-0d62d51d6685-fernet-keys" (OuterVolumeSpecName: "fernet-keys") pod "91a7b430-b4d6-49f6-8951-0d62d51d6685" (UID: "91a7b430-b4d6-49f6-8951-0d62d51d6685"). InnerVolumeSpecName "fernet-keys". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 08 18:34:13 crc kubenswrapper[4859]: I1008 18:34:13.318954 4859 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/91a7b430-b4d6-49f6-8951-0d62d51d6685-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "91a7b430-b4d6-49f6-8951-0d62d51d6685" (UID: "91a7b430-b4d6-49f6-8951-0d62d51d6685"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 08 18:34:13 crc kubenswrapper[4859]: I1008 18:34:13.319885 4859 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/ed95b639-b2c8-4879-bedb-16681e0166cf-config-data" (OuterVolumeSpecName: "config-data") pod "ed95b639-b2c8-4879-bedb-16681e0166cf" (UID: "ed95b639-b2c8-4879-bedb-16681e0166cf"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 08 18:34:13 crc kubenswrapper[4859]: I1008 18:34:13.321504 4859 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/ed95b639-b2c8-4879-bedb-16681e0166cf-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "ed95b639-b2c8-4879-bedb-16681e0166cf" (UID: "ed95b639-b2c8-4879-bedb-16681e0166cf"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 08 18:34:13 crc kubenswrapper[4859]: I1008 18:34:13.345633 4859 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/91a7b430-b4d6-49f6-8951-0d62d51d6685-config-data" (OuterVolumeSpecName: "config-data") pod "91a7b430-b4d6-49f6-8951-0d62d51d6685" (UID: "91a7b430-b4d6-49f6-8951-0d62d51d6685"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 08 18:34:13 crc kubenswrapper[4859]: I1008 18:34:13.386923 4859 reconciler_common.go:293] "Volume detached for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/91a7b430-b4d6-49f6-8951-0d62d51d6685-fernet-keys\") on node \"crc\" DevicePath \"\"" Oct 08 18:34:13 crc kubenswrapper[4859]: I1008 18:34:13.387738 4859 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/91a7b430-b4d6-49f6-8951-0d62d51d6685-scripts\") on node \"crc\" DevicePath \"\"" Oct 08 18:34:13 crc kubenswrapper[4859]: I1008 18:34:13.387748 4859 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ed95b639-b2c8-4879-bedb-16681e0166cf-config-data\") on node \"crc\" DevicePath \"\"" Oct 08 18:34:13 crc kubenswrapper[4859]: I1008 18:34:13.387758 4859 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/91a7b430-b4d6-49f6-8951-0d62d51d6685-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 08 18:34:13 crc kubenswrapper[4859]: I1008 18:34:13.387769 4859 reconciler_common.go:293] "Volume detached for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/91a7b430-b4d6-49f6-8951-0d62d51d6685-credential-keys\") on node \"crc\" DevicePath \"\"" Oct 08 18:34:13 crc kubenswrapper[4859]: I1008 18:34:13.387778 4859 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-fjb7b\" (UniqueName: \"kubernetes.io/projected/ed95b639-b2c8-4879-bedb-16681e0166cf-kube-api-access-fjb7b\") on node \"crc\" DevicePath \"\"" Oct 08 18:34:13 crc kubenswrapper[4859]: I1008 18:34:13.387786 4859 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ed95b639-b2c8-4879-bedb-16681e0166cf-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 08 18:34:13 crc kubenswrapper[4859]: I1008 18:34:13.387795 4859 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/91a7b430-b4d6-49f6-8951-0d62d51d6685-config-data\") on node \"crc\" DevicePath \"\"" Oct 08 18:34:13 crc kubenswrapper[4859]: I1008 18:34:13.387802 4859 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/ed95b639-b2c8-4879-bedb-16681e0166cf-scripts\") on node \"crc\" DevicePath \"\"" Oct 08 18:34:13 crc kubenswrapper[4859]: I1008 18:34:13.387812 4859 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-bhq6l\" (UniqueName: \"kubernetes.io/projected/91a7b430-b4d6-49f6-8951-0d62d51d6685-kube-api-access-bhq6l\") on node \"crc\" DevicePath \"\"" Oct 08 18:34:13 crc kubenswrapper[4859]: I1008 18:34:13.671769 4859 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/placement-db-sync-lsbrj" Oct 08 18:34:13 crc kubenswrapper[4859]: I1008 18:34:13.672821 4859 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-db-sync-lsbrj" event={"ID":"ed95b639-b2c8-4879-bedb-16681e0166cf","Type":"ContainerDied","Data":"df7e476e592572851bdab7449f59a5936c9d9eb10ebb99616a2965aeba03e560"} Oct 08 18:34:13 crc kubenswrapper[4859]: I1008 18:34:13.672881 4859 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="df7e476e592572851bdab7449f59a5936c9d9eb10ebb99616a2965aeba03e560" Oct 08 18:34:13 crc kubenswrapper[4859]: I1008 18:34:13.694093 4859 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-bootstrap-75br8" event={"ID":"91a7b430-b4d6-49f6-8951-0d62d51d6685","Type":"ContainerDied","Data":"fd001032ca6f5caa4b862ecbbe60f937772c5f0121bd0d6c6a8cf97142847e3f"} Oct 08 18:34:13 crc kubenswrapper[4859]: I1008 18:34:13.694146 4859 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="fd001032ca6f5caa4b862ecbbe60f937772c5f0121bd0d6c6a8cf97142847e3f" Oct 08 18:34:13 crc kubenswrapper[4859]: I1008 18:34:13.694119 4859 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-bootstrap-75br8" Oct 08 18:34:13 crc kubenswrapper[4859]: I1008 18:34:13.792541 4859 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/keystone-758ff8b98b-p6wm9"] Oct 08 18:34:13 crc kubenswrapper[4859]: E1008 18:34:13.793061 4859 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="91a7b430-b4d6-49f6-8951-0d62d51d6685" containerName="keystone-bootstrap" Oct 08 18:34:13 crc kubenswrapper[4859]: I1008 18:34:13.793085 4859 state_mem.go:107] "Deleted CPUSet assignment" podUID="91a7b430-b4d6-49f6-8951-0d62d51d6685" containerName="keystone-bootstrap" Oct 08 18:34:13 crc kubenswrapper[4859]: E1008 18:34:13.793105 4859 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ed95b639-b2c8-4879-bedb-16681e0166cf" containerName="placement-db-sync" Oct 08 18:34:13 crc kubenswrapper[4859]: I1008 18:34:13.793112 4859 state_mem.go:107] "Deleted CPUSet assignment" podUID="ed95b639-b2c8-4879-bedb-16681e0166cf" containerName="placement-db-sync" Oct 08 18:34:13 crc kubenswrapper[4859]: I1008 18:34:13.793338 4859 memory_manager.go:354] "RemoveStaleState removing state" podUID="ed95b639-b2c8-4879-bedb-16681e0166cf" containerName="placement-db-sync" Oct 08 18:34:13 crc kubenswrapper[4859]: I1008 18:34:13.793364 4859 memory_manager.go:354] "RemoveStaleState removing state" podUID="91a7b430-b4d6-49f6-8951-0d62d51d6685" containerName="keystone-bootstrap" Oct 08 18:34:13 crc kubenswrapper[4859]: I1008 18:34:13.794216 4859 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-758ff8b98b-p6wm9" Oct 08 18:34:13 crc kubenswrapper[4859]: I1008 18:34:13.796985 4859 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone" Oct 08 18:34:13 crc kubenswrapper[4859]: I1008 18:34:13.797283 4859 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-keystone-public-svc" Oct 08 18:34:13 crc kubenswrapper[4859]: I1008 18:34:13.797480 4859 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-keystone-internal-svc" Oct 08 18:34:13 crc kubenswrapper[4859]: I1008 18:34:13.797735 4859 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-config-data" Oct 08 18:34:13 crc kubenswrapper[4859]: I1008 18:34:13.810273 4859 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-scripts" Oct 08 18:34:13 crc kubenswrapper[4859]: I1008 18:34:13.810515 4859 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-keystone-dockercfg-jrkml" Oct 08 18:34:13 crc kubenswrapper[4859]: I1008 18:34:13.812009 4859 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-758ff8b98b-p6wm9"] Oct 08 18:34:13 crc kubenswrapper[4859]: I1008 18:34:13.894912 4859 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/285bdd26-6109-4191-870a-2fde949a6a28-internal-tls-certs\") pod \"keystone-758ff8b98b-p6wm9\" (UID: \"285bdd26-6109-4191-870a-2fde949a6a28\") " pod="openstack/keystone-758ff8b98b-p6wm9" Oct 08 18:34:13 crc kubenswrapper[4859]: I1008 18:34:13.895074 4859 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/285bdd26-6109-4191-870a-2fde949a6a28-config-data\") pod \"keystone-758ff8b98b-p6wm9\" (UID: \"285bdd26-6109-4191-870a-2fde949a6a28\") " pod="openstack/keystone-758ff8b98b-p6wm9" Oct 08 18:34:13 crc kubenswrapper[4859]: I1008 18:34:13.895175 4859 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-glxsx\" (UniqueName: \"kubernetes.io/projected/285bdd26-6109-4191-870a-2fde949a6a28-kube-api-access-glxsx\") pod \"keystone-758ff8b98b-p6wm9\" (UID: \"285bdd26-6109-4191-870a-2fde949a6a28\") " pod="openstack/keystone-758ff8b98b-p6wm9" Oct 08 18:34:13 crc kubenswrapper[4859]: I1008 18:34:13.895224 4859 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/285bdd26-6109-4191-870a-2fde949a6a28-combined-ca-bundle\") pod \"keystone-758ff8b98b-p6wm9\" (UID: \"285bdd26-6109-4191-870a-2fde949a6a28\") " pod="openstack/keystone-758ff8b98b-p6wm9" Oct 08 18:34:13 crc kubenswrapper[4859]: I1008 18:34:13.895312 4859 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/285bdd26-6109-4191-870a-2fde949a6a28-public-tls-certs\") pod \"keystone-758ff8b98b-p6wm9\" (UID: \"285bdd26-6109-4191-870a-2fde949a6a28\") " pod="openstack/keystone-758ff8b98b-p6wm9" Oct 08 18:34:13 crc kubenswrapper[4859]: I1008 18:34:13.895362 4859 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/285bdd26-6109-4191-870a-2fde949a6a28-fernet-keys\") pod \"keystone-758ff8b98b-p6wm9\" (UID: \"285bdd26-6109-4191-870a-2fde949a6a28\") " pod="openstack/keystone-758ff8b98b-p6wm9" Oct 08 18:34:13 crc kubenswrapper[4859]: I1008 18:34:13.895395 4859 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/285bdd26-6109-4191-870a-2fde949a6a28-credential-keys\") pod \"keystone-758ff8b98b-p6wm9\" (UID: \"285bdd26-6109-4191-870a-2fde949a6a28\") " pod="openstack/keystone-758ff8b98b-p6wm9" Oct 08 18:34:13 crc kubenswrapper[4859]: I1008 18:34:13.895621 4859 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/285bdd26-6109-4191-870a-2fde949a6a28-scripts\") pod \"keystone-758ff8b98b-p6wm9\" (UID: \"285bdd26-6109-4191-870a-2fde949a6a28\") " pod="openstack/keystone-758ff8b98b-p6wm9" Oct 08 18:34:13 crc kubenswrapper[4859]: I1008 18:34:13.996972 4859 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/285bdd26-6109-4191-870a-2fde949a6a28-fernet-keys\") pod \"keystone-758ff8b98b-p6wm9\" (UID: \"285bdd26-6109-4191-870a-2fde949a6a28\") " pod="openstack/keystone-758ff8b98b-p6wm9" Oct 08 18:34:13 crc kubenswrapper[4859]: I1008 18:34:13.997030 4859 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/285bdd26-6109-4191-870a-2fde949a6a28-credential-keys\") pod \"keystone-758ff8b98b-p6wm9\" (UID: \"285bdd26-6109-4191-870a-2fde949a6a28\") " pod="openstack/keystone-758ff8b98b-p6wm9" Oct 08 18:34:13 crc kubenswrapper[4859]: I1008 18:34:13.997116 4859 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/285bdd26-6109-4191-870a-2fde949a6a28-scripts\") pod \"keystone-758ff8b98b-p6wm9\" (UID: \"285bdd26-6109-4191-870a-2fde949a6a28\") " pod="openstack/keystone-758ff8b98b-p6wm9" Oct 08 18:34:13 crc kubenswrapper[4859]: I1008 18:34:13.997167 4859 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/285bdd26-6109-4191-870a-2fde949a6a28-internal-tls-certs\") pod \"keystone-758ff8b98b-p6wm9\" (UID: \"285bdd26-6109-4191-870a-2fde949a6a28\") " pod="openstack/keystone-758ff8b98b-p6wm9" Oct 08 18:34:13 crc kubenswrapper[4859]: I1008 18:34:13.997213 4859 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/285bdd26-6109-4191-870a-2fde949a6a28-config-data\") pod \"keystone-758ff8b98b-p6wm9\" (UID: \"285bdd26-6109-4191-870a-2fde949a6a28\") " pod="openstack/keystone-758ff8b98b-p6wm9" Oct 08 18:34:13 crc kubenswrapper[4859]: I1008 18:34:13.997260 4859 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-glxsx\" (UniqueName: \"kubernetes.io/projected/285bdd26-6109-4191-870a-2fde949a6a28-kube-api-access-glxsx\") pod \"keystone-758ff8b98b-p6wm9\" (UID: \"285bdd26-6109-4191-870a-2fde949a6a28\") " pod="openstack/keystone-758ff8b98b-p6wm9" Oct 08 18:34:13 crc kubenswrapper[4859]: I1008 18:34:13.997290 4859 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/285bdd26-6109-4191-870a-2fde949a6a28-combined-ca-bundle\") pod \"keystone-758ff8b98b-p6wm9\" (UID: \"285bdd26-6109-4191-870a-2fde949a6a28\") " pod="openstack/keystone-758ff8b98b-p6wm9" Oct 08 18:34:13 crc kubenswrapper[4859]: I1008 18:34:13.997346 4859 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/285bdd26-6109-4191-870a-2fde949a6a28-public-tls-certs\") pod \"keystone-758ff8b98b-p6wm9\" (UID: \"285bdd26-6109-4191-870a-2fde949a6a28\") " pod="openstack/keystone-758ff8b98b-p6wm9" Oct 08 18:34:14 crc kubenswrapper[4859]: I1008 18:34:14.002571 4859 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/285bdd26-6109-4191-870a-2fde949a6a28-public-tls-certs\") pod \"keystone-758ff8b98b-p6wm9\" (UID: \"285bdd26-6109-4191-870a-2fde949a6a28\") " pod="openstack/keystone-758ff8b98b-p6wm9" Oct 08 18:34:14 crc kubenswrapper[4859]: I1008 18:34:14.002599 4859 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/285bdd26-6109-4191-870a-2fde949a6a28-fernet-keys\") pod \"keystone-758ff8b98b-p6wm9\" (UID: \"285bdd26-6109-4191-870a-2fde949a6a28\") " pod="openstack/keystone-758ff8b98b-p6wm9" Oct 08 18:34:14 crc kubenswrapper[4859]: I1008 18:34:14.003930 4859 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/285bdd26-6109-4191-870a-2fde949a6a28-credential-keys\") pod \"keystone-758ff8b98b-p6wm9\" (UID: \"285bdd26-6109-4191-870a-2fde949a6a28\") " pod="openstack/keystone-758ff8b98b-p6wm9" Oct 08 18:34:14 crc kubenswrapper[4859]: I1008 18:34:14.005251 4859 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/285bdd26-6109-4191-870a-2fde949a6a28-internal-tls-certs\") pod \"keystone-758ff8b98b-p6wm9\" (UID: \"285bdd26-6109-4191-870a-2fde949a6a28\") " pod="openstack/keystone-758ff8b98b-p6wm9" Oct 08 18:34:14 crc kubenswrapper[4859]: I1008 18:34:14.008402 4859 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/285bdd26-6109-4191-870a-2fde949a6a28-config-data\") pod \"keystone-758ff8b98b-p6wm9\" (UID: \"285bdd26-6109-4191-870a-2fde949a6a28\") " pod="openstack/keystone-758ff8b98b-p6wm9" Oct 08 18:34:14 crc kubenswrapper[4859]: I1008 18:34:14.010119 4859 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/285bdd26-6109-4191-870a-2fde949a6a28-scripts\") pod \"keystone-758ff8b98b-p6wm9\" (UID: \"285bdd26-6109-4191-870a-2fde949a6a28\") " pod="openstack/keystone-758ff8b98b-p6wm9" Oct 08 18:34:14 crc kubenswrapper[4859]: I1008 18:34:14.011240 4859 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/285bdd26-6109-4191-870a-2fde949a6a28-combined-ca-bundle\") pod \"keystone-758ff8b98b-p6wm9\" (UID: \"285bdd26-6109-4191-870a-2fde949a6a28\") " pod="openstack/keystone-758ff8b98b-p6wm9" Oct 08 18:34:14 crc kubenswrapper[4859]: I1008 18:34:14.024768 4859 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-glxsx\" (UniqueName: \"kubernetes.io/projected/285bdd26-6109-4191-870a-2fde949a6a28-kube-api-access-glxsx\") pod \"keystone-758ff8b98b-p6wm9\" (UID: \"285bdd26-6109-4191-870a-2fde949a6a28\") " pod="openstack/keystone-758ff8b98b-p6wm9" Oct 08 18:34:14 crc kubenswrapper[4859]: I1008 18:34:14.120771 4859 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-758ff8b98b-p6wm9" Oct 08 18:34:14 crc kubenswrapper[4859]: I1008 18:34:14.309292 4859 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/placement-58c5f664f4-5lvrs"] Oct 08 18:34:14 crc kubenswrapper[4859]: I1008 18:34:14.311226 4859 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/placement-58c5f664f4-5lvrs" Oct 08 18:34:14 crc kubenswrapper[4859]: I1008 18:34:14.313435 4859 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"placement-scripts" Oct 08 18:34:14 crc kubenswrapper[4859]: I1008 18:34:14.313587 4859 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-placement-public-svc" Oct 08 18:34:14 crc kubenswrapper[4859]: I1008 18:34:14.314183 4859 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"placement-config-data" Oct 08 18:34:14 crc kubenswrapper[4859]: I1008 18:34:14.315497 4859 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"placement-placement-dockercfg-gwsrg" Oct 08 18:34:14 crc kubenswrapper[4859]: I1008 18:34:14.315764 4859 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-placement-internal-svc" Oct 08 18:34:14 crc kubenswrapper[4859]: I1008 18:34:14.322761 4859 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/placement-58c5f664f4-5lvrs"] Oct 08 18:34:14 crc kubenswrapper[4859]: I1008 18:34:14.404589 4859 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-hdpbt\" (UniqueName: \"kubernetes.io/projected/01262af8-82d8-4771-9fe8-472dc77e4b60-kube-api-access-hdpbt\") pod \"placement-58c5f664f4-5lvrs\" (UID: \"01262af8-82d8-4771-9fe8-472dc77e4b60\") " pod="openstack/placement-58c5f664f4-5lvrs" Oct 08 18:34:14 crc kubenswrapper[4859]: I1008 18:34:14.404769 4859 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/01262af8-82d8-4771-9fe8-472dc77e4b60-internal-tls-certs\") pod \"placement-58c5f664f4-5lvrs\" (UID: \"01262af8-82d8-4771-9fe8-472dc77e4b60\") " pod="openstack/placement-58c5f664f4-5lvrs" Oct 08 18:34:14 crc kubenswrapper[4859]: I1008 18:34:14.404800 4859 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/01262af8-82d8-4771-9fe8-472dc77e4b60-config-data\") pod \"placement-58c5f664f4-5lvrs\" (UID: \"01262af8-82d8-4771-9fe8-472dc77e4b60\") " pod="openstack/placement-58c5f664f4-5lvrs" Oct 08 18:34:14 crc kubenswrapper[4859]: I1008 18:34:14.404839 4859 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/01262af8-82d8-4771-9fe8-472dc77e4b60-logs\") pod \"placement-58c5f664f4-5lvrs\" (UID: \"01262af8-82d8-4771-9fe8-472dc77e4b60\") " pod="openstack/placement-58c5f664f4-5lvrs" Oct 08 18:34:14 crc kubenswrapper[4859]: I1008 18:34:14.404870 4859 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/01262af8-82d8-4771-9fe8-472dc77e4b60-scripts\") pod \"placement-58c5f664f4-5lvrs\" (UID: \"01262af8-82d8-4771-9fe8-472dc77e4b60\") " pod="openstack/placement-58c5f664f4-5lvrs" Oct 08 18:34:14 crc kubenswrapper[4859]: I1008 18:34:14.404892 4859 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/01262af8-82d8-4771-9fe8-472dc77e4b60-public-tls-certs\") pod \"placement-58c5f664f4-5lvrs\" (UID: \"01262af8-82d8-4771-9fe8-472dc77e4b60\") " pod="openstack/placement-58c5f664f4-5lvrs" Oct 08 18:34:14 crc kubenswrapper[4859]: I1008 18:34:14.405047 4859 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/01262af8-82d8-4771-9fe8-472dc77e4b60-combined-ca-bundle\") pod \"placement-58c5f664f4-5lvrs\" (UID: \"01262af8-82d8-4771-9fe8-472dc77e4b60\") " pod="openstack/placement-58c5f664f4-5lvrs" Oct 08 18:34:14 crc kubenswrapper[4859]: I1008 18:34:14.506011 4859 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-hdpbt\" (UniqueName: \"kubernetes.io/projected/01262af8-82d8-4771-9fe8-472dc77e4b60-kube-api-access-hdpbt\") pod \"placement-58c5f664f4-5lvrs\" (UID: \"01262af8-82d8-4771-9fe8-472dc77e4b60\") " pod="openstack/placement-58c5f664f4-5lvrs" Oct 08 18:34:14 crc kubenswrapper[4859]: I1008 18:34:14.506350 4859 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/01262af8-82d8-4771-9fe8-472dc77e4b60-internal-tls-certs\") pod \"placement-58c5f664f4-5lvrs\" (UID: \"01262af8-82d8-4771-9fe8-472dc77e4b60\") " pod="openstack/placement-58c5f664f4-5lvrs" Oct 08 18:34:14 crc kubenswrapper[4859]: I1008 18:34:14.506489 4859 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/01262af8-82d8-4771-9fe8-472dc77e4b60-config-data\") pod \"placement-58c5f664f4-5lvrs\" (UID: \"01262af8-82d8-4771-9fe8-472dc77e4b60\") " pod="openstack/placement-58c5f664f4-5lvrs" Oct 08 18:34:14 crc kubenswrapper[4859]: I1008 18:34:14.506591 4859 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/01262af8-82d8-4771-9fe8-472dc77e4b60-logs\") pod \"placement-58c5f664f4-5lvrs\" (UID: \"01262af8-82d8-4771-9fe8-472dc77e4b60\") " pod="openstack/placement-58c5f664f4-5lvrs" Oct 08 18:34:14 crc kubenswrapper[4859]: I1008 18:34:14.506722 4859 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/01262af8-82d8-4771-9fe8-472dc77e4b60-scripts\") pod \"placement-58c5f664f4-5lvrs\" (UID: \"01262af8-82d8-4771-9fe8-472dc77e4b60\") " pod="openstack/placement-58c5f664f4-5lvrs" Oct 08 18:34:14 crc kubenswrapper[4859]: I1008 18:34:14.506819 4859 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/01262af8-82d8-4771-9fe8-472dc77e4b60-public-tls-certs\") pod \"placement-58c5f664f4-5lvrs\" (UID: \"01262af8-82d8-4771-9fe8-472dc77e4b60\") " pod="openstack/placement-58c5f664f4-5lvrs" Oct 08 18:34:14 crc kubenswrapper[4859]: I1008 18:34:14.506960 4859 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/01262af8-82d8-4771-9fe8-472dc77e4b60-combined-ca-bundle\") pod \"placement-58c5f664f4-5lvrs\" (UID: \"01262af8-82d8-4771-9fe8-472dc77e4b60\") " pod="openstack/placement-58c5f664f4-5lvrs" Oct 08 18:34:14 crc kubenswrapper[4859]: I1008 18:34:14.508857 4859 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/01262af8-82d8-4771-9fe8-472dc77e4b60-logs\") pod \"placement-58c5f664f4-5lvrs\" (UID: \"01262af8-82d8-4771-9fe8-472dc77e4b60\") " pod="openstack/placement-58c5f664f4-5lvrs" Oct 08 18:34:14 crc kubenswrapper[4859]: I1008 18:34:14.511200 4859 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/01262af8-82d8-4771-9fe8-472dc77e4b60-internal-tls-certs\") pod \"placement-58c5f664f4-5lvrs\" (UID: \"01262af8-82d8-4771-9fe8-472dc77e4b60\") " pod="openstack/placement-58c5f664f4-5lvrs" Oct 08 18:34:14 crc kubenswrapper[4859]: I1008 18:34:14.512540 4859 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/01262af8-82d8-4771-9fe8-472dc77e4b60-combined-ca-bundle\") pod \"placement-58c5f664f4-5lvrs\" (UID: \"01262af8-82d8-4771-9fe8-472dc77e4b60\") " pod="openstack/placement-58c5f664f4-5lvrs" Oct 08 18:34:14 crc kubenswrapper[4859]: I1008 18:34:14.512987 4859 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/01262af8-82d8-4771-9fe8-472dc77e4b60-scripts\") pod \"placement-58c5f664f4-5lvrs\" (UID: \"01262af8-82d8-4771-9fe8-472dc77e4b60\") " pod="openstack/placement-58c5f664f4-5lvrs" Oct 08 18:34:14 crc kubenswrapper[4859]: I1008 18:34:14.518139 4859 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/01262af8-82d8-4771-9fe8-472dc77e4b60-public-tls-certs\") pod \"placement-58c5f664f4-5lvrs\" (UID: \"01262af8-82d8-4771-9fe8-472dc77e4b60\") " pod="openstack/placement-58c5f664f4-5lvrs" Oct 08 18:34:14 crc kubenswrapper[4859]: I1008 18:34:14.523446 4859 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/01262af8-82d8-4771-9fe8-472dc77e4b60-config-data\") pod \"placement-58c5f664f4-5lvrs\" (UID: \"01262af8-82d8-4771-9fe8-472dc77e4b60\") " pod="openstack/placement-58c5f664f4-5lvrs" Oct 08 18:34:14 crc kubenswrapper[4859]: I1008 18:34:14.524327 4859 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-hdpbt\" (UniqueName: \"kubernetes.io/projected/01262af8-82d8-4771-9fe8-472dc77e4b60-kube-api-access-hdpbt\") pod \"placement-58c5f664f4-5lvrs\" (UID: \"01262af8-82d8-4771-9fe8-472dc77e4b60\") " pod="openstack/placement-58c5f664f4-5lvrs" Oct 08 18:34:14 crc kubenswrapper[4859]: I1008 18:34:14.631786 4859 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/placement-58c5f664f4-5lvrs" Oct 08 18:34:15 crc kubenswrapper[4859]: I1008 18:34:15.410049 4859 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-58799d59b9-5kp6v" Oct 08 18:34:15 crc kubenswrapper[4859]: I1008 18:34:15.470308 4859 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-74776f5dd7-89sbw"] Oct 08 18:34:15 crc kubenswrapper[4859]: I1008 18:34:15.471467 4859 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-74776f5dd7-89sbw" podUID="f08c21e6-a74c-4ad9-a3fa-4e56ff2c6432" containerName="dnsmasq-dns" containerID="cri-o://f4cadb69d4dbae11278e1c5580f08f34fd3b6a5866e898ae3481fabd25c41b30" gracePeriod=10 Oct 08 18:34:15 crc kubenswrapper[4859]: I1008 18:34:15.721117 4859 generic.go:334] "Generic (PLEG): container finished" podID="f08c21e6-a74c-4ad9-a3fa-4e56ff2c6432" containerID="f4cadb69d4dbae11278e1c5580f08f34fd3b6a5866e898ae3481fabd25c41b30" exitCode=0 Oct 08 18:34:15 crc kubenswrapper[4859]: I1008 18:34:15.721445 4859 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-74776f5dd7-89sbw" event={"ID":"f08c21e6-a74c-4ad9-a3fa-4e56ff2c6432","Type":"ContainerDied","Data":"f4cadb69d4dbae11278e1c5580f08f34fd3b6a5866e898ae3481fabd25c41b30"} Oct 08 18:34:16 crc kubenswrapper[4859]: I1008 18:34:16.056413 4859 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-74776f5dd7-89sbw" Oct 08 18:34:16 crc kubenswrapper[4859]: W1008 18:34:16.217586 4859 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod285bdd26_6109_4191_870a_2fde949a6a28.slice/crio-d624fed55ddf0106b6dca5f33345a4e18b856125229a517d5dedc82f9f9392f4 WatchSource:0}: Error finding container d624fed55ddf0106b6dca5f33345a4e18b856125229a517d5dedc82f9f9392f4: Status 404 returned error can't find the container with id d624fed55ddf0106b6dca5f33345a4e18b856125229a517d5dedc82f9f9392f4 Oct 08 18:34:16 crc kubenswrapper[4859]: I1008 18:34:16.218393 4859 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-758ff8b98b-p6wm9"] Oct 08 18:34:16 crc kubenswrapper[4859]: I1008 18:34:16.242247 4859 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/f08c21e6-a74c-4ad9-a3fa-4e56ff2c6432-ovsdbserver-sb\") pod \"f08c21e6-a74c-4ad9-a3fa-4e56ff2c6432\" (UID: \"f08c21e6-a74c-4ad9-a3fa-4e56ff2c6432\") " Oct 08 18:34:16 crc kubenswrapper[4859]: I1008 18:34:16.242667 4859 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-4czsn\" (UniqueName: \"kubernetes.io/projected/f08c21e6-a74c-4ad9-a3fa-4e56ff2c6432-kube-api-access-4czsn\") pod \"f08c21e6-a74c-4ad9-a3fa-4e56ff2c6432\" (UID: \"f08c21e6-a74c-4ad9-a3fa-4e56ff2c6432\") " Oct 08 18:34:16 crc kubenswrapper[4859]: I1008 18:34:16.242740 4859 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/f08c21e6-a74c-4ad9-a3fa-4e56ff2c6432-ovsdbserver-nb\") pod \"f08c21e6-a74c-4ad9-a3fa-4e56ff2c6432\" (UID: \"f08c21e6-a74c-4ad9-a3fa-4e56ff2c6432\") " Oct 08 18:34:16 crc kubenswrapper[4859]: I1008 18:34:16.242771 4859 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/f08c21e6-a74c-4ad9-a3fa-4e56ff2c6432-config\") pod \"f08c21e6-a74c-4ad9-a3fa-4e56ff2c6432\" (UID: \"f08c21e6-a74c-4ad9-a3fa-4e56ff2c6432\") " Oct 08 18:34:16 crc kubenswrapper[4859]: I1008 18:34:16.242824 4859 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/f08c21e6-a74c-4ad9-a3fa-4e56ff2c6432-dns-svc\") pod \"f08c21e6-a74c-4ad9-a3fa-4e56ff2c6432\" (UID: \"f08c21e6-a74c-4ad9-a3fa-4e56ff2c6432\") " Oct 08 18:34:16 crc kubenswrapper[4859]: I1008 18:34:16.242864 4859 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/f08c21e6-a74c-4ad9-a3fa-4e56ff2c6432-dns-swift-storage-0\") pod \"f08c21e6-a74c-4ad9-a3fa-4e56ff2c6432\" (UID: \"f08c21e6-a74c-4ad9-a3fa-4e56ff2c6432\") " Oct 08 18:34:16 crc kubenswrapper[4859]: I1008 18:34:16.252885 4859 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/f08c21e6-a74c-4ad9-a3fa-4e56ff2c6432-kube-api-access-4czsn" (OuterVolumeSpecName: "kube-api-access-4czsn") pod "f08c21e6-a74c-4ad9-a3fa-4e56ff2c6432" (UID: "f08c21e6-a74c-4ad9-a3fa-4e56ff2c6432"). InnerVolumeSpecName "kube-api-access-4czsn". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 08 18:34:16 crc kubenswrapper[4859]: I1008 18:34:16.291111 4859 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/f08c21e6-a74c-4ad9-a3fa-4e56ff2c6432-dns-swift-storage-0" (OuterVolumeSpecName: "dns-swift-storage-0") pod "f08c21e6-a74c-4ad9-a3fa-4e56ff2c6432" (UID: "f08c21e6-a74c-4ad9-a3fa-4e56ff2c6432"). InnerVolumeSpecName "dns-swift-storage-0". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 08 18:34:16 crc kubenswrapper[4859]: I1008 18:34:16.294562 4859 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/f08c21e6-a74c-4ad9-a3fa-4e56ff2c6432-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "f08c21e6-a74c-4ad9-a3fa-4e56ff2c6432" (UID: "f08c21e6-a74c-4ad9-a3fa-4e56ff2c6432"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 08 18:34:16 crc kubenswrapper[4859]: I1008 18:34:16.297202 4859 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/f08c21e6-a74c-4ad9-a3fa-4e56ff2c6432-config" (OuterVolumeSpecName: "config") pod "f08c21e6-a74c-4ad9-a3fa-4e56ff2c6432" (UID: "f08c21e6-a74c-4ad9-a3fa-4e56ff2c6432"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 08 18:34:16 crc kubenswrapper[4859]: I1008 18:34:16.314731 4859 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/f08c21e6-a74c-4ad9-a3fa-4e56ff2c6432-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "f08c21e6-a74c-4ad9-a3fa-4e56ff2c6432" (UID: "f08c21e6-a74c-4ad9-a3fa-4e56ff2c6432"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 08 18:34:16 crc kubenswrapper[4859]: I1008 18:34:16.318515 4859 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/f08c21e6-a74c-4ad9-a3fa-4e56ff2c6432-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "f08c21e6-a74c-4ad9-a3fa-4e56ff2c6432" (UID: "f08c21e6-a74c-4ad9-a3fa-4e56ff2c6432"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 08 18:34:16 crc kubenswrapper[4859]: I1008 18:34:16.346157 4859 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/f08c21e6-a74c-4ad9-a3fa-4e56ff2c6432-dns-svc\") on node \"crc\" DevicePath \"\"" Oct 08 18:34:16 crc kubenswrapper[4859]: I1008 18:34:16.346187 4859 reconciler_common.go:293] "Volume detached for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/f08c21e6-a74c-4ad9-a3fa-4e56ff2c6432-dns-swift-storage-0\") on node \"crc\" DevicePath \"\"" Oct 08 18:34:16 crc kubenswrapper[4859]: I1008 18:34:16.346200 4859 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/f08c21e6-a74c-4ad9-a3fa-4e56ff2c6432-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Oct 08 18:34:16 crc kubenswrapper[4859]: I1008 18:34:16.346210 4859 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-4czsn\" (UniqueName: \"kubernetes.io/projected/f08c21e6-a74c-4ad9-a3fa-4e56ff2c6432-kube-api-access-4czsn\") on node \"crc\" DevicePath \"\"" Oct 08 18:34:16 crc kubenswrapper[4859]: I1008 18:34:16.346218 4859 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/f08c21e6-a74c-4ad9-a3fa-4e56ff2c6432-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Oct 08 18:34:16 crc kubenswrapper[4859]: I1008 18:34:16.346226 4859 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/f08c21e6-a74c-4ad9-a3fa-4e56ff2c6432-config\") on node \"crc\" DevicePath \"\"" Oct 08 18:34:16 crc kubenswrapper[4859]: I1008 18:34:16.396501 4859 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/placement-58c5f664f4-5lvrs"] Oct 08 18:34:16 crc kubenswrapper[4859]: W1008 18:34:16.398935 4859 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod01262af8_82d8_4771_9fe8_472dc77e4b60.slice/crio-b6ae911fb5c551aa67031c14d4d59a600ae1007b364854319dc15915b8908163 WatchSource:0}: Error finding container b6ae911fb5c551aa67031c14d4d59a600ae1007b364854319dc15915b8908163: Status 404 returned error can't find the container with id b6ae911fb5c551aa67031c14d4d59a600ae1007b364854319dc15915b8908163 Oct 08 18:34:16 crc kubenswrapper[4859]: I1008 18:34:16.748915 4859 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"87f0eecb-178a-438a-8643-e29c19390884","Type":"ContainerStarted","Data":"d4f40d2b1c23ce93761ada628d8d73a724dc6702e886ec2cd59ef35c8b2b69d8"} Oct 08 18:34:16 crc kubenswrapper[4859]: I1008 18:34:16.752332 4859 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-58c5f664f4-5lvrs" event={"ID":"01262af8-82d8-4771-9fe8-472dc77e4b60","Type":"ContainerStarted","Data":"a9e0d2ddf78918f18cacf63ec509dfc6304462204ac14be2d89d8c0c8db008f9"} Oct 08 18:34:16 crc kubenswrapper[4859]: I1008 18:34:16.752375 4859 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-58c5f664f4-5lvrs" event={"ID":"01262af8-82d8-4771-9fe8-472dc77e4b60","Type":"ContainerStarted","Data":"b6ae911fb5c551aa67031c14d4d59a600ae1007b364854319dc15915b8908163"} Oct 08 18:34:16 crc kubenswrapper[4859]: I1008 18:34:16.757453 4859 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-74776f5dd7-89sbw" event={"ID":"f08c21e6-a74c-4ad9-a3fa-4e56ff2c6432","Type":"ContainerDied","Data":"4a13491c7c60974dd907317f0881466a73694103f0829dc45b85c65b1f0e8bdb"} Oct 08 18:34:16 crc kubenswrapper[4859]: I1008 18:34:16.757501 4859 scope.go:117] "RemoveContainer" containerID="f4cadb69d4dbae11278e1c5580f08f34fd3b6a5866e898ae3481fabd25c41b30" Oct 08 18:34:16 crc kubenswrapper[4859]: I1008 18:34:16.757644 4859 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-74776f5dd7-89sbw" Oct 08 18:34:16 crc kubenswrapper[4859]: I1008 18:34:16.777392 4859 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-758ff8b98b-p6wm9" event={"ID":"285bdd26-6109-4191-870a-2fde949a6a28","Type":"ContainerStarted","Data":"4edc77153a8a5a55fb9449fbdc466518579cc599e6d7affd73663752b2864032"} Oct 08 18:34:16 crc kubenswrapper[4859]: I1008 18:34:16.777650 4859 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-758ff8b98b-p6wm9" event={"ID":"285bdd26-6109-4191-870a-2fde949a6a28","Type":"ContainerStarted","Data":"d624fed55ddf0106b6dca5f33345a4e18b856125229a517d5dedc82f9f9392f4"} Oct 08 18:34:16 crc kubenswrapper[4859]: I1008 18:34:16.777764 4859 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/keystone-758ff8b98b-p6wm9" Oct 08 18:34:16 crc kubenswrapper[4859]: I1008 18:34:16.804499 4859 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/keystone-758ff8b98b-p6wm9" podStartSLOduration=3.804480597 podStartE2EDuration="3.804480597s" podCreationTimestamp="2025-10-08 18:34:13 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-08 18:34:16.803529379 +0000 UTC m=+1027.050368758" watchObservedRunningTime="2025-10-08 18:34:16.804480597 +0000 UTC m=+1027.051319976" Oct 08 18:34:16 crc kubenswrapper[4859]: I1008 18:34:16.841972 4859 scope.go:117] "RemoveContainer" containerID="f0557242a2e57c4c6dfb47d6597787e847f098ef8c3c670c0d077683329bae49" Oct 08 18:34:16 crc kubenswrapper[4859]: I1008 18:34:16.849620 4859 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-74776f5dd7-89sbw"] Oct 08 18:34:16 crc kubenswrapper[4859]: I1008 18:34:16.861723 4859 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-74776f5dd7-89sbw"] Oct 08 18:34:17 crc kubenswrapper[4859]: I1008 18:34:17.786590 4859 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-db-sync-x99jc" event={"ID":"3b2953d1-3ebe-4d7d-b60c-6ad07d02af5d","Type":"ContainerStarted","Data":"8647fd3df54b81ebc93e3e14fe2b59cbd0e77836f881418a614ac5f2c3c4c10f"} Oct 08 18:34:17 crc kubenswrapper[4859]: I1008 18:34:17.790126 4859 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-58c5f664f4-5lvrs" event={"ID":"01262af8-82d8-4771-9fe8-472dc77e4b60","Type":"ContainerStarted","Data":"3e861a4ad9406fb0900f2e02ffa0e1bba7175d7b11900092c78dbc2a6a95f1da"} Oct 08 18:34:17 crc kubenswrapper[4859]: I1008 18:34:17.790896 4859 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/placement-58c5f664f4-5lvrs" Oct 08 18:34:17 crc kubenswrapper[4859]: I1008 18:34:17.790928 4859 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/placement-58c5f664f4-5lvrs" Oct 08 18:34:17 crc kubenswrapper[4859]: I1008 18:34:17.810323 4859 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/barbican-db-sync-x99jc" podStartSLOduration=2.324650599 podStartE2EDuration="41.810303063s" podCreationTimestamp="2025-10-08 18:33:36 +0000 UTC" firstStartedPulling="2025-10-08 18:33:37.127996574 +0000 UTC m=+987.374835953" lastFinishedPulling="2025-10-08 18:34:16.613649038 +0000 UTC m=+1026.860488417" observedRunningTime="2025-10-08 18:34:17.802108623 +0000 UTC m=+1028.048948022" watchObservedRunningTime="2025-10-08 18:34:17.810303063 +0000 UTC m=+1028.057142442" Oct 08 18:34:17 crc kubenswrapper[4859]: I1008 18:34:17.844345 4859 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/placement-58c5f664f4-5lvrs" podStartSLOduration=3.844325154 podStartE2EDuration="3.844325154s" podCreationTimestamp="2025-10-08 18:34:14 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-08 18:34:17.840273804 +0000 UTC m=+1028.087113213" watchObservedRunningTime="2025-10-08 18:34:17.844325154 +0000 UTC m=+1028.091164533" Oct 08 18:34:17 crc kubenswrapper[4859]: I1008 18:34:17.925269 4859 patch_prober.go:28] interesting pod/machine-config-daemon-82s52 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 08 18:34:17 crc kubenswrapper[4859]: I1008 18:34:17.925328 4859 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-82s52" podUID="b23a6a6c-9d92-4e7b-840e-55cfda873a2d" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 08 18:34:18 crc kubenswrapper[4859]: I1008 18:34:18.480058 4859 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="f08c21e6-a74c-4ad9-a3fa-4e56ff2c6432" path="/var/lib/kubelet/pods/f08c21e6-a74c-4ad9-a3fa-4e56ff2c6432/volumes" Oct 08 18:34:18 crc kubenswrapper[4859]: I1008 18:34:18.630538 4859 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/glance-default-external-api-0" Oct 08 18:34:18 crc kubenswrapper[4859]: I1008 18:34:18.630607 4859 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/glance-default-external-api-0" Oct 08 18:34:18 crc kubenswrapper[4859]: I1008 18:34:18.630618 4859 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/glance-default-external-api-0" Oct 08 18:34:18 crc kubenswrapper[4859]: I1008 18:34:18.630629 4859 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/glance-default-external-api-0" Oct 08 18:34:18 crc kubenswrapper[4859]: I1008 18:34:18.647727 4859 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/glance-default-internal-api-0" Oct 08 18:34:18 crc kubenswrapper[4859]: I1008 18:34:18.647781 4859 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/glance-default-internal-api-0" Oct 08 18:34:18 crc kubenswrapper[4859]: I1008 18:34:18.647794 4859 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/glance-default-internal-api-0" Oct 08 18:34:18 crc kubenswrapper[4859]: I1008 18:34:18.647803 4859 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/glance-default-internal-api-0" Oct 08 18:34:18 crc kubenswrapper[4859]: I1008 18:34:18.672512 4859 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/glance-default-external-api-0" Oct 08 18:34:18 crc kubenswrapper[4859]: I1008 18:34:18.680795 4859 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/glance-default-external-api-0" Oct 08 18:34:18 crc kubenswrapper[4859]: I1008 18:34:18.690285 4859 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/glance-default-internal-api-0" Oct 08 18:34:18 crc kubenswrapper[4859]: I1008 18:34:18.693632 4859 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/glance-default-internal-api-0" Oct 08 18:34:20 crc kubenswrapper[4859]: I1008 18:34:20.962033 4859 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/glance-default-external-api-0" Oct 08 18:34:20 crc kubenswrapper[4859]: I1008 18:34:20.962359 4859 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" Oct 08 18:34:20 crc kubenswrapper[4859]: I1008 18:34:20.963808 4859 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/glance-default-external-api-0" Oct 08 18:34:21 crc kubenswrapper[4859]: I1008 18:34:21.058641 4859 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/glance-default-internal-api-0" Oct 08 18:34:21 crc kubenswrapper[4859]: I1008 18:34:21.058991 4859 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" Oct 08 18:34:21 crc kubenswrapper[4859]: I1008 18:34:21.107081 4859 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/glance-default-internal-api-0" Oct 08 18:34:21 crc kubenswrapper[4859]: I1008 18:34:21.849861 4859 generic.go:334] "Generic (PLEG): container finished" podID="3b2953d1-3ebe-4d7d-b60c-6ad07d02af5d" containerID="8647fd3df54b81ebc93e3e14fe2b59cbd0e77836f881418a614ac5f2c3c4c10f" exitCode=0 Oct 08 18:34:21 crc kubenswrapper[4859]: I1008 18:34:21.850679 4859 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-db-sync-x99jc" event={"ID":"3b2953d1-3ebe-4d7d-b60c-6ad07d02af5d","Type":"ContainerDied","Data":"8647fd3df54b81ebc93e3e14fe2b59cbd0e77836f881418a614ac5f2c3c4c10f"} Oct 08 18:34:22 crc kubenswrapper[4859]: I1008 18:34:22.311284 4859 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/horizon-7d58687c8-fhctx" podUID="1a92487e-3309-41c4-8c82-cd57f517ab4b" containerName="horizon" probeResult="failure" output="Get \"https://10.217.0.147:8443/dashboard/auth/login/?next=/dashboard/\": dial tcp 10.217.0.147:8443: connect: connection refused" Oct 08 18:34:22 crc kubenswrapper[4859]: I1008 18:34:22.446626 4859 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/horizon-6f569d5894-mx8v5" podUID="50b5de31-90f6-43cf-9e16-7b89b09f1e57" containerName="horizon" probeResult="failure" output="Get \"https://10.217.0.148:8443/dashboard/auth/login/?next=/dashboard/\": dial tcp 10.217.0.148:8443: connect: connection refused" Oct 08 18:34:25 crc kubenswrapper[4859]: I1008 18:34:25.310447 4859 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-db-sync-x99jc" Oct 08 18:34:25 crc kubenswrapper[4859]: I1008 18:34:25.510444 4859 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3b2953d1-3ebe-4d7d-b60c-6ad07d02af5d-combined-ca-bundle\") pod \"3b2953d1-3ebe-4d7d-b60c-6ad07d02af5d\" (UID: \"3b2953d1-3ebe-4d7d-b60c-6ad07d02af5d\") " Oct 08 18:34:25 crc kubenswrapper[4859]: I1008 18:34:25.510885 4859 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-llvcl\" (UniqueName: \"kubernetes.io/projected/3b2953d1-3ebe-4d7d-b60c-6ad07d02af5d-kube-api-access-llvcl\") pod \"3b2953d1-3ebe-4d7d-b60c-6ad07d02af5d\" (UID: \"3b2953d1-3ebe-4d7d-b60c-6ad07d02af5d\") " Oct 08 18:34:25 crc kubenswrapper[4859]: I1008 18:34:25.511043 4859 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/3b2953d1-3ebe-4d7d-b60c-6ad07d02af5d-db-sync-config-data\") pod \"3b2953d1-3ebe-4d7d-b60c-6ad07d02af5d\" (UID: \"3b2953d1-3ebe-4d7d-b60c-6ad07d02af5d\") " Oct 08 18:34:25 crc kubenswrapper[4859]: I1008 18:34:25.516606 4859 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/3b2953d1-3ebe-4d7d-b60c-6ad07d02af5d-db-sync-config-data" (OuterVolumeSpecName: "db-sync-config-data") pod "3b2953d1-3ebe-4d7d-b60c-6ad07d02af5d" (UID: "3b2953d1-3ebe-4d7d-b60c-6ad07d02af5d"). InnerVolumeSpecName "db-sync-config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 08 18:34:25 crc kubenswrapper[4859]: I1008 18:34:25.521937 4859 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/3b2953d1-3ebe-4d7d-b60c-6ad07d02af5d-kube-api-access-llvcl" (OuterVolumeSpecName: "kube-api-access-llvcl") pod "3b2953d1-3ebe-4d7d-b60c-6ad07d02af5d" (UID: "3b2953d1-3ebe-4d7d-b60c-6ad07d02af5d"). InnerVolumeSpecName "kube-api-access-llvcl". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 08 18:34:25 crc kubenswrapper[4859]: I1008 18:34:25.557345 4859 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/3b2953d1-3ebe-4d7d-b60c-6ad07d02af5d-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "3b2953d1-3ebe-4d7d-b60c-6ad07d02af5d" (UID: "3b2953d1-3ebe-4d7d-b60c-6ad07d02af5d"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 08 18:34:25 crc kubenswrapper[4859]: I1008 18:34:25.613618 4859 reconciler_common.go:293] "Volume detached for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/3b2953d1-3ebe-4d7d-b60c-6ad07d02af5d-db-sync-config-data\") on node \"crc\" DevicePath \"\"" Oct 08 18:34:25 crc kubenswrapper[4859]: I1008 18:34:25.613670 4859 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3b2953d1-3ebe-4d7d-b60c-6ad07d02af5d-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 08 18:34:25 crc kubenswrapper[4859]: I1008 18:34:25.613698 4859 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-llvcl\" (UniqueName: \"kubernetes.io/projected/3b2953d1-3ebe-4d7d-b60c-6ad07d02af5d-kube-api-access-llvcl\") on node \"crc\" DevicePath \"\"" Oct 08 18:34:25 crc kubenswrapper[4859]: I1008 18:34:25.893520 4859 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"87f0eecb-178a-438a-8643-e29c19390884","Type":"ContainerStarted","Data":"a912aba480685bc65f1310b409cf7397c5278c100778aa597be6b341c5f06d10"} Oct 08 18:34:25 crc kubenswrapper[4859]: I1008 18:34:25.893619 4859 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="87f0eecb-178a-438a-8643-e29c19390884" containerName="ceilometer-central-agent" containerID="cri-o://2a699d4ef1d913cf94815cb3b9d6ef92b6f7bcdf4ce58dbd6b5b6b78663afd86" gracePeriod=30 Oct 08 18:34:25 crc kubenswrapper[4859]: I1008 18:34:25.893711 4859 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="87f0eecb-178a-438a-8643-e29c19390884" containerName="sg-core" containerID="cri-o://d4f40d2b1c23ce93761ada628d8d73a724dc6702e886ec2cd59ef35c8b2b69d8" gracePeriod=30 Oct 08 18:34:25 crc kubenswrapper[4859]: I1008 18:34:25.893793 4859 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="87f0eecb-178a-438a-8643-e29c19390884" containerName="proxy-httpd" containerID="cri-o://a912aba480685bc65f1310b409cf7397c5278c100778aa597be6b341c5f06d10" gracePeriod=30 Oct 08 18:34:25 crc kubenswrapper[4859]: I1008 18:34:25.893799 4859 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ceilometer-0" Oct 08 18:34:25 crc kubenswrapper[4859]: I1008 18:34:25.893798 4859 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="87f0eecb-178a-438a-8643-e29c19390884" containerName="ceilometer-notification-agent" containerID="cri-o://c76586e5e90236f90858e341d25b5dde536c874c2bca5a632ffc704ce10dd611" gracePeriod=30 Oct 08 18:34:25 crc kubenswrapper[4859]: I1008 18:34:25.905725 4859 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-db-sync-x99jc" Oct 08 18:34:25 crc kubenswrapper[4859]: I1008 18:34:25.905734 4859 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-db-sync-x99jc" event={"ID":"3b2953d1-3ebe-4d7d-b60c-6ad07d02af5d","Type":"ContainerDied","Data":"5326bbd772abed3f9cc2690e3fb5c558ee843f4383d3dd7134f5126d43cf5d02"} Oct 08 18:34:25 crc kubenswrapper[4859]: I1008 18:34:25.905810 4859 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="5326bbd772abed3f9cc2690e3fb5c558ee843f4383d3dd7134f5126d43cf5d02" Oct 08 18:34:25 crc kubenswrapper[4859]: I1008 18:34:25.919449 4859 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ceilometer-0" podStartSLOduration=2.306646667 podStartE2EDuration="52.919433981s" podCreationTimestamp="2025-10-08 18:33:33 +0000 UTC" firstStartedPulling="2025-10-08 18:33:34.895891032 +0000 UTC m=+985.142730411" lastFinishedPulling="2025-10-08 18:34:25.508678356 +0000 UTC m=+1035.755517725" observedRunningTime="2025-10-08 18:34:25.916652669 +0000 UTC m=+1036.163492048" watchObservedRunningTime="2025-10-08 18:34:25.919433981 +0000 UTC m=+1036.166273360" Oct 08 18:34:26 crc kubenswrapper[4859]: E1008 18:34:26.328422 4859 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod87f0eecb_178a_438a_8643_e29c19390884.slice/crio-2a699d4ef1d913cf94815cb3b9d6ef92b6f7bcdf4ce58dbd6b5b6b78663afd86.scope\": RecentStats: unable to find data in memory cache]" Oct 08 18:34:26 crc kubenswrapper[4859]: I1008 18:34:26.651424 4859 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/barbican-keystone-listener-d798f6bcd-5zw8p"] Oct 08 18:34:26 crc kubenswrapper[4859]: E1008 18:34:26.651917 4859 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f08c21e6-a74c-4ad9-a3fa-4e56ff2c6432" containerName="dnsmasq-dns" Oct 08 18:34:26 crc kubenswrapper[4859]: I1008 18:34:26.651938 4859 state_mem.go:107] "Deleted CPUSet assignment" podUID="f08c21e6-a74c-4ad9-a3fa-4e56ff2c6432" containerName="dnsmasq-dns" Oct 08 18:34:26 crc kubenswrapper[4859]: E1008 18:34:26.651953 4859 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f08c21e6-a74c-4ad9-a3fa-4e56ff2c6432" containerName="init" Oct 08 18:34:26 crc kubenswrapper[4859]: I1008 18:34:26.651961 4859 state_mem.go:107] "Deleted CPUSet assignment" podUID="f08c21e6-a74c-4ad9-a3fa-4e56ff2c6432" containerName="init" Oct 08 18:34:26 crc kubenswrapper[4859]: E1008 18:34:26.651989 4859 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3b2953d1-3ebe-4d7d-b60c-6ad07d02af5d" containerName="barbican-db-sync" Oct 08 18:34:26 crc kubenswrapper[4859]: I1008 18:34:26.651998 4859 state_mem.go:107] "Deleted CPUSet assignment" podUID="3b2953d1-3ebe-4d7d-b60c-6ad07d02af5d" containerName="barbican-db-sync" Oct 08 18:34:26 crc kubenswrapper[4859]: I1008 18:34:26.652198 4859 memory_manager.go:354] "RemoveStaleState removing state" podUID="f08c21e6-a74c-4ad9-a3fa-4e56ff2c6432" containerName="dnsmasq-dns" Oct 08 18:34:26 crc kubenswrapper[4859]: I1008 18:34:26.652230 4859 memory_manager.go:354] "RemoveStaleState removing state" podUID="3b2953d1-3ebe-4d7d-b60c-6ad07d02af5d" containerName="barbican-db-sync" Oct 08 18:34:26 crc kubenswrapper[4859]: I1008 18:34:26.653332 4859 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-keystone-listener-d798f6bcd-5zw8p" Oct 08 18:34:26 crc kubenswrapper[4859]: I1008 18:34:26.656899 4859 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"barbican-keystone-listener-config-data" Oct 08 18:34:26 crc kubenswrapper[4859]: I1008 18:34:26.662344 4859 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"barbican-barbican-dockercfg-tkhlt" Oct 08 18:34:26 crc kubenswrapper[4859]: I1008 18:34:26.669994 4859 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-keystone-listener-d798f6bcd-5zw8p"] Oct 08 18:34:26 crc kubenswrapper[4859]: I1008 18:34:26.674409 4859 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"barbican-config-data" Oct 08 18:34:26 crc kubenswrapper[4859]: I1008 18:34:26.684763 4859 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/barbican-worker-7849cf5875-6srbt"] Oct 08 18:34:26 crc kubenswrapper[4859]: I1008 18:34:26.690502 4859 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-worker-7849cf5875-6srbt" Oct 08 18:34:26 crc kubenswrapper[4859]: I1008 18:34:26.696060 4859 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"barbican-worker-config-data" Oct 08 18:34:26 crc kubenswrapper[4859]: I1008 18:34:26.704790 4859 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-worker-7849cf5875-6srbt"] Oct 08 18:34:26 crc kubenswrapper[4859]: I1008 18:34:26.784226 4859 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-744fd954cc-hv6xb"] Oct 08 18:34:26 crc kubenswrapper[4859]: I1008 18:34:26.787941 4859 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-744fd954cc-hv6xb" Oct 08 18:34:26 crc kubenswrapper[4859]: I1008 18:34:26.806778 4859 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-744fd954cc-hv6xb"] Oct 08 18:34:26 crc kubenswrapper[4859]: I1008 18:34:26.833508 4859 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/80660054-bd62-4742-a13d-5193b317b036-config-data\") pod \"barbican-keystone-listener-d798f6bcd-5zw8p\" (UID: \"80660054-bd62-4742-a13d-5193b317b036\") " pod="openstack/barbican-keystone-listener-d798f6bcd-5zw8p" Oct 08 18:34:26 crc kubenswrapper[4859]: I1008 18:34:26.833566 4859 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-7hffh\" (UniqueName: \"kubernetes.io/projected/578aad66-818f-412f-9023-97ed6ae95de3-kube-api-access-7hffh\") pod \"barbican-worker-7849cf5875-6srbt\" (UID: \"578aad66-818f-412f-9023-97ed6ae95de3\") " pod="openstack/barbican-worker-7849cf5875-6srbt" Oct 08 18:34:26 crc kubenswrapper[4859]: I1008 18:34:26.833605 4859 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/578aad66-818f-412f-9023-97ed6ae95de3-config-data-custom\") pod \"barbican-worker-7849cf5875-6srbt\" (UID: \"578aad66-818f-412f-9023-97ed6ae95de3\") " pod="openstack/barbican-worker-7849cf5875-6srbt" Oct 08 18:34:26 crc kubenswrapper[4859]: I1008 18:34:26.833633 4859 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/578aad66-818f-412f-9023-97ed6ae95de3-logs\") pod \"barbican-worker-7849cf5875-6srbt\" (UID: \"578aad66-818f-412f-9023-97ed6ae95de3\") " pod="openstack/barbican-worker-7849cf5875-6srbt" Oct 08 18:34:26 crc kubenswrapper[4859]: I1008 18:34:26.833654 4859 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/578aad66-818f-412f-9023-97ed6ae95de3-combined-ca-bundle\") pod \"barbican-worker-7849cf5875-6srbt\" (UID: \"578aad66-818f-412f-9023-97ed6ae95de3\") " pod="openstack/barbican-worker-7849cf5875-6srbt" Oct 08 18:34:26 crc kubenswrapper[4859]: I1008 18:34:26.833704 4859 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/578aad66-818f-412f-9023-97ed6ae95de3-config-data\") pod \"barbican-worker-7849cf5875-6srbt\" (UID: \"578aad66-818f-412f-9023-97ed6ae95de3\") " pod="openstack/barbican-worker-7849cf5875-6srbt" Oct 08 18:34:26 crc kubenswrapper[4859]: I1008 18:34:26.833734 4859 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/80660054-bd62-4742-a13d-5193b317b036-config-data-custom\") pod \"barbican-keystone-listener-d798f6bcd-5zw8p\" (UID: \"80660054-bd62-4742-a13d-5193b317b036\") " pod="openstack/barbican-keystone-listener-d798f6bcd-5zw8p" Oct 08 18:34:26 crc kubenswrapper[4859]: I1008 18:34:26.833752 4859 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/80660054-bd62-4742-a13d-5193b317b036-combined-ca-bundle\") pod \"barbican-keystone-listener-d798f6bcd-5zw8p\" (UID: \"80660054-bd62-4742-a13d-5193b317b036\") " pod="openstack/barbican-keystone-listener-d798f6bcd-5zw8p" Oct 08 18:34:26 crc kubenswrapper[4859]: I1008 18:34:26.833779 4859 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-nrwvz\" (UniqueName: \"kubernetes.io/projected/80660054-bd62-4742-a13d-5193b317b036-kube-api-access-nrwvz\") pod \"barbican-keystone-listener-d798f6bcd-5zw8p\" (UID: \"80660054-bd62-4742-a13d-5193b317b036\") " pod="openstack/barbican-keystone-listener-d798f6bcd-5zw8p" Oct 08 18:34:26 crc kubenswrapper[4859]: I1008 18:34:26.833847 4859 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/80660054-bd62-4742-a13d-5193b317b036-logs\") pod \"barbican-keystone-listener-d798f6bcd-5zw8p\" (UID: \"80660054-bd62-4742-a13d-5193b317b036\") " pod="openstack/barbican-keystone-listener-d798f6bcd-5zw8p" Oct 08 18:34:26 crc kubenswrapper[4859]: I1008 18:34:26.905072 4859 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/barbican-api-96f8997cd-z7vl4"] Oct 08 18:34:26 crc kubenswrapper[4859]: I1008 18:34:26.906966 4859 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-api-96f8997cd-z7vl4" Oct 08 18:34:26 crc kubenswrapper[4859]: I1008 18:34:26.910867 4859 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"barbican-api-config-data" Oct 08 18:34:26 crc kubenswrapper[4859]: I1008 18:34:26.935880 4859 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/80660054-bd62-4742-a13d-5193b317b036-logs\") pod \"barbican-keystone-listener-d798f6bcd-5zw8p\" (UID: \"80660054-bd62-4742-a13d-5193b317b036\") " pod="openstack/barbican-keystone-listener-d798f6bcd-5zw8p" Oct 08 18:34:26 crc kubenswrapper[4859]: I1008 18:34:26.935969 4859 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/80660054-bd62-4742-a13d-5193b317b036-config-data\") pod \"barbican-keystone-listener-d798f6bcd-5zw8p\" (UID: \"80660054-bd62-4742-a13d-5193b317b036\") " pod="openstack/barbican-keystone-listener-d798f6bcd-5zw8p" Oct 08 18:34:26 crc kubenswrapper[4859]: I1008 18:34:26.936022 4859 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/8252c83d-1fdd-4d33-87a0-7a797251a0b4-dns-swift-storage-0\") pod \"dnsmasq-dns-744fd954cc-hv6xb\" (UID: \"8252c83d-1fdd-4d33-87a0-7a797251a0b4\") " pod="openstack/dnsmasq-dns-744fd954cc-hv6xb" Oct 08 18:34:26 crc kubenswrapper[4859]: I1008 18:34:26.936031 4859 generic.go:334] "Generic (PLEG): container finished" podID="87f0eecb-178a-438a-8643-e29c19390884" containerID="a912aba480685bc65f1310b409cf7397c5278c100778aa597be6b341c5f06d10" exitCode=0 Oct 08 18:34:26 crc kubenswrapper[4859]: I1008 18:34:26.936050 4859 generic.go:334] "Generic (PLEG): container finished" podID="87f0eecb-178a-438a-8643-e29c19390884" containerID="d4f40d2b1c23ce93761ada628d8d73a724dc6702e886ec2cd59ef35c8b2b69d8" exitCode=2 Oct 08 18:34:26 crc kubenswrapper[4859]: I1008 18:34:26.936051 4859 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/8252c83d-1fdd-4d33-87a0-7a797251a0b4-ovsdbserver-sb\") pod \"dnsmasq-dns-744fd954cc-hv6xb\" (UID: \"8252c83d-1fdd-4d33-87a0-7a797251a0b4\") " pod="openstack/dnsmasq-dns-744fd954cc-hv6xb" Oct 08 18:34:26 crc kubenswrapper[4859]: I1008 18:34:26.936083 4859 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"87f0eecb-178a-438a-8643-e29c19390884","Type":"ContainerDied","Data":"a912aba480685bc65f1310b409cf7397c5278c100778aa597be6b341c5f06d10"} Oct 08 18:34:26 crc kubenswrapper[4859]: I1008 18:34:26.936089 4859 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-7hffh\" (UniqueName: \"kubernetes.io/projected/578aad66-818f-412f-9023-97ed6ae95de3-kube-api-access-7hffh\") pod \"barbican-worker-7849cf5875-6srbt\" (UID: \"578aad66-818f-412f-9023-97ed6ae95de3\") " pod="openstack/barbican-worker-7849cf5875-6srbt" Oct 08 18:34:26 crc kubenswrapper[4859]: I1008 18:34:26.936107 4859 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"87f0eecb-178a-438a-8643-e29c19390884","Type":"ContainerDied","Data":"d4f40d2b1c23ce93761ada628d8d73a724dc6702e886ec2cd59ef35c8b2b69d8"} Oct 08 18:34:26 crc kubenswrapper[4859]: I1008 18:34:26.936116 4859 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"87f0eecb-178a-438a-8643-e29c19390884","Type":"ContainerDied","Data":"2a699d4ef1d913cf94815cb3b9d6ef92b6f7bcdf4ce58dbd6b5b6b78663afd86"} Oct 08 18:34:26 crc kubenswrapper[4859]: I1008 18:34:26.936128 4859 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/8252c83d-1fdd-4d33-87a0-7a797251a0b4-dns-svc\") pod \"dnsmasq-dns-744fd954cc-hv6xb\" (UID: \"8252c83d-1fdd-4d33-87a0-7a797251a0b4\") " pod="openstack/dnsmasq-dns-744fd954cc-hv6xb" Oct 08 18:34:26 crc kubenswrapper[4859]: I1008 18:34:26.936161 4859 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/578aad66-818f-412f-9023-97ed6ae95de3-config-data-custom\") pod \"barbican-worker-7849cf5875-6srbt\" (UID: \"578aad66-818f-412f-9023-97ed6ae95de3\") " pod="openstack/barbican-worker-7849cf5875-6srbt" Oct 08 18:34:26 crc kubenswrapper[4859]: I1008 18:34:26.936185 4859 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/8252c83d-1fdd-4d33-87a0-7a797251a0b4-config\") pod \"dnsmasq-dns-744fd954cc-hv6xb\" (UID: \"8252c83d-1fdd-4d33-87a0-7a797251a0b4\") " pod="openstack/dnsmasq-dns-744fd954cc-hv6xb" Oct 08 18:34:26 crc kubenswrapper[4859]: I1008 18:34:26.936228 4859 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/578aad66-818f-412f-9023-97ed6ae95de3-logs\") pod \"barbican-worker-7849cf5875-6srbt\" (UID: \"578aad66-818f-412f-9023-97ed6ae95de3\") " pod="openstack/barbican-worker-7849cf5875-6srbt" Oct 08 18:34:26 crc kubenswrapper[4859]: I1008 18:34:26.936259 4859 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/578aad66-818f-412f-9023-97ed6ae95de3-combined-ca-bundle\") pod \"barbican-worker-7849cf5875-6srbt\" (UID: \"578aad66-818f-412f-9023-97ed6ae95de3\") " pod="openstack/barbican-worker-7849cf5875-6srbt" Oct 08 18:34:26 crc kubenswrapper[4859]: I1008 18:34:26.936289 4859 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/578aad66-818f-412f-9023-97ed6ae95de3-config-data\") pod \"barbican-worker-7849cf5875-6srbt\" (UID: \"578aad66-818f-412f-9023-97ed6ae95de3\") " pod="openstack/barbican-worker-7849cf5875-6srbt" Oct 08 18:34:26 crc kubenswrapper[4859]: I1008 18:34:26.936322 4859 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/80660054-bd62-4742-a13d-5193b317b036-config-data-custom\") pod \"barbican-keystone-listener-d798f6bcd-5zw8p\" (UID: \"80660054-bd62-4742-a13d-5193b317b036\") " pod="openstack/barbican-keystone-listener-d798f6bcd-5zw8p" Oct 08 18:34:26 crc kubenswrapper[4859]: I1008 18:34:26.936349 4859 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/8252c83d-1fdd-4d33-87a0-7a797251a0b4-ovsdbserver-nb\") pod \"dnsmasq-dns-744fd954cc-hv6xb\" (UID: \"8252c83d-1fdd-4d33-87a0-7a797251a0b4\") " pod="openstack/dnsmasq-dns-744fd954cc-hv6xb" Oct 08 18:34:26 crc kubenswrapper[4859]: I1008 18:34:26.936377 4859 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/80660054-bd62-4742-a13d-5193b317b036-combined-ca-bundle\") pod \"barbican-keystone-listener-d798f6bcd-5zw8p\" (UID: \"80660054-bd62-4742-a13d-5193b317b036\") " pod="openstack/barbican-keystone-listener-d798f6bcd-5zw8p" Oct 08 18:34:26 crc kubenswrapper[4859]: I1008 18:34:26.936424 4859 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-nrwvz\" (UniqueName: \"kubernetes.io/projected/80660054-bd62-4742-a13d-5193b317b036-kube-api-access-nrwvz\") pod \"barbican-keystone-listener-d798f6bcd-5zw8p\" (UID: \"80660054-bd62-4742-a13d-5193b317b036\") " pod="openstack/barbican-keystone-listener-d798f6bcd-5zw8p" Oct 08 18:34:26 crc kubenswrapper[4859]: I1008 18:34:26.936467 4859 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-jkx9v\" (UniqueName: \"kubernetes.io/projected/8252c83d-1fdd-4d33-87a0-7a797251a0b4-kube-api-access-jkx9v\") pod \"dnsmasq-dns-744fd954cc-hv6xb\" (UID: \"8252c83d-1fdd-4d33-87a0-7a797251a0b4\") " pod="openstack/dnsmasq-dns-744fd954cc-hv6xb" Oct 08 18:34:26 crc kubenswrapper[4859]: I1008 18:34:26.936524 4859 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/80660054-bd62-4742-a13d-5193b317b036-logs\") pod \"barbican-keystone-listener-d798f6bcd-5zw8p\" (UID: \"80660054-bd62-4742-a13d-5193b317b036\") " pod="openstack/barbican-keystone-listener-d798f6bcd-5zw8p" Oct 08 18:34:26 crc kubenswrapper[4859]: I1008 18:34:26.936059 4859 generic.go:334] "Generic (PLEG): container finished" podID="87f0eecb-178a-438a-8643-e29c19390884" containerID="2a699d4ef1d913cf94815cb3b9d6ef92b6f7bcdf4ce58dbd6b5b6b78663afd86" exitCode=0 Oct 08 18:34:26 crc kubenswrapper[4859]: I1008 18:34:26.938856 4859 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/578aad66-818f-412f-9023-97ed6ae95de3-logs\") pod \"barbican-worker-7849cf5875-6srbt\" (UID: \"578aad66-818f-412f-9023-97ed6ae95de3\") " pod="openstack/barbican-worker-7849cf5875-6srbt" Oct 08 18:34:26 crc kubenswrapper[4859]: I1008 18:34:26.939282 4859 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-api-96f8997cd-z7vl4"] Oct 08 18:34:26 crc kubenswrapper[4859]: I1008 18:34:26.942449 4859 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-db-sync-jqwrh" event={"ID":"720db6c0-f18f-429b-9589-fa503d212139","Type":"ContainerStarted","Data":"f705521531b392a51ca4b6b14b60c8d34d3ec8f7dbebf2794d813c0cd105ff2c"} Oct 08 18:34:26 crc kubenswrapper[4859]: I1008 18:34:26.959116 4859 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/578aad66-818f-412f-9023-97ed6ae95de3-combined-ca-bundle\") pod \"barbican-worker-7849cf5875-6srbt\" (UID: \"578aad66-818f-412f-9023-97ed6ae95de3\") " pod="openstack/barbican-worker-7849cf5875-6srbt" Oct 08 18:34:26 crc kubenswrapper[4859]: I1008 18:34:26.969944 4859 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/578aad66-818f-412f-9023-97ed6ae95de3-config-data-custom\") pod \"barbican-worker-7849cf5875-6srbt\" (UID: \"578aad66-818f-412f-9023-97ed6ae95de3\") " pod="openstack/barbican-worker-7849cf5875-6srbt" Oct 08 18:34:26 crc kubenswrapper[4859]: I1008 18:34:26.970375 4859 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/578aad66-818f-412f-9023-97ed6ae95de3-config-data\") pod \"barbican-worker-7849cf5875-6srbt\" (UID: \"578aad66-818f-412f-9023-97ed6ae95de3\") " pod="openstack/barbican-worker-7849cf5875-6srbt" Oct 08 18:34:26 crc kubenswrapper[4859]: I1008 18:34:26.970916 4859 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/80660054-bd62-4742-a13d-5193b317b036-combined-ca-bundle\") pod \"barbican-keystone-listener-d798f6bcd-5zw8p\" (UID: \"80660054-bd62-4742-a13d-5193b317b036\") " pod="openstack/barbican-keystone-listener-d798f6bcd-5zw8p" Oct 08 18:34:26 crc kubenswrapper[4859]: I1008 18:34:26.972657 4859 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/80660054-bd62-4742-a13d-5193b317b036-config-data\") pod \"barbican-keystone-listener-d798f6bcd-5zw8p\" (UID: \"80660054-bd62-4742-a13d-5193b317b036\") " pod="openstack/barbican-keystone-listener-d798f6bcd-5zw8p" Oct 08 18:34:26 crc kubenswrapper[4859]: I1008 18:34:26.976923 4859 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-nrwvz\" (UniqueName: \"kubernetes.io/projected/80660054-bd62-4742-a13d-5193b317b036-kube-api-access-nrwvz\") pod \"barbican-keystone-listener-d798f6bcd-5zw8p\" (UID: \"80660054-bd62-4742-a13d-5193b317b036\") " pod="openstack/barbican-keystone-listener-d798f6bcd-5zw8p" Oct 08 18:34:26 crc kubenswrapper[4859]: I1008 18:34:26.979506 4859 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/80660054-bd62-4742-a13d-5193b317b036-config-data-custom\") pod \"barbican-keystone-listener-d798f6bcd-5zw8p\" (UID: \"80660054-bd62-4742-a13d-5193b317b036\") " pod="openstack/barbican-keystone-listener-d798f6bcd-5zw8p" Oct 08 18:34:26 crc kubenswrapper[4859]: I1008 18:34:26.980203 4859 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-7hffh\" (UniqueName: \"kubernetes.io/projected/578aad66-818f-412f-9023-97ed6ae95de3-kube-api-access-7hffh\") pod \"barbican-worker-7849cf5875-6srbt\" (UID: \"578aad66-818f-412f-9023-97ed6ae95de3\") " pod="openstack/barbican-worker-7849cf5875-6srbt" Oct 08 18:34:26 crc kubenswrapper[4859]: I1008 18:34:26.987079 4859 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-keystone-listener-d798f6bcd-5zw8p" Oct 08 18:34:27 crc kubenswrapper[4859]: I1008 18:34:27.013255 4859 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-worker-7849cf5875-6srbt" Oct 08 18:34:27 crc kubenswrapper[4859]: I1008 18:34:27.029190 4859 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/cinder-db-sync-jqwrh" podStartSLOduration=3.419222841 podStartE2EDuration="52.029166611s" podCreationTimestamp="2025-10-08 18:33:35 +0000 UTC" firstStartedPulling="2025-10-08 18:33:36.86450202 +0000 UTC m=+987.111341399" lastFinishedPulling="2025-10-08 18:34:25.4744458 +0000 UTC m=+1035.721285169" observedRunningTime="2025-10-08 18:34:26.971848986 +0000 UTC m=+1037.218688385" watchObservedRunningTime="2025-10-08 18:34:27.029166611 +0000 UTC m=+1037.276005990" Oct 08 18:34:27 crc kubenswrapper[4859]: I1008 18:34:27.038301 4859 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/8252c83d-1fdd-4d33-87a0-7a797251a0b4-ovsdbserver-nb\") pod \"dnsmasq-dns-744fd954cc-hv6xb\" (UID: \"8252c83d-1fdd-4d33-87a0-7a797251a0b4\") " pod="openstack/dnsmasq-dns-744fd954cc-hv6xb" Oct 08 18:34:27 crc kubenswrapper[4859]: I1008 18:34:27.038730 4859 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-jkx9v\" (UniqueName: \"kubernetes.io/projected/8252c83d-1fdd-4d33-87a0-7a797251a0b4-kube-api-access-jkx9v\") pod \"dnsmasq-dns-744fd954cc-hv6xb\" (UID: \"8252c83d-1fdd-4d33-87a0-7a797251a0b4\") " pod="openstack/dnsmasq-dns-744fd954cc-hv6xb" Oct 08 18:34:27 crc kubenswrapper[4859]: I1008 18:34:27.038850 4859 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/b6cd546e-9054-4150-a5f3-7bd709d15560-logs\") pod \"barbican-api-96f8997cd-z7vl4\" (UID: \"b6cd546e-9054-4150-a5f3-7bd709d15560\") " pod="openstack/barbican-api-96f8997cd-z7vl4" Oct 08 18:34:27 crc kubenswrapper[4859]: I1008 18:34:27.038948 4859 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b6cd546e-9054-4150-a5f3-7bd709d15560-combined-ca-bundle\") pod \"barbican-api-96f8997cd-z7vl4\" (UID: \"b6cd546e-9054-4150-a5f3-7bd709d15560\") " pod="openstack/barbican-api-96f8997cd-z7vl4" Oct 08 18:34:27 crc kubenswrapper[4859]: I1008 18:34:27.039063 4859 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/b6cd546e-9054-4150-a5f3-7bd709d15560-config-data-custom\") pod \"barbican-api-96f8997cd-z7vl4\" (UID: \"b6cd546e-9054-4150-a5f3-7bd709d15560\") " pod="openstack/barbican-api-96f8997cd-z7vl4" Oct 08 18:34:27 crc kubenswrapper[4859]: I1008 18:34:27.039165 4859 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/8252c83d-1fdd-4d33-87a0-7a797251a0b4-dns-swift-storage-0\") pod \"dnsmasq-dns-744fd954cc-hv6xb\" (UID: \"8252c83d-1fdd-4d33-87a0-7a797251a0b4\") " pod="openstack/dnsmasq-dns-744fd954cc-hv6xb" Oct 08 18:34:27 crc kubenswrapper[4859]: I1008 18:34:27.039253 4859 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/8252c83d-1fdd-4d33-87a0-7a797251a0b4-ovsdbserver-sb\") pod \"dnsmasq-dns-744fd954cc-hv6xb\" (UID: \"8252c83d-1fdd-4d33-87a0-7a797251a0b4\") " pod="openstack/dnsmasq-dns-744fd954cc-hv6xb" Oct 08 18:34:27 crc kubenswrapper[4859]: I1008 18:34:27.039374 4859 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/8252c83d-1fdd-4d33-87a0-7a797251a0b4-dns-svc\") pod \"dnsmasq-dns-744fd954cc-hv6xb\" (UID: \"8252c83d-1fdd-4d33-87a0-7a797251a0b4\") " pod="openstack/dnsmasq-dns-744fd954cc-hv6xb" Oct 08 18:34:27 crc kubenswrapper[4859]: I1008 18:34:27.039464 4859 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-lwkz8\" (UniqueName: \"kubernetes.io/projected/b6cd546e-9054-4150-a5f3-7bd709d15560-kube-api-access-lwkz8\") pod \"barbican-api-96f8997cd-z7vl4\" (UID: \"b6cd546e-9054-4150-a5f3-7bd709d15560\") " pod="openstack/barbican-api-96f8997cd-z7vl4" Oct 08 18:34:27 crc kubenswrapper[4859]: I1008 18:34:27.039576 4859 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/8252c83d-1fdd-4d33-87a0-7a797251a0b4-config\") pod \"dnsmasq-dns-744fd954cc-hv6xb\" (UID: \"8252c83d-1fdd-4d33-87a0-7a797251a0b4\") " pod="openstack/dnsmasq-dns-744fd954cc-hv6xb" Oct 08 18:34:27 crc kubenswrapper[4859]: I1008 18:34:27.039429 4859 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/8252c83d-1fdd-4d33-87a0-7a797251a0b4-ovsdbserver-nb\") pod \"dnsmasq-dns-744fd954cc-hv6xb\" (UID: \"8252c83d-1fdd-4d33-87a0-7a797251a0b4\") " pod="openstack/dnsmasq-dns-744fd954cc-hv6xb" Oct 08 18:34:27 crc kubenswrapper[4859]: I1008 18:34:27.040164 4859 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/b6cd546e-9054-4150-a5f3-7bd709d15560-config-data\") pod \"barbican-api-96f8997cd-z7vl4\" (UID: \"b6cd546e-9054-4150-a5f3-7bd709d15560\") " pod="openstack/barbican-api-96f8997cd-z7vl4" Oct 08 18:34:27 crc kubenswrapper[4859]: I1008 18:34:27.040385 4859 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/8252c83d-1fdd-4d33-87a0-7a797251a0b4-ovsdbserver-sb\") pod \"dnsmasq-dns-744fd954cc-hv6xb\" (UID: \"8252c83d-1fdd-4d33-87a0-7a797251a0b4\") " pod="openstack/dnsmasq-dns-744fd954cc-hv6xb" Oct 08 18:34:27 crc kubenswrapper[4859]: I1008 18:34:27.040591 4859 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/8252c83d-1fdd-4d33-87a0-7a797251a0b4-dns-swift-storage-0\") pod \"dnsmasq-dns-744fd954cc-hv6xb\" (UID: \"8252c83d-1fdd-4d33-87a0-7a797251a0b4\") " pod="openstack/dnsmasq-dns-744fd954cc-hv6xb" Oct 08 18:34:27 crc kubenswrapper[4859]: I1008 18:34:27.041063 4859 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/8252c83d-1fdd-4d33-87a0-7a797251a0b4-dns-svc\") pod \"dnsmasq-dns-744fd954cc-hv6xb\" (UID: \"8252c83d-1fdd-4d33-87a0-7a797251a0b4\") " pod="openstack/dnsmasq-dns-744fd954cc-hv6xb" Oct 08 18:34:27 crc kubenswrapper[4859]: I1008 18:34:27.041510 4859 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/8252c83d-1fdd-4d33-87a0-7a797251a0b4-config\") pod \"dnsmasq-dns-744fd954cc-hv6xb\" (UID: \"8252c83d-1fdd-4d33-87a0-7a797251a0b4\") " pod="openstack/dnsmasq-dns-744fd954cc-hv6xb" Oct 08 18:34:27 crc kubenswrapper[4859]: I1008 18:34:27.061680 4859 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-jkx9v\" (UniqueName: \"kubernetes.io/projected/8252c83d-1fdd-4d33-87a0-7a797251a0b4-kube-api-access-jkx9v\") pod \"dnsmasq-dns-744fd954cc-hv6xb\" (UID: \"8252c83d-1fdd-4d33-87a0-7a797251a0b4\") " pod="openstack/dnsmasq-dns-744fd954cc-hv6xb" Oct 08 18:34:27 crc kubenswrapper[4859]: I1008 18:34:27.116592 4859 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-744fd954cc-hv6xb" Oct 08 18:34:27 crc kubenswrapper[4859]: I1008 18:34:27.142902 4859 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b6cd546e-9054-4150-a5f3-7bd709d15560-combined-ca-bundle\") pod \"barbican-api-96f8997cd-z7vl4\" (UID: \"b6cd546e-9054-4150-a5f3-7bd709d15560\") " pod="openstack/barbican-api-96f8997cd-z7vl4" Oct 08 18:34:27 crc kubenswrapper[4859]: I1008 18:34:27.142988 4859 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/b6cd546e-9054-4150-a5f3-7bd709d15560-config-data-custom\") pod \"barbican-api-96f8997cd-z7vl4\" (UID: \"b6cd546e-9054-4150-a5f3-7bd709d15560\") " pod="openstack/barbican-api-96f8997cd-z7vl4" Oct 08 18:34:27 crc kubenswrapper[4859]: I1008 18:34:27.143057 4859 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-lwkz8\" (UniqueName: \"kubernetes.io/projected/b6cd546e-9054-4150-a5f3-7bd709d15560-kube-api-access-lwkz8\") pod \"barbican-api-96f8997cd-z7vl4\" (UID: \"b6cd546e-9054-4150-a5f3-7bd709d15560\") " pod="openstack/barbican-api-96f8997cd-z7vl4" Oct 08 18:34:27 crc kubenswrapper[4859]: I1008 18:34:27.143100 4859 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/b6cd546e-9054-4150-a5f3-7bd709d15560-config-data\") pod \"barbican-api-96f8997cd-z7vl4\" (UID: \"b6cd546e-9054-4150-a5f3-7bd709d15560\") " pod="openstack/barbican-api-96f8997cd-z7vl4" Oct 08 18:34:27 crc kubenswrapper[4859]: I1008 18:34:27.143251 4859 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/b6cd546e-9054-4150-a5f3-7bd709d15560-logs\") pod \"barbican-api-96f8997cd-z7vl4\" (UID: \"b6cd546e-9054-4150-a5f3-7bd709d15560\") " pod="openstack/barbican-api-96f8997cd-z7vl4" Oct 08 18:34:27 crc kubenswrapper[4859]: I1008 18:34:27.143806 4859 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/b6cd546e-9054-4150-a5f3-7bd709d15560-logs\") pod \"barbican-api-96f8997cd-z7vl4\" (UID: \"b6cd546e-9054-4150-a5f3-7bd709d15560\") " pod="openstack/barbican-api-96f8997cd-z7vl4" Oct 08 18:34:27 crc kubenswrapper[4859]: I1008 18:34:27.149283 4859 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/b6cd546e-9054-4150-a5f3-7bd709d15560-config-data-custom\") pod \"barbican-api-96f8997cd-z7vl4\" (UID: \"b6cd546e-9054-4150-a5f3-7bd709d15560\") " pod="openstack/barbican-api-96f8997cd-z7vl4" Oct 08 18:34:27 crc kubenswrapper[4859]: I1008 18:34:27.150195 4859 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b6cd546e-9054-4150-a5f3-7bd709d15560-combined-ca-bundle\") pod \"barbican-api-96f8997cd-z7vl4\" (UID: \"b6cd546e-9054-4150-a5f3-7bd709d15560\") " pod="openstack/barbican-api-96f8997cd-z7vl4" Oct 08 18:34:27 crc kubenswrapper[4859]: I1008 18:34:27.155306 4859 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/b6cd546e-9054-4150-a5f3-7bd709d15560-config-data\") pod \"barbican-api-96f8997cd-z7vl4\" (UID: \"b6cd546e-9054-4150-a5f3-7bd709d15560\") " pod="openstack/barbican-api-96f8997cd-z7vl4" Oct 08 18:34:27 crc kubenswrapper[4859]: I1008 18:34:27.174616 4859 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-lwkz8\" (UniqueName: \"kubernetes.io/projected/b6cd546e-9054-4150-a5f3-7bd709d15560-kube-api-access-lwkz8\") pod \"barbican-api-96f8997cd-z7vl4\" (UID: \"b6cd546e-9054-4150-a5f3-7bd709d15560\") " pod="openstack/barbican-api-96f8997cd-z7vl4" Oct 08 18:34:27 crc kubenswrapper[4859]: I1008 18:34:27.237749 4859 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-api-96f8997cd-z7vl4" Oct 08 18:34:27 crc kubenswrapper[4859]: I1008 18:34:27.523927 4859 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-keystone-listener-d798f6bcd-5zw8p"] Oct 08 18:34:27 crc kubenswrapper[4859]: I1008 18:34:27.655570 4859 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-worker-7849cf5875-6srbt"] Oct 08 18:34:27 crc kubenswrapper[4859]: W1008 18:34:27.782433 4859 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod8252c83d_1fdd_4d33_87a0_7a797251a0b4.slice/crio-a20343aff730d6979c82829bba208a71d6c9b51409c7e6e6318db3ba888fc8b8 WatchSource:0}: Error finding container a20343aff730d6979c82829bba208a71d6c9b51409c7e6e6318db3ba888fc8b8: Status 404 returned error can't find the container with id a20343aff730d6979c82829bba208a71d6c9b51409c7e6e6318db3ba888fc8b8 Oct 08 18:34:27 crc kubenswrapper[4859]: I1008 18:34:27.791328 4859 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-744fd954cc-hv6xb"] Oct 08 18:34:27 crc kubenswrapper[4859]: I1008 18:34:27.869154 4859 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-api-96f8997cd-z7vl4"] Oct 08 18:34:27 crc kubenswrapper[4859]: W1008 18:34:27.872196 4859 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podb6cd546e_9054_4150_a5f3_7bd709d15560.slice/crio-335733d7843c6fc9d74b64c45ad4ab2c6b5e04222f1d6841fea11f46f57d5ba8 WatchSource:0}: Error finding container 335733d7843c6fc9d74b64c45ad4ab2c6b5e04222f1d6841fea11f46f57d5ba8: Status 404 returned error can't find the container with id 335733d7843c6fc9d74b64c45ad4ab2c6b5e04222f1d6841fea11f46f57d5ba8 Oct 08 18:34:27 crc kubenswrapper[4859]: I1008 18:34:27.958303 4859 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-api-96f8997cd-z7vl4" event={"ID":"b6cd546e-9054-4150-a5f3-7bd709d15560","Type":"ContainerStarted","Data":"335733d7843c6fc9d74b64c45ad4ab2c6b5e04222f1d6841fea11f46f57d5ba8"} Oct 08 18:34:27 crc kubenswrapper[4859]: I1008 18:34:27.960167 4859 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-744fd954cc-hv6xb" event={"ID":"8252c83d-1fdd-4d33-87a0-7a797251a0b4","Type":"ContainerStarted","Data":"a20343aff730d6979c82829bba208a71d6c9b51409c7e6e6318db3ba888fc8b8"} Oct 08 18:34:27 crc kubenswrapper[4859]: I1008 18:34:27.965166 4859 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-keystone-listener-d798f6bcd-5zw8p" event={"ID":"80660054-bd62-4742-a13d-5193b317b036","Type":"ContainerStarted","Data":"2f9a5813728b83ef43844c3fa16b149d32da22d17d6786742b0f597a63a8823d"} Oct 08 18:34:27 crc kubenswrapper[4859]: I1008 18:34:27.969217 4859 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-worker-7849cf5875-6srbt" event={"ID":"578aad66-818f-412f-9023-97ed6ae95de3","Type":"ContainerStarted","Data":"1cbc1696fdbbed35bcc2a4ecfcb6f275ad8a6c638b68c3cbc37ff31c5ddb4160"} Oct 08 18:34:28 crc kubenswrapper[4859]: I1008 18:34:28.982848 4859 generic.go:334] "Generic (PLEG): container finished" podID="8252c83d-1fdd-4d33-87a0-7a797251a0b4" containerID="b63328e8d573503c45cd4a842f1b869c293349e50d9aa5add18090b4f8c83200" exitCode=0 Oct 08 18:34:28 crc kubenswrapper[4859]: I1008 18:34:28.982895 4859 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-744fd954cc-hv6xb" event={"ID":"8252c83d-1fdd-4d33-87a0-7a797251a0b4","Type":"ContainerDied","Data":"b63328e8d573503c45cd4a842f1b869c293349e50d9aa5add18090b4f8c83200"} Oct 08 18:34:28 crc kubenswrapper[4859]: I1008 18:34:28.990020 4859 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-api-96f8997cd-z7vl4" event={"ID":"b6cd546e-9054-4150-a5f3-7bd709d15560","Type":"ContainerStarted","Data":"46929166d9610dcba7b53f0ba54350f18274f5815cacde12d63fdbf70d02c898"} Oct 08 18:34:28 crc kubenswrapper[4859]: I1008 18:34:28.990059 4859 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-api-96f8997cd-z7vl4" event={"ID":"b6cd546e-9054-4150-a5f3-7bd709d15560","Type":"ContainerStarted","Data":"f10ceaf742ba5f03074f81bdbcb2ca60481d377e7fca65e10865b51c73de052a"} Oct 08 18:34:28 crc kubenswrapper[4859]: I1008 18:34:28.990323 4859 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/barbican-api-96f8997cd-z7vl4" Oct 08 18:34:28 crc kubenswrapper[4859]: I1008 18:34:28.990844 4859 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/barbican-api-96f8997cd-z7vl4" Oct 08 18:34:29 crc kubenswrapper[4859]: I1008 18:34:29.036603 4859 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/barbican-api-96f8997cd-z7vl4" podStartSLOduration=3.036583428 podStartE2EDuration="3.036583428s" podCreationTimestamp="2025-10-08 18:34:26 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-08 18:34:29.031439857 +0000 UTC m=+1039.278279246" watchObservedRunningTime="2025-10-08 18:34:29.036583428 +0000 UTC m=+1039.283422807" Oct 08 18:34:29 crc kubenswrapper[4859]: I1008 18:34:29.802175 4859 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/barbican-api-588f98c7dd-c7l8h"] Oct 08 18:34:29 crc kubenswrapper[4859]: I1008 18:34:29.808204 4859 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-api-588f98c7dd-c7l8h" Oct 08 18:34:29 crc kubenswrapper[4859]: I1008 18:34:29.811022 4859 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-barbican-internal-svc" Oct 08 18:34:29 crc kubenswrapper[4859]: I1008 18:34:29.811395 4859 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-barbican-public-svc" Oct 08 18:34:29 crc kubenswrapper[4859]: I1008 18:34:29.815906 4859 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-api-588f98c7dd-c7l8h"] Oct 08 18:34:29 crc kubenswrapper[4859]: I1008 18:34:29.905168 4859 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/3ecae548-eaa6-4c24-a538-b2a34fa0308f-config-data\") pod \"barbican-api-588f98c7dd-c7l8h\" (UID: \"3ecae548-eaa6-4c24-a538-b2a34fa0308f\") " pod="openstack/barbican-api-588f98c7dd-c7l8h" Oct 08 18:34:29 crc kubenswrapper[4859]: I1008 18:34:29.905574 4859 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/3ecae548-eaa6-4c24-a538-b2a34fa0308f-config-data-custom\") pod \"barbican-api-588f98c7dd-c7l8h\" (UID: \"3ecae548-eaa6-4c24-a538-b2a34fa0308f\") " pod="openstack/barbican-api-588f98c7dd-c7l8h" Oct 08 18:34:29 crc kubenswrapper[4859]: I1008 18:34:29.905612 4859 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/3ecae548-eaa6-4c24-a538-b2a34fa0308f-public-tls-certs\") pod \"barbican-api-588f98c7dd-c7l8h\" (UID: \"3ecae548-eaa6-4c24-a538-b2a34fa0308f\") " pod="openstack/barbican-api-588f98c7dd-c7l8h" Oct 08 18:34:29 crc kubenswrapper[4859]: I1008 18:34:29.905652 4859 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3ecae548-eaa6-4c24-a538-b2a34fa0308f-combined-ca-bundle\") pod \"barbican-api-588f98c7dd-c7l8h\" (UID: \"3ecae548-eaa6-4c24-a538-b2a34fa0308f\") " pod="openstack/barbican-api-588f98c7dd-c7l8h" Oct 08 18:34:29 crc kubenswrapper[4859]: I1008 18:34:29.905898 4859 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/3ecae548-eaa6-4c24-a538-b2a34fa0308f-logs\") pod \"barbican-api-588f98c7dd-c7l8h\" (UID: \"3ecae548-eaa6-4c24-a538-b2a34fa0308f\") " pod="openstack/barbican-api-588f98c7dd-c7l8h" Oct 08 18:34:29 crc kubenswrapper[4859]: I1008 18:34:29.905951 4859 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-2vph8\" (UniqueName: \"kubernetes.io/projected/3ecae548-eaa6-4c24-a538-b2a34fa0308f-kube-api-access-2vph8\") pod \"barbican-api-588f98c7dd-c7l8h\" (UID: \"3ecae548-eaa6-4c24-a538-b2a34fa0308f\") " pod="openstack/barbican-api-588f98c7dd-c7l8h" Oct 08 18:34:29 crc kubenswrapper[4859]: I1008 18:34:29.906028 4859 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/3ecae548-eaa6-4c24-a538-b2a34fa0308f-internal-tls-certs\") pod \"barbican-api-588f98c7dd-c7l8h\" (UID: \"3ecae548-eaa6-4c24-a538-b2a34fa0308f\") " pod="openstack/barbican-api-588f98c7dd-c7l8h" Oct 08 18:34:30 crc kubenswrapper[4859]: I1008 18:34:30.002526 4859 generic.go:334] "Generic (PLEG): container finished" podID="87f0eecb-178a-438a-8643-e29c19390884" containerID="c76586e5e90236f90858e341d25b5dde536c874c2bca5a632ffc704ce10dd611" exitCode=0 Oct 08 18:34:30 crc kubenswrapper[4859]: I1008 18:34:30.002595 4859 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"87f0eecb-178a-438a-8643-e29c19390884","Type":"ContainerDied","Data":"c76586e5e90236f90858e341d25b5dde536c874c2bca5a632ffc704ce10dd611"} Oct 08 18:34:30 crc kubenswrapper[4859]: I1008 18:34:30.002631 4859 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"87f0eecb-178a-438a-8643-e29c19390884","Type":"ContainerDied","Data":"78f9e1079ba0c58f7bdefd70eb6b6101a88872c51f57d7f91919f7b8db0bdcb2"} Oct 08 18:34:30 crc kubenswrapper[4859]: I1008 18:34:30.002644 4859 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="78f9e1079ba0c58f7bdefd70eb6b6101a88872c51f57d7f91919f7b8db0bdcb2" Oct 08 18:34:30 crc kubenswrapper[4859]: I1008 18:34:30.007980 4859 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/3ecae548-eaa6-4c24-a538-b2a34fa0308f-logs\") pod \"barbican-api-588f98c7dd-c7l8h\" (UID: \"3ecae548-eaa6-4c24-a538-b2a34fa0308f\") " pod="openstack/barbican-api-588f98c7dd-c7l8h" Oct 08 18:34:30 crc kubenswrapper[4859]: I1008 18:34:30.008338 4859 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-2vph8\" (UniqueName: \"kubernetes.io/projected/3ecae548-eaa6-4c24-a538-b2a34fa0308f-kube-api-access-2vph8\") pod \"barbican-api-588f98c7dd-c7l8h\" (UID: \"3ecae548-eaa6-4c24-a538-b2a34fa0308f\") " pod="openstack/barbican-api-588f98c7dd-c7l8h" Oct 08 18:34:30 crc kubenswrapper[4859]: I1008 18:34:30.008383 4859 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/3ecae548-eaa6-4c24-a538-b2a34fa0308f-internal-tls-certs\") pod \"barbican-api-588f98c7dd-c7l8h\" (UID: \"3ecae548-eaa6-4c24-a538-b2a34fa0308f\") " pod="openstack/barbican-api-588f98c7dd-c7l8h" Oct 08 18:34:30 crc kubenswrapper[4859]: I1008 18:34:30.008409 4859 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/3ecae548-eaa6-4c24-a538-b2a34fa0308f-config-data\") pod \"barbican-api-588f98c7dd-c7l8h\" (UID: \"3ecae548-eaa6-4c24-a538-b2a34fa0308f\") " pod="openstack/barbican-api-588f98c7dd-c7l8h" Oct 08 18:34:30 crc kubenswrapper[4859]: I1008 18:34:30.008537 4859 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/3ecae548-eaa6-4c24-a538-b2a34fa0308f-config-data-custom\") pod \"barbican-api-588f98c7dd-c7l8h\" (UID: \"3ecae548-eaa6-4c24-a538-b2a34fa0308f\") " pod="openstack/barbican-api-588f98c7dd-c7l8h" Oct 08 18:34:30 crc kubenswrapper[4859]: I1008 18:34:30.008555 4859 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/3ecae548-eaa6-4c24-a538-b2a34fa0308f-logs\") pod \"barbican-api-588f98c7dd-c7l8h\" (UID: \"3ecae548-eaa6-4c24-a538-b2a34fa0308f\") " pod="openstack/barbican-api-588f98c7dd-c7l8h" Oct 08 18:34:30 crc kubenswrapper[4859]: I1008 18:34:30.008589 4859 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/3ecae548-eaa6-4c24-a538-b2a34fa0308f-public-tls-certs\") pod \"barbican-api-588f98c7dd-c7l8h\" (UID: \"3ecae548-eaa6-4c24-a538-b2a34fa0308f\") " pod="openstack/barbican-api-588f98c7dd-c7l8h" Oct 08 18:34:30 crc kubenswrapper[4859]: I1008 18:34:30.008631 4859 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3ecae548-eaa6-4c24-a538-b2a34fa0308f-combined-ca-bundle\") pod \"barbican-api-588f98c7dd-c7l8h\" (UID: \"3ecae548-eaa6-4c24-a538-b2a34fa0308f\") " pod="openstack/barbican-api-588f98c7dd-c7l8h" Oct 08 18:34:30 crc kubenswrapper[4859]: I1008 18:34:30.013913 4859 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/3ecae548-eaa6-4c24-a538-b2a34fa0308f-config-data-custom\") pod \"barbican-api-588f98c7dd-c7l8h\" (UID: \"3ecae548-eaa6-4c24-a538-b2a34fa0308f\") " pod="openstack/barbican-api-588f98c7dd-c7l8h" Oct 08 18:34:30 crc kubenswrapper[4859]: I1008 18:34:30.019314 4859 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/3ecae548-eaa6-4c24-a538-b2a34fa0308f-internal-tls-certs\") pod \"barbican-api-588f98c7dd-c7l8h\" (UID: \"3ecae548-eaa6-4c24-a538-b2a34fa0308f\") " pod="openstack/barbican-api-588f98c7dd-c7l8h" Oct 08 18:34:30 crc kubenswrapper[4859]: I1008 18:34:30.019463 4859 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/3ecae548-eaa6-4c24-a538-b2a34fa0308f-config-data\") pod \"barbican-api-588f98c7dd-c7l8h\" (UID: \"3ecae548-eaa6-4c24-a538-b2a34fa0308f\") " pod="openstack/barbican-api-588f98c7dd-c7l8h" Oct 08 18:34:30 crc kubenswrapper[4859]: I1008 18:34:30.019576 4859 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3ecae548-eaa6-4c24-a538-b2a34fa0308f-combined-ca-bundle\") pod \"barbican-api-588f98c7dd-c7l8h\" (UID: \"3ecae548-eaa6-4c24-a538-b2a34fa0308f\") " pod="openstack/barbican-api-588f98c7dd-c7l8h" Oct 08 18:34:30 crc kubenswrapper[4859]: I1008 18:34:30.019726 4859 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/3ecae548-eaa6-4c24-a538-b2a34fa0308f-public-tls-certs\") pod \"barbican-api-588f98c7dd-c7l8h\" (UID: \"3ecae548-eaa6-4c24-a538-b2a34fa0308f\") " pod="openstack/barbican-api-588f98c7dd-c7l8h" Oct 08 18:34:30 crc kubenswrapper[4859]: I1008 18:34:30.028908 4859 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-2vph8\" (UniqueName: \"kubernetes.io/projected/3ecae548-eaa6-4c24-a538-b2a34fa0308f-kube-api-access-2vph8\") pod \"barbican-api-588f98c7dd-c7l8h\" (UID: \"3ecae548-eaa6-4c24-a538-b2a34fa0308f\") " pod="openstack/barbican-api-588f98c7dd-c7l8h" Oct 08 18:34:30 crc kubenswrapper[4859]: I1008 18:34:30.118930 4859 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-api-588f98c7dd-c7l8h" Oct 08 18:34:30 crc kubenswrapper[4859]: I1008 18:34:30.122179 4859 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Oct 08 18:34:30 crc kubenswrapper[4859]: I1008 18:34:30.225430 4859 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/87f0eecb-178a-438a-8643-e29c19390884-log-httpd\") pod \"87f0eecb-178a-438a-8643-e29c19390884\" (UID: \"87f0eecb-178a-438a-8643-e29c19390884\") " Oct 08 18:34:30 crc kubenswrapper[4859]: I1008 18:34:30.225490 4859 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/87f0eecb-178a-438a-8643-e29c19390884-config-data\") pod \"87f0eecb-178a-438a-8643-e29c19390884\" (UID: \"87f0eecb-178a-438a-8643-e29c19390884\") " Oct 08 18:34:30 crc kubenswrapper[4859]: I1008 18:34:30.225554 4859 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/87f0eecb-178a-438a-8643-e29c19390884-combined-ca-bundle\") pod \"87f0eecb-178a-438a-8643-e29c19390884\" (UID: \"87f0eecb-178a-438a-8643-e29c19390884\") " Oct 08 18:34:30 crc kubenswrapper[4859]: I1008 18:34:30.225608 4859 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/87f0eecb-178a-438a-8643-e29c19390884-scripts\") pod \"87f0eecb-178a-438a-8643-e29c19390884\" (UID: \"87f0eecb-178a-438a-8643-e29c19390884\") " Oct 08 18:34:30 crc kubenswrapper[4859]: I1008 18:34:30.225636 4859 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/87f0eecb-178a-438a-8643-e29c19390884-sg-core-conf-yaml\") pod \"87f0eecb-178a-438a-8643-e29c19390884\" (UID: \"87f0eecb-178a-438a-8643-e29c19390884\") " Oct 08 18:34:30 crc kubenswrapper[4859]: I1008 18:34:30.225733 4859 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/87f0eecb-178a-438a-8643-e29c19390884-run-httpd\") pod \"87f0eecb-178a-438a-8643-e29c19390884\" (UID: \"87f0eecb-178a-438a-8643-e29c19390884\") " Oct 08 18:34:30 crc kubenswrapper[4859]: I1008 18:34:30.225809 4859 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-nx5wn\" (UniqueName: \"kubernetes.io/projected/87f0eecb-178a-438a-8643-e29c19390884-kube-api-access-nx5wn\") pod \"87f0eecb-178a-438a-8643-e29c19390884\" (UID: \"87f0eecb-178a-438a-8643-e29c19390884\") " Oct 08 18:34:30 crc kubenswrapper[4859]: I1008 18:34:30.232862 4859 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/87f0eecb-178a-438a-8643-e29c19390884-scripts" (OuterVolumeSpecName: "scripts") pod "87f0eecb-178a-438a-8643-e29c19390884" (UID: "87f0eecb-178a-438a-8643-e29c19390884"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 08 18:34:30 crc kubenswrapper[4859]: I1008 18:34:30.233254 4859 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/87f0eecb-178a-438a-8643-e29c19390884-log-httpd" (OuterVolumeSpecName: "log-httpd") pod "87f0eecb-178a-438a-8643-e29c19390884" (UID: "87f0eecb-178a-438a-8643-e29c19390884"). InnerVolumeSpecName "log-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 08 18:34:30 crc kubenswrapper[4859]: I1008 18:34:30.234710 4859 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/87f0eecb-178a-438a-8643-e29c19390884-run-httpd" (OuterVolumeSpecName: "run-httpd") pod "87f0eecb-178a-438a-8643-e29c19390884" (UID: "87f0eecb-178a-438a-8643-e29c19390884"). InnerVolumeSpecName "run-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 08 18:34:30 crc kubenswrapper[4859]: I1008 18:34:30.242633 4859 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/87f0eecb-178a-438a-8643-e29c19390884-kube-api-access-nx5wn" (OuterVolumeSpecName: "kube-api-access-nx5wn") pod "87f0eecb-178a-438a-8643-e29c19390884" (UID: "87f0eecb-178a-438a-8643-e29c19390884"). InnerVolumeSpecName "kube-api-access-nx5wn". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 08 18:34:30 crc kubenswrapper[4859]: I1008 18:34:30.261594 4859 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/87f0eecb-178a-438a-8643-e29c19390884-sg-core-conf-yaml" (OuterVolumeSpecName: "sg-core-conf-yaml") pod "87f0eecb-178a-438a-8643-e29c19390884" (UID: "87f0eecb-178a-438a-8643-e29c19390884"). InnerVolumeSpecName "sg-core-conf-yaml". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 08 18:34:30 crc kubenswrapper[4859]: I1008 18:34:30.334025 4859 reconciler_common.go:293] "Volume detached for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/87f0eecb-178a-438a-8643-e29c19390884-run-httpd\") on node \"crc\" DevicePath \"\"" Oct 08 18:34:30 crc kubenswrapper[4859]: I1008 18:34:30.334073 4859 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-nx5wn\" (UniqueName: \"kubernetes.io/projected/87f0eecb-178a-438a-8643-e29c19390884-kube-api-access-nx5wn\") on node \"crc\" DevicePath \"\"" Oct 08 18:34:30 crc kubenswrapper[4859]: I1008 18:34:30.334088 4859 reconciler_common.go:293] "Volume detached for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/87f0eecb-178a-438a-8643-e29c19390884-log-httpd\") on node \"crc\" DevicePath \"\"" Oct 08 18:34:30 crc kubenswrapper[4859]: I1008 18:34:30.334100 4859 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/87f0eecb-178a-438a-8643-e29c19390884-scripts\") on node \"crc\" DevicePath \"\"" Oct 08 18:34:30 crc kubenswrapper[4859]: I1008 18:34:30.334110 4859 reconciler_common.go:293] "Volume detached for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/87f0eecb-178a-438a-8643-e29c19390884-sg-core-conf-yaml\") on node \"crc\" DevicePath \"\"" Oct 08 18:34:30 crc kubenswrapper[4859]: I1008 18:34:30.437385 4859 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/87f0eecb-178a-438a-8643-e29c19390884-config-data" (OuterVolumeSpecName: "config-data") pod "87f0eecb-178a-438a-8643-e29c19390884" (UID: "87f0eecb-178a-438a-8643-e29c19390884"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 08 18:34:30 crc kubenswrapper[4859]: I1008 18:34:30.462860 4859 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/87f0eecb-178a-438a-8643-e29c19390884-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "87f0eecb-178a-438a-8643-e29c19390884" (UID: "87f0eecb-178a-438a-8643-e29c19390884"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 08 18:34:30 crc kubenswrapper[4859]: I1008 18:34:30.537478 4859 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/87f0eecb-178a-438a-8643-e29c19390884-config-data\") on node \"crc\" DevicePath \"\"" Oct 08 18:34:30 crc kubenswrapper[4859]: I1008 18:34:30.538654 4859 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/87f0eecb-178a-438a-8643-e29c19390884-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 08 18:34:30 crc kubenswrapper[4859]: I1008 18:34:30.681344 4859 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-api-588f98c7dd-c7l8h"] Oct 08 18:34:30 crc kubenswrapper[4859]: W1008 18:34:30.684874 4859 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod3ecae548_eaa6_4c24_a538_b2a34fa0308f.slice/crio-b14b7481aa3cc30f9629c6a0ec5b73e86034b444322e4f3172b707870f27429a WatchSource:0}: Error finding container b14b7481aa3cc30f9629c6a0ec5b73e86034b444322e4f3172b707870f27429a: Status 404 returned error can't find the container with id b14b7481aa3cc30f9629c6a0ec5b73e86034b444322e4f3172b707870f27429a Oct 08 18:34:31 crc kubenswrapper[4859]: I1008 18:34:31.010883 4859 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-744fd954cc-hv6xb" event={"ID":"8252c83d-1fdd-4d33-87a0-7a797251a0b4","Type":"ContainerStarted","Data":"080d2b87a50a3198eeb74c3989be6bedec81fb60d37a7c683e9630dc72474e73"} Oct 08 18:34:31 crc kubenswrapper[4859]: I1008 18:34:31.011039 4859 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-744fd954cc-hv6xb" Oct 08 18:34:31 crc kubenswrapper[4859]: I1008 18:34:31.013517 4859 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-api-588f98c7dd-c7l8h" event={"ID":"3ecae548-eaa6-4c24-a538-b2a34fa0308f","Type":"ContainerStarted","Data":"006d1c905b3b825b739c89788456b9db6545bb24fad677fade94d0d6aa193180"} Oct 08 18:34:31 crc kubenswrapper[4859]: I1008 18:34:31.013554 4859 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-api-588f98c7dd-c7l8h" event={"ID":"3ecae548-eaa6-4c24-a538-b2a34fa0308f","Type":"ContainerStarted","Data":"b14b7481aa3cc30f9629c6a0ec5b73e86034b444322e4f3172b707870f27429a"} Oct 08 18:34:31 crc kubenswrapper[4859]: I1008 18:34:31.015197 4859 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-keystone-listener-d798f6bcd-5zw8p" event={"ID":"80660054-bd62-4742-a13d-5193b317b036","Type":"ContainerStarted","Data":"6acc6550472aa8cbcfe5e421e001159e4a454b68eed0d972de3345d610c1694b"} Oct 08 18:34:31 crc kubenswrapper[4859]: I1008 18:34:31.015226 4859 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-keystone-listener-d798f6bcd-5zw8p" event={"ID":"80660054-bd62-4742-a13d-5193b317b036","Type":"ContainerStarted","Data":"7b9ddd84f5b7595b2e97a4ad3e49442ce4cc096e8c30b99ff8f31a2ded0e034a"} Oct 08 18:34:31 crc kubenswrapper[4859]: I1008 18:34:31.016696 4859 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-worker-7849cf5875-6srbt" event={"ID":"578aad66-818f-412f-9023-97ed6ae95de3","Type":"ContainerStarted","Data":"feea4ccd8bb9548f9e92f1a1fbbea83def65d8f9aa7fade25ce7088c912b5f84"} Oct 08 18:34:31 crc kubenswrapper[4859]: I1008 18:34:31.016725 4859 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-worker-7849cf5875-6srbt" event={"ID":"578aad66-818f-412f-9023-97ed6ae95de3","Type":"ContainerStarted","Data":"0526a6ef25fa627b6402ab942796d84fcf8fb713850a1f02219ee07d6d1afa76"} Oct 08 18:34:31 crc kubenswrapper[4859]: I1008 18:34:31.017031 4859 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Oct 08 18:34:31 crc kubenswrapper[4859]: I1008 18:34:31.040062 4859 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-744fd954cc-hv6xb" podStartSLOduration=5.040047539 podStartE2EDuration="5.040047539s" podCreationTimestamp="2025-10-08 18:34:26 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-08 18:34:31.038238416 +0000 UTC m=+1041.285077795" watchObservedRunningTime="2025-10-08 18:34:31.040047539 +0000 UTC m=+1041.286886918" Oct 08 18:34:31 crc kubenswrapper[4859]: I1008 18:34:31.074747 4859 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Oct 08 18:34:31 crc kubenswrapper[4859]: I1008 18:34:31.095775 4859 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/ceilometer-0"] Oct 08 18:34:31 crc kubenswrapper[4859]: I1008 18:34:31.124165 4859 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/barbican-keystone-listener-d798f6bcd-5zw8p" podStartSLOduration=3.016761686 podStartE2EDuration="5.124125575s" podCreationTimestamp="2025-10-08 18:34:26 +0000 UTC" firstStartedPulling="2025-10-08 18:34:27.540649145 +0000 UTC m=+1037.787488524" lastFinishedPulling="2025-10-08 18:34:29.648013034 +0000 UTC m=+1039.894852413" observedRunningTime="2025-10-08 18:34:31.07831374 +0000 UTC m=+1041.325153129" watchObservedRunningTime="2025-10-08 18:34:31.124125575 +0000 UTC m=+1041.370964954" Oct 08 18:34:31 crc kubenswrapper[4859]: I1008 18:34:31.153288 4859 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ceilometer-0"] Oct 08 18:34:31 crc kubenswrapper[4859]: E1008 18:34:31.153767 4859 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="87f0eecb-178a-438a-8643-e29c19390884" containerName="ceilometer-notification-agent" Oct 08 18:34:31 crc kubenswrapper[4859]: I1008 18:34:31.153787 4859 state_mem.go:107] "Deleted CPUSet assignment" podUID="87f0eecb-178a-438a-8643-e29c19390884" containerName="ceilometer-notification-agent" Oct 08 18:34:31 crc kubenswrapper[4859]: E1008 18:34:31.153807 4859 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="87f0eecb-178a-438a-8643-e29c19390884" containerName="sg-core" Oct 08 18:34:31 crc kubenswrapper[4859]: I1008 18:34:31.153815 4859 state_mem.go:107] "Deleted CPUSet assignment" podUID="87f0eecb-178a-438a-8643-e29c19390884" containerName="sg-core" Oct 08 18:34:31 crc kubenswrapper[4859]: E1008 18:34:31.153824 4859 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="87f0eecb-178a-438a-8643-e29c19390884" containerName="ceilometer-central-agent" Oct 08 18:34:31 crc kubenswrapper[4859]: I1008 18:34:31.153831 4859 state_mem.go:107] "Deleted CPUSet assignment" podUID="87f0eecb-178a-438a-8643-e29c19390884" containerName="ceilometer-central-agent" Oct 08 18:34:31 crc kubenswrapper[4859]: E1008 18:34:31.153861 4859 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="87f0eecb-178a-438a-8643-e29c19390884" containerName="proxy-httpd" Oct 08 18:34:31 crc kubenswrapper[4859]: I1008 18:34:31.153871 4859 state_mem.go:107] "Deleted CPUSet assignment" podUID="87f0eecb-178a-438a-8643-e29c19390884" containerName="proxy-httpd" Oct 08 18:34:31 crc kubenswrapper[4859]: I1008 18:34:31.154083 4859 memory_manager.go:354] "RemoveStaleState removing state" podUID="87f0eecb-178a-438a-8643-e29c19390884" containerName="ceilometer-central-agent" Oct 08 18:34:31 crc kubenswrapper[4859]: I1008 18:34:31.154108 4859 memory_manager.go:354] "RemoveStaleState removing state" podUID="87f0eecb-178a-438a-8643-e29c19390884" containerName="sg-core" Oct 08 18:34:31 crc kubenswrapper[4859]: I1008 18:34:31.154127 4859 memory_manager.go:354] "RemoveStaleState removing state" podUID="87f0eecb-178a-438a-8643-e29c19390884" containerName="ceilometer-notification-agent" Oct 08 18:34:31 crc kubenswrapper[4859]: I1008 18:34:31.154140 4859 memory_manager.go:354] "RemoveStaleState removing state" podUID="87f0eecb-178a-438a-8643-e29c19390884" containerName="proxy-httpd" Oct 08 18:34:31 crc kubenswrapper[4859]: I1008 18:34:31.156472 4859 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Oct 08 18:34:31 crc kubenswrapper[4859]: I1008 18:34:31.175377 4859 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-scripts" Oct 08 18:34:31 crc kubenswrapper[4859]: I1008 18:34:31.175631 4859 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-config-data" Oct 08 18:34:31 crc kubenswrapper[4859]: I1008 18:34:31.181049 4859 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Oct 08 18:34:31 crc kubenswrapper[4859]: I1008 18:34:31.266662 4859 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/7ad402c0-68c6-4c69-b1ad-0fd0d5c69dff-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"7ad402c0-68c6-4c69-b1ad-0fd0d5c69dff\") " pod="openstack/ceilometer-0" Oct 08 18:34:31 crc kubenswrapper[4859]: I1008 18:34:31.266734 4859 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/7ad402c0-68c6-4c69-b1ad-0fd0d5c69dff-scripts\") pod \"ceilometer-0\" (UID: \"7ad402c0-68c6-4c69-b1ad-0fd0d5c69dff\") " pod="openstack/ceilometer-0" Oct 08 18:34:31 crc kubenswrapper[4859]: I1008 18:34:31.266761 4859 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/7ad402c0-68c6-4c69-b1ad-0fd0d5c69dff-log-httpd\") pod \"ceilometer-0\" (UID: \"7ad402c0-68c6-4c69-b1ad-0fd0d5c69dff\") " pod="openstack/ceilometer-0" Oct 08 18:34:31 crc kubenswrapper[4859]: I1008 18:34:31.267030 4859 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/7ad402c0-68c6-4c69-b1ad-0fd0d5c69dff-run-httpd\") pod \"ceilometer-0\" (UID: \"7ad402c0-68c6-4c69-b1ad-0fd0d5c69dff\") " pod="openstack/ceilometer-0" Oct 08 18:34:31 crc kubenswrapper[4859]: I1008 18:34:31.267174 4859 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7ad402c0-68c6-4c69-b1ad-0fd0d5c69dff-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"7ad402c0-68c6-4c69-b1ad-0fd0d5c69dff\") " pod="openstack/ceilometer-0" Oct 08 18:34:31 crc kubenswrapper[4859]: I1008 18:34:31.267253 4859 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-lzscf\" (UniqueName: \"kubernetes.io/projected/7ad402c0-68c6-4c69-b1ad-0fd0d5c69dff-kube-api-access-lzscf\") pod \"ceilometer-0\" (UID: \"7ad402c0-68c6-4c69-b1ad-0fd0d5c69dff\") " pod="openstack/ceilometer-0" Oct 08 18:34:31 crc kubenswrapper[4859]: I1008 18:34:31.267315 4859 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/7ad402c0-68c6-4c69-b1ad-0fd0d5c69dff-config-data\") pod \"ceilometer-0\" (UID: \"7ad402c0-68c6-4c69-b1ad-0fd0d5c69dff\") " pod="openstack/ceilometer-0" Oct 08 18:34:31 crc kubenswrapper[4859]: I1008 18:34:31.369479 4859 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/7ad402c0-68c6-4c69-b1ad-0fd0d5c69dff-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"7ad402c0-68c6-4c69-b1ad-0fd0d5c69dff\") " pod="openstack/ceilometer-0" Oct 08 18:34:31 crc kubenswrapper[4859]: I1008 18:34:31.369533 4859 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/7ad402c0-68c6-4c69-b1ad-0fd0d5c69dff-scripts\") pod \"ceilometer-0\" (UID: \"7ad402c0-68c6-4c69-b1ad-0fd0d5c69dff\") " pod="openstack/ceilometer-0" Oct 08 18:34:31 crc kubenswrapper[4859]: I1008 18:34:31.369553 4859 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/7ad402c0-68c6-4c69-b1ad-0fd0d5c69dff-log-httpd\") pod \"ceilometer-0\" (UID: \"7ad402c0-68c6-4c69-b1ad-0fd0d5c69dff\") " pod="openstack/ceilometer-0" Oct 08 18:34:31 crc kubenswrapper[4859]: I1008 18:34:31.369605 4859 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/7ad402c0-68c6-4c69-b1ad-0fd0d5c69dff-run-httpd\") pod \"ceilometer-0\" (UID: \"7ad402c0-68c6-4c69-b1ad-0fd0d5c69dff\") " pod="openstack/ceilometer-0" Oct 08 18:34:31 crc kubenswrapper[4859]: I1008 18:34:31.369647 4859 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7ad402c0-68c6-4c69-b1ad-0fd0d5c69dff-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"7ad402c0-68c6-4c69-b1ad-0fd0d5c69dff\") " pod="openstack/ceilometer-0" Oct 08 18:34:31 crc kubenswrapper[4859]: I1008 18:34:31.369677 4859 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-lzscf\" (UniqueName: \"kubernetes.io/projected/7ad402c0-68c6-4c69-b1ad-0fd0d5c69dff-kube-api-access-lzscf\") pod \"ceilometer-0\" (UID: \"7ad402c0-68c6-4c69-b1ad-0fd0d5c69dff\") " pod="openstack/ceilometer-0" Oct 08 18:34:31 crc kubenswrapper[4859]: I1008 18:34:31.369723 4859 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/7ad402c0-68c6-4c69-b1ad-0fd0d5c69dff-config-data\") pod \"ceilometer-0\" (UID: \"7ad402c0-68c6-4c69-b1ad-0fd0d5c69dff\") " pod="openstack/ceilometer-0" Oct 08 18:34:31 crc kubenswrapper[4859]: I1008 18:34:31.370117 4859 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/7ad402c0-68c6-4c69-b1ad-0fd0d5c69dff-run-httpd\") pod \"ceilometer-0\" (UID: \"7ad402c0-68c6-4c69-b1ad-0fd0d5c69dff\") " pod="openstack/ceilometer-0" Oct 08 18:34:31 crc kubenswrapper[4859]: I1008 18:34:31.370138 4859 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/7ad402c0-68c6-4c69-b1ad-0fd0d5c69dff-log-httpd\") pod \"ceilometer-0\" (UID: \"7ad402c0-68c6-4c69-b1ad-0fd0d5c69dff\") " pod="openstack/ceilometer-0" Oct 08 18:34:31 crc kubenswrapper[4859]: I1008 18:34:31.372952 4859 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/7ad402c0-68c6-4c69-b1ad-0fd0d5c69dff-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"7ad402c0-68c6-4c69-b1ad-0fd0d5c69dff\") " pod="openstack/ceilometer-0" Oct 08 18:34:31 crc kubenswrapper[4859]: I1008 18:34:31.373725 4859 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/7ad402c0-68c6-4c69-b1ad-0fd0d5c69dff-scripts\") pod \"ceilometer-0\" (UID: \"7ad402c0-68c6-4c69-b1ad-0fd0d5c69dff\") " pod="openstack/ceilometer-0" Oct 08 18:34:31 crc kubenswrapper[4859]: I1008 18:34:31.373961 4859 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/7ad402c0-68c6-4c69-b1ad-0fd0d5c69dff-config-data\") pod \"ceilometer-0\" (UID: \"7ad402c0-68c6-4c69-b1ad-0fd0d5c69dff\") " pod="openstack/ceilometer-0" Oct 08 18:34:31 crc kubenswrapper[4859]: I1008 18:34:31.378181 4859 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7ad402c0-68c6-4c69-b1ad-0fd0d5c69dff-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"7ad402c0-68c6-4c69-b1ad-0fd0d5c69dff\") " pod="openstack/ceilometer-0" Oct 08 18:34:31 crc kubenswrapper[4859]: I1008 18:34:31.387329 4859 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-lzscf\" (UniqueName: \"kubernetes.io/projected/7ad402c0-68c6-4c69-b1ad-0fd0d5c69dff-kube-api-access-lzscf\") pod \"ceilometer-0\" (UID: \"7ad402c0-68c6-4c69-b1ad-0fd0d5c69dff\") " pod="openstack/ceilometer-0" Oct 08 18:34:31 crc kubenswrapper[4859]: I1008 18:34:31.500339 4859 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Oct 08 18:34:32 crc kubenswrapper[4859]: I1008 18:34:32.009434 4859 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Oct 08 18:34:32 crc kubenswrapper[4859]: I1008 18:34:32.026812 4859 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"7ad402c0-68c6-4c69-b1ad-0fd0d5c69dff","Type":"ContainerStarted","Data":"5005b74fa8e9eb369194edce4800584323343c0e0f5e64adbc4cb80fb161fbb5"} Oct 08 18:34:32 crc kubenswrapper[4859]: I1008 18:34:32.030805 4859 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-api-588f98c7dd-c7l8h" event={"ID":"3ecae548-eaa6-4c24-a538-b2a34fa0308f","Type":"ContainerStarted","Data":"f9de87d868c2ec3b852a8b9c3242543c78423232cfe817f0af3df915cf1c4423"} Oct 08 18:34:32 crc kubenswrapper[4859]: I1008 18:34:32.030848 4859 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/barbican-api-588f98c7dd-c7l8h" Oct 08 18:34:32 crc kubenswrapper[4859]: I1008 18:34:32.031204 4859 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/barbican-api-588f98c7dd-c7l8h" Oct 08 18:34:32 crc kubenswrapper[4859]: I1008 18:34:32.053981 4859 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/barbican-api-588f98c7dd-c7l8h" podStartSLOduration=3.053961299 podStartE2EDuration="3.053961299s" podCreationTimestamp="2025-10-08 18:34:29 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-08 18:34:32.05299317 +0000 UTC m=+1042.299832559" watchObservedRunningTime="2025-10-08 18:34:32.053961299 +0000 UTC m=+1042.300800678" Oct 08 18:34:32 crc kubenswrapper[4859]: I1008 18:34:32.482720 4859 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="87f0eecb-178a-438a-8643-e29c19390884" path="/var/lib/kubelet/pods/87f0eecb-178a-438a-8643-e29c19390884/volumes" Oct 08 18:34:33 crc kubenswrapper[4859]: I1008 18:34:33.047946 4859 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"7ad402c0-68c6-4c69-b1ad-0fd0d5c69dff","Type":"ContainerStarted","Data":"00b2cd92e3299a8a3ff2ab92560ece56a4acf64fd255489ac76be208726ef44c"} Oct 08 18:34:34 crc kubenswrapper[4859]: I1008 18:34:34.073078 4859 generic.go:334] "Generic (PLEG): container finished" podID="720db6c0-f18f-429b-9589-fa503d212139" containerID="f705521531b392a51ca4b6b14b60c8d34d3ec8f7dbebf2794d813c0cd105ff2c" exitCode=0 Oct 08 18:34:34 crc kubenswrapper[4859]: I1008 18:34:34.074708 4859 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-db-sync-jqwrh" event={"ID":"720db6c0-f18f-429b-9589-fa503d212139","Type":"ContainerDied","Data":"f705521531b392a51ca4b6b14b60c8d34d3ec8f7dbebf2794d813c0cd105ff2c"} Oct 08 18:34:34 crc kubenswrapper[4859]: I1008 18:34:34.078131 4859 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"7ad402c0-68c6-4c69-b1ad-0fd0d5c69dff","Type":"ContainerStarted","Data":"bd3b5549c52a699ff8bf5a14fe7e680fd8f522b93cfaf90e614d985a95dc1393"} Oct 08 18:34:34 crc kubenswrapper[4859]: I1008 18:34:34.104960 4859 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/barbican-worker-7849cf5875-6srbt" podStartSLOduration=6.121182277 podStartE2EDuration="8.104937923s" podCreationTimestamp="2025-10-08 18:34:26 +0000 UTC" firstStartedPulling="2025-10-08 18:34:27.664451754 +0000 UTC m=+1037.911291133" lastFinishedPulling="2025-10-08 18:34:29.6482074 +0000 UTC m=+1039.895046779" observedRunningTime="2025-10-08 18:34:32.077228497 +0000 UTC m=+1042.324067896" watchObservedRunningTime="2025-10-08 18:34:34.104937923 +0000 UTC m=+1044.351777312" Oct 08 18:34:34 crc kubenswrapper[4859]: I1008 18:34:34.275005 4859 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/horizon-7d58687c8-fhctx" Oct 08 18:34:34 crc kubenswrapper[4859]: I1008 18:34:34.402639 4859 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/horizon-6f569d5894-mx8v5" Oct 08 18:34:35 crc kubenswrapper[4859]: I1008 18:34:35.088801 4859 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"7ad402c0-68c6-4c69-b1ad-0fd0d5c69dff","Type":"ContainerStarted","Data":"e5ecb40eb581178fdd97b35c9692c88e1ad16ec8e8acf8286e1cf25aa9b4268c"} Oct 08 18:34:35 crc kubenswrapper[4859]: I1008 18:34:35.512577 4859 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-db-sync-jqwrh" Oct 08 18:34:35 crc kubenswrapper[4859]: I1008 18:34:35.594664 4859 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/neutron-7755787cf6-5gvlv" Oct 08 18:34:35 crc kubenswrapper[4859]: I1008 18:34:35.649526 4859 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/720db6c0-f18f-429b-9589-fa503d212139-scripts\") pod \"720db6c0-f18f-429b-9589-fa503d212139\" (UID: \"720db6c0-f18f-429b-9589-fa503d212139\") " Oct 08 18:34:35 crc kubenswrapper[4859]: I1008 18:34:35.649615 4859 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/720db6c0-f18f-429b-9589-fa503d212139-db-sync-config-data\") pod \"720db6c0-f18f-429b-9589-fa503d212139\" (UID: \"720db6c0-f18f-429b-9589-fa503d212139\") " Oct 08 18:34:35 crc kubenswrapper[4859]: I1008 18:34:35.649711 4859 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/720db6c0-f18f-429b-9589-fa503d212139-combined-ca-bundle\") pod \"720db6c0-f18f-429b-9589-fa503d212139\" (UID: \"720db6c0-f18f-429b-9589-fa503d212139\") " Oct 08 18:34:35 crc kubenswrapper[4859]: I1008 18:34:35.649772 4859 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-bdbmw\" (UniqueName: \"kubernetes.io/projected/720db6c0-f18f-429b-9589-fa503d212139-kube-api-access-bdbmw\") pod \"720db6c0-f18f-429b-9589-fa503d212139\" (UID: \"720db6c0-f18f-429b-9589-fa503d212139\") " Oct 08 18:34:35 crc kubenswrapper[4859]: I1008 18:34:35.649835 4859 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/720db6c0-f18f-429b-9589-fa503d212139-etc-machine-id\") pod \"720db6c0-f18f-429b-9589-fa503d212139\" (UID: \"720db6c0-f18f-429b-9589-fa503d212139\") " Oct 08 18:34:35 crc kubenswrapper[4859]: I1008 18:34:35.649953 4859 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/720db6c0-f18f-429b-9589-fa503d212139-config-data\") pod \"720db6c0-f18f-429b-9589-fa503d212139\" (UID: \"720db6c0-f18f-429b-9589-fa503d212139\") " Oct 08 18:34:35 crc kubenswrapper[4859]: I1008 18:34:35.650014 4859 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/720db6c0-f18f-429b-9589-fa503d212139-etc-machine-id" (OuterVolumeSpecName: "etc-machine-id") pod "720db6c0-f18f-429b-9589-fa503d212139" (UID: "720db6c0-f18f-429b-9589-fa503d212139"). InnerVolumeSpecName "etc-machine-id". PluginName "kubernetes.io/host-path", VolumeGidValue "" Oct 08 18:34:35 crc kubenswrapper[4859]: I1008 18:34:35.651083 4859 reconciler_common.go:293] "Volume detached for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/720db6c0-f18f-429b-9589-fa503d212139-etc-machine-id\") on node \"crc\" DevicePath \"\"" Oct 08 18:34:35 crc kubenswrapper[4859]: I1008 18:34:35.658824 4859 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/720db6c0-f18f-429b-9589-fa503d212139-scripts" (OuterVolumeSpecName: "scripts") pod "720db6c0-f18f-429b-9589-fa503d212139" (UID: "720db6c0-f18f-429b-9589-fa503d212139"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 08 18:34:35 crc kubenswrapper[4859]: I1008 18:34:35.659432 4859 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/720db6c0-f18f-429b-9589-fa503d212139-kube-api-access-bdbmw" (OuterVolumeSpecName: "kube-api-access-bdbmw") pod "720db6c0-f18f-429b-9589-fa503d212139" (UID: "720db6c0-f18f-429b-9589-fa503d212139"). InnerVolumeSpecName "kube-api-access-bdbmw". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 08 18:34:35 crc kubenswrapper[4859]: I1008 18:34:35.664809 4859 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/720db6c0-f18f-429b-9589-fa503d212139-db-sync-config-data" (OuterVolumeSpecName: "db-sync-config-data") pod "720db6c0-f18f-429b-9589-fa503d212139" (UID: "720db6c0-f18f-429b-9589-fa503d212139"). InnerVolumeSpecName "db-sync-config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 08 18:34:35 crc kubenswrapper[4859]: I1008 18:34:35.693389 4859 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/720db6c0-f18f-429b-9589-fa503d212139-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "720db6c0-f18f-429b-9589-fa503d212139" (UID: "720db6c0-f18f-429b-9589-fa503d212139"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 08 18:34:35 crc kubenswrapper[4859]: I1008 18:34:35.734787 4859 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/720db6c0-f18f-429b-9589-fa503d212139-config-data" (OuterVolumeSpecName: "config-data") pod "720db6c0-f18f-429b-9589-fa503d212139" (UID: "720db6c0-f18f-429b-9589-fa503d212139"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 08 18:34:35 crc kubenswrapper[4859]: I1008 18:34:35.754008 4859 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/720db6c0-f18f-429b-9589-fa503d212139-scripts\") on node \"crc\" DevicePath \"\"" Oct 08 18:34:35 crc kubenswrapper[4859]: I1008 18:34:35.754048 4859 reconciler_common.go:293] "Volume detached for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/720db6c0-f18f-429b-9589-fa503d212139-db-sync-config-data\") on node \"crc\" DevicePath \"\"" Oct 08 18:34:35 crc kubenswrapper[4859]: I1008 18:34:35.754063 4859 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/720db6c0-f18f-429b-9589-fa503d212139-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 08 18:34:35 crc kubenswrapper[4859]: I1008 18:34:35.754076 4859 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-bdbmw\" (UniqueName: \"kubernetes.io/projected/720db6c0-f18f-429b-9589-fa503d212139-kube-api-access-bdbmw\") on node \"crc\" DevicePath \"\"" Oct 08 18:34:35 crc kubenswrapper[4859]: I1008 18:34:35.754089 4859 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/720db6c0-f18f-429b-9589-fa503d212139-config-data\") on node \"crc\" DevicePath \"\"" Oct 08 18:34:36 crc kubenswrapper[4859]: I1008 18:34:36.073190 4859 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/horizon-7d58687c8-fhctx" Oct 08 18:34:36 crc kubenswrapper[4859]: I1008 18:34:36.099738 4859 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-db-sync-jqwrh" event={"ID":"720db6c0-f18f-429b-9589-fa503d212139","Type":"ContainerDied","Data":"6158396a7a7a7bb9ea51437d23c0b6b2e9b2ef400482d5e49cc2cee88629328c"} Oct 08 18:34:36 crc kubenswrapper[4859]: I1008 18:34:36.099782 4859 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="6158396a7a7a7bb9ea51437d23c0b6b2e9b2ef400482d5e49cc2cee88629328c" Oct 08 18:34:36 crc kubenswrapper[4859]: I1008 18:34:36.099818 4859 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-db-sync-jqwrh" Oct 08 18:34:36 crc kubenswrapper[4859]: I1008 18:34:36.106193 4859 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"7ad402c0-68c6-4c69-b1ad-0fd0d5c69dff","Type":"ContainerStarted","Data":"ec6a14d0c9861094bc0a93ff618867a826b6ddf9021a5b4ed58abcf176899896"} Oct 08 18:34:36 crc kubenswrapper[4859]: I1008 18:34:36.106514 4859 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ceilometer-0" Oct 08 18:34:36 crc kubenswrapper[4859]: I1008 18:34:36.280731 4859 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/horizon-6f569d5894-mx8v5" Oct 08 18:34:36 crc kubenswrapper[4859]: I1008 18:34:36.305622 4859 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ceilometer-0" podStartSLOduration=1.6305956830000001 podStartE2EDuration="5.30559856s" podCreationTimestamp="2025-10-08 18:34:31 +0000 UTC" firstStartedPulling="2025-10-08 18:34:32.011951957 +0000 UTC m=+1042.258791336" lastFinishedPulling="2025-10-08 18:34:35.686954834 +0000 UTC m=+1045.933794213" observedRunningTime="2025-10-08 18:34:36.191875098 +0000 UTC m=+1046.438714497" watchObservedRunningTime="2025-10-08 18:34:36.30559856 +0000 UTC m=+1046.552437949" Oct 08 18:34:36 crc kubenswrapper[4859]: I1008 18:34:36.340120 4859 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/horizon-7d58687c8-fhctx"] Oct 08 18:34:36 crc kubenswrapper[4859]: I1008 18:34:36.340577 4859 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/horizon-7d58687c8-fhctx" podUID="1a92487e-3309-41c4-8c82-cd57f517ab4b" containerName="horizon-log" containerID="cri-o://fb897bd998eaff60715c21b1cde92b421d7bf9ae9f4afad9a1ed7b20f7cfd7b3" gracePeriod=30 Oct 08 18:34:36 crc kubenswrapper[4859]: I1008 18:34:36.340616 4859 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/horizon-7d58687c8-fhctx" podUID="1a92487e-3309-41c4-8c82-cd57f517ab4b" containerName="horizon" containerID="cri-o://648c622ce9eb0bdf3d9c43d8313db6636c307abe8d840bde30cf40e0039e99c7" gracePeriod=30 Oct 08 18:34:36 crc kubenswrapper[4859]: I1008 18:34:36.495252 4859 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/cinder-scheduler-0"] Oct 08 18:34:36 crc kubenswrapper[4859]: E1008 18:34:36.496151 4859 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="720db6c0-f18f-429b-9589-fa503d212139" containerName="cinder-db-sync" Oct 08 18:34:36 crc kubenswrapper[4859]: I1008 18:34:36.496171 4859 state_mem.go:107] "Deleted CPUSet assignment" podUID="720db6c0-f18f-429b-9589-fa503d212139" containerName="cinder-db-sync" Oct 08 18:34:36 crc kubenswrapper[4859]: I1008 18:34:36.496409 4859 memory_manager.go:354] "RemoveStaleState removing state" podUID="720db6c0-f18f-429b-9589-fa503d212139" containerName="cinder-db-sync" Oct 08 18:34:36 crc kubenswrapper[4859]: I1008 18:34:36.497363 4859 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-scheduler-0" Oct 08 18:34:36 crc kubenswrapper[4859]: I1008 18:34:36.507633 4859 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-cinder-dockercfg-gldt6" Oct 08 18:34:36 crc kubenswrapper[4859]: I1008 18:34:36.507856 4859 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-config-data" Oct 08 18:34:36 crc kubenswrapper[4859]: I1008 18:34:36.507961 4859 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-scheduler-config-data" Oct 08 18:34:36 crc kubenswrapper[4859]: I1008 18:34:36.518818 4859 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-scripts" Oct 08 18:34:36 crc kubenswrapper[4859]: I1008 18:34:36.543163 4859 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-scheduler-0"] Oct 08 18:34:36 crc kubenswrapper[4859]: I1008 18:34:36.612079 4859 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-744fd954cc-hv6xb"] Oct 08 18:34:36 crc kubenswrapper[4859]: I1008 18:34:36.612114 4859 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/085aabba-1f1f-4037-b5cf-036d5b519b02-etc-machine-id\") pod \"cinder-scheduler-0\" (UID: \"085aabba-1f1f-4037-b5cf-036d5b519b02\") " pod="openstack/cinder-scheduler-0" Oct 08 18:34:36 crc kubenswrapper[4859]: I1008 18:34:36.612461 4859 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/085aabba-1f1f-4037-b5cf-036d5b519b02-scripts\") pod \"cinder-scheduler-0\" (UID: \"085aabba-1f1f-4037-b5cf-036d5b519b02\") " pod="openstack/cinder-scheduler-0" Oct 08 18:34:36 crc kubenswrapper[4859]: I1008 18:34:36.612638 4859 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/085aabba-1f1f-4037-b5cf-036d5b519b02-combined-ca-bundle\") pod \"cinder-scheduler-0\" (UID: \"085aabba-1f1f-4037-b5cf-036d5b519b02\") " pod="openstack/cinder-scheduler-0" Oct 08 18:34:36 crc kubenswrapper[4859]: I1008 18:34:36.612781 4859 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-dqz5f\" (UniqueName: \"kubernetes.io/projected/085aabba-1f1f-4037-b5cf-036d5b519b02-kube-api-access-dqz5f\") pod \"cinder-scheduler-0\" (UID: \"085aabba-1f1f-4037-b5cf-036d5b519b02\") " pod="openstack/cinder-scheduler-0" Oct 08 18:34:36 crc kubenswrapper[4859]: I1008 18:34:36.612891 4859 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/085aabba-1f1f-4037-b5cf-036d5b519b02-config-data\") pod \"cinder-scheduler-0\" (UID: \"085aabba-1f1f-4037-b5cf-036d5b519b02\") " pod="openstack/cinder-scheduler-0" Oct 08 18:34:36 crc kubenswrapper[4859]: I1008 18:34:36.613010 4859 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/085aabba-1f1f-4037-b5cf-036d5b519b02-config-data-custom\") pod \"cinder-scheduler-0\" (UID: \"085aabba-1f1f-4037-b5cf-036d5b519b02\") " pod="openstack/cinder-scheduler-0" Oct 08 18:34:36 crc kubenswrapper[4859]: I1008 18:34:36.628735 4859 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-5f74b5f5cc-tl7pk"] Oct 08 18:34:36 crc kubenswrapper[4859]: I1008 18:34:36.634290 4859 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-744fd954cc-hv6xb" podUID="8252c83d-1fdd-4d33-87a0-7a797251a0b4" containerName="dnsmasq-dns" containerID="cri-o://080d2b87a50a3198eeb74c3989be6bedec81fb60d37a7c683e9630dc72474e73" gracePeriod=10 Oct 08 18:34:36 crc kubenswrapper[4859]: I1008 18:34:36.637129 4859 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-5f74b5f5cc-tl7pk" Oct 08 18:34:36 crc kubenswrapper[4859]: I1008 18:34:36.638309 4859 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-744fd954cc-hv6xb" Oct 08 18:34:36 crc kubenswrapper[4859]: I1008 18:34:36.639955 4859 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-5f74b5f5cc-tl7pk"] Oct 08 18:34:36 crc kubenswrapper[4859]: I1008 18:34:36.716707 4859 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/085aabba-1f1f-4037-b5cf-036d5b519b02-combined-ca-bundle\") pod \"cinder-scheduler-0\" (UID: \"085aabba-1f1f-4037-b5cf-036d5b519b02\") " pod="openstack/cinder-scheduler-0" Oct 08 18:34:36 crc kubenswrapper[4859]: I1008 18:34:36.717040 4859 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-dqz5f\" (UniqueName: \"kubernetes.io/projected/085aabba-1f1f-4037-b5cf-036d5b519b02-kube-api-access-dqz5f\") pod \"cinder-scheduler-0\" (UID: \"085aabba-1f1f-4037-b5cf-036d5b519b02\") " pod="openstack/cinder-scheduler-0" Oct 08 18:34:36 crc kubenswrapper[4859]: I1008 18:34:36.717122 4859 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/085aabba-1f1f-4037-b5cf-036d5b519b02-config-data\") pod \"cinder-scheduler-0\" (UID: \"085aabba-1f1f-4037-b5cf-036d5b519b02\") " pod="openstack/cinder-scheduler-0" Oct 08 18:34:36 crc kubenswrapper[4859]: I1008 18:34:36.717782 4859 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/085aabba-1f1f-4037-b5cf-036d5b519b02-config-data-custom\") pod \"cinder-scheduler-0\" (UID: \"085aabba-1f1f-4037-b5cf-036d5b519b02\") " pod="openstack/cinder-scheduler-0" Oct 08 18:34:36 crc kubenswrapper[4859]: I1008 18:34:36.717895 4859 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/085aabba-1f1f-4037-b5cf-036d5b519b02-etc-machine-id\") pod \"cinder-scheduler-0\" (UID: \"085aabba-1f1f-4037-b5cf-036d5b519b02\") " pod="openstack/cinder-scheduler-0" Oct 08 18:34:36 crc kubenswrapper[4859]: I1008 18:34:36.717971 4859 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/085aabba-1f1f-4037-b5cf-036d5b519b02-scripts\") pod \"cinder-scheduler-0\" (UID: \"085aabba-1f1f-4037-b5cf-036d5b519b02\") " pod="openstack/cinder-scheduler-0" Oct 08 18:34:36 crc kubenswrapper[4859]: I1008 18:34:36.735112 4859 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/085aabba-1f1f-4037-b5cf-036d5b519b02-etc-machine-id\") pod \"cinder-scheduler-0\" (UID: \"085aabba-1f1f-4037-b5cf-036d5b519b02\") " pod="openstack/cinder-scheduler-0" Oct 08 18:34:36 crc kubenswrapper[4859]: I1008 18:34:36.747152 4859 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/085aabba-1f1f-4037-b5cf-036d5b519b02-config-data-custom\") pod \"cinder-scheduler-0\" (UID: \"085aabba-1f1f-4037-b5cf-036d5b519b02\") " pod="openstack/cinder-scheduler-0" Oct 08 18:34:36 crc kubenswrapper[4859]: I1008 18:34:36.747536 4859 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/085aabba-1f1f-4037-b5cf-036d5b519b02-combined-ca-bundle\") pod \"cinder-scheduler-0\" (UID: \"085aabba-1f1f-4037-b5cf-036d5b519b02\") " pod="openstack/cinder-scheduler-0" Oct 08 18:34:36 crc kubenswrapper[4859]: I1008 18:34:36.751281 4859 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/085aabba-1f1f-4037-b5cf-036d5b519b02-config-data\") pod \"cinder-scheduler-0\" (UID: \"085aabba-1f1f-4037-b5cf-036d5b519b02\") " pod="openstack/cinder-scheduler-0" Oct 08 18:34:36 crc kubenswrapper[4859]: I1008 18:34:36.753625 4859 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/085aabba-1f1f-4037-b5cf-036d5b519b02-scripts\") pod \"cinder-scheduler-0\" (UID: \"085aabba-1f1f-4037-b5cf-036d5b519b02\") " pod="openstack/cinder-scheduler-0" Oct 08 18:34:36 crc kubenswrapper[4859]: I1008 18:34:36.778271 4859 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/cinder-api-0"] Oct 08 18:34:36 crc kubenswrapper[4859]: I1008 18:34:36.780359 4859 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-api-0" Oct 08 18:34:36 crc kubenswrapper[4859]: I1008 18:34:36.780451 4859 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-dqz5f\" (UniqueName: \"kubernetes.io/projected/085aabba-1f1f-4037-b5cf-036d5b519b02-kube-api-access-dqz5f\") pod \"cinder-scheduler-0\" (UID: \"085aabba-1f1f-4037-b5cf-036d5b519b02\") " pod="openstack/cinder-scheduler-0" Oct 08 18:34:36 crc kubenswrapper[4859]: I1008 18:34:36.783557 4859 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-api-config-data" Oct 08 18:34:36 crc kubenswrapper[4859]: I1008 18:34:36.823533 4859 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-api-0"] Oct 08 18:34:36 crc kubenswrapper[4859]: I1008 18:34:36.829480 4859 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/a03169fb-e503-427f-bafe-2f8767a23482-ovsdbserver-nb\") pod \"dnsmasq-dns-5f74b5f5cc-tl7pk\" (UID: \"a03169fb-e503-427f-bafe-2f8767a23482\") " pod="openstack/dnsmasq-dns-5f74b5f5cc-tl7pk" Oct 08 18:34:36 crc kubenswrapper[4859]: I1008 18:34:36.829550 4859 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/a03169fb-e503-427f-bafe-2f8767a23482-config\") pod \"dnsmasq-dns-5f74b5f5cc-tl7pk\" (UID: \"a03169fb-e503-427f-bafe-2f8767a23482\") " pod="openstack/dnsmasq-dns-5f74b5f5cc-tl7pk" Oct 08 18:34:36 crc kubenswrapper[4859]: I1008 18:34:36.829578 4859 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/a03169fb-e503-427f-bafe-2f8767a23482-dns-swift-storage-0\") pod \"dnsmasq-dns-5f74b5f5cc-tl7pk\" (UID: \"a03169fb-e503-427f-bafe-2f8767a23482\") " pod="openstack/dnsmasq-dns-5f74b5f5cc-tl7pk" Oct 08 18:34:36 crc kubenswrapper[4859]: I1008 18:34:36.829669 4859 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/a03169fb-e503-427f-bafe-2f8767a23482-dns-svc\") pod \"dnsmasq-dns-5f74b5f5cc-tl7pk\" (UID: \"a03169fb-e503-427f-bafe-2f8767a23482\") " pod="openstack/dnsmasq-dns-5f74b5f5cc-tl7pk" Oct 08 18:34:36 crc kubenswrapper[4859]: I1008 18:34:36.829776 4859 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/a03169fb-e503-427f-bafe-2f8767a23482-ovsdbserver-sb\") pod \"dnsmasq-dns-5f74b5f5cc-tl7pk\" (UID: \"a03169fb-e503-427f-bafe-2f8767a23482\") " pod="openstack/dnsmasq-dns-5f74b5f5cc-tl7pk" Oct 08 18:34:36 crc kubenswrapper[4859]: I1008 18:34:36.829843 4859 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-77d9b\" (UniqueName: \"kubernetes.io/projected/a03169fb-e503-427f-bafe-2f8767a23482-kube-api-access-77d9b\") pod \"dnsmasq-dns-5f74b5f5cc-tl7pk\" (UID: \"a03169fb-e503-427f-bafe-2f8767a23482\") " pod="openstack/dnsmasq-dns-5f74b5f5cc-tl7pk" Oct 08 18:34:36 crc kubenswrapper[4859]: I1008 18:34:36.839240 4859 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-scheduler-0" Oct 08 18:34:36 crc kubenswrapper[4859]: I1008 18:34:36.932131 4859 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/a03169fb-e503-427f-bafe-2f8767a23482-ovsdbserver-sb\") pod \"dnsmasq-dns-5f74b5f5cc-tl7pk\" (UID: \"a03169fb-e503-427f-bafe-2f8767a23482\") " pod="openstack/dnsmasq-dns-5f74b5f5cc-tl7pk" Oct 08 18:34:36 crc kubenswrapper[4859]: I1008 18:34:36.932219 4859 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-77d9b\" (UniqueName: \"kubernetes.io/projected/a03169fb-e503-427f-bafe-2f8767a23482-kube-api-access-77d9b\") pod \"dnsmasq-dns-5f74b5f5cc-tl7pk\" (UID: \"a03169fb-e503-427f-bafe-2f8767a23482\") " pod="openstack/dnsmasq-dns-5f74b5f5cc-tl7pk" Oct 08 18:34:36 crc kubenswrapper[4859]: I1008 18:34:36.932259 4859 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/65e6bdd0-9ffe-48d7-9b02-444ce9b204dd-config-data\") pod \"cinder-api-0\" (UID: \"65e6bdd0-9ffe-48d7-9b02-444ce9b204dd\") " pod="openstack/cinder-api-0" Oct 08 18:34:36 crc kubenswrapper[4859]: I1008 18:34:36.932294 4859 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/65e6bdd0-9ffe-48d7-9b02-444ce9b204dd-combined-ca-bundle\") pod \"cinder-api-0\" (UID: \"65e6bdd0-9ffe-48d7-9b02-444ce9b204dd\") " pod="openstack/cinder-api-0" Oct 08 18:34:36 crc kubenswrapper[4859]: I1008 18:34:36.932412 4859 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/a03169fb-e503-427f-bafe-2f8767a23482-ovsdbserver-nb\") pod \"dnsmasq-dns-5f74b5f5cc-tl7pk\" (UID: \"a03169fb-e503-427f-bafe-2f8767a23482\") " pod="openstack/dnsmasq-dns-5f74b5f5cc-tl7pk" Oct 08 18:34:36 crc kubenswrapper[4859]: I1008 18:34:36.932455 4859 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/a03169fb-e503-427f-bafe-2f8767a23482-config\") pod \"dnsmasq-dns-5f74b5f5cc-tl7pk\" (UID: \"a03169fb-e503-427f-bafe-2f8767a23482\") " pod="openstack/dnsmasq-dns-5f74b5f5cc-tl7pk" Oct 08 18:34:36 crc kubenswrapper[4859]: I1008 18:34:36.932481 4859 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/a03169fb-e503-427f-bafe-2f8767a23482-dns-swift-storage-0\") pod \"dnsmasq-dns-5f74b5f5cc-tl7pk\" (UID: \"a03169fb-e503-427f-bafe-2f8767a23482\") " pod="openstack/dnsmasq-dns-5f74b5f5cc-tl7pk" Oct 08 18:34:36 crc kubenswrapper[4859]: I1008 18:34:36.932722 4859 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-fhgh7\" (UniqueName: \"kubernetes.io/projected/65e6bdd0-9ffe-48d7-9b02-444ce9b204dd-kube-api-access-fhgh7\") pod \"cinder-api-0\" (UID: \"65e6bdd0-9ffe-48d7-9b02-444ce9b204dd\") " pod="openstack/cinder-api-0" Oct 08 18:34:36 crc kubenswrapper[4859]: I1008 18:34:36.932759 4859 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/65e6bdd0-9ffe-48d7-9b02-444ce9b204dd-etc-machine-id\") pod \"cinder-api-0\" (UID: \"65e6bdd0-9ffe-48d7-9b02-444ce9b204dd\") " pod="openstack/cinder-api-0" Oct 08 18:34:36 crc kubenswrapper[4859]: I1008 18:34:36.932797 4859 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/65e6bdd0-9ffe-48d7-9b02-444ce9b204dd-scripts\") pod \"cinder-api-0\" (UID: \"65e6bdd0-9ffe-48d7-9b02-444ce9b204dd\") " pod="openstack/cinder-api-0" Oct 08 18:34:36 crc kubenswrapper[4859]: I1008 18:34:36.932826 4859 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/65e6bdd0-9ffe-48d7-9b02-444ce9b204dd-config-data-custom\") pod \"cinder-api-0\" (UID: \"65e6bdd0-9ffe-48d7-9b02-444ce9b204dd\") " pod="openstack/cinder-api-0" Oct 08 18:34:36 crc kubenswrapper[4859]: I1008 18:34:36.932858 4859 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/65e6bdd0-9ffe-48d7-9b02-444ce9b204dd-logs\") pod \"cinder-api-0\" (UID: \"65e6bdd0-9ffe-48d7-9b02-444ce9b204dd\") " pod="openstack/cinder-api-0" Oct 08 18:34:36 crc kubenswrapper[4859]: I1008 18:34:36.932888 4859 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/a03169fb-e503-427f-bafe-2f8767a23482-dns-svc\") pod \"dnsmasq-dns-5f74b5f5cc-tl7pk\" (UID: \"a03169fb-e503-427f-bafe-2f8767a23482\") " pod="openstack/dnsmasq-dns-5f74b5f5cc-tl7pk" Oct 08 18:34:36 crc kubenswrapper[4859]: I1008 18:34:36.938560 4859 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/a03169fb-e503-427f-bafe-2f8767a23482-dns-svc\") pod \"dnsmasq-dns-5f74b5f5cc-tl7pk\" (UID: \"a03169fb-e503-427f-bafe-2f8767a23482\") " pod="openstack/dnsmasq-dns-5f74b5f5cc-tl7pk" Oct 08 18:34:36 crc kubenswrapper[4859]: I1008 18:34:36.938866 4859 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/a03169fb-e503-427f-bafe-2f8767a23482-config\") pod \"dnsmasq-dns-5f74b5f5cc-tl7pk\" (UID: \"a03169fb-e503-427f-bafe-2f8767a23482\") " pod="openstack/dnsmasq-dns-5f74b5f5cc-tl7pk" Oct 08 18:34:36 crc kubenswrapper[4859]: I1008 18:34:36.940118 4859 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/a03169fb-e503-427f-bafe-2f8767a23482-dns-swift-storage-0\") pod \"dnsmasq-dns-5f74b5f5cc-tl7pk\" (UID: \"a03169fb-e503-427f-bafe-2f8767a23482\") " pod="openstack/dnsmasq-dns-5f74b5f5cc-tl7pk" Oct 08 18:34:36 crc kubenswrapper[4859]: I1008 18:34:36.940388 4859 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/a03169fb-e503-427f-bafe-2f8767a23482-ovsdbserver-sb\") pod \"dnsmasq-dns-5f74b5f5cc-tl7pk\" (UID: \"a03169fb-e503-427f-bafe-2f8767a23482\") " pod="openstack/dnsmasq-dns-5f74b5f5cc-tl7pk" Oct 08 18:34:36 crc kubenswrapper[4859]: I1008 18:34:36.940825 4859 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/a03169fb-e503-427f-bafe-2f8767a23482-ovsdbserver-nb\") pod \"dnsmasq-dns-5f74b5f5cc-tl7pk\" (UID: \"a03169fb-e503-427f-bafe-2f8767a23482\") " pod="openstack/dnsmasq-dns-5f74b5f5cc-tl7pk" Oct 08 18:34:36 crc kubenswrapper[4859]: I1008 18:34:36.969063 4859 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-77d9b\" (UniqueName: \"kubernetes.io/projected/a03169fb-e503-427f-bafe-2f8767a23482-kube-api-access-77d9b\") pod \"dnsmasq-dns-5f74b5f5cc-tl7pk\" (UID: \"a03169fb-e503-427f-bafe-2f8767a23482\") " pod="openstack/dnsmasq-dns-5f74b5f5cc-tl7pk" Oct 08 18:34:37 crc kubenswrapper[4859]: I1008 18:34:37.035653 4859 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/65e6bdd0-9ffe-48d7-9b02-444ce9b204dd-config-data\") pod \"cinder-api-0\" (UID: \"65e6bdd0-9ffe-48d7-9b02-444ce9b204dd\") " pod="openstack/cinder-api-0" Oct 08 18:34:37 crc kubenswrapper[4859]: I1008 18:34:37.035825 4859 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/65e6bdd0-9ffe-48d7-9b02-444ce9b204dd-combined-ca-bundle\") pod \"cinder-api-0\" (UID: \"65e6bdd0-9ffe-48d7-9b02-444ce9b204dd\") " pod="openstack/cinder-api-0" Oct 08 18:34:37 crc kubenswrapper[4859]: I1008 18:34:37.035958 4859 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-fhgh7\" (UniqueName: \"kubernetes.io/projected/65e6bdd0-9ffe-48d7-9b02-444ce9b204dd-kube-api-access-fhgh7\") pod \"cinder-api-0\" (UID: \"65e6bdd0-9ffe-48d7-9b02-444ce9b204dd\") " pod="openstack/cinder-api-0" Oct 08 18:34:37 crc kubenswrapper[4859]: I1008 18:34:37.035997 4859 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/65e6bdd0-9ffe-48d7-9b02-444ce9b204dd-etc-machine-id\") pod \"cinder-api-0\" (UID: \"65e6bdd0-9ffe-48d7-9b02-444ce9b204dd\") " pod="openstack/cinder-api-0" Oct 08 18:34:37 crc kubenswrapper[4859]: I1008 18:34:37.036060 4859 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/65e6bdd0-9ffe-48d7-9b02-444ce9b204dd-scripts\") pod \"cinder-api-0\" (UID: \"65e6bdd0-9ffe-48d7-9b02-444ce9b204dd\") " pod="openstack/cinder-api-0" Oct 08 18:34:37 crc kubenswrapper[4859]: I1008 18:34:37.036089 4859 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/65e6bdd0-9ffe-48d7-9b02-444ce9b204dd-config-data-custom\") pod \"cinder-api-0\" (UID: \"65e6bdd0-9ffe-48d7-9b02-444ce9b204dd\") " pod="openstack/cinder-api-0" Oct 08 18:34:37 crc kubenswrapper[4859]: I1008 18:34:37.036126 4859 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/65e6bdd0-9ffe-48d7-9b02-444ce9b204dd-logs\") pod \"cinder-api-0\" (UID: \"65e6bdd0-9ffe-48d7-9b02-444ce9b204dd\") " pod="openstack/cinder-api-0" Oct 08 18:34:37 crc kubenswrapper[4859]: I1008 18:34:37.036805 4859 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/65e6bdd0-9ffe-48d7-9b02-444ce9b204dd-logs\") pod \"cinder-api-0\" (UID: \"65e6bdd0-9ffe-48d7-9b02-444ce9b204dd\") " pod="openstack/cinder-api-0" Oct 08 18:34:37 crc kubenswrapper[4859]: I1008 18:34:37.038477 4859 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/65e6bdd0-9ffe-48d7-9b02-444ce9b204dd-etc-machine-id\") pod \"cinder-api-0\" (UID: \"65e6bdd0-9ffe-48d7-9b02-444ce9b204dd\") " pod="openstack/cinder-api-0" Oct 08 18:34:37 crc kubenswrapper[4859]: I1008 18:34:37.040914 4859 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/65e6bdd0-9ffe-48d7-9b02-444ce9b204dd-config-data\") pod \"cinder-api-0\" (UID: \"65e6bdd0-9ffe-48d7-9b02-444ce9b204dd\") " pod="openstack/cinder-api-0" Oct 08 18:34:37 crc kubenswrapper[4859]: I1008 18:34:37.043191 4859 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/65e6bdd0-9ffe-48d7-9b02-444ce9b204dd-combined-ca-bundle\") pod \"cinder-api-0\" (UID: \"65e6bdd0-9ffe-48d7-9b02-444ce9b204dd\") " pod="openstack/cinder-api-0" Oct 08 18:34:37 crc kubenswrapper[4859]: I1008 18:34:37.043301 4859 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/65e6bdd0-9ffe-48d7-9b02-444ce9b204dd-scripts\") pod \"cinder-api-0\" (UID: \"65e6bdd0-9ffe-48d7-9b02-444ce9b204dd\") " pod="openstack/cinder-api-0" Oct 08 18:34:37 crc kubenswrapper[4859]: I1008 18:34:37.044471 4859 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-5f74b5f5cc-tl7pk" Oct 08 18:34:37 crc kubenswrapper[4859]: I1008 18:34:37.045456 4859 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/65e6bdd0-9ffe-48d7-9b02-444ce9b204dd-config-data-custom\") pod \"cinder-api-0\" (UID: \"65e6bdd0-9ffe-48d7-9b02-444ce9b204dd\") " pod="openstack/cinder-api-0" Oct 08 18:34:37 crc kubenswrapper[4859]: I1008 18:34:37.059368 4859 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-fhgh7\" (UniqueName: \"kubernetes.io/projected/65e6bdd0-9ffe-48d7-9b02-444ce9b204dd-kube-api-access-fhgh7\") pod \"cinder-api-0\" (UID: \"65e6bdd0-9ffe-48d7-9b02-444ce9b204dd\") " pod="openstack/cinder-api-0" Oct 08 18:34:37 crc kubenswrapper[4859]: I1008 18:34:37.156699 4859 generic.go:334] "Generic (PLEG): container finished" podID="8252c83d-1fdd-4d33-87a0-7a797251a0b4" containerID="080d2b87a50a3198eeb74c3989be6bedec81fb60d37a7c683e9630dc72474e73" exitCode=0 Oct 08 18:34:37 crc kubenswrapper[4859]: I1008 18:34:37.156748 4859 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-744fd954cc-hv6xb" event={"ID":"8252c83d-1fdd-4d33-87a0-7a797251a0b4","Type":"ContainerDied","Data":"080d2b87a50a3198eeb74c3989be6bedec81fb60d37a7c683e9630dc72474e73"} Oct 08 18:34:37 crc kubenswrapper[4859]: I1008 18:34:37.170093 4859 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-api-0" Oct 08 18:34:37 crc kubenswrapper[4859]: I1008 18:34:37.234317 4859 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-744fd954cc-hv6xb" Oct 08 18:34:37 crc kubenswrapper[4859]: I1008 18:34:37.346453 4859 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/8252c83d-1fdd-4d33-87a0-7a797251a0b4-dns-svc\") pod \"8252c83d-1fdd-4d33-87a0-7a797251a0b4\" (UID: \"8252c83d-1fdd-4d33-87a0-7a797251a0b4\") " Oct 08 18:34:37 crc kubenswrapper[4859]: I1008 18:34:37.346557 4859 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/8252c83d-1fdd-4d33-87a0-7a797251a0b4-dns-swift-storage-0\") pod \"8252c83d-1fdd-4d33-87a0-7a797251a0b4\" (UID: \"8252c83d-1fdd-4d33-87a0-7a797251a0b4\") " Oct 08 18:34:37 crc kubenswrapper[4859]: I1008 18:34:37.346619 4859 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-jkx9v\" (UniqueName: \"kubernetes.io/projected/8252c83d-1fdd-4d33-87a0-7a797251a0b4-kube-api-access-jkx9v\") pod \"8252c83d-1fdd-4d33-87a0-7a797251a0b4\" (UID: \"8252c83d-1fdd-4d33-87a0-7a797251a0b4\") " Oct 08 18:34:37 crc kubenswrapper[4859]: I1008 18:34:37.346662 4859 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/8252c83d-1fdd-4d33-87a0-7a797251a0b4-ovsdbserver-nb\") pod \"8252c83d-1fdd-4d33-87a0-7a797251a0b4\" (UID: \"8252c83d-1fdd-4d33-87a0-7a797251a0b4\") " Oct 08 18:34:37 crc kubenswrapper[4859]: I1008 18:34:37.346722 4859 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/8252c83d-1fdd-4d33-87a0-7a797251a0b4-config\") pod \"8252c83d-1fdd-4d33-87a0-7a797251a0b4\" (UID: \"8252c83d-1fdd-4d33-87a0-7a797251a0b4\") " Oct 08 18:34:37 crc kubenswrapper[4859]: I1008 18:34:37.346852 4859 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/8252c83d-1fdd-4d33-87a0-7a797251a0b4-ovsdbserver-sb\") pod \"8252c83d-1fdd-4d33-87a0-7a797251a0b4\" (UID: \"8252c83d-1fdd-4d33-87a0-7a797251a0b4\") " Oct 08 18:34:37 crc kubenswrapper[4859]: I1008 18:34:37.365081 4859 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8252c83d-1fdd-4d33-87a0-7a797251a0b4-kube-api-access-jkx9v" (OuterVolumeSpecName: "kube-api-access-jkx9v") pod "8252c83d-1fdd-4d33-87a0-7a797251a0b4" (UID: "8252c83d-1fdd-4d33-87a0-7a797251a0b4"). InnerVolumeSpecName "kube-api-access-jkx9v". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 08 18:34:37 crc kubenswrapper[4859]: I1008 18:34:37.410626 4859 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/8252c83d-1fdd-4d33-87a0-7a797251a0b4-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "8252c83d-1fdd-4d33-87a0-7a797251a0b4" (UID: "8252c83d-1fdd-4d33-87a0-7a797251a0b4"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 08 18:34:37 crc kubenswrapper[4859]: I1008 18:34:37.465628 4859 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/8252c83d-1fdd-4d33-87a0-7a797251a0b4-dns-swift-storage-0" (OuterVolumeSpecName: "dns-swift-storage-0") pod "8252c83d-1fdd-4d33-87a0-7a797251a0b4" (UID: "8252c83d-1fdd-4d33-87a0-7a797251a0b4"). InnerVolumeSpecName "dns-swift-storage-0". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 08 18:34:37 crc kubenswrapper[4859]: I1008 18:34:37.467081 4859 reconciler_common.go:293] "Volume detached for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/8252c83d-1fdd-4d33-87a0-7a797251a0b4-dns-swift-storage-0\") on node \"crc\" DevicePath \"\"" Oct 08 18:34:37 crc kubenswrapper[4859]: I1008 18:34:37.467135 4859 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-jkx9v\" (UniqueName: \"kubernetes.io/projected/8252c83d-1fdd-4d33-87a0-7a797251a0b4-kube-api-access-jkx9v\") on node \"crc\" DevicePath \"\"" Oct 08 18:34:37 crc kubenswrapper[4859]: I1008 18:34:37.467149 4859 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/8252c83d-1fdd-4d33-87a0-7a797251a0b4-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Oct 08 18:34:37 crc kubenswrapper[4859]: I1008 18:34:37.479272 4859 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/8252c83d-1fdd-4d33-87a0-7a797251a0b4-config" (OuterVolumeSpecName: "config") pod "8252c83d-1fdd-4d33-87a0-7a797251a0b4" (UID: "8252c83d-1fdd-4d33-87a0-7a797251a0b4"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 08 18:34:37 crc kubenswrapper[4859]: I1008 18:34:37.508252 4859 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/8252c83d-1fdd-4d33-87a0-7a797251a0b4-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "8252c83d-1fdd-4d33-87a0-7a797251a0b4" (UID: "8252c83d-1fdd-4d33-87a0-7a797251a0b4"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 08 18:34:37 crc kubenswrapper[4859]: I1008 18:34:37.529454 4859 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/8252c83d-1fdd-4d33-87a0-7a797251a0b4-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "8252c83d-1fdd-4d33-87a0-7a797251a0b4" (UID: "8252c83d-1fdd-4d33-87a0-7a797251a0b4"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 08 18:34:37 crc kubenswrapper[4859]: I1008 18:34:37.543052 4859 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-scheduler-0"] Oct 08 18:34:37 crc kubenswrapper[4859]: I1008 18:34:37.575392 4859 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/8252c83d-1fdd-4d33-87a0-7a797251a0b4-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Oct 08 18:34:37 crc kubenswrapper[4859]: I1008 18:34:37.575419 4859 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/8252c83d-1fdd-4d33-87a0-7a797251a0b4-dns-svc\") on node \"crc\" DevicePath \"\"" Oct 08 18:34:37 crc kubenswrapper[4859]: I1008 18:34:37.575429 4859 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/8252c83d-1fdd-4d33-87a0-7a797251a0b4-config\") on node \"crc\" DevicePath \"\"" Oct 08 18:34:37 crc kubenswrapper[4859]: I1008 18:34:37.727968 4859 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-5f74b5f5cc-tl7pk"] Oct 08 18:34:37 crc kubenswrapper[4859]: I1008 18:34:37.869454 4859 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-api-0"] Oct 08 18:34:38 crc kubenswrapper[4859]: I1008 18:34:38.071672 4859 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/barbican-api-588f98c7dd-c7l8h" Oct 08 18:34:38 crc kubenswrapper[4859]: I1008 18:34:38.164658 4859 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/neutron-5f9b9b6f45-wr5fx" Oct 08 18:34:38 crc kubenswrapper[4859]: I1008 18:34:38.187347 4859 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-scheduler-0" event={"ID":"085aabba-1f1f-4037-b5cf-036d5b519b02","Type":"ContainerStarted","Data":"02c84458f7fd46ebb947c2b80a0b6fbcd7f6f1abf7683c0e94eb6f254a0fa634"} Oct 08 18:34:38 crc kubenswrapper[4859]: I1008 18:34:38.202141 4859 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-api-0" event={"ID":"65e6bdd0-9ffe-48d7-9b02-444ce9b204dd","Type":"ContainerStarted","Data":"1c6c50170c1f93a8ab17006c51930faa05132506dadfdda58eb6611fde6c2efe"} Oct 08 18:34:38 crc kubenswrapper[4859]: I1008 18:34:38.221993 4859 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-744fd954cc-hv6xb" event={"ID":"8252c83d-1fdd-4d33-87a0-7a797251a0b4","Type":"ContainerDied","Data":"a20343aff730d6979c82829bba208a71d6c9b51409c7e6e6318db3ba888fc8b8"} Oct 08 18:34:38 crc kubenswrapper[4859]: I1008 18:34:38.222049 4859 scope.go:117] "RemoveContainer" containerID="080d2b87a50a3198eeb74c3989be6bedec81fb60d37a7c683e9630dc72474e73" Oct 08 18:34:38 crc kubenswrapper[4859]: I1008 18:34:38.222194 4859 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-744fd954cc-hv6xb" Oct 08 18:34:38 crc kubenswrapper[4859]: I1008 18:34:38.226779 4859 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/neutron-7755787cf6-5gvlv"] Oct 08 18:34:38 crc kubenswrapper[4859]: I1008 18:34:38.226994 4859 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/neutron-7755787cf6-5gvlv" podUID="08efd21e-517e-46dc-bf88-60cfb1943208" containerName="neutron-api" containerID="cri-o://9c658a235d8e9170379bfc9b21179e87097616e2d09b4f9696d1fc6548e1ac2c" gracePeriod=30 Oct 08 18:34:38 crc kubenswrapper[4859]: I1008 18:34:38.227291 4859 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/neutron-7755787cf6-5gvlv" podUID="08efd21e-517e-46dc-bf88-60cfb1943208" containerName="neutron-httpd" containerID="cri-o://2822847f53055edb0f62021465c0ea6cda6373ec67373966a4dcf91c5c45ddef" gracePeriod=30 Oct 08 18:34:38 crc kubenswrapper[4859]: I1008 18:34:38.254301 4859 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-5f74b5f5cc-tl7pk" event={"ID":"a03169fb-e503-427f-bafe-2f8767a23482","Type":"ContainerStarted","Data":"d5ffe6df38528c3cd0c30ee428245d2135e36d1f8e40e3dc7bf93bccb5962981"} Oct 08 18:34:38 crc kubenswrapper[4859]: I1008 18:34:38.254347 4859 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-5f74b5f5cc-tl7pk" event={"ID":"a03169fb-e503-427f-bafe-2f8767a23482","Type":"ContainerStarted","Data":"e5814d1fa8bffbeb39ac72b0c31e8f6c976fa8b2b107e84302e99df288c6fc98"} Oct 08 18:34:38 crc kubenswrapper[4859]: I1008 18:34:38.279386 4859 scope.go:117] "RemoveContainer" containerID="b63328e8d573503c45cd4a842f1b869c293349e50d9aa5add18090b4f8c83200" Oct 08 18:34:38 crc kubenswrapper[4859]: I1008 18:34:38.287454 4859 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-744fd954cc-hv6xb"] Oct 08 18:34:38 crc kubenswrapper[4859]: I1008 18:34:38.306375 4859 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-744fd954cc-hv6xb"] Oct 08 18:34:38 crc kubenswrapper[4859]: I1008 18:34:38.525239 4859 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="8252c83d-1fdd-4d33-87a0-7a797251a0b4" path="/var/lib/kubelet/pods/8252c83d-1fdd-4d33-87a0-7a797251a0b4/volumes" Oct 08 18:34:39 crc kubenswrapper[4859]: I1008 18:34:39.275513 4859 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-api-0" event={"ID":"65e6bdd0-9ffe-48d7-9b02-444ce9b204dd","Type":"ContainerStarted","Data":"852ca86a7fece5a8f498b03bbc9d181549eb5372cbff049da3663b702b20ec91"} Oct 08 18:34:39 crc kubenswrapper[4859]: I1008 18:34:39.308408 4859 generic.go:334] "Generic (PLEG): container finished" podID="08efd21e-517e-46dc-bf88-60cfb1943208" containerID="2822847f53055edb0f62021465c0ea6cda6373ec67373966a4dcf91c5c45ddef" exitCode=0 Oct 08 18:34:39 crc kubenswrapper[4859]: I1008 18:34:39.308522 4859 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-7755787cf6-5gvlv" event={"ID":"08efd21e-517e-46dc-bf88-60cfb1943208","Type":"ContainerDied","Data":"2822847f53055edb0f62021465c0ea6cda6373ec67373966a4dcf91c5c45ddef"} Oct 08 18:34:39 crc kubenswrapper[4859]: I1008 18:34:39.331950 4859 generic.go:334] "Generic (PLEG): container finished" podID="a03169fb-e503-427f-bafe-2f8767a23482" containerID="d5ffe6df38528c3cd0c30ee428245d2135e36d1f8e40e3dc7bf93bccb5962981" exitCode=0 Oct 08 18:34:39 crc kubenswrapper[4859]: I1008 18:34:39.332021 4859 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-5f74b5f5cc-tl7pk" event={"ID":"a03169fb-e503-427f-bafe-2f8767a23482","Type":"ContainerDied","Data":"d5ffe6df38528c3cd0c30ee428245d2135e36d1f8e40e3dc7bf93bccb5962981"} Oct 08 18:34:39 crc kubenswrapper[4859]: I1008 18:34:39.427973 4859 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/cinder-api-0"] Oct 08 18:34:39 crc kubenswrapper[4859]: I1008 18:34:39.925006 4859 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/barbican-api-96f8997cd-z7vl4" Oct 08 18:34:40 crc kubenswrapper[4859]: I1008 18:34:40.223580 4859 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/barbican-api-96f8997cd-z7vl4" Oct 08 18:34:40 crc kubenswrapper[4859]: I1008 18:34:40.385935 4859 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-api-0" event={"ID":"65e6bdd0-9ffe-48d7-9b02-444ce9b204dd","Type":"ContainerStarted","Data":"8299c2ef48c3db54ed4f65c7fe5f11ea6a9f2987dae9dc6db90360dbb8a676e6"} Oct 08 18:34:40 crc kubenswrapper[4859]: I1008 18:34:40.386178 4859 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/cinder-api-0" podUID="65e6bdd0-9ffe-48d7-9b02-444ce9b204dd" containerName="cinder-api-log" containerID="cri-o://852ca86a7fece5a8f498b03bbc9d181549eb5372cbff049da3663b702b20ec91" gracePeriod=30 Oct 08 18:34:40 crc kubenswrapper[4859]: I1008 18:34:40.386326 4859 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/cinder-api-0" Oct 08 18:34:40 crc kubenswrapper[4859]: I1008 18:34:40.386437 4859 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/cinder-api-0" podUID="65e6bdd0-9ffe-48d7-9b02-444ce9b204dd" containerName="cinder-api" containerID="cri-o://8299c2ef48c3db54ed4f65c7fe5f11ea6a9f2987dae9dc6db90360dbb8a676e6" gracePeriod=30 Oct 08 18:34:40 crc kubenswrapper[4859]: I1008 18:34:40.401071 4859 generic.go:334] "Generic (PLEG): container finished" podID="1a92487e-3309-41c4-8c82-cd57f517ab4b" containerID="648c622ce9eb0bdf3d9c43d8313db6636c307abe8d840bde30cf40e0039e99c7" exitCode=0 Oct 08 18:34:40 crc kubenswrapper[4859]: I1008 18:34:40.401153 4859 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-7d58687c8-fhctx" event={"ID":"1a92487e-3309-41c4-8c82-cd57f517ab4b","Type":"ContainerDied","Data":"648c622ce9eb0bdf3d9c43d8313db6636c307abe8d840bde30cf40e0039e99c7"} Oct 08 18:34:40 crc kubenswrapper[4859]: I1008 18:34:40.403830 4859 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-5f74b5f5cc-tl7pk" event={"ID":"a03169fb-e503-427f-bafe-2f8767a23482","Type":"ContainerStarted","Data":"a17c560ce5bf3618ca0c82f026671dacb150ab5c919db4a2fb2daccfe125f76e"} Oct 08 18:34:40 crc kubenswrapper[4859]: I1008 18:34:40.404655 4859 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-5f74b5f5cc-tl7pk" Oct 08 18:34:40 crc kubenswrapper[4859]: I1008 18:34:40.418594 4859 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-scheduler-0" event={"ID":"085aabba-1f1f-4037-b5cf-036d5b519b02","Type":"ContainerStarted","Data":"9d376e0b797b164a6ab89d03ed49939fc5de488f40250b13210b65732559c4ce"} Oct 08 18:34:40 crc kubenswrapper[4859]: I1008 18:34:40.445666 4859 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/cinder-api-0" podStartSLOduration=4.445646222 podStartE2EDuration="4.445646222s" podCreationTimestamp="2025-10-08 18:34:36 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-08 18:34:40.414009267 +0000 UTC m=+1050.660848666" watchObservedRunningTime="2025-10-08 18:34:40.445646222 +0000 UTC m=+1050.692485601" Oct 08 18:34:40 crc kubenswrapper[4859]: I1008 18:34:40.462658 4859 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-5f74b5f5cc-tl7pk" podStartSLOduration=4.462636484 podStartE2EDuration="4.462636484s" podCreationTimestamp="2025-10-08 18:34:36 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-08 18:34:40.450021941 +0000 UTC m=+1050.696861320" watchObservedRunningTime="2025-10-08 18:34:40.462636484 +0000 UTC m=+1050.709475863" Oct 08 18:34:40 crc kubenswrapper[4859]: I1008 18:34:40.539716 4859 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/barbican-api-588f98c7dd-c7l8h" Oct 08 18:34:40 crc kubenswrapper[4859]: I1008 18:34:40.679059 4859 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/barbican-api-96f8997cd-z7vl4"] Oct 08 18:34:41 crc kubenswrapper[4859]: I1008 18:34:41.432256 4859 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-scheduler-0" event={"ID":"085aabba-1f1f-4037-b5cf-036d5b519b02","Type":"ContainerStarted","Data":"f308d17d64205c33abf59dfc9b5158deb3f442727d16a6d1e0e9136ca7322737"} Oct 08 18:34:41 crc kubenswrapper[4859]: I1008 18:34:41.436238 4859 generic.go:334] "Generic (PLEG): container finished" podID="65e6bdd0-9ffe-48d7-9b02-444ce9b204dd" containerID="852ca86a7fece5a8f498b03bbc9d181549eb5372cbff049da3663b702b20ec91" exitCode=143 Oct 08 18:34:41 crc kubenswrapper[4859]: I1008 18:34:41.436303 4859 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-api-0" event={"ID":"65e6bdd0-9ffe-48d7-9b02-444ce9b204dd","Type":"ContainerDied","Data":"852ca86a7fece5a8f498b03bbc9d181549eb5372cbff049da3663b702b20ec91"} Oct 08 18:34:41 crc kubenswrapper[4859]: I1008 18:34:41.440207 4859 generic.go:334] "Generic (PLEG): container finished" podID="08efd21e-517e-46dc-bf88-60cfb1943208" containerID="9c658a235d8e9170379bfc9b21179e87097616e2d09b4f9696d1fc6548e1ac2c" exitCode=0 Oct 08 18:34:41 crc kubenswrapper[4859]: I1008 18:34:41.440258 4859 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-7755787cf6-5gvlv" event={"ID":"08efd21e-517e-46dc-bf88-60cfb1943208","Type":"ContainerDied","Data":"9c658a235d8e9170379bfc9b21179e87097616e2d09b4f9696d1fc6548e1ac2c"} Oct 08 18:34:41 crc kubenswrapper[4859]: I1008 18:34:41.440667 4859 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/barbican-api-96f8997cd-z7vl4" podUID="b6cd546e-9054-4150-a5f3-7bd709d15560" containerName="barbican-api-log" containerID="cri-o://f10ceaf742ba5f03074f81bdbcb2ca60481d377e7fca65e10865b51c73de052a" gracePeriod=30 Oct 08 18:34:41 crc kubenswrapper[4859]: I1008 18:34:41.440667 4859 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/barbican-api-96f8997cd-z7vl4" podUID="b6cd546e-9054-4150-a5f3-7bd709d15560" containerName="barbican-api" containerID="cri-o://46929166d9610dcba7b53f0ba54350f18274f5815cacde12d63fdbf70d02c898" gracePeriod=30 Oct 08 18:34:41 crc kubenswrapper[4859]: I1008 18:34:41.449340 4859 prober.go:107] "Probe failed" probeType="Liveness" pod="openstack/barbican-api-96f8997cd-z7vl4" podUID="b6cd546e-9054-4150-a5f3-7bd709d15560" containerName="barbican-api" probeResult="failure" output="Get \"http://10.217.0.160:9311/healthcheck\": EOF" Oct 08 18:34:41 crc kubenswrapper[4859]: I1008 18:34:41.449639 4859 prober.go:107] "Probe failed" probeType="Liveness" pod="openstack/barbican-api-96f8997cd-z7vl4" podUID="b6cd546e-9054-4150-a5f3-7bd709d15560" containerName="barbican-api-log" probeResult="failure" output="Get \"http://10.217.0.160:9311/healthcheck\": EOF" Oct 08 18:34:41 crc kubenswrapper[4859]: I1008 18:34:41.460346 4859 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/cinder-scheduler-0" podStartSLOduration=3.949941329 podStartE2EDuration="5.460325823s" podCreationTimestamp="2025-10-08 18:34:36 +0000 UTC" firstStartedPulling="2025-10-08 18:34:37.548608981 +0000 UTC m=+1047.795448360" lastFinishedPulling="2025-10-08 18:34:39.058993485 +0000 UTC m=+1049.305832854" observedRunningTime="2025-10-08 18:34:41.454512272 +0000 UTC m=+1051.701351651" watchObservedRunningTime="2025-10-08 18:34:41.460325823 +0000 UTC m=+1051.707165202" Oct 08 18:34:41 crc kubenswrapper[4859]: I1008 18:34:41.840928 4859 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/cinder-scheduler-0" Oct 08 18:34:42 crc kubenswrapper[4859]: I1008 18:34:42.117156 4859 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/dnsmasq-dns-744fd954cc-hv6xb" podUID="8252c83d-1fdd-4d33-87a0-7a797251a0b4" containerName="dnsmasq-dns" probeResult="failure" output="dial tcp 10.217.0.159:5353: i/o timeout" Oct 08 18:34:42 crc kubenswrapper[4859]: I1008 18:34:42.227306 4859 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-7755787cf6-5gvlv" Oct 08 18:34:42 crc kubenswrapper[4859]: I1008 18:34:42.307417 4859 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/horizon-7d58687c8-fhctx" podUID="1a92487e-3309-41c4-8c82-cd57f517ab4b" containerName="horizon" probeResult="failure" output="Get \"https://10.217.0.147:8443/dashboard/auth/login/?next=/dashboard/\": dial tcp 10.217.0.147:8443: connect: connection refused" Oct 08 18:34:42 crc kubenswrapper[4859]: I1008 18:34:42.332840 4859 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/08efd21e-517e-46dc-bf88-60cfb1943208-combined-ca-bundle\") pod \"08efd21e-517e-46dc-bf88-60cfb1943208\" (UID: \"08efd21e-517e-46dc-bf88-60cfb1943208\") " Oct 08 18:34:42 crc kubenswrapper[4859]: I1008 18:34:42.332900 4859 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-wgpl2\" (UniqueName: \"kubernetes.io/projected/08efd21e-517e-46dc-bf88-60cfb1943208-kube-api-access-wgpl2\") pod \"08efd21e-517e-46dc-bf88-60cfb1943208\" (UID: \"08efd21e-517e-46dc-bf88-60cfb1943208\") " Oct 08 18:34:42 crc kubenswrapper[4859]: I1008 18:34:42.333025 4859 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovndb-tls-certs\" (UniqueName: \"kubernetes.io/secret/08efd21e-517e-46dc-bf88-60cfb1943208-ovndb-tls-certs\") pod \"08efd21e-517e-46dc-bf88-60cfb1943208\" (UID: \"08efd21e-517e-46dc-bf88-60cfb1943208\") " Oct 08 18:34:42 crc kubenswrapper[4859]: I1008 18:34:42.333080 4859 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/08efd21e-517e-46dc-bf88-60cfb1943208-config\") pod \"08efd21e-517e-46dc-bf88-60cfb1943208\" (UID: \"08efd21e-517e-46dc-bf88-60cfb1943208\") " Oct 08 18:34:42 crc kubenswrapper[4859]: I1008 18:34:42.333168 4859 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"httpd-config\" (UniqueName: \"kubernetes.io/secret/08efd21e-517e-46dc-bf88-60cfb1943208-httpd-config\") pod \"08efd21e-517e-46dc-bf88-60cfb1943208\" (UID: \"08efd21e-517e-46dc-bf88-60cfb1943208\") " Oct 08 18:34:42 crc kubenswrapper[4859]: I1008 18:34:42.341882 4859 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/08efd21e-517e-46dc-bf88-60cfb1943208-httpd-config" (OuterVolumeSpecName: "httpd-config") pod "08efd21e-517e-46dc-bf88-60cfb1943208" (UID: "08efd21e-517e-46dc-bf88-60cfb1943208"). InnerVolumeSpecName "httpd-config". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 08 18:34:42 crc kubenswrapper[4859]: I1008 18:34:42.345127 4859 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/08efd21e-517e-46dc-bf88-60cfb1943208-kube-api-access-wgpl2" (OuterVolumeSpecName: "kube-api-access-wgpl2") pod "08efd21e-517e-46dc-bf88-60cfb1943208" (UID: "08efd21e-517e-46dc-bf88-60cfb1943208"). InnerVolumeSpecName "kube-api-access-wgpl2". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 08 18:34:42 crc kubenswrapper[4859]: I1008 18:34:42.395794 4859 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/08efd21e-517e-46dc-bf88-60cfb1943208-config" (OuterVolumeSpecName: "config") pod "08efd21e-517e-46dc-bf88-60cfb1943208" (UID: "08efd21e-517e-46dc-bf88-60cfb1943208"). InnerVolumeSpecName "config". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 08 18:34:42 crc kubenswrapper[4859]: I1008 18:34:42.410473 4859 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/08efd21e-517e-46dc-bf88-60cfb1943208-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "08efd21e-517e-46dc-bf88-60cfb1943208" (UID: "08efd21e-517e-46dc-bf88-60cfb1943208"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 08 18:34:42 crc kubenswrapper[4859]: I1008 18:34:42.414651 4859 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/08efd21e-517e-46dc-bf88-60cfb1943208-ovndb-tls-certs" (OuterVolumeSpecName: "ovndb-tls-certs") pod "08efd21e-517e-46dc-bf88-60cfb1943208" (UID: "08efd21e-517e-46dc-bf88-60cfb1943208"). InnerVolumeSpecName "ovndb-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 08 18:34:42 crc kubenswrapper[4859]: I1008 18:34:42.435431 4859 reconciler_common.go:293] "Volume detached for volume \"httpd-config\" (UniqueName: \"kubernetes.io/secret/08efd21e-517e-46dc-bf88-60cfb1943208-httpd-config\") on node \"crc\" DevicePath \"\"" Oct 08 18:34:42 crc kubenswrapper[4859]: I1008 18:34:42.435467 4859 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/08efd21e-517e-46dc-bf88-60cfb1943208-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 08 18:34:42 crc kubenswrapper[4859]: I1008 18:34:42.435482 4859 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-wgpl2\" (UniqueName: \"kubernetes.io/projected/08efd21e-517e-46dc-bf88-60cfb1943208-kube-api-access-wgpl2\") on node \"crc\" DevicePath \"\"" Oct 08 18:34:42 crc kubenswrapper[4859]: I1008 18:34:42.435494 4859 reconciler_common.go:293] "Volume detached for volume \"ovndb-tls-certs\" (UniqueName: \"kubernetes.io/secret/08efd21e-517e-46dc-bf88-60cfb1943208-ovndb-tls-certs\") on node \"crc\" DevicePath \"\"" Oct 08 18:34:42 crc kubenswrapper[4859]: I1008 18:34:42.435507 4859 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/secret/08efd21e-517e-46dc-bf88-60cfb1943208-config\") on node \"crc\" DevicePath \"\"" Oct 08 18:34:42 crc kubenswrapper[4859]: I1008 18:34:42.465643 4859 generic.go:334] "Generic (PLEG): container finished" podID="b6cd546e-9054-4150-a5f3-7bd709d15560" containerID="f10ceaf742ba5f03074f81bdbcb2ca60481d377e7fca65e10865b51c73de052a" exitCode=143 Oct 08 18:34:42 crc kubenswrapper[4859]: I1008 18:34:42.465753 4859 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-api-96f8997cd-z7vl4" event={"ID":"b6cd546e-9054-4150-a5f3-7bd709d15560","Type":"ContainerDied","Data":"f10ceaf742ba5f03074f81bdbcb2ca60481d377e7fca65e10865b51c73de052a"} Oct 08 18:34:42 crc kubenswrapper[4859]: I1008 18:34:42.469581 4859 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-7755787cf6-5gvlv" Oct 08 18:34:42 crc kubenswrapper[4859]: I1008 18:34:42.469854 4859 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-7755787cf6-5gvlv" event={"ID":"08efd21e-517e-46dc-bf88-60cfb1943208","Type":"ContainerDied","Data":"90e37068995838dc25b4e6fb72406bd39a3a12e4c276b08c48491e71c0dc8d00"} Oct 08 18:34:42 crc kubenswrapper[4859]: I1008 18:34:42.470029 4859 scope.go:117] "RemoveContainer" containerID="2822847f53055edb0f62021465c0ea6cda6373ec67373966a4dcf91c5c45ddef" Oct 08 18:34:42 crc kubenswrapper[4859]: I1008 18:34:42.536422 4859 scope.go:117] "RemoveContainer" containerID="9c658a235d8e9170379bfc9b21179e87097616e2d09b4f9696d1fc6548e1ac2c" Oct 08 18:34:42 crc kubenswrapper[4859]: I1008 18:34:42.548588 4859 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/neutron-7755787cf6-5gvlv"] Oct 08 18:34:42 crc kubenswrapper[4859]: I1008 18:34:42.563510 4859 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/neutron-7755787cf6-5gvlv"] Oct 08 18:34:44 crc kubenswrapper[4859]: I1008 18:34:44.552384 4859 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="08efd21e-517e-46dc-bf88-60cfb1943208" path="/var/lib/kubelet/pods/08efd21e-517e-46dc-bf88-60cfb1943208/volumes" Oct 08 18:34:44 crc kubenswrapper[4859]: I1008 18:34:44.890888 4859 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/barbican-api-96f8997cd-z7vl4" podUID="b6cd546e-9054-4150-a5f3-7bd709d15560" containerName="barbican-api-log" probeResult="failure" output="Get \"http://10.217.0.160:9311/healthcheck\": read tcp 10.217.0.2:50742->10.217.0.160:9311: read: connection reset by peer" Oct 08 18:34:44 crc kubenswrapper[4859]: I1008 18:34:44.890976 4859 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/barbican-api-96f8997cd-z7vl4" podUID="b6cd546e-9054-4150-a5f3-7bd709d15560" containerName="barbican-api" probeResult="failure" output="Get \"http://10.217.0.160:9311/healthcheck\": read tcp 10.217.0.2:50754->10.217.0.160:9311: read: connection reset by peer" Oct 08 18:34:45 crc kubenswrapper[4859]: I1008 18:34:45.393794 4859 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-api-96f8997cd-z7vl4" Oct 08 18:34:45 crc kubenswrapper[4859]: I1008 18:34:45.529145 4859 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/b6cd546e-9054-4150-a5f3-7bd709d15560-logs\") pod \"b6cd546e-9054-4150-a5f3-7bd709d15560\" (UID: \"b6cd546e-9054-4150-a5f3-7bd709d15560\") " Oct 08 18:34:45 crc kubenswrapper[4859]: I1008 18:34:45.529206 4859 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/b6cd546e-9054-4150-a5f3-7bd709d15560-config-data\") pod \"b6cd546e-9054-4150-a5f3-7bd709d15560\" (UID: \"b6cd546e-9054-4150-a5f3-7bd709d15560\") " Oct 08 18:34:45 crc kubenswrapper[4859]: I1008 18:34:45.529262 4859 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-lwkz8\" (UniqueName: \"kubernetes.io/projected/b6cd546e-9054-4150-a5f3-7bd709d15560-kube-api-access-lwkz8\") pod \"b6cd546e-9054-4150-a5f3-7bd709d15560\" (UID: \"b6cd546e-9054-4150-a5f3-7bd709d15560\") " Oct 08 18:34:45 crc kubenswrapper[4859]: I1008 18:34:45.529896 4859 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/b6cd546e-9054-4150-a5f3-7bd709d15560-logs" (OuterVolumeSpecName: "logs") pod "b6cd546e-9054-4150-a5f3-7bd709d15560" (UID: "b6cd546e-9054-4150-a5f3-7bd709d15560"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 08 18:34:45 crc kubenswrapper[4859]: I1008 18:34:45.530152 4859 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b6cd546e-9054-4150-a5f3-7bd709d15560-combined-ca-bundle\") pod \"b6cd546e-9054-4150-a5f3-7bd709d15560\" (UID: \"b6cd546e-9054-4150-a5f3-7bd709d15560\") " Oct 08 18:34:45 crc kubenswrapper[4859]: I1008 18:34:45.530293 4859 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/b6cd546e-9054-4150-a5f3-7bd709d15560-config-data-custom\") pod \"b6cd546e-9054-4150-a5f3-7bd709d15560\" (UID: \"b6cd546e-9054-4150-a5f3-7bd709d15560\") " Oct 08 18:34:45 crc kubenswrapper[4859]: I1008 18:34:45.530789 4859 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/b6cd546e-9054-4150-a5f3-7bd709d15560-logs\") on node \"crc\" DevicePath \"\"" Oct 08 18:34:45 crc kubenswrapper[4859]: I1008 18:34:45.538885 4859 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b6cd546e-9054-4150-a5f3-7bd709d15560-kube-api-access-lwkz8" (OuterVolumeSpecName: "kube-api-access-lwkz8") pod "b6cd546e-9054-4150-a5f3-7bd709d15560" (UID: "b6cd546e-9054-4150-a5f3-7bd709d15560"). InnerVolumeSpecName "kube-api-access-lwkz8". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 08 18:34:45 crc kubenswrapper[4859]: I1008 18:34:45.542551 4859 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b6cd546e-9054-4150-a5f3-7bd709d15560-config-data-custom" (OuterVolumeSpecName: "config-data-custom") pod "b6cd546e-9054-4150-a5f3-7bd709d15560" (UID: "b6cd546e-9054-4150-a5f3-7bd709d15560"). InnerVolumeSpecName "config-data-custom". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 08 18:34:45 crc kubenswrapper[4859]: I1008 18:34:45.563644 4859 generic.go:334] "Generic (PLEG): container finished" podID="b6cd546e-9054-4150-a5f3-7bd709d15560" containerID="46929166d9610dcba7b53f0ba54350f18274f5815cacde12d63fdbf70d02c898" exitCode=0 Oct 08 18:34:45 crc kubenswrapper[4859]: I1008 18:34:45.563716 4859 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-api-96f8997cd-z7vl4" event={"ID":"b6cd546e-9054-4150-a5f3-7bd709d15560","Type":"ContainerDied","Data":"46929166d9610dcba7b53f0ba54350f18274f5815cacde12d63fdbf70d02c898"} Oct 08 18:34:45 crc kubenswrapper[4859]: I1008 18:34:45.563750 4859 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-api-96f8997cd-z7vl4" event={"ID":"b6cd546e-9054-4150-a5f3-7bd709d15560","Type":"ContainerDied","Data":"335733d7843c6fc9d74b64c45ad4ab2c6b5e04222f1d6841fea11f46f57d5ba8"} Oct 08 18:34:45 crc kubenswrapper[4859]: I1008 18:34:45.563771 4859 scope.go:117] "RemoveContainer" containerID="46929166d9610dcba7b53f0ba54350f18274f5815cacde12d63fdbf70d02c898" Oct 08 18:34:45 crc kubenswrapper[4859]: I1008 18:34:45.563941 4859 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-api-96f8997cd-z7vl4" Oct 08 18:34:45 crc kubenswrapper[4859]: I1008 18:34:45.579289 4859 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b6cd546e-9054-4150-a5f3-7bd709d15560-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "b6cd546e-9054-4150-a5f3-7bd709d15560" (UID: "b6cd546e-9054-4150-a5f3-7bd709d15560"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 08 18:34:45 crc kubenswrapper[4859]: I1008 18:34:45.597945 4859 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b6cd546e-9054-4150-a5f3-7bd709d15560-config-data" (OuterVolumeSpecName: "config-data") pod "b6cd546e-9054-4150-a5f3-7bd709d15560" (UID: "b6cd546e-9054-4150-a5f3-7bd709d15560"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 08 18:34:45 crc kubenswrapper[4859]: I1008 18:34:45.633407 4859 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/b6cd546e-9054-4150-a5f3-7bd709d15560-config-data\") on node \"crc\" DevicePath \"\"" Oct 08 18:34:45 crc kubenswrapper[4859]: I1008 18:34:45.633447 4859 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-lwkz8\" (UniqueName: \"kubernetes.io/projected/b6cd546e-9054-4150-a5f3-7bd709d15560-kube-api-access-lwkz8\") on node \"crc\" DevicePath \"\"" Oct 08 18:34:45 crc kubenswrapper[4859]: I1008 18:34:45.633459 4859 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b6cd546e-9054-4150-a5f3-7bd709d15560-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 08 18:34:45 crc kubenswrapper[4859]: I1008 18:34:45.633471 4859 reconciler_common.go:293] "Volume detached for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/b6cd546e-9054-4150-a5f3-7bd709d15560-config-data-custom\") on node \"crc\" DevicePath \"\"" Oct 08 18:34:45 crc kubenswrapper[4859]: I1008 18:34:45.667261 4859 scope.go:117] "RemoveContainer" containerID="f10ceaf742ba5f03074f81bdbcb2ca60481d377e7fca65e10865b51c73de052a" Oct 08 18:34:45 crc kubenswrapper[4859]: I1008 18:34:45.686867 4859 scope.go:117] "RemoveContainer" containerID="46929166d9610dcba7b53f0ba54350f18274f5815cacde12d63fdbf70d02c898" Oct 08 18:34:45 crc kubenswrapper[4859]: E1008 18:34:45.687478 4859 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"46929166d9610dcba7b53f0ba54350f18274f5815cacde12d63fdbf70d02c898\": container with ID starting with 46929166d9610dcba7b53f0ba54350f18274f5815cacde12d63fdbf70d02c898 not found: ID does not exist" containerID="46929166d9610dcba7b53f0ba54350f18274f5815cacde12d63fdbf70d02c898" Oct 08 18:34:45 crc kubenswrapper[4859]: I1008 18:34:45.687540 4859 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"46929166d9610dcba7b53f0ba54350f18274f5815cacde12d63fdbf70d02c898"} err="failed to get container status \"46929166d9610dcba7b53f0ba54350f18274f5815cacde12d63fdbf70d02c898\": rpc error: code = NotFound desc = could not find container \"46929166d9610dcba7b53f0ba54350f18274f5815cacde12d63fdbf70d02c898\": container with ID starting with 46929166d9610dcba7b53f0ba54350f18274f5815cacde12d63fdbf70d02c898 not found: ID does not exist" Oct 08 18:34:45 crc kubenswrapper[4859]: I1008 18:34:45.687572 4859 scope.go:117] "RemoveContainer" containerID="f10ceaf742ba5f03074f81bdbcb2ca60481d377e7fca65e10865b51c73de052a" Oct 08 18:34:45 crc kubenswrapper[4859]: E1008 18:34:45.687997 4859 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"f10ceaf742ba5f03074f81bdbcb2ca60481d377e7fca65e10865b51c73de052a\": container with ID starting with f10ceaf742ba5f03074f81bdbcb2ca60481d377e7fca65e10865b51c73de052a not found: ID does not exist" containerID="f10ceaf742ba5f03074f81bdbcb2ca60481d377e7fca65e10865b51c73de052a" Oct 08 18:34:45 crc kubenswrapper[4859]: I1008 18:34:45.688039 4859 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"f10ceaf742ba5f03074f81bdbcb2ca60481d377e7fca65e10865b51c73de052a"} err="failed to get container status \"f10ceaf742ba5f03074f81bdbcb2ca60481d377e7fca65e10865b51c73de052a\": rpc error: code = NotFound desc = could not find container \"f10ceaf742ba5f03074f81bdbcb2ca60481d377e7fca65e10865b51c73de052a\": container with ID starting with f10ceaf742ba5f03074f81bdbcb2ca60481d377e7fca65e10865b51c73de052a not found: ID does not exist" Oct 08 18:34:45 crc kubenswrapper[4859]: I1008 18:34:45.740148 4859 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/placement-58c5f664f4-5lvrs" Oct 08 18:34:45 crc kubenswrapper[4859]: I1008 18:34:45.764098 4859 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/placement-58c5f664f4-5lvrs" Oct 08 18:34:45 crc kubenswrapper[4859]: I1008 18:34:45.880078 4859 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/keystone-758ff8b98b-p6wm9" Oct 08 18:34:45 crc kubenswrapper[4859]: I1008 18:34:45.906994 4859 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/barbican-api-96f8997cd-z7vl4"] Oct 08 18:34:45 crc kubenswrapper[4859]: I1008 18:34:45.924916 4859 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/barbican-api-96f8997cd-z7vl4"] Oct 08 18:34:46 crc kubenswrapper[4859]: I1008 18:34:46.494379 4859 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="b6cd546e-9054-4150-a5f3-7bd709d15560" path="/var/lib/kubelet/pods/b6cd546e-9054-4150-a5f3-7bd709d15560/volumes" Oct 08 18:34:47 crc kubenswrapper[4859]: I1008 18:34:47.047595 4859 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-5f74b5f5cc-tl7pk" Oct 08 18:34:47 crc kubenswrapper[4859]: I1008 18:34:47.066969 4859 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/cinder-scheduler-0" Oct 08 18:34:47 crc kubenswrapper[4859]: I1008 18:34:47.199889 4859 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/cinder-scheduler-0"] Oct 08 18:34:47 crc kubenswrapper[4859]: I1008 18:34:47.217109 4859 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-58799d59b9-5kp6v"] Oct 08 18:34:47 crc kubenswrapper[4859]: I1008 18:34:47.217584 4859 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-58799d59b9-5kp6v" podUID="881d266c-6730-404d-8243-ea0df0d3d03f" containerName="dnsmasq-dns" containerID="cri-o://30a2952f16a49c92b8051f736d0387ff2fd3296b5cd3c22fcca6085b250996db" gracePeriod=10 Oct 08 18:34:47 crc kubenswrapper[4859]: I1008 18:34:47.589965 4859 generic.go:334] "Generic (PLEG): container finished" podID="881d266c-6730-404d-8243-ea0df0d3d03f" containerID="30a2952f16a49c92b8051f736d0387ff2fd3296b5cd3c22fcca6085b250996db" exitCode=0 Oct 08 18:34:47 crc kubenswrapper[4859]: I1008 18:34:47.590423 4859 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/cinder-scheduler-0" podUID="085aabba-1f1f-4037-b5cf-036d5b519b02" containerName="cinder-scheduler" containerID="cri-o://9d376e0b797b164a6ab89d03ed49939fc5de488f40250b13210b65732559c4ce" gracePeriod=30 Oct 08 18:34:47 crc kubenswrapper[4859]: I1008 18:34:47.590517 4859 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/cinder-scheduler-0" podUID="085aabba-1f1f-4037-b5cf-036d5b519b02" containerName="probe" containerID="cri-o://f308d17d64205c33abf59dfc9b5158deb3f442727d16a6d1e0e9136ca7322737" gracePeriod=30 Oct 08 18:34:47 crc kubenswrapper[4859]: I1008 18:34:47.590374 4859 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-58799d59b9-5kp6v" event={"ID":"881d266c-6730-404d-8243-ea0df0d3d03f","Type":"ContainerDied","Data":"30a2952f16a49c92b8051f736d0387ff2fd3296b5cd3c22fcca6085b250996db"} Oct 08 18:34:47 crc kubenswrapper[4859]: I1008 18:34:47.920590 4859 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-58799d59b9-5kp6v" Oct 08 18:34:47 crc kubenswrapper[4859]: I1008 18:34:47.924274 4859 patch_prober.go:28] interesting pod/machine-config-daemon-82s52 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 08 18:34:47 crc kubenswrapper[4859]: I1008 18:34:47.924322 4859 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-82s52" podUID="b23a6a6c-9d92-4e7b-840e-55cfda873a2d" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 08 18:34:47 crc kubenswrapper[4859]: I1008 18:34:47.924359 4859 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-82s52" Oct 08 18:34:47 crc kubenswrapper[4859]: I1008 18:34:47.925070 4859 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"091916d0315d4cd4d10d5ce64ba3d175b9444c9b90ffd7170b4b5fd2fb8bb62d"} pod="openshift-machine-config-operator/machine-config-daemon-82s52" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Oct 08 18:34:47 crc kubenswrapper[4859]: I1008 18:34:47.925131 4859 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-82s52" podUID="b23a6a6c-9d92-4e7b-840e-55cfda873a2d" containerName="machine-config-daemon" containerID="cri-o://091916d0315d4cd4d10d5ce64ba3d175b9444c9b90ffd7170b4b5fd2fb8bb62d" gracePeriod=600 Oct 08 18:34:47 crc kubenswrapper[4859]: I1008 18:34:47.980757 4859 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/881d266c-6730-404d-8243-ea0df0d3d03f-ovsdbserver-sb\") pod \"881d266c-6730-404d-8243-ea0df0d3d03f\" (UID: \"881d266c-6730-404d-8243-ea0df0d3d03f\") " Oct 08 18:34:47 crc kubenswrapper[4859]: I1008 18:34:47.980901 4859 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/881d266c-6730-404d-8243-ea0df0d3d03f-config\") pod \"881d266c-6730-404d-8243-ea0df0d3d03f\" (UID: \"881d266c-6730-404d-8243-ea0df0d3d03f\") " Oct 08 18:34:47 crc kubenswrapper[4859]: I1008 18:34:47.980951 4859 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-2gc55\" (UniqueName: \"kubernetes.io/projected/881d266c-6730-404d-8243-ea0df0d3d03f-kube-api-access-2gc55\") pod \"881d266c-6730-404d-8243-ea0df0d3d03f\" (UID: \"881d266c-6730-404d-8243-ea0df0d3d03f\") " Oct 08 18:34:47 crc kubenswrapper[4859]: I1008 18:34:47.980983 4859 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/881d266c-6730-404d-8243-ea0df0d3d03f-dns-svc\") pod \"881d266c-6730-404d-8243-ea0df0d3d03f\" (UID: \"881d266c-6730-404d-8243-ea0df0d3d03f\") " Oct 08 18:34:47 crc kubenswrapper[4859]: I1008 18:34:47.981119 4859 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/881d266c-6730-404d-8243-ea0df0d3d03f-ovsdbserver-nb\") pod \"881d266c-6730-404d-8243-ea0df0d3d03f\" (UID: \"881d266c-6730-404d-8243-ea0df0d3d03f\") " Oct 08 18:34:47 crc kubenswrapper[4859]: I1008 18:34:47.981145 4859 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/881d266c-6730-404d-8243-ea0df0d3d03f-dns-swift-storage-0\") pod \"881d266c-6730-404d-8243-ea0df0d3d03f\" (UID: \"881d266c-6730-404d-8243-ea0df0d3d03f\") " Oct 08 18:34:47 crc kubenswrapper[4859]: I1008 18:34:47.992001 4859 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/881d266c-6730-404d-8243-ea0df0d3d03f-kube-api-access-2gc55" (OuterVolumeSpecName: "kube-api-access-2gc55") pod "881d266c-6730-404d-8243-ea0df0d3d03f" (UID: "881d266c-6730-404d-8243-ea0df0d3d03f"). InnerVolumeSpecName "kube-api-access-2gc55". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 08 18:34:48 crc kubenswrapper[4859]: I1008 18:34:48.083977 4859 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/881d266c-6730-404d-8243-ea0df0d3d03f-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "881d266c-6730-404d-8243-ea0df0d3d03f" (UID: "881d266c-6730-404d-8243-ea0df0d3d03f"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 08 18:34:48 crc kubenswrapper[4859]: I1008 18:34:48.084088 4859 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/881d266c-6730-404d-8243-ea0df0d3d03f-ovsdbserver-nb\") pod \"881d266c-6730-404d-8243-ea0df0d3d03f\" (UID: \"881d266c-6730-404d-8243-ea0df0d3d03f\") " Oct 08 18:34:48 crc kubenswrapper[4859]: I1008 18:34:48.084616 4859 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-2gc55\" (UniqueName: \"kubernetes.io/projected/881d266c-6730-404d-8243-ea0df0d3d03f-kube-api-access-2gc55\") on node \"crc\" DevicePath \"\"" Oct 08 18:34:48 crc kubenswrapper[4859]: W1008 18:34:48.084957 4859 empty_dir.go:500] Warning: Unmount skipped because path does not exist: /var/lib/kubelet/pods/881d266c-6730-404d-8243-ea0df0d3d03f/volumes/kubernetes.io~configmap/ovsdbserver-nb Oct 08 18:34:48 crc kubenswrapper[4859]: I1008 18:34:48.084968 4859 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/881d266c-6730-404d-8243-ea0df0d3d03f-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "881d266c-6730-404d-8243-ea0df0d3d03f" (UID: "881d266c-6730-404d-8243-ea0df0d3d03f"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 08 18:34:48 crc kubenswrapper[4859]: I1008 18:34:48.091502 4859 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/881d266c-6730-404d-8243-ea0df0d3d03f-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "881d266c-6730-404d-8243-ea0df0d3d03f" (UID: "881d266c-6730-404d-8243-ea0df0d3d03f"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 08 18:34:48 crc kubenswrapper[4859]: I1008 18:34:48.098010 4859 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/881d266c-6730-404d-8243-ea0df0d3d03f-dns-swift-storage-0" (OuterVolumeSpecName: "dns-swift-storage-0") pod "881d266c-6730-404d-8243-ea0df0d3d03f" (UID: "881d266c-6730-404d-8243-ea0df0d3d03f"). InnerVolumeSpecName "dns-swift-storage-0". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 08 18:34:48 crc kubenswrapper[4859]: I1008 18:34:48.104351 4859 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/881d266c-6730-404d-8243-ea0df0d3d03f-config" (OuterVolumeSpecName: "config") pod "881d266c-6730-404d-8243-ea0df0d3d03f" (UID: "881d266c-6730-404d-8243-ea0df0d3d03f"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 08 18:34:48 crc kubenswrapper[4859]: I1008 18:34:48.120942 4859 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/881d266c-6730-404d-8243-ea0df0d3d03f-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "881d266c-6730-404d-8243-ea0df0d3d03f" (UID: "881d266c-6730-404d-8243-ea0df0d3d03f"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 08 18:34:48 crc kubenswrapper[4859]: I1008 18:34:48.187191 4859 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/881d266c-6730-404d-8243-ea0df0d3d03f-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Oct 08 18:34:48 crc kubenswrapper[4859]: I1008 18:34:48.187465 4859 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/881d266c-6730-404d-8243-ea0df0d3d03f-config\") on node \"crc\" DevicePath \"\"" Oct 08 18:34:48 crc kubenswrapper[4859]: I1008 18:34:48.187589 4859 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/881d266c-6730-404d-8243-ea0df0d3d03f-dns-svc\") on node \"crc\" DevicePath \"\"" Oct 08 18:34:48 crc kubenswrapper[4859]: I1008 18:34:48.187699 4859 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/881d266c-6730-404d-8243-ea0df0d3d03f-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Oct 08 18:34:48 crc kubenswrapper[4859]: I1008 18:34:48.194029 4859 reconciler_common.go:293] "Volume detached for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/881d266c-6730-404d-8243-ea0df0d3d03f-dns-swift-storage-0\") on node \"crc\" DevicePath \"\"" Oct 08 18:34:48 crc kubenswrapper[4859]: I1008 18:34:48.599852 4859 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-58799d59b9-5kp6v" event={"ID":"881d266c-6730-404d-8243-ea0df0d3d03f","Type":"ContainerDied","Data":"5361fd0c4f9aa7d00088e20917cbe941dd6b0518e8bed856f763ca12404f277f"} Oct 08 18:34:48 crc kubenswrapper[4859]: I1008 18:34:48.600167 4859 scope.go:117] "RemoveContainer" containerID="30a2952f16a49c92b8051f736d0387ff2fd3296b5cd3c22fcca6085b250996db" Oct 08 18:34:48 crc kubenswrapper[4859]: I1008 18:34:48.599901 4859 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-58799d59b9-5kp6v" Oct 08 18:34:48 crc kubenswrapper[4859]: I1008 18:34:48.602587 4859 generic.go:334] "Generic (PLEG): container finished" podID="b23a6a6c-9d92-4e7b-840e-55cfda873a2d" containerID="091916d0315d4cd4d10d5ce64ba3d175b9444c9b90ffd7170b4b5fd2fb8bb62d" exitCode=0 Oct 08 18:34:48 crc kubenswrapper[4859]: I1008 18:34:48.602634 4859 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-82s52" event={"ID":"b23a6a6c-9d92-4e7b-840e-55cfda873a2d","Type":"ContainerDied","Data":"091916d0315d4cd4d10d5ce64ba3d175b9444c9b90ffd7170b4b5fd2fb8bb62d"} Oct 08 18:34:48 crc kubenswrapper[4859]: I1008 18:34:48.602664 4859 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-82s52" event={"ID":"b23a6a6c-9d92-4e7b-840e-55cfda873a2d","Type":"ContainerStarted","Data":"7ed110461431af172203421fc6cf54521a9bc279642133f6f29a4bb290304252"} Oct 08 18:34:48 crc kubenswrapper[4859]: I1008 18:34:48.666453 4859 scope.go:117] "RemoveContainer" containerID="0079be4d4c960d2bcd1f78238bedd74e2b721bf336a7f9bd857e41e20ed129f6" Oct 08 18:34:48 crc kubenswrapper[4859]: I1008 18:34:48.680831 4859 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-58799d59b9-5kp6v"] Oct 08 18:34:48 crc kubenswrapper[4859]: I1008 18:34:48.686739 4859 scope.go:117] "RemoveContainer" containerID="6ca72342b251d747b10d72d072b3e566097d846a52f53943f7da38fe2a34e100" Oct 08 18:34:48 crc kubenswrapper[4859]: I1008 18:34:48.686951 4859 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-58799d59b9-5kp6v"] Oct 08 18:34:49 crc kubenswrapper[4859]: I1008 18:34:49.547254 4859 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/openstackclient"] Oct 08 18:34:49 crc kubenswrapper[4859]: E1008 18:34:49.548105 4859 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8252c83d-1fdd-4d33-87a0-7a797251a0b4" containerName="dnsmasq-dns" Oct 08 18:34:49 crc kubenswrapper[4859]: I1008 18:34:49.548123 4859 state_mem.go:107] "Deleted CPUSet assignment" podUID="8252c83d-1fdd-4d33-87a0-7a797251a0b4" containerName="dnsmasq-dns" Oct 08 18:34:49 crc kubenswrapper[4859]: E1008 18:34:49.548139 4859 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="08efd21e-517e-46dc-bf88-60cfb1943208" containerName="neutron-httpd" Oct 08 18:34:49 crc kubenswrapper[4859]: I1008 18:34:49.548147 4859 state_mem.go:107] "Deleted CPUSet assignment" podUID="08efd21e-517e-46dc-bf88-60cfb1943208" containerName="neutron-httpd" Oct 08 18:34:49 crc kubenswrapper[4859]: E1008 18:34:49.548160 4859 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b6cd546e-9054-4150-a5f3-7bd709d15560" containerName="barbican-api" Oct 08 18:34:49 crc kubenswrapper[4859]: I1008 18:34:49.548167 4859 state_mem.go:107] "Deleted CPUSet assignment" podUID="b6cd546e-9054-4150-a5f3-7bd709d15560" containerName="barbican-api" Oct 08 18:34:49 crc kubenswrapper[4859]: E1008 18:34:49.548190 4859 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="08efd21e-517e-46dc-bf88-60cfb1943208" containerName="neutron-api" Oct 08 18:34:49 crc kubenswrapper[4859]: I1008 18:34:49.548196 4859 state_mem.go:107] "Deleted CPUSet assignment" podUID="08efd21e-517e-46dc-bf88-60cfb1943208" containerName="neutron-api" Oct 08 18:34:49 crc kubenswrapper[4859]: E1008 18:34:49.548221 4859 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="881d266c-6730-404d-8243-ea0df0d3d03f" containerName="dnsmasq-dns" Oct 08 18:34:49 crc kubenswrapper[4859]: I1008 18:34:49.548229 4859 state_mem.go:107] "Deleted CPUSet assignment" podUID="881d266c-6730-404d-8243-ea0df0d3d03f" containerName="dnsmasq-dns" Oct 08 18:34:49 crc kubenswrapper[4859]: E1008 18:34:49.548242 4859 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8252c83d-1fdd-4d33-87a0-7a797251a0b4" containerName="init" Oct 08 18:34:49 crc kubenswrapper[4859]: I1008 18:34:49.548249 4859 state_mem.go:107] "Deleted CPUSet assignment" podUID="8252c83d-1fdd-4d33-87a0-7a797251a0b4" containerName="init" Oct 08 18:34:49 crc kubenswrapper[4859]: E1008 18:34:49.548271 4859 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="881d266c-6730-404d-8243-ea0df0d3d03f" containerName="init" Oct 08 18:34:49 crc kubenswrapper[4859]: I1008 18:34:49.548279 4859 state_mem.go:107] "Deleted CPUSet assignment" podUID="881d266c-6730-404d-8243-ea0df0d3d03f" containerName="init" Oct 08 18:34:49 crc kubenswrapper[4859]: E1008 18:34:49.548293 4859 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b6cd546e-9054-4150-a5f3-7bd709d15560" containerName="barbican-api-log" Oct 08 18:34:49 crc kubenswrapper[4859]: I1008 18:34:49.548300 4859 state_mem.go:107] "Deleted CPUSet assignment" podUID="b6cd546e-9054-4150-a5f3-7bd709d15560" containerName="barbican-api-log" Oct 08 18:34:49 crc kubenswrapper[4859]: I1008 18:34:49.548491 4859 memory_manager.go:354] "RemoveStaleState removing state" podUID="b6cd546e-9054-4150-a5f3-7bd709d15560" containerName="barbican-api" Oct 08 18:34:49 crc kubenswrapper[4859]: I1008 18:34:49.548506 4859 memory_manager.go:354] "RemoveStaleState removing state" podUID="8252c83d-1fdd-4d33-87a0-7a797251a0b4" containerName="dnsmasq-dns" Oct 08 18:34:49 crc kubenswrapper[4859]: I1008 18:34:49.548519 4859 memory_manager.go:354] "RemoveStaleState removing state" podUID="08efd21e-517e-46dc-bf88-60cfb1943208" containerName="neutron-httpd" Oct 08 18:34:49 crc kubenswrapper[4859]: I1008 18:34:49.548538 4859 memory_manager.go:354] "RemoveStaleState removing state" podUID="b6cd546e-9054-4150-a5f3-7bd709d15560" containerName="barbican-api-log" Oct 08 18:34:49 crc kubenswrapper[4859]: I1008 18:34:49.548547 4859 memory_manager.go:354] "RemoveStaleState removing state" podUID="881d266c-6730-404d-8243-ea0df0d3d03f" containerName="dnsmasq-dns" Oct 08 18:34:49 crc kubenswrapper[4859]: I1008 18:34:49.548563 4859 memory_manager.go:354] "RemoveStaleState removing state" podUID="08efd21e-517e-46dc-bf88-60cfb1943208" containerName="neutron-api" Oct 08 18:34:49 crc kubenswrapper[4859]: I1008 18:34:49.549374 4859 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/openstackclient" Oct 08 18:34:49 crc kubenswrapper[4859]: I1008 18:34:49.551871 4859 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstackclient-openstackclient-dockercfg-v79s9" Oct 08 18:34:49 crc kubenswrapper[4859]: I1008 18:34:49.551896 4859 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-config-secret" Oct 08 18:34:49 crc kubenswrapper[4859]: I1008 18:34:49.552591 4859 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-config" Oct 08 18:34:49 crc kubenswrapper[4859]: I1008 18:34:49.560556 4859 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/openstackclient"] Oct 08 18:34:49 crc kubenswrapper[4859]: I1008 18:34:49.630049 4859 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/435bf36d-2c04-41b5-a7dd-de54ccd0ee06-combined-ca-bundle\") pod \"openstackclient\" (UID: \"435bf36d-2c04-41b5-a7dd-de54ccd0ee06\") " pod="openstack/openstackclient" Oct 08 18:34:49 crc kubenswrapper[4859]: I1008 18:34:49.630088 4859 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/435bf36d-2c04-41b5-a7dd-de54ccd0ee06-openstack-config-secret\") pod \"openstackclient\" (UID: \"435bf36d-2c04-41b5-a7dd-de54ccd0ee06\") " pod="openstack/openstackclient" Oct 08 18:34:49 crc kubenswrapper[4859]: I1008 18:34:49.630136 4859 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/435bf36d-2c04-41b5-a7dd-de54ccd0ee06-openstack-config\") pod \"openstackclient\" (UID: \"435bf36d-2c04-41b5-a7dd-de54ccd0ee06\") " pod="openstack/openstackclient" Oct 08 18:34:49 crc kubenswrapper[4859]: I1008 18:34:49.630351 4859 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-66dqq\" (UniqueName: \"kubernetes.io/projected/435bf36d-2c04-41b5-a7dd-de54ccd0ee06-kube-api-access-66dqq\") pod \"openstackclient\" (UID: \"435bf36d-2c04-41b5-a7dd-de54ccd0ee06\") " pod="openstack/openstackclient" Oct 08 18:34:49 crc kubenswrapper[4859]: I1008 18:34:49.668202 4859 generic.go:334] "Generic (PLEG): container finished" podID="085aabba-1f1f-4037-b5cf-036d5b519b02" containerID="f308d17d64205c33abf59dfc9b5158deb3f442727d16a6d1e0e9136ca7322737" exitCode=0 Oct 08 18:34:49 crc kubenswrapper[4859]: I1008 18:34:49.668249 4859 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-scheduler-0" event={"ID":"085aabba-1f1f-4037-b5cf-036d5b519b02","Type":"ContainerDied","Data":"f308d17d64205c33abf59dfc9b5158deb3f442727d16a6d1e0e9136ca7322737"} Oct 08 18:34:49 crc kubenswrapper[4859]: I1008 18:34:49.679910 4859 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/cinder-api-0" Oct 08 18:34:49 crc kubenswrapper[4859]: I1008 18:34:49.732465 4859 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/435bf36d-2c04-41b5-a7dd-de54ccd0ee06-combined-ca-bundle\") pod \"openstackclient\" (UID: \"435bf36d-2c04-41b5-a7dd-de54ccd0ee06\") " pod="openstack/openstackclient" Oct 08 18:34:49 crc kubenswrapper[4859]: I1008 18:34:49.732974 4859 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/435bf36d-2c04-41b5-a7dd-de54ccd0ee06-openstack-config-secret\") pod \"openstackclient\" (UID: \"435bf36d-2c04-41b5-a7dd-de54ccd0ee06\") " pod="openstack/openstackclient" Oct 08 18:34:49 crc kubenswrapper[4859]: I1008 18:34:49.733033 4859 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/435bf36d-2c04-41b5-a7dd-de54ccd0ee06-openstack-config\") pod \"openstackclient\" (UID: \"435bf36d-2c04-41b5-a7dd-de54ccd0ee06\") " pod="openstack/openstackclient" Oct 08 18:34:49 crc kubenswrapper[4859]: I1008 18:34:49.733123 4859 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-66dqq\" (UniqueName: \"kubernetes.io/projected/435bf36d-2c04-41b5-a7dd-de54ccd0ee06-kube-api-access-66dqq\") pod \"openstackclient\" (UID: \"435bf36d-2c04-41b5-a7dd-de54ccd0ee06\") " pod="openstack/openstackclient" Oct 08 18:34:49 crc kubenswrapper[4859]: I1008 18:34:49.744148 4859 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/435bf36d-2c04-41b5-a7dd-de54ccd0ee06-openstack-config\") pod \"openstackclient\" (UID: \"435bf36d-2c04-41b5-a7dd-de54ccd0ee06\") " pod="openstack/openstackclient" Oct 08 18:34:49 crc kubenswrapper[4859]: I1008 18:34:49.744312 4859 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/435bf36d-2c04-41b5-a7dd-de54ccd0ee06-openstack-config-secret\") pod \"openstackclient\" (UID: \"435bf36d-2c04-41b5-a7dd-de54ccd0ee06\") " pod="openstack/openstackclient" Oct 08 18:34:49 crc kubenswrapper[4859]: I1008 18:34:49.747908 4859 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/435bf36d-2c04-41b5-a7dd-de54ccd0ee06-combined-ca-bundle\") pod \"openstackclient\" (UID: \"435bf36d-2c04-41b5-a7dd-de54ccd0ee06\") " pod="openstack/openstackclient" Oct 08 18:34:49 crc kubenswrapper[4859]: I1008 18:34:49.755597 4859 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-66dqq\" (UniqueName: \"kubernetes.io/projected/435bf36d-2c04-41b5-a7dd-de54ccd0ee06-kube-api-access-66dqq\") pod \"openstackclient\" (UID: \"435bf36d-2c04-41b5-a7dd-de54ccd0ee06\") " pod="openstack/openstackclient" Oct 08 18:34:49 crc kubenswrapper[4859]: I1008 18:34:49.813363 4859 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/openstackclient"] Oct 08 18:34:49 crc kubenswrapper[4859]: I1008 18:34:49.814100 4859 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/openstackclient" Oct 08 18:34:49 crc kubenswrapper[4859]: I1008 18:34:49.829600 4859 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/openstackclient"] Oct 08 18:34:49 crc kubenswrapper[4859]: I1008 18:34:49.869662 4859 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/openstackclient"] Oct 08 18:34:49 crc kubenswrapper[4859]: I1008 18:34:49.871297 4859 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/openstackclient" Oct 08 18:34:49 crc kubenswrapper[4859]: I1008 18:34:49.889959 4859 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/openstackclient"] Oct 08 18:34:49 crc kubenswrapper[4859]: I1008 18:34:49.937991 4859 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1a631d33-4c0c-46c9-a3d5-4d9dc9f6de84-combined-ca-bundle\") pod \"openstackclient\" (UID: \"1a631d33-4c0c-46c9-a3d5-4d9dc9f6de84\") " pod="openstack/openstackclient" Oct 08 18:34:49 crc kubenswrapper[4859]: I1008 18:34:49.938035 4859 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-w95f9\" (UniqueName: \"kubernetes.io/projected/1a631d33-4c0c-46c9-a3d5-4d9dc9f6de84-kube-api-access-w95f9\") pod \"openstackclient\" (UID: \"1a631d33-4c0c-46c9-a3d5-4d9dc9f6de84\") " pod="openstack/openstackclient" Oct 08 18:34:49 crc kubenswrapper[4859]: I1008 18:34:49.938322 4859 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/1a631d33-4c0c-46c9-a3d5-4d9dc9f6de84-openstack-config-secret\") pod \"openstackclient\" (UID: \"1a631d33-4c0c-46c9-a3d5-4d9dc9f6de84\") " pod="openstack/openstackclient" Oct 08 18:34:49 crc kubenswrapper[4859]: I1008 18:34:49.938449 4859 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/1a631d33-4c0c-46c9-a3d5-4d9dc9f6de84-openstack-config\") pod \"openstackclient\" (UID: \"1a631d33-4c0c-46c9-a3d5-4d9dc9f6de84\") " pod="openstack/openstackclient" Oct 08 18:34:49 crc kubenswrapper[4859]: E1008 18:34:49.961967 4859 log.go:32] "RunPodSandbox from runtime service failed" err=< Oct 08 18:34:49 crc kubenswrapper[4859]: rpc error: code = Unknown desc = failed to create pod network sandbox k8s_openstackclient_openstack_435bf36d-2c04-41b5-a7dd-de54ccd0ee06_0(ce1d3c2382f29328f4e08cd65aeefa3854230797fde45a5a5387c8fce43cecd1): error adding pod openstack_openstackclient to CNI network "multus-cni-network": plugin type="multus-shim" name="multus-cni-network" failed (add): CmdAdd (shim): CNI request failed with status 400: 'ContainerID:"ce1d3c2382f29328f4e08cd65aeefa3854230797fde45a5a5387c8fce43cecd1" Netns:"/var/run/netns/c21d6295-d707-4b42-bff1-94d42cdf9789" IfName:"eth0" Args:"IgnoreUnknown=1;K8S_POD_NAMESPACE=openstack;K8S_POD_NAME=openstackclient;K8S_POD_INFRA_CONTAINER_ID=ce1d3c2382f29328f4e08cd65aeefa3854230797fde45a5a5387c8fce43cecd1;K8S_POD_UID=435bf36d-2c04-41b5-a7dd-de54ccd0ee06" Path:"" ERRORED: error configuring pod [openstack/openstackclient] networking: Multus: [openstack/openstackclient/435bf36d-2c04-41b5-a7dd-de54ccd0ee06]: expected pod UID "435bf36d-2c04-41b5-a7dd-de54ccd0ee06" but got "1a631d33-4c0c-46c9-a3d5-4d9dc9f6de84" from Kube API Oct 08 18:34:49 crc kubenswrapper[4859]: ': StdinData: {"binDir":"/var/lib/cni/bin","clusterNetwork":"/host/run/multus/cni/net.d/10-ovn-kubernetes.conf","cniVersion":"0.3.1","daemonSocketDir":"/run/multus/socket","globalNamespaces":"default,openshift-multus,openshift-sriov-network-operator,openshift-cnv","logLevel":"verbose","logToStderr":true,"name":"multus-cni-network","namespaceIsolation":true,"type":"multus-shim"} Oct 08 18:34:49 crc kubenswrapper[4859]: > Oct 08 18:34:49 crc kubenswrapper[4859]: E1008 18:34:49.962246 4859 kuberuntime_sandbox.go:72] "Failed to create sandbox for pod" err=< Oct 08 18:34:49 crc kubenswrapper[4859]: rpc error: code = Unknown desc = failed to create pod network sandbox k8s_openstackclient_openstack_435bf36d-2c04-41b5-a7dd-de54ccd0ee06_0(ce1d3c2382f29328f4e08cd65aeefa3854230797fde45a5a5387c8fce43cecd1): error adding pod openstack_openstackclient to CNI network "multus-cni-network": plugin type="multus-shim" name="multus-cni-network" failed (add): CmdAdd (shim): CNI request failed with status 400: 'ContainerID:"ce1d3c2382f29328f4e08cd65aeefa3854230797fde45a5a5387c8fce43cecd1" Netns:"/var/run/netns/c21d6295-d707-4b42-bff1-94d42cdf9789" IfName:"eth0" Args:"IgnoreUnknown=1;K8S_POD_NAMESPACE=openstack;K8S_POD_NAME=openstackclient;K8S_POD_INFRA_CONTAINER_ID=ce1d3c2382f29328f4e08cd65aeefa3854230797fde45a5a5387c8fce43cecd1;K8S_POD_UID=435bf36d-2c04-41b5-a7dd-de54ccd0ee06" Path:"" ERRORED: error configuring pod [openstack/openstackclient] networking: Multus: [openstack/openstackclient/435bf36d-2c04-41b5-a7dd-de54ccd0ee06]: expected pod UID "435bf36d-2c04-41b5-a7dd-de54ccd0ee06" but got "1a631d33-4c0c-46c9-a3d5-4d9dc9f6de84" from Kube API Oct 08 18:34:49 crc kubenswrapper[4859]: ': StdinData: {"binDir":"/var/lib/cni/bin","clusterNetwork":"/host/run/multus/cni/net.d/10-ovn-kubernetes.conf","cniVersion":"0.3.1","daemonSocketDir":"/run/multus/socket","globalNamespaces":"default,openshift-multus,openshift-sriov-network-operator,openshift-cnv","logLevel":"verbose","logToStderr":true,"name":"multus-cni-network","namespaceIsolation":true,"type":"multus-shim"} Oct 08 18:34:49 crc kubenswrapper[4859]: > pod="openstack/openstackclient" Oct 08 18:34:50 crc kubenswrapper[4859]: I1008 18:34:50.040100 4859 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1a631d33-4c0c-46c9-a3d5-4d9dc9f6de84-combined-ca-bundle\") pod \"openstackclient\" (UID: \"1a631d33-4c0c-46c9-a3d5-4d9dc9f6de84\") " pod="openstack/openstackclient" Oct 08 18:34:50 crc kubenswrapper[4859]: I1008 18:34:50.040150 4859 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-w95f9\" (UniqueName: \"kubernetes.io/projected/1a631d33-4c0c-46c9-a3d5-4d9dc9f6de84-kube-api-access-w95f9\") pod \"openstackclient\" (UID: \"1a631d33-4c0c-46c9-a3d5-4d9dc9f6de84\") " pod="openstack/openstackclient" Oct 08 18:34:50 crc kubenswrapper[4859]: I1008 18:34:50.040244 4859 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/1a631d33-4c0c-46c9-a3d5-4d9dc9f6de84-openstack-config-secret\") pod \"openstackclient\" (UID: \"1a631d33-4c0c-46c9-a3d5-4d9dc9f6de84\") " pod="openstack/openstackclient" Oct 08 18:34:50 crc kubenswrapper[4859]: I1008 18:34:50.040297 4859 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/1a631d33-4c0c-46c9-a3d5-4d9dc9f6de84-openstack-config\") pod \"openstackclient\" (UID: \"1a631d33-4c0c-46c9-a3d5-4d9dc9f6de84\") " pod="openstack/openstackclient" Oct 08 18:34:50 crc kubenswrapper[4859]: I1008 18:34:50.041732 4859 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/1a631d33-4c0c-46c9-a3d5-4d9dc9f6de84-openstack-config\") pod \"openstackclient\" (UID: \"1a631d33-4c0c-46c9-a3d5-4d9dc9f6de84\") " pod="openstack/openstackclient" Oct 08 18:34:50 crc kubenswrapper[4859]: I1008 18:34:50.049230 4859 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1a631d33-4c0c-46c9-a3d5-4d9dc9f6de84-combined-ca-bundle\") pod \"openstackclient\" (UID: \"1a631d33-4c0c-46c9-a3d5-4d9dc9f6de84\") " pod="openstack/openstackclient" Oct 08 18:34:50 crc kubenswrapper[4859]: I1008 18:34:50.052931 4859 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/1a631d33-4c0c-46c9-a3d5-4d9dc9f6de84-openstack-config-secret\") pod \"openstackclient\" (UID: \"1a631d33-4c0c-46c9-a3d5-4d9dc9f6de84\") " pod="openstack/openstackclient" Oct 08 18:34:50 crc kubenswrapper[4859]: I1008 18:34:50.058092 4859 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-w95f9\" (UniqueName: \"kubernetes.io/projected/1a631d33-4c0c-46c9-a3d5-4d9dc9f6de84-kube-api-access-w95f9\") pod \"openstackclient\" (UID: \"1a631d33-4c0c-46c9-a3d5-4d9dc9f6de84\") " pod="openstack/openstackclient" Oct 08 18:34:50 crc kubenswrapper[4859]: I1008 18:34:50.311667 4859 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/openstackclient" Oct 08 18:34:50 crc kubenswrapper[4859]: I1008 18:34:50.486016 4859 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="881d266c-6730-404d-8243-ea0df0d3d03f" path="/var/lib/kubelet/pods/881d266c-6730-404d-8243-ea0df0d3d03f/volumes" Oct 08 18:34:50 crc kubenswrapper[4859]: I1008 18:34:50.677047 4859 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/openstackclient" Oct 08 18:34:50 crc kubenswrapper[4859]: I1008 18:34:50.690153 4859 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/openstackclient" Oct 08 18:34:50 crc kubenswrapper[4859]: I1008 18:34:50.692757 4859 status_manager.go:861] "Pod was deleted and then recreated, skipping status update" pod="openstack/openstackclient" oldPodUID="435bf36d-2c04-41b5-a7dd-de54ccd0ee06" podUID="1a631d33-4c0c-46c9-a3d5-4d9dc9f6de84" Oct 08 18:34:50 crc kubenswrapper[4859]: I1008 18:34:50.759587 4859 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/435bf36d-2c04-41b5-a7dd-de54ccd0ee06-openstack-config-secret\") pod \"435bf36d-2c04-41b5-a7dd-de54ccd0ee06\" (UID: \"435bf36d-2c04-41b5-a7dd-de54ccd0ee06\") " Oct 08 18:34:50 crc kubenswrapper[4859]: I1008 18:34:50.759640 4859 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/435bf36d-2c04-41b5-a7dd-de54ccd0ee06-combined-ca-bundle\") pod \"435bf36d-2c04-41b5-a7dd-de54ccd0ee06\" (UID: \"435bf36d-2c04-41b5-a7dd-de54ccd0ee06\") " Oct 08 18:34:50 crc kubenswrapper[4859]: I1008 18:34:50.759718 4859 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-66dqq\" (UniqueName: \"kubernetes.io/projected/435bf36d-2c04-41b5-a7dd-de54ccd0ee06-kube-api-access-66dqq\") pod \"435bf36d-2c04-41b5-a7dd-de54ccd0ee06\" (UID: \"435bf36d-2c04-41b5-a7dd-de54ccd0ee06\") " Oct 08 18:34:50 crc kubenswrapper[4859]: I1008 18:34:50.759795 4859 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/435bf36d-2c04-41b5-a7dd-de54ccd0ee06-openstack-config\") pod \"435bf36d-2c04-41b5-a7dd-de54ccd0ee06\" (UID: \"435bf36d-2c04-41b5-a7dd-de54ccd0ee06\") " Oct 08 18:34:50 crc kubenswrapper[4859]: I1008 18:34:50.760639 4859 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/435bf36d-2c04-41b5-a7dd-de54ccd0ee06-openstack-config" (OuterVolumeSpecName: "openstack-config") pod "435bf36d-2c04-41b5-a7dd-de54ccd0ee06" (UID: "435bf36d-2c04-41b5-a7dd-de54ccd0ee06"). InnerVolumeSpecName "openstack-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 08 18:34:50 crc kubenswrapper[4859]: I1008 18:34:50.766406 4859 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/435bf36d-2c04-41b5-a7dd-de54ccd0ee06-kube-api-access-66dqq" (OuterVolumeSpecName: "kube-api-access-66dqq") pod "435bf36d-2c04-41b5-a7dd-de54ccd0ee06" (UID: "435bf36d-2c04-41b5-a7dd-de54ccd0ee06"). InnerVolumeSpecName "kube-api-access-66dqq". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 08 18:34:50 crc kubenswrapper[4859]: I1008 18:34:50.766473 4859 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/435bf36d-2c04-41b5-a7dd-de54ccd0ee06-openstack-config-secret" (OuterVolumeSpecName: "openstack-config-secret") pod "435bf36d-2c04-41b5-a7dd-de54ccd0ee06" (UID: "435bf36d-2c04-41b5-a7dd-de54ccd0ee06"). InnerVolumeSpecName "openstack-config-secret". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 08 18:34:50 crc kubenswrapper[4859]: I1008 18:34:50.766960 4859 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/435bf36d-2c04-41b5-a7dd-de54ccd0ee06-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "435bf36d-2c04-41b5-a7dd-de54ccd0ee06" (UID: "435bf36d-2c04-41b5-a7dd-de54ccd0ee06"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 08 18:34:50 crc kubenswrapper[4859]: I1008 18:34:50.818463 4859 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/openstackclient"] Oct 08 18:34:50 crc kubenswrapper[4859]: I1008 18:34:50.862228 4859 reconciler_common.go:293] "Volume detached for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/435bf36d-2c04-41b5-a7dd-de54ccd0ee06-openstack-config-secret\") on node \"crc\" DevicePath \"\"" Oct 08 18:34:50 crc kubenswrapper[4859]: I1008 18:34:50.862266 4859 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/435bf36d-2c04-41b5-a7dd-de54ccd0ee06-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 08 18:34:50 crc kubenswrapper[4859]: I1008 18:34:50.862278 4859 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-66dqq\" (UniqueName: \"kubernetes.io/projected/435bf36d-2c04-41b5-a7dd-de54ccd0ee06-kube-api-access-66dqq\") on node \"crc\" DevicePath \"\"" Oct 08 18:34:50 crc kubenswrapper[4859]: I1008 18:34:50.862289 4859 reconciler_common.go:293] "Volume detached for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/435bf36d-2c04-41b5-a7dd-de54ccd0ee06-openstack-config\") on node \"crc\" DevicePath \"\"" Oct 08 18:34:51 crc kubenswrapper[4859]: I1008 18:34:51.698655 4859 generic.go:334] "Generic (PLEG): container finished" podID="085aabba-1f1f-4037-b5cf-036d5b519b02" containerID="9d376e0b797b164a6ab89d03ed49939fc5de488f40250b13210b65732559c4ce" exitCode=0 Oct 08 18:34:51 crc kubenswrapper[4859]: I1008 18:34:51.698737 4859 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-scheduler-0" event={"ID":"085aabba-1f1f-4037-b5cf-036d5b519b02","Type":"ContainerDied","Data":"9d376e0b797b164a6ab89d03ed49939fc5de488f40250b13210b65732559c4ce"} Oct 08 18:34:51 crc kubenswrapper[4859]: I1008 18:34:51.703966 4859 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/openstackclient" Oct 08 18:34:51 crc kubenswrapper[4859]: I1008 18:34:51.705260 4859 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstackclient" event={"ID":"1a631d33-4c0c-46c9-a3d5-4d9dc9f6de84","Type":"ContainerStarted","Data":"5d40301f0902afbaf07d88c390bac8c5a5a7aebc30a87d610c0970d4948ebb43"} Oct 08 18:34:51 crc kubenswrapper[4859]: I1008 18:34:51.729104 4859 status_manager.go:861] "Pod was deleted and then recreated, skipping status update" pod="openstack/openstackclient" oldPodUID="435bf36d-2c04-41b5-a7dd-de54ccd0ee06" podUID="1a631d33-4c0c-46c9-a3d5-4d9dc9f6de84" Oct 08 18:34:52 crc kubenswrapper[4859]: I1008 18:34:52.105372 4859 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-scheduler-0" Oct 08 18:34:52 crc kubenswrapper[4859]: I1008 18:34:52.191556 4859 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-dqz5f\" (UniqueName: \"kubernetes.io/projected/085aabba-1f1f-4037-b5cf-036d5b519b02-kube-api-access-dqz5f\") pod \"085aabba-1f1f-4037-b5cf-036d5b519b02\" (UID: \"085aabba-1f1f-4037-b5cf-036d5b519b02\") " Oct 08 18:34:52 crc kubenswrapper[4859]: I1008 18:34:52.191653 4859 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/085aabba-1f1f-4037-b5cf-036d5b519b02-config-data\") pod \"085aabba-1f1f-4037-b5cf-036d5b519b02\" (UID: \"085aabba-1f1f-4037-b5cf-036d5b519b02\") " Oct 08 18:34:52 crc kubenswrapper[4859]: I1008 18:34:52.191885 4859 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/085aabba-1f1f-4037-b5cf-036d5b519b02-config-data-custom\") pod \"085aabba-1f1f-4037-b5cf-036d5b519b02\" (UID: \"085aabba-1f1f-4037-b5cf-036d5b519b02\") " Oct 08 18:34:52 crc kubenswrapper[4859]: I1008 18:34:52.192036 4859 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/085aabba-1f1f-4037-b5cf-036d5b519b02-etc-machine-id\") pod \"085aabba-1f1f-4037-b5cf-036d5b519b02\" (UID: \"085aabba-1f1f-4037-b5cf-036d5b519b02\") " Oct 08 18:34:52 crc kubenswrapper[4859]: I1008 18:34:52.192173 4859 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/085aabba-1f1f-4037-b5cf-036d5b519b02-etc-machine-id" (OuterVolumeSpecName: "etc-machine-id") pod "085aabba-1f1f-4037-b5cf-036d5b519b02" (UID: "085aabba-1f1f-4037-b5cf-036d5b519b02"). InnerVolumeSpecName "etc-machine-id". PluginName "kubernetes.io/host-path", VolumeGidValue "" Oct 08 18:34:52 crc kubenswrapper[4859]: I1008 18:34:52.192283 4859 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/085aabba-1f1f-4037-b5cf-036d5b519b02-combined-ca-bundle\") pod \"085aabba-1f1f-4037-b5cf-036d5b519b02\" (UID: \"085aabba-1f1f-4037-b5cf-036d5b519b02\") " Oct 08 18:34:52 crc kubenswrapper[4859]: I1008 18:34:52.192909 4859 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/085aabba-1f1f-4037-b5cf-036d5b519b02-scripts\") pod \"085aabba-1f1f-4037-b5cf-036d5b519b02\" (UID: \"085aabba-1f1f-4037-b5cf-036d5b519b02\") " Oct 08 18:34:52 crc kubenswrapper[4859]: I1008 18:34:52.195541 4859 reconciler_common.go:293] "Volume detached for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/085aabba-1f1f-4037-b5cf-036d5b519b02-etc-machine-id\") on node \"crc\" DevicePath \"\"" Oct 08 18:34:52 crc kubenswrapper[4859]: I1008 18:34:52.198993 4859 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/085aabba-1f1f-4037-b5cf-036d5b519b02-config-data-custom" (OuterVolumeSpecName: "config-data-custom") pod "085aabba-1f1f-4037-b5cf-036d5b519b02" (UID: "085aabba-1f1f-4037-b5cf-036d5b519b02"). InnerVolumeSpecName "config-data-custom". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 08 18:34:52 crc kubenswrapper[4859]: I1008 18:34:52.199124 4859 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/085aabba-1f1f-4037-b5cf-036d5b519b02-scripts" (OuterVolumeSpecName: "scripts") pod "085aabba-1f1f-4037-b5cf-036d5b519b02" (UID: "085aabba-1f1f-4037-b5cf-036d5b519b02"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 08 18:34:52 crc kubenswrapper[4859]: I1008 18:34:52.199814 4859 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/085aabba-1f1f-4037-b5cf-036d5b519b02-kube-api-access-dqz5f" (OuterVolumeSpecName: "kube-api-access-dqz5f") pod "085aabba-1f1f-4037-b5cf-036d5b519b02" (UID: "085aabba-1f1f-4037-b5cf-036d5b519b02"). InnerVolumeSpecName "kube-api-access-dqz5f". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 08 18:34:52 crc kubenswrapper[4859]: I1008 18:34:52.263782 4859 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/085aabba-1f1f-4037-b5cf-036d5b519b02-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "085aabba-1f1f-4037-b5cf-036d5b519b02" (UID: "085aabba-1f1f-4037-b5cf-036d5b519b02"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 08 18:34:52 crc kubenswrapper[4859]: I1008 18:34:52.296948 4859 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/085aabba-1f1f-4037-b5cf-036d5b519b02-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 08 18:34:52 crc kubenswrapper[4859]: I1008 18:34:52.297204 4859 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/085aabba-1f1f-4037-b5cf-036d5b519b02-scripts\") on node \"crc\" DevicePath \"\"" Oct 08 18:34:52 crc kubenswrapper[4859]: I1008 18:34:52.297307 4859 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-dqz5f\" (UniqueName: \"kubernetes.io/projected/085aabba-1f1f-4037-b5cf-036d5b519b02-kube-api-access-dqz5f\") on node \"crc\" DevicePath \"\"" Oct 08 18:34:52 crc kubenswrapper[4859]: I1008 18:34:52.297407 4859 reconciler_common.go:293] "Volume detached for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/085aabba-1f1f-4037-b5cf-036d5b519b02-config-data-custom\") on node \"crc\" DevicePath \"\"" Oct 08 18:34:52 crc kubenswrapper[4859]: I1008 18:34:52.307665 4859 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/horizon-7d58687c8-fhctx" podUID="1a92487e-3309-41c4-8c82-cd57f517ab4b" containerName="horizon" probeResult="failure" output="Get \"https://10.217.0.147:8443/dashboard/auth/login/?next=/dashboard/\": dial tcp 10.217.0.147:8443: connect: connection refused" Oct 08 18:34:52 crc kubenswrapper[4859]: I1008 18:34:52.313234 4859 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/085aabba-1f1f-4037-b5cf-036d5b519b02-config-data" (OuterVolumeSpecName: "config-data") pod "085aabba-1f1f-4037-b5cf-036d5b519b02" (UID: "085aabba-1f1f-4037-b5cf-036d5b519b02"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 08 18:34:52 crc kubenswrapper[4859]: I1008 18:34:52.399266 4859 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/085aabba-1f1f-4037-b5cf-036d5b519b02-config-data\") on node \"crc\" DevicePath \"\"" Oct 08 18:34:52 crc kubenswrapper[4859]: I1008 18:34:52.485172 4859 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="435bf36d-2c04-41b5-a7dd-de54ccd0ee06" path="/var/lib/kubelet/pods/435bf36d-2c04-41b5-a7dd-de54ccd0ee06/volumes" Oct 08 18:34:52 crc kubenswrapper[4859]: I1008 18:34:52.716641 4859 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-scheduler-0" event={"ID":"085aabba-1f1f-4037-b5cf-036d5b519b02","Type":"ContainerDied","Data":"02c84458f7fd46ebb947c2b80a0b6fbcd7f6f1abf7683c0e94eb6f254a0fa634"} Oct 08 18:34:52 crc kubenswrapper[4859]: I1008 18:34:52.716817 4859 scope.go:117] "RemoveContainer" containerID="f308d17d64205c33abf59dfc9b5158deb3f442727d16a6d1e0e9136ca7322737" Oct 08 18:34:52 crc kubenswrapper[4859]: I1008 18:34:52.716856 4859 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-scheduler-0" Oct 08 18:34:52 crc kubenswrapper[4859]: I1008 18:34:52.748805 4859 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/cinder-scheduler-0"] Oct 08 18:34:52 crc kubenswrapper[4859]: I1008 18:34:52.764364 4859 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/cinder-scheduler-0"] Oct 08 18:34:52 crc kubenswrapper[4859]: I1008 18:34:52.767307 4859 scope.go:117] "RemoveContainer" containerID="9d376e0b797b164a6ab89d03ed49939fc5de488f40250b13210b65732559c4ce" Oct 08 18:34:52 crc kubenswrapper[4859]: I1008 18:34:52.776559 4859 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/cinder-scheduler-0"] Oct 08 18:34:52 crc kubenswrapper[4859]: E1008 18:34:52.777033 4859 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="085aabba-1f1f-4037-b5cf-036d5b519b02" containerName="cinder-scheduler" Oct 08 18:34:52 crc kubenswrapper[4859]: I1008 18:34:52.777057 4859 state_mem.go:107] "Deleted CPUSet assignment" podUID="085aabba-1f1f-4037-b5cf-036d5b519b02" containerName="cinder-scheduler" Oct 08 18:34:52 crc kubenswrapper[4859]: E1008 18:34:52.777087 4859 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="085aabba-1f1f-4037-b5cf-036d5b519b02" containerName="probe" Oct 08 18:34:52 crc kubenswrapper[4859]: I1008 18:34:52.777094 4859 state_mem.go:107] "Deleted CPUSet assignment" podUID="085aabba-1f1f-4037-b5cf-036d5b519b02" containerName="probe" Oct 08 18:34:52 crc kubenswrapper[4859]: I1008 18:34:52.777266 4859 memory_manager.go:354] "RemoveStaleState removing state" podUID="085aabba-1f1f-4037-b5cf-036d5b519b02" containerName="cinder-scheduler" Oct 08 18:34:52 crc kubenswrapper[4859]: I1008 18:34:52.777288 4859 memory_manager.go:354] "RemoveStaleState removing state" podUID="085aabba-1f1f-4037-b5cf-036d5b519b02" containerName="probe" Oct 08 18:34:52 crc kubenswrapper[4859]: I1008 18:34:52.778286 4859 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-scheduler-0" Oct 08 18:34:52 crc kubenswrapper[4859]: I1008 18:34:52.780164 4859 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-scheduler-config-data" Oct 08 18:34:52 crc kubenswrapper[4859]: I1008 18:34:52.785010 4859 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-scheduler-0"] Oct 08 18:34:52 crc kubenswrapper[4859]: I1008 18:34:52.908667 4859 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/f20f7045-6768-4824-9295-6e5ba74da387-config-data-custom\") pod \"cinder-scheduler-0\" (UID: \"f20f7045-6768-4824-9295-6e5ba74da387\") " pod="openstack/cinder-scheduler-0" Oct 08 18:34:52 crc kubenswrapper[4859]: I1008 18:34:52.908728 4859 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/f20f7045-6768-4824-9295-6e5ba74da387-etc-machine-id\") pod \"cinder-scheduler-0\" (UID: \"f20f7045-6768-4824-9295-6e5ba74da387\") " pod="openstack/cinder-scheduler-0" Oct 08 18:34:52 crc kubenswrapper[4859]: I1008 18:34:52.908767 4859 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/f20f7045-6768-4824-9295-6e5ba74da387-scripts\") pod \"cinder-scheduler-0\" (UID: \"f20f7045-6768-4824-9295-6e5ba74da387\") " pod="openstack/cinder-scheduler-0" Oct 08 18:34:52 crc kubenswrapper[4859]: I1008 18:34:52.908918 4859 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-66tmk\" (UniqueName: \"kubernetes.io/projected/f20f7045-6768-4824-9295-6e5ba74da387-kube-api-access-66tmk\") pod \"cinder-scheduler-0\" (UID: \"f20f7045-6768-4824-9295-6e5ba74da387\") " pod="openstack/cinder-scheduler-0" Oct 08 18:34:52 crc kubenswrapper[4859]: I1008 18:34:52.908997 4859 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f20f7045-6768-4824-9295-6e5ba74da387-config-data\") pod \"cinder-scheduler-0\" (UID: \"f20f7045-6768-4824-9295-6e5ba74da387\") " pod="openstack/cinder-scheduler-0" Oct 08 18:34:52 crc kubenswrapper[4859]: I1008 18:34:52.909018 4859 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f20f7045-6768-4824-9295-6e5ba74da387-combined-ca-bundle\") pod \"cinder-scheduler-0\" (UID: \"f20f7045-6768-4824-9295-6e5ba74da387\") " pod="openstack/cinder-scheduler-0" Oct 08 18:34:52 crc kubenswrapper[4859]: I1008 18:34:52.959088 4859 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/swift-proxy-776bbffb49-bdzz5"] Oct 08 18:34:52 crc kubenswrapper[4859]: I1008 18:34:52.961458 4859 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/swift-proxy-776bbffb49-bdzz5" Oct 08 18:34:52 crc kubenswrapper[4859]: I1008 18:34:52.970094 4859 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-swift-public-svc" Oct 08 18:34:52 crc kubenswrapper[4859]: I1008 18:34:52.970321 4859 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"swift-proxy-config-data" Oct 08 18:34:52 crc kubenswrapper[4859]: I1008 18:34:52.970354 4859 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-swift-internal-svc" Oct 08 18:34:52 crc kubenswrapper[4859]: I1008 18:34:52.995215 4859 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/swift-proxy-776bbffb49-bdzz5"] Oct 08 18:34:53 crc kubenswrapper[4859]: I1008 18:34:53.021949 4859 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-6qnzw\" (UniqueName: \"kubernetes.io/projected/465dc520-34ae-41b8-869c-5d4ead3b16ab-kube-api-access-6qnzw\") pod \"swift-proxy-776bbffb49-bdzz5\" (UID: \"465dc520-34ae-41b8-869c-5d4ead3b16ab\") " pod="openstack/swift-proxy-776bbffb49-bdzz5" Oct 08 18:34:53 crc kubenswrapper[4859]: I1008 18:34:53.022059 4859 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/f20f7045-6768-4824-9295-6e5ba74da387-scripts\") pod \"cinder-scheduler-0\" (UID: \"f20f7045-6768-4824-9295-6e5ba74da387\") " pod="openstack/cinder-scheduler-0" Oct 08 18:34:53 crc kubenswrapper[4859]: I1008 18:34:53.022093 4859 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/465dc520-34ae-41b8-869c-5d4ead3b16ab-log-httpd\") pod \"swift-proxy-776bbffb49-bdzz5\" (UID: \"465dc520-34ae-41b8-869c-5d4ead3b16ab\") " pod="openstack/swift-proxy-776bbffb49-bdzz5" Oct 08 18:34:53 crc kubenswrapper[4859]: I1008 18:34:53.022248 4859 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-66tmk\" (UniqueName: \"kubernetes.io/projected/f20f7045-6768-4824-9295-6e5ba74da387-kube-api-access-66tmk\") pod \"cinder-scheduler-0\" (UID: \"f20f7045-6768-4824-9295-6e5ba74da387\") " pod="openstack/cinder-scheduler-0" Oct 08 18:34:53 crc kubenswrapper[4859]: I1008 18:34:53.022318 4859 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f20f7045-6768-4824-9295-6e5ba74da387-config-data\") pod \"cinder-scheduler-0\" (UID: \"f20f7045-6768-4824-9295-6e5ba74da387\") " pod="openstack/cinder-scheduler-0" Oct 08 18:34:53 crc kubenswrapper[4859]: I1008 18:34:53.022351 4859 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f20f7045-6768-4824-9295-6e5ba74da387-combined-ca-bundle\") pod \"cinder-scheduler-0\" (UID: \"f20f7045-6768-4824-9295-6e5ba74da387\") " pod="openstack/cinder-scheduler-0" Oct 08 18:34:53 crc kubenswrapper[4859]: I1008 18:34:53.022433 4859 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/465dc520-34ae-41b8-869c-5d4ead3b16ab-internal-tls-certs\") pod \"swift-proxy-776bbffb49-bdzz5\" (UID: \"465dc520-34ae-41b8-869c-5d4ead3b16ab\") " pod="openstack/swift-proxy-776bbffb49-bdzz5" Oct 08 18:34:53 crc kubenswrapper[4859]: I1008 18:34:53.022494 4859 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/465dc520-34ae-41b8-869c-5d4ead3b16ab-combined-ca-bundle\") pod \"swift-proxy-776bbffb49-bdzz5\" (UID: \"465dc520-34ae-41b8-869c-5d4ead3b16ab\") " pod="openstack/swift-proxy-776bbffb49-bdzz5" Oct 08 18:34:53 crc kubenswrapper[4859]: I1008 18:34:53.022668 4859 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/465dc520-34ae-41b8-869c-5d4ead3b16ab-etc-swift\") pod \"swift-proxy-776bbffb49-bdzz5\" (UID: \"465dc520-34ae-41b8-869c-5d4ead3b16ab\") " pod="openstack/swift-proxy-776bbffb49-bdzz5" Oct 08 18:34:53 crc kubenswrapper[4859]: I1008 18:34:53.022736 4859 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/465dc520-34ae-41b8-869c-5d4ead3b16ab-config-data\") pod \"swift-proxy-776bbffb49-bdzz5\" (UID: \"465dc520-34ae-41b8-869c-5d4ead3b16ab\") " pod="openstack/swift-proxy-776bbffb49-bdzz5" Oct 08 18:34:53 crc kubenswrapper[4859]: I1008 18:34:53.022783 4859 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/465dc520-34ae-41b8-869c-5d4ead3b16ab-run-httpd\") pod \"swift-proxy-776bbffb49-bdzz5\" (UID: \"465dc520-34ae-41b8-869c-5d4ead3b16ab\") " pod="openstack/swift-proxy-776bbffb49-bdzz5" Oct 08 18:34:53 crc kubenswrapper[4859]: I1008 18:34:53.022813 4859 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/f20f7045-6768-4824-9295-6e5ba74da387-config-data-custom\") pod \"cinder-scheduler-0\" (UID: \"f20f7045-6768-4824-9295-6e5ba74da387\") " pod="openstack/cinder-scheduler-0" Oct 08 18:34:53 crc kubenswrapper[4859]: I1008 18:34:53.022882 4859 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/f20f7045-6768-4824-9295-6e5ba74da387-etc-machine-id\") pod \"cinder-scheduler-0\" (UID: \"f20f7045-6768-4824-9295-6e5ba74da387\") " pod="openstack/cinder-scheduler-0" Oct 08 18:34:53 crc kubenswrapper[4859]: I1008 18:34:53.022918 4859 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/465dc520-34ae-41b8-869c-5d4ead3b16ab-public-tls-certs\") pod \"swift-proxy-776bbffb49-bdzz5\" (UID: \"465dc520-34ae-41b8-869c-5d4ead3b16ab\") " pod="openstack/swift-proxy-776bbffb49-bdzz5" Oct 08 18:34:53 crc kubenswrapper[4859]: I1008 18:34:53.025933 4859 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/f20f7045-6768-4824-9295-6e5ba74da387-etc-machine-id\") pod \"cinder-scheduler-0\" (UID: \"f20f7045-6768-4824-9295-6e5ba74da387\") " pod="openstack/cinder-scheduler-0" Oct 08 18:34:53 crc kubenswrapper[4859]: I1008 18:34:53.030516 4859 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f20f7045-6768-4824-9295-6e5ba74da387-combined-ca-bundle\") pod \"cinder-scheduler-0\" (UID: \"f20f7045-6768-4824-9295-6e5ba74da387\") " pod="openstack/cinder-scheduler-0" Oct 08 18:34:53 crc kubenswrapper[4859]: I1008 18:34:53.039739 4859 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f20f7045-6768-4824-9295-6e5ba74da387-config-data\") pod \"cinder-scheduler-0\" (UID: \"f20f7045-6768-4824-9295-6e5ba74da387\") " pod="openstack/cinder-scheduler-0" Oct 08 18:34:53 crc kubenswrapper[4859]: I1008 18:34:53.044132 4859 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-66tmk\" (UniqueName: \"kubernetes.io/projected/f20f7045-6768-4824-9295-6e5ba74da387-kube-api-access-66tmk\") pod \"cinder-scheduler-0\" (UID: \"f20f7045-6768-4824-9295-6e5ba74da387\") " pod="openstack/cinder-scheduler-0" Oct 08 18:34:53 crc kubenswrapper[4859]: I1008 18:34:53.053478 4859 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/f20f7045-6768-4824-9295-6e5ba74da387-config-data-custom\") pod \"cinder-scheduler-0\" (UID: \"f20f7045-6768-4824-9295-6e5ba74da387\") " pod="openstack/cinder-scheduler-0" Oct 08 18:34:53 crc kubenswrapper[4859]: I1008 18:34:53.059909 4859 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/f20f7045-6768-4824-9295-6e5ba74da387-scripts\") pod \"cinder-scheduler-0\" (UID: \"f20f7045-6768-4824-9295-6e5ba74da387\") " pod="openstack/cinder-scheduler-0" Oct 08 18:34:53 crc kubenswrapper[4859]: I1008 18:34:53.095651 4859 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-scheduler-0" Oct 08 18:34:53 crc kubenswrapper[4859]: I1008 18:34:53.126464 4859 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/465dc520-34ae-41b8-869c-5d4ead3b16ab-config-data\") pod \"swift-proxy-776bbffb49-bdzz5\" (UID: \"465dc520-34ae-41b8-869c-5d4ead3b16ab\") " pod="openstack/swift-proxy-776bbffb49-bdzz5" Oct 08 18:34:53 crc kubenswrapper[4859]: I1008 18:34:53.126534 4859 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/465dc520-34ae-41b8-869c-5d4ead3b16ab-run-httpd\") pod \"swift-proxy-776bbffb49-bdzz5\" (UID: \"465dc520-34ae-41b8-869c-5d4ead3b16ab\") " pod="openstack/swift-proxy-776bbffb49-bdzz5" Oct 08 18:34:53 crc kubenswrapper[4859]: I1008 18:34:53.126588 4859 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/465dc520-34ae-41b8-869c-5d4ead3b16ab-public-tls-certs\") pod \"swift-proxy-776bbffb49-bdzz5\" (UID: \"465dc520-34ae-41b8-869c-5d4ead3b16ab\") " pod="openstack/swift-proxy-776bbffb49-bdzz5" Oct 08 18:34:53 crc kubenswrapper[4859]: I1008 18:34:53.126636 4859 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-6qnzw\" (UniqueName: \"kubernetes.io/projected/465dc520-34ae-41b8-869c-5d4ead3b16ab-kube-api-access-6qnzw\") pod \"swift-proxy-776bbffb49-bdzz5\" (UID: \"465dc520-34ae-41b8-869c-5d4ead3b16ab\") " pod="openstack/swift-proxy-776bbffb49-bdzz5" Oct 08 18:34:53 crc kubenswrapper[4859]: I1008 18:34:53.127361 4859 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/465dc520-34ae-41b8-869c-5d4ead3b16ab-log-httpd\") pod \"swift-proxy-776bbffb49-bdzz5\" (UID: \"465dc520-34ae-41b8-869c-5d4ead3b16ab\") " pod="openstack/swift-proxy-776bbffb49-bdzz5" Oct 08 18:34:53 crc kubenswrapper[4859]: I1008 18:34:53.127531 4859 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/465dc520-34ae-41b8-869c-5d4ead3b16ab-internal-tls-certs\") pod \"swift-proxy-776bbffb49-bdzz5\" (UID: \"465dc520-34ae-41b8-869c-5d4ead3b16ab\") " pod="openstack/swift-proxy-776bbffb49-bdzz5" Oct 08 18:34:53 crc kubenswrapper[4859]: I1008 18:34:53.127584 4859 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/465dc520-34ae-41b8-869c-5d4ead3b16ab-combined-ca-bundle\") pod \"swift-proxy-776bbffb49-bdzz5\" (UID: \"465dc520-34ae-41b8-869c-5d4ead3b16ab\") " pod="openstack/swift-proxy-776bbffb49-bdzz5" Oct 08 18:34:53 crc kubenswrapper[4859]: I1008 18:34:53.127680 4859 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/465dc520-34ae-41b8-869c-5d4ead3b16ab-etc-swift\") pod \"swift-proxy-776bbffb49-bdzz5\" (UID: \"465dc520-34ae-41b8-869c-5d4ead3b16ab\") " pod="openstack/swift-proxy-776bbffb49-bdzz5" Oct 08 18:34:53 crc kubenswrapper[4859]: I1008 18:34:53.129300 4859 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/465dc520-34ae-41b8-869c-5d4ead3b16ab-run-httpd\") pod \"swift-proxy-776bbffb49-bdzz5\" (UID: \"465dc520-34ae-41b8-869c-5d4ead3b16ab\") " pod="openstack/swift-proxy-776bbffb49-bdzz5" Oct 08 18:34:53 crc kubenswrapper[4859]: I1008 18:34:53.130544 4859 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/465dc520-34ae-41b8-869c-5d4ead3b16ab-log-httpd\") pod \"swift-proxy-776bbffb49-bdzz5\" (UID: \"465dc520-34ae-41b8-869c-5d4ead3b16ab\") " pod="openstack/swift-proxy-776bbffb49-bdzz5" Oct 08 18:34:53 crc kubenswrapper[4859]: I1008 18:34:53.132602 4859 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/465dc520-34ae-41b8-869c-5d4ead3b16ab-public-tls-certs\") pod \"swift-proxy-776bbffb49-bdzz5\" (UID: \"465dc520-34ae-41b8-869c-5d4ead3b16ab\") " pod="openstack/swift-proxy-776bbffb49-bdzz5" Oct 08 18:34:53 crc kubenswrapper[4859]: I1008 18:34:53.136219 4859 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/465dc520-34ae-41b8-869c-5d4ead3b16ab-combined-ca-bundle\") pod \"swift-proxy-776bbffb49-bdzz5\" (UID: \"465dc520-34ae-41b8-869c-5d4ead3b16ab\") " pod="openstack/swift-proxy-776bbffb49-bdzz5" Oct 08 18:34:53 crc kubenswrapper[4859]: I1008 18:34:53.138380 4859 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/465dc520-34ae-41b8-869c-5d4ead3b16ab-internal-tls-certs\") pod \"swift-proxy-776bbffb49-bdzz5\" (UID: \"465dc520-34ae-41b8-869c-5d4ead3b16ab\") " pod="openstack/swift-proxy-776bbffb49-bdzz5" Oct 08 18:34:53 crc kubenswrapper[4859]: I1008 18:34:53.140240 4859 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/465dc520-34ae-41b8-869c-5d4ead3b16ab-etc-swift\") pod \"swift-proxy-776bbffb49-bdzz5\" (UID: \"465dc520-34ae-41b8-869c-5d4ead3b16ab\") " pod="openstack/swift-proxy-776bbffb49-bdzz5" Oct 08 18:34:53 crc kubenswrapper[4859]: I1008 18:34:53.140432 4859 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/465dc520-34ae-41b8-869c-5d4ead3b16ab-config-data\") pod \"swift-proxy-776bbffb49-bdzz5\" (UID: \"465dc520-34ae-41b8-869c-5d4ead3b16ab\") " pod="openstack/swift-proxy-776bbffb49-bdzz5" Oct 08 18:34:53 crc kubenswrapper[4859]: I1008 18:34:53.148024 4859 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-6qnzw\" (UniqueName: \"kubernetes.io/projected/465dc520-34ae-41b8-869c-5d4ead3b16ab-kube-api-access-6qnzw\") pod \"swift-proxy-776bbffb49-bdzz5\" (UID: \"465dc520-34ae-41b8-869c-5d4ead3b16ab\") " pod="openstack/swift-proxy-776bbffb49-bdzz5" Oct 08 18:34:53 crc kubenswrapper[4859]: I1008 18:34:53.236521 4859 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/swift-proxy-776bbffb49-bdzz5" Oct 08 18:34:53 crc kubenswrapper[4859]: I1008 18:34:53.594764 4859 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-scheduler-0"] Oct 08 18:34:53 crc kubenswrapper[4859]: W1008 18:34:53.608920 4859 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podf20f7045_6768_4824_9295_6e5ba74da387.slice/crio-890b12d9653a4ed35c379e0bb7f6372a8934c3f6b8495fcd4790ab64ccb3ef7f WatchSource:0}: Error finding container 890b12d9653a4ed35c379e0bb7f6372a8934c3f6b8495fcd4790ab64ccb3ef7f: Status 404 returned error can't find the container with id 890b12d9653a4ed35c379e0bb7f6372a8934c3f6b8495fcd4790ab64ccb3ef7f Oct 08 18:34:53 crc kubenswrapper[4859]: I1008 18:34:53.737274 4859 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-scheduler-0" event={"ID":"f20f7045-6768-4824-9295-6e5ba74da387","Type":"ContainerStarted","Data":"890b12d9653a4ed35c379e0bb7f6372a8934c3f6b8495fcd4790ab64ccb3ef7f"} Oct 08 18:34:53 crc kubenswrapper[4859]: I1008 18:34:53.794227 4859 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/swift-proxy-776bbffb49-bdzz5"] Oct 08 18:34:54 crc kubenswrapper[4859]: I1008 18:34:54.482330 4859 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="085aabba-1f1f-4037-b5cf-036d5b519b02" path="/var/lib/kubelet/pods/085aabba-1f1f-4037-b5cf-036d5b519b02/volumes" Oct 08 18:34:54 crc kubenswrapper[4859]: I1008 18:34:54.649223 4859 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Oct 08 18:34:54 crc kubenswrapper[4859]: I1008 18:34:54.649501 4859 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="7ad402c0-68c6-4c69-b1ad-0fd0d5c69dff" containerName="ceilometer-central-agent" containerID="cri-o://00b2cd92e3299a8a3ff2ab92560ece56a4acf64fd255489ac76be208726ef44c" gracePeriod=30 Oct 08 18:34:54 crc kubenswrapper[4859]: I1008 18:34:54.649650 4859 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="7ad402c0-68c6-4c69-b1ad-0fd0d5c69dff" containerName="proxy-httpd" containerID="cri-o://ec6a14d0c9861094bc0a93ff618867a826b6ddf9021a5b4ed58abcf176899896" gracePeriod=30 Oct 08 18:34:54 crc kubenswrapper[4859]: I1008 18:34:54.649708 4859 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="7ad402c0-68c6-4c69-b1ad-0fd0d5c69dff" containerName="sg-core" containerID="cri-o://e5ecb40eb581178fdd97b35c9692c88e1ad16ec8e8acf8286e1cf25aa9b4268c" gracePeriod=30 Oct 08 18:34:54 crc kubenswrapper[4859]: I1008 18:34:54.649739 4859 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="7ad402c0-68c6-4c69-b1ad-0fd0d5c69dff" containerName="ceilometer-notification-agent" containerID="cri-o://bd3b5549c52a699ff8bf5a14fe7e680fd8f522b93cfaf90e614d985a95dc1393" gracePeriod=30 Oct 08 18:34:54 crc kubenswrapper[4859]: I1008 18:34:54.667829 4859 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/ceilometer-0" Oct 08 18:34:54 crc kubenswrapper[4859]: I1008 18:34:54.766379 4859 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-proxy-776bbffb49-bdzz5" event={"ID":"465dc520-34ae-41b8-869c-5d4ead3b16ab","Type":"ContainerStarted","Data":"c95549f0c3ca6fb78a5395ecc03cadbe4a3fbd0b734b115326da7c9cbef8c452"} Oct 08 18:34:54 crc kubenswrapper[4859]: I1008 18:34:54.766675 4859 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-proxy-776bbffb49-bdzz5" event={"ID":"465dc520-34ae-41b8-869c-5d4ead3b16ab","Type":"ContainerStarted","Data":"c06efa88ecad5caf8186e09c2f2adf0ba9ee99a9bb0d92069336ed703d0705c2"} Oct 08 18:34:54 crc kubenswrapper[4859]: I1008 18:34:54.766717 4859 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/swift-proxy-776bbffb49-bdzz5" Oct 08 18:34:54 crc kubenswrapper[4859]: I1008 18:34:54.766728 4859 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/swift-proxy-776bbffb49-bdzz5" Oct 08 18:34:54 crc kubenswrapper[4859]: I1008 18:34:54.766738 4859 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-proxy-776bbffb49-bdzz5" event={"ID":"465dc520-34ae-41b8-869c-5d4ead3b16ab","Type":"ContainerStarted","Data":"e9e57fc61821a14c40fe03584c4b93bed7b0bcb92e57de3d6215557d3be467d9"} Oct 08 18:34:54 crc kubenswrapper[4859]: I1008 18:34:54.773976 4859 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-scheduler-0" event={"ID":"f20f7045-6768-4824-9295-6e5ba74da387","Type":"ContainerStarted","Data":"1f18b0aba11f8d00ecce281256fae25578b969b3e3d6d5dcbe4d4eb0121e7b3e"} Oct 08 18:34:54 crc kubenswrapper[4859]: I1008 18:34:54.806883 4859 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/swift-proxy-776bbffb49-bdzz5" podStartSLOduration=2.806866983 podStartE2EDuration="2.806866983s" podCreationTimestamp="2025-10-08 18:34:52 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-08 18:34:54.799408763 +0000 UTC m=+1065.046248142" watchObservedRunningTime="2025-10-08 18:34:54.806866983 +0000 UTC m=+1065.053706362" Oct 08 18:34:55 crc kubenswrapper[4859]: I1008 18:34:55.798566 4859 generic.go:334] "Generic (PLEG): container finished" podID="7ad402c0-68c6-4c69-b1ad-0fd0d5c69dff" containerID="ec6a14d0c9861094bc0a93ff618867a826b6ddf9021a5b4ed58abcf176899896" exitCode=0 Oct 08 18:34:55 crc kubenswrapper[4859]: I1008 18:34:55.798874 4859 generic.go:334] "Generic (PLEG): container finished" podID="7ad402c0-68c6-4c69-b1ad-0fd0d5c69dff" containerID="e5ecb40eb581178fdd97b35c9692c88e1ad16ec8e8acf8286e1cf25aa9b4268c" exitCode=2 Oct 08 18:34:55 crc kubenswrapper[4859]: I1008 18:34:55.798884 4859 generic.go:334] "Generic (PLEG): container finished" podID="7ad402c0-68c6-4c69-b1ad-0fd0d5c69dff" containerID="00b2cd92e3299a8a3ff2ab92560ece56a4acf64fd255489ac76be208726ef44c" exitCode=0 Oct 08 18:34:55 crc kubenswrapper[4859]: I1008 18:34:55.798638 4859 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"7ad402c0-68c6-4c69-b1ad-0fd0d5c69dff","Type":"ContainerDied","Data":"ec6a14d0c9861094bc0a93ff618867a826b6ddf9021a5b4ed58abcf176899896"} Oct 08 18:34:55 crc kubenswrapper[4859]: I1008 18:34:55.798952 4859 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"7ad402c0-68c6-4c69-b1ad-0fd0d5c69dff","Type":"ContainerDied","Data":"e5ecb40eb581178fdd97b35c9692c88e1ad16ec8e8acf8286e1cf25aa9b4268c"} Oct 08 18:34:55 crc kubenswrapper[4859]: I1008 18:34:55.798966 4859 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"7ad402c0-68c6-4c69-b1ad-0fd0d5c69dff","Type":"ContainerDied","Data":"00b2cd92e3299a8a3ff2ab92560ece56a4acf64fd255489ac76be208726ef44c"} Oct 08 18:34:55 crc kubenswrapper[4859]: I1008 18:34:55.801290 4859 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-scheduler-0" event={"ID":"f20f7045-6768-4824-9295-6e5ba74da387","Type":"ContainerStarted","Data":"f53e7e31071f99c249e2eb20f6176b42d44d2bf9037f559f3c69451e23e7e95d"} Oct 08 18:34:55 crc kubenswrapper[4859]: I1008 18:34:55.840734 4859 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/cinder-scheduler-0" podStartSLOduration=3.84067437 podStartE2EDuration="3.84067437s" podCreationTimestamp="2025-10-08 18:34:52 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-08 18:34:55.827270944 +0000 UTC m=+1066.074110323" watchObservedRunningTime="2025-10-08 18:34:55.84067437 +0000 UTC m=+1066.087513749" Oct 08 18:34:56 crc kubenswrapper[4859]: I1008 18:34:56.816618 4859 generic.go:334] "Generic (PLEG): container finished" podID="7ad402c0-68c6-4c69-b1ad-0fd0d5c69dff" containerID="bd3b5549c52a699ff8bf5a14fe7e680fd8f522b93cfaf90e614d985a95dc1393" exitCode=0 Oct 08 18:34:56 crc kubenswrapper[4859]: I1008 18:34:56.816737 4859 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"7ad402c0-68c6-4c69-b1ad-0fd0d5c69dff","Type":"ContainerDied","Data":"bd3b5549c52a699ff8bf5a14fe7e680fd8f522b93cfaf90e614d985a95dc1393"} Oct 08 18:34:58 crc kubenswrapper[4859]: I1008 18:34:58.096644 4859 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/cinder-scheduler-0" Oct 08 18:34:58 crc kubenswrapper[4859]: I1008 18:34:58.438070 4859 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-api-db-create-2t4xd"] Oct 08 18:34:58 crc kubenswrapper[4859]: I1008 18:34:58.439191 4859 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-db-create-2t4xd" Oct 08 18:34:58 crc kubenswrapper[4859]: I1008 18:34:58.447401 4859 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-db-create-2t4xd"] Oct 08 18:34:58 crc kubenswrapper[4859]: I1008 18:34:58.536327 4859 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell0-db-create-drcrc"] Oct 08 18:34:58 crc kubenswrapper[4859]: I1008 18:34:58.539050 4859 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-db-create-drcrc" Oct 08 18:34:58 crc kubenswrapper[4859]: I1008 18:34:58.545947 4859 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-db-create-drcrc"] Oct 08 18:34:58 crc kubenswrapper[4859]: I1008 18:34:58.549229 4859 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-p8sfk\" (UniqueName: \"kubernetes.io/projected/3d62765f-7919-4338-b4c7-abc341837b2e-kube-api-access-p8sfk\") pod \"nova-api-db-create-2t4xd\" (UID: \"3d62765f-7919-4338-b4c7-abc341837b2e\") " pod="openstack/nova-api-db-create-2t4xd" Oct 08 18:34:58 crc kubenswrapper[4859]: I1008 18:34:58.637309 4859 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell1-db-create-h7mqh"] Oct 08 18:34:58 crc kubenswrapper[4859]: I1008 18:34:58.638421 4859 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-db-create-h7mqh" Oct 08 18:34:58 crc kubenswrapper[4859]: I1008 18:34:58.647277 4859 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-db-create-h7mqh"] Oct 08 18:34:58 crc kubenswrapper[4859]: I1008 18:34:58.654044 4859 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-p8sfk\" (UniqueName: \"kubernetes.io/projected/3d62765f-7919-4338-b4c7-abc341837b2e-kube-api-access-p8sfk\") pod \"nova-api-db-create-2t4xd\" (UID: \"3d62765f-7919-4338-b4c7-abc341837b2e\") " pod="openstack/nova-api-db-create-2t4xd" Oct 08 18:34:58 crc kubenswrapper[4859]: I1008 18:34:58.654165 4859 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-4pc2p\" (UniqueName: \"kubernetes.io/projected/fa8367eb-d900-45d2-90c0-4401b7d9cc89-kube-api-access-4pc2p\") pod \"nova-cell0-db-create-drcrc\" (UID: \"fa8367eb-d900-45d2-90c0-4401b7d9cc89\") " pod="openstack/nova-cell0-db-create-drcrc" Oct 08 18:34:58 crc kubenswrapper[4859]: I1008 18:34:58.699527 4859 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-p8sfk\" (UniqueName: \"kubernetes.io/projected/3d62765f-7919-4338-b4c7-abc341837b2e-kube-api-access-p8sfk\") pod \"nova-api-db-create-2t4xd\" (UID: \"3d62765f-7919-4338-b4c7-abc341837b2e\") " pod="openstack/nova-api-db-create-2t4xd" Oct 08 18:34:58 crc kubenswrapper[4859]: I1008 18:34:58.756012 4859 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-4pc2p\" (UniqueName: \"kubernetes.io/projected/fa8367eb-d900-45d2-90c0-4401b7d9cc89-kube-api-access-4pc2p\") pod \"nova-cell0-db-create-drcrc\" (UID: \"fa8367eb-d900-45d2-90c0-4401b7d9cc89\") " pod="openstack/nova-cell0-db-create-drcrc" Oct 08 18:34:58 crc kubenswrapper[4859]: I1008 18:34:58.756148 4859 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-rbzlv\" (UniqueName: \"kubernetes.io/projected/d6886827-bf3f-4530-be76-a21a757a4c7f-kube-api-access-rbzlv\") pod \"nova-cell1-db-create-h7mqh\" (UID: \"d6886827-bf3f-4530-be76-a21a757a4c7f\") " pod="openstack/nova-cell1-db-create-h7mqh" Oct 08 18:34:58 crc kubenswrapper[4859]: I1008 18:34:58.774559 4859 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-4pc2p\" (UniqueName: \"kubernetes.io/projected/fa8367eb-d900-45d2-90c0-4401b7d9cc89-kube-api-access-4pc2p\") pod \"nova-cell0-db-create-drcrc\" (UID: \"fa8367eb-d900-45d2-90c0-4401b7d9cc89\") " pod="openstack/nova-cell0-db-create-drcrc" Oct 08 18:34:58 crc kubenswrapper[4859]: I1008 18:34:58.782094 4859 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-db-create-2t4xd" Oct 08 18:34:58 crc kubenswrapper[4859]: I1008 18:34:58.857643 4859 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rbzlv\" (UniqueName: \"kubernetes.io/projected/d6886827-bf3f-4530-be76-a21a757a4c7f-kube-api-access-rbzlv\") pod \"nova-cell1-db-create-h7mqh\" (UID: \"d6886827-bf3f-4530-be76-a21a757a4c7f\") " pod="openstack/nova-cell1-db-create-h7mqh" Oct 08 18:34:58 crc kubenswrapper[4859]: I1008 18:34:58.864107 4859 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-db-create-drcrc" Oct 08 18:34:58 crc kubenswrapper[4859]: I1008 18:34:58.875389 4859 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rbzlv\" (UniqueName: \"kubernetes.io/projected/d6886827-bf3f-4530-be76-a21a757a4c7f-kube-api-access-rbzlv\") pod \"nova-cell1-db-create-h7mqh\" (UID: \"d6886827-bf3f-4530-be76-a21a757a4c7f\") " pod="openstack/nova-cell1-db-create-h7mqh" Oct 08 18:34:58 crc kubenswrapper[4859]: I1008 18:34:58.956469 4859 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-db-create-h7mqh" Oct 08 18:35:00 crc kubenswrapper[4859]: I1008 18:35:00.700744 4859 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/kube-state-metrics-0"] Oct 08 18:35:00 crc kubenswrapper[4859]: I1008 18:35:00.701719 4859 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/kube-state-metrics-0" podUID="54254e29-9af5-4b28-8664-be427c4c4f3a" containerName="kube-state-metrics" containerID="cri-o://428e4f197e15b371774b895f204f030f0983940198b04559e4433b1c1cf15bb3" gracePeriod=30 Oct 08 18:35:00 crc kubenswrapper[4859]: I1008 18:35:00.882800 4859 generic.go:334] "Generic (PLEG): container finished" podID="54254e29-9af5-4b28-8664-be427c4c4f3a" containerID="428e4f197e15b371774b895f204f030f0983940198b04559e4433b1c1cf15bb3" exitCode=2 Oct 08 18:35:00 crc kubenswrapper[4859]: I1008 18:35:00.882849 4859 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/kube-state-metrics-0" event={"ID":"54254e29-9af5-4b28-8664-be427c4c4f3a","Type":"ContainerDied","Data":"428e4f197e15b371774b895f204f030f0983940198b04559e4433b1c1cf15bb3"} Oct 08 18:35:01 crc kubenswrapper[4859]: I1008 18:35:01.501660 4859 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/ceilometer-0" podUID="7ad402c0-68c6-4c69-b1ad-0fd0d5c69dff" containerName="proxy-httpd" probeResult="failure" output="Get \"http://10.217.0.162:3000/\": dial tcp 10.217.0.162:3000: connect: connection refused" Oct 08 18:35:02 crc kubenswrapper[4859]: I1008 18:35:02.308287 4859 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/horizon-7d58687c8-fhctx" podUID="1a92487e-3309-41c4-8c82-cd57f517ab4b" containerName="horizon" probeResult="failure" output="Get \"https://10.217.0.147:8443/dashboard/auth/login/?next=/dashboard/\": dial tcp 10.217.0.147:8443: connect: connection refused" Oct 08 18:35:02 crc kubenswrapper[4859]: I1008 18:35:02.308493 4859 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/horizon-7d58687c8-fhctx" Oct 08 18:35:03 crc kubenswrapper[4859]: I1008 18:35:03.242794 4859 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/swift-proxy-776bbffb49-bdzz5" Oct 08 18:35:03 crc kubenswrapper[4859]: I1008 18:35:03.246894 4859 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/swift-proxy-776bbffb49-bdzz5" Oct 08 18:35:03 crc kubenswrapper[4859]: I1008 18:35:03.382470 4859 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/cinder-scheduler-0" Oct 08 18:35:04 crc kubenswrapper[4859]: I1008 18:35:04.163428 4859 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-db-create-h7mqh"] Oct 08 18:35:04 crc kubenswrapper[4859]: I1008 18:35:04.283383 4859 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-db-create-drcrc"] Oct 08 18:35:04 crc kubenswrapper[4859]: W1008 18:35:04.288008 4859 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podfa8367eb_d900_45d2_90c0_4401b7d9cc89.slice/crio-b6adb6cae27b5fa6417ef31cc9f9c3e65f413021437028d6147f4300167e8bd0 WatchSource:0}: Error finding container b6adb6cae27b5fa6417ef31cc9f9c3e65f413021437028d6147f4300167e8bd0: Status 404 returned error can't find the container with id b6adb6cae27b5fa6417ef31cc9f9c3e65f413021437028d6147f4300167e8bd0 Oct 08 18:35:04 crc kubenswrapper[4859]: I1008 18:35:04.377378 4859 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/kube-state-metrics-0" Oct 08 18:35:04 crc kubenswrapper[4859]: I1008 18:35:04.394873 4859 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-db-create-2t4xd"] Oct 08 18:35:04 crc kubenswrapper[4859]: I1008 18:35:04.486510 4859 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-sh54v\" (UniqueName: \"kubernetes.io/projected/54254e29-9af5-4b28-8664-be427c4c4f3a-kube-api-access-sh54v\") pod \"54254e29-9af5-4b28-8664-be427c4c4f3a\" (UID: \"54254e29-9af5-4b28-8664-be427c4c4f3a\") " Oct 08 18:35:04 crc kubenswrapper[4859]: I1008 18:35:04.498888 4859 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/54254e29-9af5-4b28-8664-be427c4c4f3a-kube-api-access-sh54v" (OuterVolumeSpecName: "kube-api-access-sh54v") pod "54254e29-9af5-4b28-8664-be427c4c4f3a" (UID: "54254e29-9af5-4b28-8664-be427c4c4f3a"). InnerVolumeSpecName "kube-api-access-sh54v". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 08 18:35:04 crc kubenswrapper[4859]: I1008 18:35:04.589538 4859 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-sh54v\" (UniqueName: \"kubernetes.io/projected/54254e29-9af5-4b28-8664-be427c4c4f3a-kube-api-access-sh54v\") on node \"crc\" DevicePath \"\"" Oct 08 18:35:04 crc kubenswrapper[4859]: I1008 18:35:04.920444 4859 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-db-create-h7mqh" event={"ID":"d6886827-bf3f-4530-be76-a21a757a4c7f","Type":"ContainerStarted","Data":"588f280ec9b7b2a0e1c06aeef03cd5b1088654dadd0f63f256a0d0a4de198212"} Oct 08 18:35:04 crc kubenswrapper[4859]: I1008 18:35:04.920861 4859 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-db-create-h7mqh" event={"ID":"d6886827-bf3f-4530-be76-a21a757a4c7f","Type":"ContainerStarted","Data":"4bda2ea27b1f9efb0721b1baec5204f689ef2077226b7a587f3305821562b94a"} Oct 08 18:35:04 crc kubenswrapper[4859]: I1008 18:35:04.922102 4859 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/kube-state-metrics-0" event={"ID":"54254e29-9af5-4b28-8664-be427c4c4f3a","Type":"ContainerDied","Data":"ab6edf00f86dbbad4da7ade8f9bcefe588dafcbb7c10cc34277b15e6639b3874"} Oct 08 18:35:04 crc kubenswrapper[4859]: I1008 18:35:04.922125 4859 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/kube-state-metrics-0" Oct 08 18:35:04 crc kubenswrapper[4859]: I1008 18:35:04.922183 4859 scope.go:117] "RemoveContainer" containerID="428e4f197e15b371774b895f204f030f0983940198b04559e4433b1c1cf15bb3" Oct 08 18:35:04 crc kubenswrapper[4859]: I1008 18:35:04.923433 4859 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-db-create-2t4xd" event={"ID":"3d62765f-7919-4338-b4c7-abc341837b2e","Type":"ContainerStarted","Data":"db410a22e0d0d0475a68cb37ff024bd3d3082bc523f44a930e66e252f41b521b"} Oct 08 18:35:04 crc kubenswrapper[4859]: I1008 18:35:04.926275 4859 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-db-create-drcrc" event={"ID":"fa8367eb-d900-45d2-90c0-4401b7d9cc89","Type":"ContainerStarted","Data":"b6adb6cae27b5fa6417ef31cc9f9c3e65f413021437028d6147f4300167e8bd0"} Oct 08 18:35:04 crc kubenswrapper[4859]: I1008 18:35:04.961740 4859 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/kube-state-metrics-0"] Oct 08 18:35:04 crc kubenswrapper[4859]: I1008 18:35:04.968785 4859 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/kube-state-metrics-0"] Oct 08 18:35:04 crc kubenswrapper[4859]: I1008 18:35:04.981714 4859 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/kube-state-metrics-0"] Oct 08 18:35:04 crc kubenswrapper[4859]: E1008 18:35:04.982073 4859 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="54254e29-9af5-4b28-8664-be427c4c4f3a" containerName="kube-state-metrics" Oct 08 18:35:04 crc kubenswrapper[4859]: I1008 18:35:04.982089 4859 state_mem.go:107] "Deleted CPUSet assignment" podUID="54254e29-9af5-4b28-8664-be427c4c4f3a" containerName="kube-state-metrics" Oct 08 18:35:04 crc kubenswrapper[4859]: I1008 18:35:04.982274 4859 memory_manager.go:354] "RemoveStaleState removing state" podUID="54254e29-9af5-4b28-8664-be427c4c4f3a" containerName="kube-state-metrics" Oct 08 18:35:04 crc kubenswrapper[4859]: I1008 18:35:04.982878 4859 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/kube-state-metrics-0" Oct 08 18:35:04 crc kubenswrapper[4859]: I1008 18:35:04.986161 4859 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"kube-state-metrics-tls-config" Oct 08 18:35:04 crc kubenswrapper[4859]: I1008 18:35:04.986457 4859 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-kube-state-metrics-svc" Oct 08 18:35:04 crc kubenswrapper[4859]: I1008 18:35:04.992542 4859 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/kube-state-metrics-0"] Oct 08 18:35:05 crc kubenswrapper[4859]: I1008 18:35:05.098598 4859 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-state-metrics-tls-certs\" (UniqueName: \"kubernetes.io/secret/3652de4f-2bd2-4fe2-b138-63dd20b42a61-kube-state-metrics-tls-certs\") pod \"kube-state-metrics-0\" (UID: \"3652de4f-2bd2-4fe2-b138-63dd20b42a61\") " pod="openstack/kube-state-metrics-0" Oct 08 18:35:05 crc kubenswrapper[4859]: I1008 18:35:05.098761 4859 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-bvdks\" (UniqueName: \"kubernetes.io/projected/3652de4f-2bd2-4fe2-b138-63dd20b42a61-kube-api-access-bvdks\") pod \"kube-state-metrics-0\" (UID: \"3652de4f-2bd2-4fe2-b138-63dd20b42a61\") " pod="openstack/kube-state-metrics-0" Oct 08 18:35:05 crc kubenswrapper[4859]: I1008 18:35:05.098828 4859 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3652de4f-2bd2-4fe2-b138-63dd20b42a61-combined-ca-bundle\") pod \"kube-state-metrics-0\" (UID: \"3652de4f-2bd2-4fe2-b138-63dd20b42a61\") " pod="openstack/kube-state-metrics-0" Oct 08 18:35:05 crc kubenswrapper[4859]: I1008 18:35:05.098858 4859 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-state-metrics-tls-config\" (UniqueName: \"kubernetes.io/secret/3652de4f-2bd2-4fe2-b138-63dd20b42a61-kube-state-metrics-tls-config\") pod \"kube-state-metrics-0\" (UID: \"3652de4f-2bd2-4fe2-b138-63dd20b42a61\") " pod="openstack/kube-state-metrics-0" Oct 08 18:35:05 crc kubenswrapper[4859]: I1008 18:35:05.200343 4859 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3652de4f-2bd2-4fe2-b138-63dd20b42a61-combined-ca-bundle\") pod \"kube-state-metrics-0\" (UID: \"3652de4f-2bd2-4fe2-b138-63dd20b42a61\") " pod="openstack/kube-state-metrics-0" Oct 08 18:35:05 crc kubenswrapper[4859]: I1008 18:35:05.200411 4859 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-state-metrics-tls-config\" (UniqueName: \"kubernetes.io/secret/3652de4f-2bd2-4fe2-b138-63dd20b42a61-kube-state-metrics-tls-config\") pod \"kube-state-metrics-0\" (UID: \"3652de4f-2bd2-4fe2-b138-63dd20b42a61\") " pod="openstack/kube-state-metrics-0" Oct 08 18:35:05 crc kubenswrapper[4859]: I1008 18:35:05.200465 4859 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-state-metrics-tls-certs\" (UniqueName: \"kubernetes.io/secret/3652de4f-2bd2-4fe2-b138-63dd20b42a61-kube-state-metrics-tls-certs\") pod \"kube-state-metrics-0\" (UID: \"3652de4f-2bd2-4fe2-b138-63dd20b42a61\") " pod="openstack/kube-state-metrics-0" Oct 08 18:35:05 crc kubenswrapper[4859]: I1008 18:35:05.200586 4859 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-bvdks\" (UniqueName: \"kubernetes.io/projected/3652de4f-2bd2-4fe2-b138-63dd20b42a61-kube-api-access-bvdks\") pod \"kube-state-metrics-0\" (UID: \"3652de4f-2bd2-4fe2-b138-63dd20b42a61\") " pod="openstack/kube-state-metrics-0" Oct 08 18:35:05 crc kubenswrapper[4859]: I1008 18:35:05.205920 4859 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-state-metrics-tls-config\" (UniqueName: \"kubernetes.io/secret/3652de4f-2bd2-4fe2-b138-63dd20b42a61-kube-state-metrics-tls-config\") pod \"kube-state-metrics-0\" (UID: \"3652de4f-2bd2-4fe2-b138-63dd20b42a61\") " pod="openstack/kube-state-metrics-0" Oct 08 18:35:05 crc kubenswrapper[4859]: I1008 18:35:05.206025 4859 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3652de4f-2bd2-4fe2-b138-63dd20b42a61-combined-ca-bundle\") pod \"kube-state-metrics-0\" (UID: \"3652de4f-2bd2-4fe2-b138-63dd20b42a61\") " pod="openstack/kube-state-metrics-0" Oct 08 18:35:05 crc kubenswrapper[4859]: I1008 18:35:05.209577 4859 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-state-metrics-tls-certs\" (UniqueName: \"kubernetes.io/secret/3652de4f-2bd2-4fe2-b138-63dd20b42a61-kube-state-metrics-tls-certs\") pod \"kube-state-metrics-0\" (UID: \"3652de4f-2bd2-4fe2-b138-63dd20b42a61\") " pod="openstack/kube-state-metrics-0" Oct 08 18:35:05 crc kubenswrapper[4859]: I1008 18:35:05.227538 4859 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-bvdks\" (UniqueName: \"kubernetes.io/projected/3652de4f-2bd2-4fe2-b138-63dd20b42a61-kube-api-access-bvdks\") pod \"kube-state-metrics-0\" (UID: \"3652de4f-2bd2-4fe2-b138-63dd20b42a61\") " pod="openstack/kube-state-metrics-0" Oct 08 18:35:05 crc kubenswrapper[4859]: I1008 18:35:05.308883 4859 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/kube-state-metrics-0" Oct 08 18:35:05 crc kubenswrapper[4859]: I1008 18:35:05.448437 4859 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Oct 08 18:35:05 crc kubenswrapper[4859]: I1008 18:35:05.505409 4859 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/7ad402c0-68c6-4c69-b1ad-0fd0d5c69dff-config-data\") pod \"7ad402c0-68c6-4c69-b1ad-0fd0d5c69dff\" (UID: \"7ad402c0-68c6-4c69-b1ad-0fd0d5c69dff\") " Oct 08 18:35:05 crc kubenswrapper[4859]: I1008 18:35:05.505527 4859 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/7ad402c0-68c6-4c69-b1ad-0fd0d5c69dff-scripts\") pod \"7ad402c0-68c6-4c69-b1ad-0fd0d5c69dff\" (UID: \"7ad402c0-68c6-4c69-b1ad-0fd0d5c69dff\") " Oct 08 18:35:05 crc kubenswrapper[4859]: I1008 18:35:05.505703 4859 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/7ad402c0-68c6-4c69-b1ad-0fd0d5c69dff-run-httpd\") pod \"7ad402c0-68c6-4c69-b1ad-0fd0d5c69dff\" (UID: \"7ad402c0-68c6-4c69-b1ad-0fd0d5c69dff\") " Oct 08 18:35:05 crc kubenswrapper[4859]: I1008 18:35:05.505746 4859 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/7ad402c0-68c6-4c69-b1ad-0fd0d5c69dff-log-httpd\") pod \"7ad402c0-68c6-4c69-b1ad-0fd0d5c69dff\" (UID: \"7ad402c0-68c6-4c69-b1ad-0fd0d5c69dff\") " Oct 08 18:35:05 crc kubenswrapper[4859]: I1008 18:35:05.505793 4859 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-lzscf\" (UniqueName: \"kubernetes.io/projected/7ad402c0-68c6-4c69-b1ad-0fd0d5c69dff-kube-api-access-lzscf\") pod \"7ad402c0-68c6-4c69-b1ad-0fd0d5c69dff\" (UID: \"7ad402c0-68c6-4c69-b1ad-0fd0d5c69dff\") " Oct 08 18:35:05 crc kubenswrapper[4859]: I1008 18:35:05.505828 4859 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7ad402c0-68c6-4c69-b1ad-0fd0d5c69dff-combined-ca-bundle\") pod \"7ad402c0-68c6-4c69-b1ad-0fd0d5c69dff\" (UID: \"7ad402c0-68c6-4c69-b1ad-0fd0d5c69dff\") " Oct 08 18:35:05 crc kubenswrapper[4859]: I1008 18:35:05.505923 4859 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/7ad402c0-68c6-4c69-b1ad-0fd0d5c69dff-sg-core-conf-yaml\") pod \"7ad402c0-68c6-4c69-b1ad-0fd0d5c69dff\" (UID: \"7ad402c0-68c6-4c69-b1ad-0fd0d5c69dff\") " Oct 08 18:35:05 crc kubenswrapper[4859]: I1008 18:35:05.507696 4859 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/7ad402c0-68c6-4c69-b1ad-0fd0d5c69dff-log-httpd" (OuterVolumeSpecName: "log-httpd") pod "7ad402c0-68c6-4c69-b1ad-0fd0d5c69dff" (UID: "7ad402c0-68c6-4c69-b1ad-0fd0d5c69dff"). InnerVolumeSpecName "log-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 08 18:35:05 crc kubenswrapper[4859]: I1008 18:35:05.508099 4859 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/7ad402c0-68c6-4c69-b1ad-0fd0d5c69dff-run-httpd" (OuterVolumeSpecName: "run-httpd") pod "7ad402c0-68c6-4c69-b1ad-0fd0d5c69dff" (UID: "7ad402c0-68c6-4c69-b1ad-0fd0d5c69dff"). InnerVolumeSpecName "run-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 08 18:35:05 crc kubenswrapper[4859]: I1008 18:35:05.511961 4859 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/7ad402c0-68c6-4c69-b1ad-0fd0d5c69dff-scripts" (OuterVolumeSpecName: "scripts") pod "7ad402c0-68c6-4c69-b1ad-0fd0d5c69dff" (UID: "7ad402c0-68c6-4c69-b1ad-0fd0d5c69dff"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 08 18:35:05 crc kubenswrapper[4859]: I1008 18:35:05.532830 4859 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/7ad402c0-68c6-4c69-b1ad-0fd0d5c69dff-kube-api-access-lzscf" (OuterVolumeSpecName: "kube-api-access-lzscf") pod "7ad402c0-68c6-4c69-b1ad-0fd0d5c69dff" (UID: "7ad402c0-68c6-4c69-b1ad-0fd0d5c69dff"). InnerVolumeSpecName "kube-api-access-lzscf". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 08 18:35:05 crc kubenswrapper[4859]: I1008 18:35:05.546252 4859 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/7ad402c0-68c6-4c69-b1ad-0fd0d5c69dff-sg-core-conf-yaml" (OuterVolumeSpecName: "sg-core-conf-yaml") pod "7ad402c0-68c6-4c69-b1ad-0fd0d5c69dff" (UID: "7ad402c0-68c6-4c69-b1ad-0fd0d5c69dff"). InnerVolumeSpecName "sg-core-conf-yaml". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 08 18:35:05 crc kubenswrapper[4859]: I1008 18:35:05.599133 4859 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/7ad402c0-68c6-4c69-b1ad-0fd0d5c69dff-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "7ad402c0-68c6-4c69-b1ad-0fd0d5c69dff" (UID: "7ad402c0-68c6-4c69-b1ad-0fd0d5c69dff"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 08 18:35:05 crc kubenswrapper[4859]: I1008 18:35:05.608793 4859 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/7ad402c0-68c6-4c69-b1ad-0fd0d5c69dff-scripts\") on node \"crc\" DevicePath \"\"" Oct 08 18:35:05 crc kubenswrapper[4859]: I1008 18:35:05.609035 4859 reconciler_common.go:293] "Volume detached for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/7ad402c0-68c6-4c69-b1ad-0fd0d5c69dff-run-httpd\") on node \"crc\" DevicePath \"\"" Oct 08 18:35:05 crc kubenswrapper[4859]: I1008 18:35:05.609113 4859 reconciler_common.go:293] "Volume detached for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/7ad402c0-68c6-4c69-b1ad-0fd0d5c69dff-log-httpd\") on node \"crc\" DevicePath \"\"" Oct 08 18:35:05 crc kubenswrapper[4859]: I1008 18:35:05.609194 4859 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-lzscf\" (UniqueName: \"kubernetes.io/projected/7ad402c0-68c6-4c69-b1ad-0fd0d5c69dff-kube-api-access-lzscf\") on node \"crc\" DevicePath \"\"" Oct 08 18:35:05 crc kubenswrapper[4859]: I1008 18:35:05.609272 4859 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7ad402c0-68c6-4c69-b1ad-0fd0d5c69dff-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 08 18:35:05 crc kubenswrapper[4859]: I1008 18:35:05.609624 4859 reconciler_common.go:293] "Volume detached for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/7ad402c0-68c6-4c69-b1ad-0fd0d5c69dff-sg-core-conf-yaml\") on node \"crc\" DevicePath \"\"" Oct 08 18:35:05 crc kubenswrapper[4859]: E1008 18:35:05.627585 4859 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/podified-antelope-centos9/openstack-openstackclient@sha256:b8bff6857fec93c3c1521f1a8c23de21bcb86fc0f960972e81f6c3f95d4185be" Oct 08 18:35:05 crc kubenswrapper[4859]: E1008 18:35:05.627790 4859 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:openstackclient,Image:quay.io/podified-antelope-centos9/openstack-openstackclient@sha256:b8bff6857fec93c3c1521f1a8c23de21bcb86fc0f960972e81f6c3f95d4185be,Command:[/bin/sleep],Args:[infinity],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:CONFIG_HASH,Value:nd8hbch87hb8h78hdh5dch9ch694h55h564h5d4h66ch5f9h64h687h5d4h58fhfdh88h669h5cbh548h96h5h57h57dh5fchf4h585hbfh67bq,ValueFrom:nil,},EnvVar{Name:OS_CLOUD,Value:default,ValueFrom:nil,},EnvVar{Name:PROMETHEUS_HOST,Value:metric-storage-prometheus.openstack.svc,ValueFrom:nil,},EnvVar{Name:PROMETHEUS_PORT,Value:9090,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:openstack-config,ReadOnly:false,MountPath:/home/cloud-admin/.config/openstack/clouds.yaml,SubPath:clouds.yaml,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:openstack-config-secret,ReadOnly:false,MountPath:/home/cloud-admin/.config/openstack/secure.yaml,SubPath:secure.yaml,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:openstack-config-secret,ReadOnly:false,MountPath:/home/cloud-admin/cloudrc,SubPath:cloudrc,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:combined-ca-bundle,ReadOnly:true,MountPath:/etc/pki/ca-trust/extracted/pem/tls-ca-bundle.pem,SubPath:tls-ca-bundle.pem,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-w95f9,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*42401,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:*42401,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod openstackclient_openstack(1a631d33-4c0c-46c9-a3d5-4d9dc9f6de84): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Oct 08 18:35:05 crc kubenswrapper[4859]: E1008 18:35:05.629128 4859 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"openstackclient\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack/openstackclient" podUID="1a631d33-4c0c-46c9-a3d5-4d9dc9f6de84" Oct 08 18:35:05 crc kubenswrapper[4859]: I1008 18:35:05.698953 4859 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/7ad402c0-68c6-4c69-b1ad-0fd0d5c69dff-config-data" (OuterVolumeSpecName: "config-data") pod "7ad402c0-68c6-4c69-b1ad-0fd0d5c69dff" (UID: "7ad402c0-68c6-4c69-b1ad-0fd0d5c69dff"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 08 18:35:05 crc kubenswrapper[4859]: I1008 18:35:05.710834 4859 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/7ad402c0-68c6-4c69-b1ad-0fd0d5c69dff-config-data\") on node \"crc\" DevicePath \"\"" Oct 08 18:35:05 crc kubenswrapper[4859]: I1008 18:35:05.816200 4859 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/kube-state-metrics-0"] Oct 08 18:35:05 crc kubenswrapper[4859]: W1008 18:35:05.819438 4859 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod3652de4f_2bd2_4fe2_b138_63dd20b42a61.slice/crio-4bb931c1bd27bf0fe3b9a7c9a43c32ce048bc510a9389db486680e2bd8806508 WatchSource:0}: Error finding container 4bb931c1bd27bf0fe3b9a7c9a43c32ce048bc510a9389db486680e2bd8806508: Status 404 returned error can't find the container with id 4bb931c1bd27bf0fe3b9a7c9a43c32ce048bc510a9389db486680e2bd8806508 Oct 08 18:35:05 crc kubenswrapper[4859]: I1008 18:35:05.937804 4859 generic.go:334] "Generic (PLEG): container finished" podID="3d62765f-7919-4338-b4c7-abc341837b2e" containerID="60e9bb3a4e7c180c91ff79487b9adc18ab0fc78b4c8d1eba75682ae1b02a173f" exitCode=0 Oct 08 18:35:05 crc kubenswrapper[4859]: I1008 18:35:05.937875 4859 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-db-create-2t4xd" event={"ID":"3d62765f-7919-4338-b4c7-abc341837b2e","Type":"ContainerDied","Data":"60e9bb3a4e7c180c91ff79487b9adc18ab0fc78b4c8d1eba75682ae1b02a173f"} Oct 08 18:35:05 crc kubenswrapper[4859]: I1008 18:35:05.939928 4859 generic.go:334] "Generic (PLEG): container finished" podID="fa8367eb-d900-45d2-90c0-4401b7d9cc89" containerID="0986204b68ccd47e6d46c16f7ae883db00acc964ad708ef4f4f5b7c984d726f7" exitCode=0 Oct 08 18:35:05 crc kubenswrapper[4859]: I1008 18:35:05.940032 4859 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-db-create-drcrc" event={"ID":"fa8367eb-d900-45d2-90c0-4401b7d9cc89","Type":"ContainerDied","Data":"0986204b68ccd47e6d46c16f7ae883db00acc964ad708ef4f4f5b7c984d726f7"} Oct 08 18:35:05 crc kubenswrapper[4859]: I1008 18:35:05.942284 4859 generic.go:334] "Generic (PLEG): container finished" podID="d6886827-bf3f-4530-be76-a21a757a4c7f" containerID="588f280ec9b7b2a0e1c06aeef03cd5b1088654dadd0f63f256a0d0a4de198212" exitCode=0 Oct 08 18:35:05 crc kubenswrapper[4859]: I1008 18:35:05.942328 4859 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-db-create-h7mqh" event={"ID":"d6886827-bf3f-4530-be76-a21a757a4c7f","Type":"ContainerDied","Data":"588f280ec9b7b2a0e1c06aeef03cd5b1088654dadd0f63f256a0d0a4de198212"} Oct 08 18:35:05 crc kubenswrapper[4859]: I1008 18:35:05.945619 4859 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"7ad402c0-68c6-4c69-b1ad-0fd0d5c69dff","Type":"ContainerDied","Data":"5005b74fa8e9eb369194edce4800584323343c0e0f5e64adbc4cb80fb161fbb5"} Oct 08 18:35:05 crc kubenswrapper[4859]: I1008 18:35:05.945656 4859 scope.go:117] "RemoveContainer" containerID="ec6a14d0c9861094bc0a93ff618867a826b6ddf9021a5b4ed58abcf176899896" Oct 08 18:35:05 crc kubenswrapper[4859]: I1008 18:35:05.945776 4859 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Oct 08 18:35:05 crc kubenswrapper[4859]: I1008 18:35:05.951872 4859 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/kube-state-metrics-0" event={"ID":"3652de4f-2bd2-4fe2-b138-63dd20b42a61","Type":"ContainerStarted","Data":"4bb931c1bd27bf0fe3b9a7c9a43c32ce048bc510a9389db486680e2bd8806508"} Oct 08 18:35:05 crc kubenswrapper[4859]: E1008 18:35:05.957851 4859 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"openstackclient\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/podified-antelope-centos9/openstack-openstackclient@sha256:b8bff6857fec93c3c1521f1a8c23de21bcb86fc0f960972e81f6c3f95d4185be\\\"\"" pod="openstack/openstackclient" podUID="1a631d33-4c0c-46c9-a3d5-4d9dc9f6de84" Oct 08 18:35:06 crc kubenswrapper[4859]: I1008 18:35:06.045844 4859 scope.go:117] "RemoveContainer" containerID="e5ecb40eb581178fdd97b35c9692c88e1ad16ec8e8acf8286e1cf25aa9b4268c" Oct 08 18:35:06 crc kubenswrapper[4859]: I1008 18:35:06.069807 4859 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Oct 08 18:35:06 crc kubenswrapper[4859]: I1008 18:35:06.073113 4859 scope.go:117] "RemoveContainer" containerID="bd3b5549c52a699ff8bf5a14fe7e680fd8f522b93cfaf90e614d985a95dc1393" Oct 08 18:35:06 crc kubenswrapper[4859]: I1008 18:35:06.087265 4859 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/ceilometer-0"] Oct 08 18:35:06 crc kubenswrapper[4859]: I1008 18:35:06.096397 4859 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ceilometer-0"] Oct 08 18:35:06 crc kubenswrapper[4859]: E1008 18:35:06.097122 4859 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7ad402c0-68c6-4c69-b1ad-0fd0d5c69dff" containerName="ceilometer-central-agent" Oct 08 18:35:06 crc kubenswrapper[4859]: I1008 18:35:06.097146 4859 state_mem.go:107] "Deleted CPUSet assignment" podUID="7ad402c0-68c6-4c69-b1ad-0fd0d5c69dff" containerName="ceilometer-central-agent" Oct 08 18:35:06 crc kubenswrapper[4859]: E1008 18:35:06.097168 4859 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7ad402c0-68c6-4c69-b1ad-0fd0d5c69dff" containerName="sg-core" Oct 08 18:35:06 crc kubenswrapper[4859]: I1008 18:35:06.097174 4859 state_mem.go:107] "Deleted CPUSet assignment" podUID="7ad402c0-68c6-4c69-b1ad-0fd0d5c69dff" containerName="sg-core" Oct 08 18:35:06 crc kubenswrapper[4859]: E1008 18:35:06.097184 4859 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7ad402c0-68c6-4c69-b1ad-0fd0d5c69dff" containerName="ceilometer-notification-agent" Oct 08 18:35:06 crc kubenswrapper[4859]: I1008 18:35:06.097192 4859 state_mem.go:107] "Deleted CPUSet assignment" podUID="7ad402c0-68c6-4c69-b1ad-0fd0d5c69dff" containerName="ceilometer-notification-agent" Oct 08 18:35:06 crc kubenswrapper[4859]: E1008 18:35:06.097202 4859 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7ad402c0-68c6-4c69-b1ad-0fd0d5c69dff" containerName="proxy-httpd" Oct 08 18:35:06 crc kubenswrapper[4859]: I1008 18:35:06.097207 4859 state_mem.go:107] "Deleted CPUSet assignment" podUID="7ad402c0-68c6-4c69-b1ad-0fd0d5c69dff" containerName="proxy-httpd" Oct 08 18:35:06 crc kubenswrapper[4859]: I1008 18:35:06.097397 4859 memory_manager.go:354] "RemoveStaleState removing state" podUID="7ad402c0-68c6-4c69-b1ad-0fd0d5c69dff" containerName="ceilometer-notification-agent" Oct 08 18:35:06 crc kubenswrapper[4859]: I1008 18:35:06.097409 4859 memory_manager.go:354] "RemoveStaleState removing state" podUID="7ad402c0-68c6-4c69-b1ad-0fd0d5c69dff" containerName="ceilometer-central-agent" Oct 08 18:35:06 crc kubenswrapper[4859]: I1008 18:35:06.097419 4859 memory_manager.go:354] "RemoveStaleState removing state" podUID="7ad402c0-68c6-4c69-b1ad-0fd0d5c69dff" containerName="sg-core" Oct 08 18:35:06 crc kubenswrapper[4859]: I1008 18:35:06.097428 4859 memory_manager.go:354] "RemoveStaleState removing state" podUID="7ad402c0-68c6-4c69-b1ad-0fd0d5c69dff" containerName="proxy-httpd" Oct 08 18:35:06 crc kubenswrapper[4859]: I1008 18:35:06.122319 4859 scope.go:117] "RemoveContainer" containerID="00b2cd92e3299a8a3ff2ab92560ece56a4acf64fd255489ac76be208726ef44c" Oct 08 18:35:06 crc kubenswrapper[4859]: I1008 18:35:06.128320 4859 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Oct 08 18:35:06 crc kubenswrapper[4859]: I1008 18:35:06.128478 4859 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Oct 08 18:35:06 crc kubenswrapper[4859]: I1008 18:35:06.130618 4859 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-scripts" Oct 08 18:35:06 crc kubenswrapper[4859]: I1008 18:35:06.131393 4859 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-config-data" Oct 08 18:35:06 crc kubenswrapper[4859]: I1008 18:35:06.132043 4859 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-ceilometer-internal-svc" Oct 08 18:35:06 crc kubenswrapper[4859]: I1008 18:35:06.225969 4859 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/d95266a3-07ae-4c7f-9cc7-af1a632290cc-run-httpd\") pod \"ceilometer-0\" (UID: \"d95266a3-07ae-4c7f-9cc7-af1a632290cc\") " pod="openstack/ceilometer-0" Oct 08 18:35:06 crc kubenswrapper[4859]: I1008 18:35:06.226046 4859 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/d95266a3-07ae-4c7f-9cc7-af1a632290cc-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"d95266a3-07ae-4c7f-9cc7-af1a632290cc\") " pod="openstack/ceilometer-0" Oct 08 18:35:06 crc kubenswrapper[4859]: I1008 18:35:06.226092 4859 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-h2mfw\" (UniqueName: \"kubernetes.io/projected/d95266a3-07ae-4c7f-9cc7-af1a632290cc-kube-api-access-h2mfw\") pod \"ceilometer-0\" (UID: \"d95266a3-07ae-4c7f-9cc7-af1a632290cc\") " pod="openstack/ceilometer-0" Oct 08 18:35:06 crc kubenswrapper[4859]: I1008 18:35:06.226189 4859 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/d95266a3-07ae-4c7f-9cc7-af1a632290cc-scripts\") pod \"ceilometer-0\" (UID: \"d95266a3-07ae-4c7f-9cc7-af1a632290cc\") " pod="openstack/ceilometer-0" Oct 08 18:35:06 crc kubenswrapper[4859]: I1008 18:35:06.226271 4859 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d95266a3-07ae-4c7f-9cc7-af1a632290cc-config-data\") pod \"ceilometer-0\" (UID: \"d95266a3-07ae-4c7f-9cc7-af1a632290cc\") " pod="openstack/ceilometer-0" Oct 08 18:35:06 crc kubenswrapper[4859]: I1008 18:35:06.226299 4859 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d95266a3-07ae-4c7f-9cc7-af1a632290cc-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"d95266a3-07ae-4c7f-9cc7-af1a632290cc\") " pod="openstack/ceilometer-0" Oct 08 18:35:06 crc kubenswrapper[4859]: I1008 18:35:06.226336 4859 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/d95266a3-07ae-4c7f-9cc7-af1a632290cc-log-httpd\") pod \"ceilometer-0\" (UID: \"d95266a3-07ae-4c7f-9cc7-af1a632290cc\") " pod="openstack/ceilometer-0" Oct 08 18:35:06 crc kubenswrapper[4859]: I1008 18:35:06.226355 4859 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/d95266a3-07ae-4c7f-9cc7-af1a632290cc-ceilometer-tls-certs\") pod \"ceilometer-0\" (UID: \"d95266a3-07ae-4c7f-9cc7-af1a632290cc\") " pod="openstack/ceilometer-0" Oct 08 18:35:06 crc kubenswrapper[4859]: I1008 18:35:06.328899 4859 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/d95266a3-07ae-4c7f-9cc7-af1a632290cc-run-httpd\") pod \"ceilometer-0\" (UID: \"d95266a3-07ae-4c7f-9cc7-af1a632290cc\") " pod="openstack/ceilometer-0" Oct 08 18:35:06 crc kubenswrapper[4859]: I1008 18:35:06.329219 4859 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/d95266a3-07ae-4c7f-9cc7-af1a632290cc-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"d95266a3-07ae-4c7f-9cc7-af1a632290cc\") " pod="openstack/ceilometer-0" Oct 08 18:35:06 crc kubenswrapper[4859]: I1008 18:35:06.329345 4859 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-h2mfw\" (UniqueName: \"kubernetes.io/projected/d95266a3-07ae-4c7f-9cc7-af1a632290cc-kube-api-access-h2mfw\") pod \"ceilometer-0\" (UID: \"d95266a3-07ae-4c7f-9cc7-af1a632290cc\") " pod="openstack/ceilometer-0" Oct 08 18:35:06 crc kubenswrapper[4859]: I1008 18:35:06.330171 4859 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/d95266a3-07ae-4c7f-9cc7-af1a632290cc-run-httpd\") pod \"ceilometer-0\" (UID: \"d95266a3-07ae-4c7f-9cc7-af1a632290cc\") " pod="openstack/ceilometer-0" Oct 08 18:35:06 crc kubenswrapper[4859]: I1008 18:35:06.331846 4859 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/d95266a3-07ae-4c7f-9cc7-af1a632290cc-scripts\") pod \"ceilometer-0\" (UID: \"d95266a3-07ae-4c7f-9cc7-af1a632290cc\") " pod="openstack/ceilometer-0" Oct 08 18:35:06 crc kubenswrapper[4859]: I1008 18:35:06.332085 4859 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d95266a3-07ae-4c7f-9cc7-af1a632290cc-config-data\") pod \"ceilometer-0\" (UID: \"d95266a3-07ae-4c7f-9cc7-af1a632290cc\") " pod="openstack/ceilometer-0" Oct 08 18:35:06 crc kubenswrapper[4859]: I1008 18:35:06.332164 4859 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d95266a3-07ae-4c7f-9cc7-af1a632290cc-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"d95266a3-07ae-4c7f-9cc7-af1a632290cc\") " pod="openstack/ceilometer-0" Oct 08 18:35:06 crc kubenswrapper[4859]: I1008 18:35:06.332262 4859 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/d95266a3-07ae-4c7f-9cc7-af1a632290cc-log-httpd\") pod \"ceilometer-0\" (UID: \"d95266a3-07ae-4c7f-9cc7-af1a632290cc\") " pod="openstack/ceilometer-0" Oct 08 18:35:06 crc kubenswrapper[4859]: I1008 18:35:06.332310 4859 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/d95266a3-07ae-4c7f-9cc7-af1a632290cc-ceilometer-tls-certs\") pod \"ceilometer-0\" (UID: \"d95266a3-07ae-4c7f-9cc7-af1a632290cc\") " pod="openstack/ceilometer-0" Oct 08 18:35:06 crc kubenswrapper[4859]: I1008 18:35:06.333856 4859 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/d95266a3-07ae-4c7f-9cc7-af1a632290cc-log-httpd\") pod \"ceilometer-0\" (UID: \"d95266a3-07ae-4c7f-9cc7-af1a632290cc\") " pod="openstack/ceilometer-0" Oct 08 18:35:06 crc kubenswrapper[4859]: I1008 18:35:06.335857 4859 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/d95266a3-07ae-4c7f-9cc7-af1a632290cc-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"d95266a3-07ae-4c7f-9cc7-af1a632290cc\") " pod="openstack/ceilometer-0" Oct 08 18:35:06 crc kubenswrapper[4859]: I1008 18:35:06.336744 4859 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/d95266a3-07ae-4c7f-9cc7-af1a632290cc-scripts\") pod \"ceilometer-0\" (UID: \"d95266a3-07ae-4c7f-9cc7-af1a632290cc\") " pod="openstack/ceilometer-0" Oct 08 18:35:06 crc kubenswrapper[4859]: I1008 18:35:06.337201 4859 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d95266a3-07ae-4c7f-9cc7-af1a632290cc-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"d95266a3-07ae-4c7f-9cc7-af1a632290cc\") " pod="openstack/ceilometer-0" Oct 08 18:35:06 crc kubenswrapper[4859]: I1008 18:35:06.338237 4859 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d95266a3-07ae-4c7f-9cc7-af1a632290cc-config-data\") pod \"ceilometer-0\" (UID: \"d95266a3-07ae-4c7f-9cc7-af1a632290cc\") " pod="openstack/ceilometer-0" Oct 08 18:35:06 crc kubenswrapper[4859]: I1008 18:35:06.339796 4859 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/d95266a3-07ae-4c7f-9cc7-af1a632290cc-ceilometer-tls-certs\") pod \"ceilometer-0\" (UID: \"d95266a3-07ae-4c7f-9cc7-af1a632290cc\") " pod="openstack/ceilometer-0" Oct 08 18:35:06 crc kubenswrapper[4859]: I1008 18:35:06.354893 4859 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-h2mfw\" (UniqueName: \"kubernetes.io/projected/d95266a3-07ae-4c7f-9cc7-af1a632290cc-kube-api-access-h2mfw\") pod \"ceilometer-0\" (UID: \"d95266a3-07ae-4c7f-9cc7-af1a632290cc\") " pod="openstack/ceilometer-0" Oct 08 18:35:06 crc kubenswrapper[4859]: I1008 18:35:06.453623 4859 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Oct 08 18:35:06 crc kubenswrapper[4859]: I1008 18:35:06.517754 4859 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="54254e29-9af5-4b28-8664-be427c4c4f3a" path="/var/lib/kubelet/pods/54254e29-9af5-4b28-8664-be427c4c4f3a/volumes" Oct 08 18:35:06 crc kubenswrapper[4859]: I1008 18:35:06.518756 4859 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="7ad402c0-68c6-4c69-b1ad-0fd0d5c69dff" path="/var/lib/kubelet/pods/7ad402c0-68c6-4c69-b1ad-0fd0d5c69dff/volumes" Oct 08 18:35:06 crc kubenswrapper[4859]: I1008 18:35:06.858091 4859 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-7d58687c8-fhctx" Oct 08 18:35:06 crc kubenswrapper[4859]: I1008 18:35:06.951350 4859 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/1a92487e-3309-41c4-8c82-cd57f517ab4b-horizon-secret-key\") pod \"1a92487e-3309-41c4-8c82-cd57f517ab4b\" (UID: \"1a92487e-3309-41c4-8c82-cd57f517ab4b\") " Oct 08 18:35:06 crc kubenswrapper[4859]: I1008 18:35:06.951450 4859 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"horizon-tls-certs\" (UniqueName: \"kubernetes.io/secret/1a92487e-3309-41c4-8c82-cd57f517ab4b-horizon-tls-certs\") pod \"1a92487e-3309-41c4-8c82-cd57f517ab4b\" (UID: \"1a92487e-3309-41c4-8c82-cd57f517ab4b\") " Oct 08 18:35:06 crc kubenswrapper[4859]: I1008 18:35:06.951517 4859 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-7q7tg\" (UniqueName: \"kubernetes.io/projected/1a92487e-3309-41c4-8c82-cd57f517ab4b-kube-api-access-7q7tg\") pod \"1a92487e-3309-41c4-8c82-cd57f517ab4b\" (UID: \"1a92487e-3309-41c4-8c82-cd57f517ab4b\") " Oct 08 18:35:06 crc kubenswrapper[4859]: I1008 18:35:06.951582 4859 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1a92487e-3309-41c4-8c82-cd57f517ab4b-combined-ca-bundle\") pod \"1a92487e-3309-41c4-8c82-cd57f517ab4b\" (UID: \"1a92487e-3309-41c4-8c82-cd57f517ab4b\") " Oct 08 18:35:06 crc kubenswrapper[4859]: I1008 18:35:06.951704 4859 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/1a92487e-3309-41c4-8c82-cd57f517ab4b-config-data\") pod \"1a92487e-3309-41c4-8c82-cd57f517ab4b\" (UID: \"1a92487e-3309-41c4-8c82-cd57f517ab4b\") " Oct 08 18:35:06 crc kubenswrapper[4859]: I1008 18:35:06.951753 4859 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/1a92487e-3309-41c4-8c82-cd57f517ab4b-scripts\") pod \"1a92487e-3309-41c4-8c82-cd57f517ab4b\" (UID: \"1a92487e-3309-41c4-8c82-cd57f517ab4b\") " Oct 08 18:35:06 crc kubenswrapper[4859]: I1008 18:35:06.951799 4859 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/1a92487e-3309-41c4-8c82-cd57f517ab4b-logs\") pod \"1a92487e-3309-41c4-8c82-cd57f517ab4b\" (UID: \"1a92487e-3309-41c4-8c82-cd57f517ab4b\") " Oct 08 18:35:06 crc kubenswrapper[4859]: I1008 18:35:06.952493 4859 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/1a92487e-3309-41c4-8c82-cd57f517ab4b-logs" (OuterVolumeSpecName: "logs") pod "1a92487e-3309-41c4-8c82-cd57f517ab4b" (UID: "1a92487e-3309-41c4-8c82-cd57f517ab4b"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 08 18:35:06 crc kubenswrapper[4859]: I1008 18:35:06.957929 4859 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/1a92487e-3309-41c4-8c82-cd57f517ab4b-kube-api-access-7q7tg" (OuterVolumeSpecName: "kube-api-access-7q7tg") pod "1a92487e-3309-41c4-8c82-cd57f517ab4b" (UID: "1a92487e-3309-41c4-8c82-cd57f517ab4b"). InnerVolumeSpecName "kube-api-access-7q7tg". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 08 18:35:06 crc kubenswrapper[4859]: I1008 18:35:06.958526 4859 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1a92487e-3309-41c4-8c82-cd57f517ab4b-horizon-secret-key" (OuterVolumeSpecName: "horizon-secret-key") pod "1a92487e-3309-41c4-8c82-cd57f517ab4b" (UID: "1a92487e-3309-41c4-8c82-cd57f517ab4b"). InnerVolumeSpecName "horizon-secret-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 08 18:35:06 crc kubenswrapper[4859]: I1008 18:35:06.978138 4859 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/kube-state-metrics-0" event={"ID":"3652de4f-2bd2-4fe2-b138-63dd20b42a61","Type":"ContainerStarted","Data":"e3ecfa11520021e8b20a5fce3696b02412f1d4089a0a1baba0954e7c1a92481c"} Oct 08 18:35:06 crc kubenswrapper[4859]: I1008 18:35:06.978680 4859 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/kube-state-metrics-0" Oct 08 18:35:06 crc kubenswrapper[4859]: I1008 18:35:06.979158 4859 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1a92487e-3309-41c4-8c82-cd57f517ab4b-scripts" (OuterVolumeSpecName: "scripts") pod "1a92487e-3309-41c4-8c82-cd57f517ab4b" (UID: "1a92487e-3309-41c4-8c82-cd57f517ab4b"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 08 18:35:06 crc kubenswrapper[4859]: I1008 18:35:06.981392 4859 generic.go:334] "Generic (PLEG): container finished" podID="1a92487e-3309-41c4-8c82-cd57f517ab4b" containerID="fb897bd998eaff60715c21b1cde92b421d7bf9ae9f4afad9a1ed7b20f7cfd7b3" exitCode=137 Oct 08 18:35:06 crc kubenswrapper[4859]: I1008 18:35:06.981609 4859 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-7d58687c8-fhctx" Oct 08 18:35:06 crc kubenswrapper[4859]: I1008 18:35:06.982097 4859 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-7d58687c8-fhctx" event={"ID":"1a92487e-3309-41c4-8c82-cd57f517ab4b","Type":"ContainerDied","Data":"fb897bd998eaff60715c21b1cde92b421d7bf9ae9f4afad9a1ed7b20f7cfd7b3"} Oct 08 18:35:06 crc kubenswrapper[4859]: I1008 18:35:06.982132 4859 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-7d58687c8-fhctx" event={"ID":"1a92487e-3309-41c4-8c82-cd57f517ab4b","Type":"ContainerDied","Data":"c546fc2ab05e42f55afac42efd39b6705f14b4294bf04b9895522878307c7ccf"} Oct 08 18:35:06 crc kubenswrapper[4859]: I1008 18:35:06.982151 4859 scope.go:117] "RemoveContainer" containerID="648c622ce9eb0bdf3d9c43d8313db6636c307abe8d840bde30cf40e0039e99c7" Oct 08 18:35:07 crc kubenswrapper[4859]: I1008 18:35:07.024472 4859 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1a92487e-3309-41c4-8c82-cd57f517ab4b-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "1a92487e-3309-41c4-8c82-cd57f517ab4b" (UID: "1a92487e-3309-41c4-8c82-cd57f517ab4b"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 08 18:35:07 crc kubenswrapper[4859]: I1008 18:35:07.040348 4859 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1a92487e-3309-41c4-8c82-cd57f517ab4b-config-data" (OuterVolumeSpecName: "config-data") pod "1a92487e-3309-41c4-8c82-cd57f517ab4b" (UID: "1a92487e-3309-41c4-8c82-cd57f517ab4b"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 08 18:35:07 crc kubenswrapper[4859]: I1008 18:35:07.065615 4859 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1a92487e-3309-41c4-8c82-cd57f517ab4b-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 08 18:35:07 crc kubenswrapper[4859]: I1008 18:35:07.065662 4859 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/1a92487e-3309-41c4-8c82-cd57f517ab4b-config-data\") on node \"crc\" DevicePath \"\"" Oct 08 18:35:07 crc kubenswrapper[4859]: I1008 18:35:07.065674 4859 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/1a92487e-3309-41c4-8c82-cd57f517ab4b-scripts\") on node \"crc\" DevicePath \"\"" Oct 08 18:35:07 crc kubenswrapper[4859]: I1008 18:35:07.065704 4859 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/1a92487e-3309-41c4-8c82-cd57f517ab4b-logs\") on node \"crc\" DevicePath \"\"" Oct 08 18:35:07 crc kubenswrapper[4859]: I1008 18:35:07.065716 4859 reconciler_common.go:293] "Volume detached for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/1a92487e-3309-41c4-8c82-cd57f517ab4b-horizon-secret-key\") on node \"crc\" DevicePath \"\"" Oct 08 18:35:07 crc kubenswrapper[4859]: I1008 18:35:07.065727 4859 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-7q7tg\" (UniqueName: \"kubernetes.io/projected/1a92487e-3309-41c4-8c82-cd57f517ab4b-kube-api-access-7q7tg\") on node \"crc\" DevicePath \"\"" Oct 08 18:35:07 crc kubenswrapper[4859]: I1008 18:35:07.065902 4859 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1a92487e-3309-41c4-8c82-cd57f517ab4b-horizon-tls-certs" (OuterVolumeSpecName: "horizon-tls-certs") pod "1a92487e-3309-41c4-8c82-cd57f517ab4b" (UID: "1a92487e-3309-41c4-8c82-cd57f517ab4b"). InnerVolumeSpecName "horizon-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 08 18:35:07 crc kubenswrapper[4859]: I1008 18:35:07.089849 4859 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/kube-state-metrics-0" podStartSLOduration=2.724384362 podStartE2EDuration="3.089821444s" podCreationTimestamp="2025-10-08 18:35:04 +0000 UTC" firstStartedPulling="2025-10-08 18:35:05.821297469 +0000 UTC m=+1076.068136848" lastFinishedPulling="2025-10-08 18:35:06.186734551 +0000 UTC m=+1076.433573930" observedRunningTime="2025-10-08 18:35:07.024891695 +0000 UTC m=+1077.271731094" watchObservedRunningTime="2025-10-08 18:35:07.089821444 +0000 UTC m=+1077.336660823" Oct 08 18:35:07 crc kubenswrapper[4859]: I1008 18:35:07.100755 4859 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Oct 08 18:35:07 crc kubenswrapper[4859]: W1008 18:35:07.121330 4859 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podd95266a3_07ae_4c7f_9cc7_af1a632290cc.slice/crio-b77ebe586efb4734f6b2c8db2f0a9fc962bfe811a7a2985eb1ee4fe9dd8ed4b3 WatchSource:0}: Error finding container b77ebe586efb4734f6b2c8db2f0a9fc962bfe811a7a2985eb1ee4fe9dd8ed4b3: Status 404 returned error can't find the container with id b77ebe586efb4734f6b2c8db2f0a9fc962bfe811a7a2985eb1ee4fe9dd8ed4b3 Oct 08 18:35:07 crc kubenswrapper[4859]: I1008 18:35:07.166962 4859 reconciler_common.go:293] "Volume detached for volume \"horizon-tls-certs\" (UniqueName: \"kubernetes.io/secret/1a92487e-3309-41c4-8c82-cd57f517ab4b-horizon-tls-certs\") on node \"crc\" DevicePath \"\"" Oct 08 18:35:07 crc kubenswrapper[4859]: I1008 18:35:07.322262 4859 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/horizon-7d58687c8-fhctx"] Oct 08 18:35:07 crc kubenswrapper[4859]: I1008 18:35:07.335055 4859 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/horizon-7d58687c8-fhctx"] Oct 08 18:35:07 crc kubenswrapper[4859]: I1008 18:35:07.372746 4859 scope.go:117] "RemoveContainer" containerID="fb897bd998eaff60715c21b1cde92b421d7bf9ae9f4afad9a1ed7b20f7cfd7b3" Oct 08 18:35:07 crc kubenswrapper[4859]: I1008 18:35:07.390258 4859 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-db-create-2t4xd" Oct 08 18:35:07 crc kubenswrapper[4859]: I1008 18:35:07.397749 4859 scope.go:117] "RemoveContainer" containerID="648c622ce9eb0bdf3d9c43d8313db6636c307abe8d840bde30cf40e0039e99c7" Oct 08 18:35:07 crc kubenswrapper[4859]: E1008 18:35:07.398802 4859 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"648c622ce9eb0bdf3d9c43d8313db6636c307abe8d840bde30cf40e0039e99c7\": container with ID starting with 648c622ce9eb0bdf3d9c43d8313db6636c307abe8d840bde30cf40e0039e99c7 not found: ID does not exist" containerID="648c622ce9eb0bdf3d9c43d8313db6636c307abe8d840bde30cf40e0039e99c7" Oct 08 18:35:07 crc kubenswrapper[4859]: I1008 18:35:07.398844 4859 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"648c622ce9eb0bdf3d9c43d8313db6636c307abe8d840bde30cf40e0039e99c7"} err="failed to get container status \"648c622ce9eb0bdf3d9c43d8313db6636c307abe8d840bde30cf40e0039e99c7\": rpc error: code = NotFound desc = could not find container \"648c622ce9eb0bdf3d9c43d8313db6636c307abe8d840bde30cf40e0039e99c7\": container with ID starting with 648c622ce9eb0bdf3d9c43d8313db6636c307abe8d840bde30cf40e0039e99c7 not found: ID does not exist" Oct 08 18:35:07 crc kubenswrapper[4859]: I1008 18:35:07.398869 4859 scope.go:117] "RemoveContainer" containerID="fb897bd998eaff60715c21b1cde92b421d7bf9ae9f4afad9a1ed7b20f7cfd7b3" Oct 08 18:35:07 crc kubenswrapper[4859]: E1008 18:35:07.401358 4859 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"fb897bd998eaff60715c21b1cde92b421d7bf9ae9f4afad9a1ed7b20f7cfd7b3\": container with ID starting with fb897bd998eaff60715c21b1cde92b421d7bf9ae9f4afad9a1ed7b20f7cfd7b3 not found: ID does not exist" containerID="fb897bd998eaff60715c21b1cde92b421d7bf9ae9f4afad9a1ed7b20f7cfd7b3" Oct 08 18:35:07 crc kubenswrapper[4859]: I1008 18:35:07.401392 4859 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"fb897bd998eaff60715c21b1cde92b421d7bf9ae9f4afad9a1ed7b20f7cfd7b3"} err="failed to get container status \"fb897bd998eaff60715c21b1cde92b421d7bf9ae9f4afad9a1ed7b20f7cfd7b3\": rpc error: code = NotFound desc = could not find container \"fb897bd998eaff60715c21b1cde92b421d7bf9ae9f4afad9a1ed7b20f7cfd7b3\": container with ID starting with fb897bd998eaff60715c21b1cde92b421d7bf9ae9f4afad9a1ed7b20f7cfd7b3 not found: ID does not exist" Oct 08 18:35:07 crc kubenswrapper[4859]: I1008 18:35:07.475153 4859 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-p8sfk\" (UniqueName: \"kubernetes.io/projected/3d62765f-7919-4338-b4c7-abc341837b2e-kube-api-access-p8sfk\") pod \"3d62765f-7919-4338-b4c7-abc341837b2e\" (UID: \"3d62765f-7919-4338-b4c7-abc341837b2e\") " Oct 08 18:35:07 crc kubenswrapper[4859]: I1008 18:35:07.481433 4859 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/3d62765f-7919-4338-b4c7-abc341837b2e-kube-api-access-p8sfk" (OuterVolumeSpecName: "kube-api-access-p8sfk") pod "3d62765f-7919-4338-b4c7-abc341837b2e" (UID: "3d62765f-7919-4338-b4c7-abc341837b2e"). InnerVolumeSpecName "kube-api-access-p8sfk". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 08 18:35:07 crc kubenswrapper[4859]: I1008 18:35:07.581275 4859 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-p8sfk\" (UniqueName: \"kubernetes.io/projected/3d62765f-7919-4338-b4c7-abc341837b2e-kube-api-access-p8sfk\") on node \"crc\" DevicePath \"\"" Oct 08 18:35:07 crc kubenswrapper[4859]: I1008 18:35:07.791062 4859 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-db-create-drcrc" Oct 08 18:35:07 crc kubenswrapper[4859]: I1008 18:35:07.814698 4859 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-db-create-h7mqh" Oct 08 18:35:07 crc kubenswrapper[4859]: I1008 18:35:07.886081 4859 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-4pc2p\" (UniqueName: \"kubernetes.io/projected/fa8367eb-d900-45d2-90c0-4401b7d9cc89-kube-api-access-4pc2p\") pod \"fa8367eb-d900-45d2-90c0-4401b7d9cc89\" (UID: \"fa8367eb-d900-45d2-90c0-4401b7d9cc89\") " Oct 08 18:35:07 crc kubenswrapper[4859]: I1008 18:35:07.886222 4859 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-rbzlv\" (UniqueName: \"kubernetes.io/projected/d6886827-bf3f-4530-be76-a21a757a4c7f-kube-api-access-rbzlv\") pod \"d6886827-bf3f-4530-be76-a21a757a4c7f\" (UID: \"d6886827-bf3f-4530-be76-a21a757a4c7f\") " Oct 08 18:35:07 crc kubenswrapper[4859]: I1008 18:35:07.890968 4859 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/fa8367eb-d900-45d2-90c0-4401b7d9cc89-kube-api-access-4pc2p" (OuterVolumeSpecName: "kube-api-access-4pc2p") pod "fa8367eb-d900-45d2-90c0-4401b7d9cc89" (UID: "fa8367eb-d900-45d2-90c0-4401b7d9cc89"). InnerVolumeSpecName "kube-api-access-4pc2p". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 08 18:35:07 crc kubenswrapper[4859]: I1008 18:35:07.891231 4859 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/d6886827-bf3f-4530-be76-a21a757a4c7f-kube-api-access-rbzlv" (OuterVolumeSpecName: "kube-api-access-rbzlv") pod "d6886827-bf3f-4530-be76-a21a757a4c7f" (UID: "d6886827-bf3f-4530-be76-a21a757a4c7f"). InnerVolumeSpecName "kube-api-access-rbzlv". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 08 18:35:07 crc kubenswrapper[4859]: I1008 18:35:07.989651 4859 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-4pc2p\" (UniqueName: \"kubernetes.io/projected/fa8367eb-d900-45d2-90c0-4401b7d9cc89-kube-api-access-4pc2p\") on node \"crc\" DevicePath \"\"" Oct 08 18:35:07 crc kubenswrapper[4859]: I1008 18:35:07.989711 4859 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-rbzlv\" (UniqueName: \"kubernetes.io/projected/d6886827-bf3f-4530-be76-a21a757a4c7f-kube-api-access-rbzlv\") on node \"crc\" DevicePath \"\"" Oct 08 18:35:08 crc kubenswrapper[4859]: I1008 18:35:08.004109 4859 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-db-create-2t4xd" event={"ID":"3d62765f-7919-4338-b4c7-abc341837b2e","Type":"ContainerDied","Data":"db410a22e0d0d0475a68cb37ff024bd3d3082bc523f44a930e66e252f41b521b"} Oct 08 18:35:08 crc kubenswrapper[4859]: I1008 18:35:08.004357 4859 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="db410a22e0d0d0475a68cb37ff024bd3d3082bc523f44a930e66e252f41b521b" Oct 08 18:35:08 crc kubenswrapper[4859]: I1008 18:35:08.004516 4859 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-db-create-2t4xd" Oct 08 18:35:08 crc kubenswrapper[4859]: I1008 18:35:08.006937 4859 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-db-create-drcrc" Oct 08 18:35:08 crc kubenswrapper[4859]: I1008 18:35:08.006943 4859 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-db-create-drcrc" event={"ID":"fa8367eb-d900-45d2-90c0-4401b7d9cc89","Type":"ContainerDied","Data":"b6adb6cae27b5fa6417ef31cc9f9c3e65f413021437028d6147f4300167e8bd0"} Oct 08 18:35:08 crc kubenswrapper[4859]: I1008 18:35:08.006996 4859 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="b6adb6cae27b5fa6417ef31cc9f9c3e65f413021437028d6147f4300167e8bd0" Oct 08 18:35:08 crc kubenswrapper[4859]: I1008 18:35:08.021075 4859 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"d95266a3-07ae-4c7f-9cc7-af1a632290cc","Type":"ContainerStarted","Data":"9a699a097ec98f3331d90f2e9089ee5a2d6472715e420e0f0ac4c813b0f7c776"} Oct 08 18:35:08 crc kubenswrapper[4859]: I1008 18:35:08.021424 4859 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"d95266a3-07ae-4c7f-9cc7-af1a632290cc","Type":"ContainerStarted","Data":"b77ebe586efb4734f6b2c8db2f0a9fc962bfe811a7a2985eb1ee4fe9dd8ed4b3"} Oct 08 18:35:08 crc kubenswrapper[4859]: I1008 18:35:08.034491 4859 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-db-create-h7mqh" Oct 08 18:35:08 crc kubenswrapper[4859]: I1008 18:35:08.035357 4859 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-db-create-h7mqh" event={"ID":"d6886827-bf3f-4530-be76-a21a757a4c7f","Type":"ContainerDied","Data":"4bda2ea27b1f9efb0721b1baec5204f689ef2077226b7a587f3305821562b94a"} Oct 08 18:35:08 crc kubenswrapper[4859]: I1008 18:35:08.035500 4859 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="4bda2ea27b1f9efb0721b1baec5204f689ef2077226b7a587f3305821562b94a" Oct 08 18:35:08 crc kubenswrapper[4859]: I1008 18:35:08.483863 4859 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="1a92487e-3309-41c4-8c82-cd57f517ab4b" path="/var/lib/kubelet/pods/1a92487e-3309-41c4-8c82-cd57f517ab4b/volumes" Oct 08 18:35:08 crc kubenswrapper[4859]: I1008 18:35:08.671188 4859 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-api-dd49-account-create-cnv2d"] Oct 08 18:35:08 crc kubenswrapper[4859]: E1008 18:35:08.671562 4859 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="fa8367eb-d900-45d2-90c0-4401b7d9cc89" containerName="mariadb-database-create" Oct 08 18:35:08 crc kubenswrapper[4859]: I1008 18:35:08.671579 4859 state_mem.go:107] "Deleted CPUSet assignment" podUID="fa8367eb-d900-45d2-90c0-4401b7d9cc89" containerName="mariadb-database-create" Oct 08 18:35:08 crc kubenswrapper[4859]: E1008 18:35:08.671603 4859 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3d62765f-7919-4338-b4c7-abc341837b2e" containerName="mariadb-database-create" Oct 08 18:35:08 crc kubenswrapper[4859]: I1008 18:35:08.671610 4859 state_mem.go:107] "Deleted CPUSet assignment" podUID="3d62765f-7919-4338-b4c7-abc341837b2e" containerName="mariadb-database-create" Oct 08 18:35:08 crc kubenswrapper[4859]: E1008 18:35:08.671626 4859 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1a92487e-3309-41c4-8c82-cd57f517ab4b" containerName="horizon-log" Oct 08 18:35:08 crc kubenswrapper[4859]: I1008 18:35:08.671632 4859 state_mem.go:107] "Deleted CPUSet assignment" podUID="1a92487e-3309-41c4-8c82-cd57f517ab4b" containerName="horizon-log" Oct 08 18:35:08 crc kubenswrapper[4859]: E1008 18:35:08.671652 4859 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1a92487e-3309-41c4-8c82-cd57f517ab4b" containerName="horizon" Oct 08 18:35:08 crc kubenswrapper[4859]: I1008 18:35:08.671659 4859 state_mem.go:107] "Deleted CPUSet assignment" podUID="1a92487e-3309-41c4-8c82-cd57f517ab4b" containerName="horizon" Oct 08 18:35:08 crc kubenswrapper[4859]: E1008 18:35:08.671675 4859 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d6886827-bf3f-4530-be76-a21a757a4c7f" containerName="mariadb-database-create" Oct 08 18:35:08 crc kubenswrapper[4859]: I1008 18:35:08.671681 4859 state_mem.go:107] "Deleted CPUSet assignment" podUID="d6886827-bf3f-4530-be76-a21a757a4c7f" containerName="mariadb-database-create" Oct 08 18:35:08 crc kubenswrapper[4859]: I1008 18:35:08.671897 4859 memory_manager.go:354] "RemoveStaleState removing state" podUID="1a92487e-3309-41c4-8c82-cd57f517ab4b" containerName="horizon-log" Oct 08 18:35:08 crc kubenswrapper[4859]: I1008 18:35:08.671913 4859 memory_manager.go:354] "RemoveStaleState removing state" podUID="1a92487e-3309-41c4-8c82-cd57f517ab4b" containerName="horizon" Oct 08 18:35:08 crc kubenswrapper[4859]: I1008 18:35:08.671923 4859 memory_manager.go:354] "RemoveStaleState removing state" podUID="3d62765f-7919-4338-b4c7-abc341837b2e" containerName="mariadb-database-create" Oct 08 18:35:08 crc kubenswrapper[4859]: I1008 18:35:08.671935 4859 memory_manager.go:354] "RemoveStaleState removing state" podUID="d6886827-bf3f-4530-be76-a21a757a4c7f" containerName="mariadb-database-create" Oct 08 18:35:08 crc kubenswrapper[4859]: I1008 18:35:08.671950 4859 memory_manager.go:354] "RemoveStaleState removing state" podUID="fa8367eb-d900-45d2-90c0-4401b7d9cc89" containerName="mariadb-database-create" Oct 08 18:35:08 crc kubenswrapper[4859]: I1008 18:35:08.672652 4859 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-dd49-account-create-cnv2d" Oct 08 18:35:08 crc kubenswrapper[4859]: I1008 18:35:08.675050 4859 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-api-db-secret" Oct 08 18:35:08 crc kubenswrapper[4859]: I1008 18:35:08.698469 4859 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-dd49-account-create-cnv2d"] Oct 08 18:35:08 crc kubenswrapper[4859]: I1008 18:35:08.802090 4859 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-7st6z\" (UniqueName: \"kubernetes.io/projected/4ee130b1-a5ee-428e-81f9-e4669ad95b11-kube-api-access-7st6z\") pod \"nova-api-dd49-account-create-cnv2d\" (UID: \"4ee130b1-a5ee-428e-81f9-e4669ad95b11\") " pod="openstack/nova-api-dd49-account-create-cnv2d" Oct 08 18:35:08 crc kubenswrapper[4859]: I1008 18:35:08.903445 4859 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-7st6z\" (UniqueName: \"kubernetes.io/projected/4ee130b1-a5ee-428e-81f9-e4669ad95b11-kube-api-access-7st6z\") pod \"nova-api-dd49-account-create-cnv2d\" (UID: \"4ee130b1-a5ee-428e-81f9-e4669ad95b11\") " pod="openstack/nova-api-dd49-account-create-cnv2d" Oct 08 18:35:08 crc kubenswrapper[4859]: I1008 18:35:08.929511 4859 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-7st6z\" (UniqueName: \"kubernetes.io/projected/4ee130b1-a5ee-428e-81f9-e4669ad95b11-kube-api-access-7st6z\") pod \"nova-api-dd49-account-create-cnv2d\" (UID: \"4ee130b1-a5ee-428e-81f9-e4669ad95b11\") " pod="openstack/nova-api-dd49-account-create-cnv2d" Oct 08 18:35:08 crc kubenswrapper[4859]: I1008 18:35:08.995479 4859 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-dd49-account-create-cnv2d" Oct 08 18:35:09 crc kubenswrapper[4859]: I1008 18:35:09.042515 4859 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"d95266a3-07ae-4c7f-9cc7-af1a632290cc","Type":"ContainerStarted","Data":"83b10b00b3fde51de9b23a21cf635ce9b4b442e8ddab240ca122aee97845c3e0"} Oct 08 18:35:09 crc kubenswrapper[4859]: I1008 18:35:09.509046 4859 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-dd49-account-create-cnv2d"] Oct 08 18:35:09 crc kubenswrapper[4859]: W1008 18:35:09.515955 4859 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod4ee130b1_a5ee_428e_81f9_e4669ad95b11.slice/crio-816f851c1db6e609c36fd3250ebf24fdca6dda096e47b77e1b14ca8a7d0487cf WatchSource:0}: Error finding container 816f851c1db6e609c36fd3250ebf24fdca6dda096e47b77e1b14ca8a7d0487cf: Status 404 returned error can't find the container with id 816f851c1db6e609c36fd3250ebf24fdca6dda096e47b77e1b14ca8a7d0487cf Oct 08 18:35:10 crc kubenswrapper[4859]: I1008 18:35:10.055053 4859 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"d95266a3-07ae-4c7f-9cc7-af1a632290cc","Type":"ContainerStarted","Data":"f4b804fb4923d970b1a669c891e974729f108bea4442714c5bb19f4176d284ba"} Oct 08 18:35:10 crc kubenswrapper[4859]: I1008 18:35:10.056524 4859 generic.go:334] "Generic (PLEG): container finished" podID="4ee130b1-a5ee-428e-81f9-e4669ad95b11" containerID="a3312ac46e8b3275c032896bb6f261fcefd038ae943b940c85e22a90bff4e1d6" exitCode=0 Oct 08 18:35:10 crc kubenswrapper[4859]: I1008 18:35:10.056557 4859 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-dd49-account-create-cnv2d" event={"ID":"4ee130b1-a5ee-428e-81f9-e4669ad95b11","Type":"ContainerDied","Data":"a3312ac46e8b3275c032896bb6f261fcefd038ae943b940c85e22a90bff4e1d6"} Oct 08 18:35:10 crc kubenswrapper[4859]: I1008 18:35:10.056571 4859 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-dd49-account-create-cnv2d" event={"ID":"4ee130b1-a5ee-428e-81f9-e4669ad95b11","Type":"ContainerStarted","Data":"816f851c1db6e609c36fd3250ebf24fdca6dda096e47b77e1b14ca8a7d0487cf"} Oct 08 18:35:10 crc kubenswrapper[4859]: I1008 18:35:10.913703 4859 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-api-0" Oct 08 18:35:11 crc kubenswrapper[4859]: I1008 18:35:11.044980 4859 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/65e6bdd0-9ffe-48d7-9b02-444ce9b204dd-config-data-custom\") pod \"65e6bdd0-9ffe-48d7-9b02-444ce9b204dd\" (UID: \"65e6bdd0-9ffe-48d7-9b02-444ce9b204dd\") " Oct 08 18:35:11 crc kubenswrapper[4859]: I1008 18:35:11.045054 4859 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/65e6bdd0-9ffe-48d7-9b02-444ce9b204dd-combined-ca-bundle\") pod \"65e6bdd0-9ffe-48d7-9b02-444ce9b204dd\" (UID: \"65e6bdd0-9ffe-48d7-9b02-444ce9b204dd\") " Oct 08 18:35:11 crc kubenswrapper[4859]: I1008 18:35:11.045104 4859 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/65e6bdd0-9ffe-48d7-9b02-444ce9b204dd-scripts\") pod \"65e6bdd0-9ffe-48d7-9b02-444ce9b204dd\" (UID: \"65e6bdd0-9ffe-48d7-9b02-444ce9b204dd\") " Oct 08 18:35:11 crc kubenswrapper[4859]: I1008 18:35:11.045132 4859 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-fhgh7\" (UniqueName: \"kubernetes.io/projected/65e6bdd0-9ffe-48d7-9b02-444ce9b204dd-kube-api-access-fhgh7\") pod \"65e6bdd0-9ffe-48d7-9b02-444ce9b204dd\" (UID: \"65e6bdd0-9ffe-48d7-9b02-444ce9b204dd\") " Oct 08 18:35:11 crc kubenswrapper[4859]: I1008 18:35:11.045177 4859 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/65e6bdd0-9ffe-48d7-9b02-444ce9b204dd-etc-machine-id\") pod \"65e6bdd0-9ffe-48d7-9b02-444ce9b204dd\" (UID: \"65e6bdd0-9ffe-48d7-9b02-444ce9b204dd\") " Oct 08 18:35:11 crc kubenswrapper[4859]: I1008 18:35:11.045314 4859 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/65e6bdd0-9ffe-48d7-9b02-444ce9b204dd-logs\") pod \"65e6bdd0-9ffe-48d7-9b02-444ce9b204dd\" (UID: \"65e6bdd0-9ffe-48d7-9b02-444ce9b204dd\") " Oct 08 18:35:11 crc kubenswrapper[4859]: I1008 18:35:11.045893 4859 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/65e6bdd0-9ffe-48d7-9b02-444ce9b204dd-config-data\") pod \"65e6bdd0-9ffe-48d7-9b02-444ce9b204dd\" (UID: \"65e6bdd0-9ffe-48d7-9b02-444ce9b204dd\") " Oct 08 18:35:11 crc kubenswrapper[4859]: I1008 18:35:11.045915 4859 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/65e6bdd0-9ffe-48d7-9b02-444ce9b204dd-etc-machine-id" (OuterVolumeSpecName: "etc-machine-id") pod "65e6bdd0-9ffe-48d7-9b02-444ce9b204dd" (UID: "65e6bdd0-9ffe-48d7-9b02-444ce9b204dd"). InnerVolumeSpecName "etc-machine-id". PluginName "kubernetes.io/host-path", VolumeGidValue "" Oct 08 18:35:11 crc kubenswrapper[4859]: I1008 18:35:11.046701 4859 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/65e6bdd0-9ffe-48d7-9b02-444ce9b204dd-logs" (OuterVolumeSpecName: "logs") pod "65e6bdd0-9ffe-48d7-9b02-444ce9b204dd" (UID: "65e6bdd0-9ffe-48d7-9b02-444ce9b204dd"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 08 18:35:11 crc kubenswrapper[4859]: I1008 18:35:11.046872 4859 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/65e6bdd0-9ffe-48d7-9b02-444ce9b204dd-logs\") on node \"crc\" DevicePath \"\"" Oct 08 18:35:11 crc kubenswrapper[4859]: I1008 18:35:11.046902 4859 reconciler_common.go:293] "Volume detached for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/65e6bdd0-9ffe-48d7-9b02-444ce9b204dd-etc-machine-id\") on node \"crc\" DevicePath \"\"" Oct 08 18:35:11 crc kubenswrapper[4859]: I1008 18:35:11.052027 4859 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/65e6bdd0-9ffe-48d7-9b02-444ce9b204dd-config-data-custom" (OuterVolumeSpecName: "config-data-custom") pod "65e6bdd0-9ffe-48d7-9b02-444ce9b204dd" (UID: "65e6bdd0-9ffe-48d7-9b02-444ce9b204dd"). InnerVolumeSpecName "config-data-custom". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 08 18:35:11 crc kubenswrapper[4859]: I1008 18:35:11.052097 4859 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/65e6bdd0-9ffe-48d7-9b02-444ce9b204dd-scripts" (OuterVolumeSpecName: "scripts") pod "65e6bdd0-9ffe-48d7-9b02-444ce9b204dd" (UID: "65e6bdd0-9ffe-48d7-9b02-444ce9b204dd"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 08 18:35:11 crc kubenswrapper[4859]: I1008 18:35:11.052717 4859 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/65e6bdd0-9ffe-48d7-9b02-444ce9b204dd-kube-api-access-fhgh7" (OuterVolumeSpecName: "kube-api-access-fhgh7") pod "65e6bdd0-9ffe-48d7-9b02-444ce9b204dd" (UID: "65e6bdd0-9ffe-48d7-9b02-444ce9b204dd"). InnerVolumeSpecName "kube-api-access-fhgh7". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 08 18:35:11 crc kubenswrapper[4859]: I1008 18:35:11.081798 4859 generic.go:334] "Generic (PLEG): container finished" podID="65e6bdd0-9ffe-48d7-9b02-444ce9b204dd" containerID="8299c2ef48c3db54ed4f65c7fe5f11ea6a9f2987dae9dc6db90360dbb8a676e6" exitCode=137 Oct 08 18:35:11 crc kubenswrapper[4859]: I1008 18:35:11.082027 4859 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-api-0" Oct 08 18:35:11 crc kubenswrapper[4859]: I1008 18:35:11.082509 4859 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-api-0" event={"ID":"65e6bdd0-9ffe-48d7-9b02-444ce9b204dd","Type":"ContainerDied","Data":"8299c2ef48c3db54ed4f65c7fe5f11ea6a9f2987dae9dc6db90360dbb8a676e6"} Oct 08 18:35:11 crc kubenswrapper[4859]: I1008 18:35:11.082727 4859 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-api-0" event={"ID":"65e6bdd0-9ffe-48d7-9b02-444ce9b204dd","Type":"ContainerDied","Data":"1c6c50170c1f93a8ab17006c51930faa05132506dadfdda58eb6611fde6c2efe"} Oct 08 18:35:11 crc kubenswrapper[4859]: I1008 18:35:11.082829 4859 scope.go:117] "RemoveContainer" containerID="8299c2ef48c3db54ed4f65c7fe5f11ea6a9f2987dae9dc6db90360dbb8a676e6" Oct 08 18:35:11 crc kubenswrapper[4859]: I1008 18:35:11.113925 4859 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/65e6bdd0-9ffe-48d7-9b02-444ce9b204dd-config-data" (OuterVolumeSpecName: "config-data") pod "65e6bdd0-9ffe-48d7-9b02-444ce9b204dd" (UID: "65e6bdd0-9ffe-48d7-9b02-444ce9b204dd"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 08 18:35:11 crc kubenswrapper[4859]: I1008 18:35:11.124466 4859 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/65e6bdd0-9ffe-48d7-9b02-444ce9b204dd-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "65e6bdd0-9ffe-48d7-9b02-444ce9b204dd" (UID: "65e6bdd0-9ffe-48d7-9b02-444ce9b204dd"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 08 18:35:11 crc kubenswrapper[4859]: I1008 18:35:11.158003 4859 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/65e6bdd0-9ffe-48d7-9b02-444ce9b204dd-config-data\") on node \"crc\" DevicePath \"\"" Oct 08 18:35:11 crc kubenswrapper[4859]: I1008 18:35:11.158029 4859 reconciler_common.go:293] "Volume detached for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/65e6bdd0-9ffe-48d7-9b02-444ce9b204dd-config-data-custom\") on node \"crc\" DevicePath \"\"" Oct 08 18:35:11 crc kubenswrapper[4859]: I1008 18:35:11.158044 4859 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/65e6bdd0-9ffe-48d7-9b02-444ce9b204dd-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 08 18:35:11 crc kubenswrapper[4859]: I1008 18:35:11.158054 4859 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/65e6bdd0-9ffe-48d7-9b02-444ce9b204dd-scripts\") on node \"crc\" DevicePath \"\"" Oct 08 18:35:11 crc kubenswrapper[4859]: I1008 18:35:11.158064 4859 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-fhgh7\" (UniqueName: \"kubernetes.io/projected/65e6bdd0-9ffe-48d7-9b02-444ce9b204dd-kube-api-access-fhgh7\") on node \"crc\" DevicePath \"\"" Oct 08 18:35:11 crc kubenswrapper[4859]: I1008 18:35:11.230841 4859 scope.go:117] "RemoveContainer" containerID="852ca86a7fece5a8f498b03bbc9d181549eb5372cbff049da3663b702b20ec91" Oct 08 18:35:11 crc kubenswrapper[4859]: I1008 18:35:11.272836 4859 scope.go:117] "RemoveContainer" containerID="8299c2ef48c3db54ed4f65c7fe5f11ea6a9f2987dae9dc6db90360dbb8a676e6" Oct 08 18:35:11 crc kubenswrapper[4859]: E1008 18:35:11.274239 4859 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"8299c2ef48c3db54ed4f65c7fe5f11ea6a9f2987dae9dc6db90360dbb8a676e6\": container with ID starting with 8299c2ef48c3db54ed4f65c7fe5f11ea6a9f2987dae9dc6db90360dbb8a676e6 not found: ID does not exist" containerID="8299c2ef48c3db54ed4f65c7fe5f11ea6a9f2987dae9dc6db90360dbb8a676e6" Oct 08 18:35:11 crc kubenswrapper[4859]: I1008 18:35:11.274410 4859 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"8299c2ef48c3db54ed4f65c7fe5f11ea6a9f2987dae9dc6db90360dbb8a676e6"} err="failed to get container status \"8299c2ef48c3db54ed4f65c7fe5f11ea6a9f2987dae9dc6db90360dbb8a676e6\": rpc error: code = NotFound desc = could not find container \"8299c2ef48c3db54ed4f65c7fe5f11ea6a9f2987dae9dc6db90360dbb8a676e6\": container with ID starting with 8299c2ef48c3db54ed4f65c7fe5f11ea6a9f2987dae9dc6db90360dbb8a676e6 not found: ID does not exist" Oct 08 18:35:11 crc kubenswrapper[4859]: I1008 18:35:11.274520 4859 scope.go:117] "RemoveContainer" containerID="852ca86a7fece5a8f498b03bbc9d181549eb5372cbff049da3663b702b20ec91" Oct 08 18:35:11 crc kubenswrapper[4859]: E1008 18:35:11.275085 4859 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"852ca86a7fece5a8f498b03bbc9d181549eb5372cbff049da3663b702b20ec91\": container with ID starting with 852ca86a7fece5a8f498b03bbc9d181549eb5372cbff049da3663b702b20ec91 not found: ID does not exist" containerID="852ca86a7fece5a8f498b03bbc9d181549eb5372cbff049da3663b702b20ec91" Oct 08 18:35:11 crc kubenswrapper[4859]: I1008 18:35:11.275219 4859 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"852ca86a7fece5a8f498b03bbc9d181549eb5372cbff049da3663b702b20ec91"} err="failed to get container status \"852ca86a7fece5a8f498b03bbc9d181549eb5372cbff049da3663b702b20ec91\": rpc error: code = NotFound desc = could not find container \"852ca86a7fece5a8f498b03bbc9d181549eb5372cbff049da3663b702b20ec91\": container with ID starting with 852ca86a7fece5a8f498b03bbc9d181549eb5372cbff049da3663b702b20ec91 not found: ID does not exist" Oct 08 18:35:11 crc kubenswrapper[4859]: I1008 18:35:11.366266 4859 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-dd49-account-create-cnv2d" Oct 08 18:35:11 crc kubenswrapper[4859]: I1008 18:35:11.431201 4859 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/cinder-api-0"] Oct 08 18:35:11 crc kubenswrapper[4859]: I1008 18:35:11.440541 4859 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/cinder-api-0"] Oct 08 18:35:11 crc kubenswrapper[4859]: I1008 18:35:11.466097 4859 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/cinder-api-0"] Oct 08 18:35:11 crc kubenswrapper[4859]: E1008 18:35:11.466510 4859 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="65e6bdd0-9ffe-48d7-9b02-444ce9b204dd" containerName="cinder-api-log" Oct 08 18:35:11 crc kubenswrapper[4859]: I1008 18:35:11.466530 4859 state_mem.go:107] "Deleted CPUSet assignment" podUID="65e6bdd0-9ffe-48d7-9b02-444ce9b204dd" containerName="cinder-api-log" Oct 08 18:35:11 crc kubenswrapper[4859]: E1008 18:35:11.466556 4859 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="4ee130b1-a5ee-428e-81f9-e4669ad95b11" containerName="mariadb-account-create" Oct 08 18:35:11 crc kubenswrapper[4859]: I1008 18:35:11.466563 4859 state_mem.go:107] "Deleted CPUSet assignment" podUID="4ee130b1-a5ee-428e-81f9-e4669ad95b11" containerName="mariadb-account-create" Oct 08 18:35:11 crc kubenswrapper[4859]: E1008 18:35:11.466580 4859 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="65e6bdd0-9ffe-48d7-9b02-444ce9b204dd" containerName="cinder-api" Oct 08 18:35:11 crc kubenswrapper[4859]: I1008 18:35:11.466586 4859 state_mem.go:107] "Deleted CPUSet assignment" podUID="65e6bdd0-9ffe-48d7-9b02-444ce9b204dd" containerName="cinder-api" Oct 08 18:35:11 crc kubenswrapper[4859]: I1008 18:35:11.466773 4859 memory_manager.go:354] "RemoveStaleState removing state" podUID="65e6bdd0-9ffe-48d7-9b02-444ce9b204dd" containerName="cinder-api" Oct 08 18:35:11 crc kubenswrapper[4859]: I1008 18:35:11.466797 4859 memory_manager.go:354] "RemoveStaleState removing state" podUID="4ee130b1-a5ee-428e-81f9-e4669ad95b11" containerName="mariadb-account-create" Oct 08 18:35:11 crc kubenswrapper[4859]: I1008 18:35:11.466813 4859 memory_manager.go:354] "RemoveStaleState removing state" podUID="65e6bdd0-9ffe-48d7-9b02-444ce9b204dd" containerName="cinder-api-log" Oct 08 18:35:11 crc kubenswrapper[4859]: I1008 18:35:11.468016 4859 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-api-0" Oct 08 18:35:11 crc kubenswrapper[4859]: I1008 18:35:11.468542 4859 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-7st6z\" (UniqueName: \"kubernetes.io/projected/4ee130b1-a5ee-428e-81f9-e4669ad95b11-kube-api-access-7st6z\") pod \"4ee130b1-a5ee-428e-81f9-e4669ad95b11\" (UID: \"4ee130b1-a5ee-428e-81f9-e4669ad95b11\") " Oct 08 18:35:11 crc kubenswrapper[4859]: I1008 18:35:11.473390 4859 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-api-config-data" Oct 08 18:35:11 crc kubenswrapper[4859]: I1008 18:35:11.473410 4859 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-cinder-public-svc" Oct 08 18:35:11 crc kubenswrapper[4859]: I1008 18:35:11.473889 4859 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-cinder-internal-svc" Oct 08 18:35:11 crc kubenswrapper[4859]: I1008 18:35:11.476779 4859 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/4ee130b1-a5ee-428e-81f9-e4669ad95b11-kube-api-access-7st6z" (OuterVolumeSpecName: "kube-api-access-7st6z") pod "4ee130b1-a5ee-428e-81f9-e4669ad95b11" (UID: "4ee130b1-a5ee-428e-81f9-e4669ad95b11"). InnerVolumeSpecName "kube-api-access-7st6z". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 08 18:35:11 crc kubenswrapper[4859]: I1008 18:35:11.487213 4859 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-api-0"] Oct 08 18:35:11 crc kubenswrapper[4859]: I1008 18:35:11.570658 4859 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/f8eeb421-2f6d-428f-93d5-5aca080785cc-etc-machine-id\") pod \"cinder-api-0\" (UID: \"f8eeb421-2f6d-428f-93d5-5aca080785cc\") " pod="openstack/cinder-api-0" Oct 08 18:35:11 crc kubenswrapper[4859]: I1008 18:35:11.571002 4859 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f8eeb421-2f6d-428f-93d5-5aca080785cc-combined-ca-bundle\") pod \"cinder-api-0\" (UID: \"f8eeb421-2f6d-428f-93d5-5aca080785cc\") " pod="openstack/cinder-api-0" Oct 08 18:35:11 crc kubenswrapper[4859]: I1008 18:35:11.571099 4859 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f8eeb421-2f6d-428f-93d5-5aca080785cc-config-data\") pod \"cinder-api-0\" (UID: \"f8eeb421-2f6d-428f-93d5-5aca080785cc\") " pod="openstack/cinder-api-0" Oct 08 18:35:11 crc kubenswrapper[4859]: I1008 18:35:11.571173 4859 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/f8eeb421-2f6d-428f-93d5-5aca080785cc-public-tls-certs\") pod \"cinder-api-0\" (UID: \"f8eeb421-2f6d-428f-93d5-5aca080785cc\") " pod="openstack/cinder-api-0" Oct 08 18:35:11 crc kubenswrapper[4859]: I1008 18:35:11.571310 4859 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-nchql\" (UniqueName: \"kubernetes.io/projected/f8eeb421-2f6d-428f-93d5-5aca080785cc-kube-api-access-nchql\") pod \"cinder-api-0\" (UID: \"f8eeb421-2f6d-428f-93d5-5aca080785cc\") " pod="openstack/cinder-api-0" Oct 08 18:35:11 crc kubenswrapper[4859]: I1008 18:35:11.571483 4859 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/f8eeb421-2f6d-428f-93d5-5aca080785cc-scripts\") pod \"cinder-api-0\" (UID: \"f8eeb421-2f6d-428f-93d5-5aca080785cc\") " pod="openstack/cinder-api-0" Oct 08 18:35:11 crc kubenswrapper[4859]: I1008 18:35:11.571632 4859 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/f8eeb421-2f6d-428f-93d5-5aca080785cc-internal-tls-certs\") pod \"cinder-api-0\" (UID: \"f8eeb421-2f6d-428f-93d5-5aca080785cc\") " pod="openstack/cinder-api-0" Oct 08 18:35:11 crc kubenswrapper[4859]: I1008 18:35:11.571730 4859 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/f8eeb421-2f6d-428f-93d5-5aca080785cc-config-data-custom\") pod \"cinder-api-0\" (UID: \"f8eeb421-2f6d-428f-93d5-5aca080785cc\") " pod="openstack/cinder-api-0" Oct 08 18:35:11 crc kubenswrapper[4859]: I1008 18:35:11.571886 4859 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/f8eeb421-2f6d-428f-93d5-5aca080785cc-logs\") pod \"cinder-api-0\" (UID: \"f8eeb421-2f6d-428f-93d5-5aca080785cc\") " pod="openstack/cinder-api-0" Oct 08 18:35:11 crc kubenswrapper[4859]: I1008 18:35:11.572146 4859 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-7st6z\" (UniqueName: \"kubernetes.io/projected/4ee130b1-a5ee-428e-81f9-e4669ad95b11-kube-api-access-7st6z\") on node \"crc\" DevicePath \"\"" Oct 08 18:35:11 crc kubenswrapper[4859]: I1008 18:35:11.673154 4859 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/f8eeb421-2f6d-428f-93d5-5aca080785cc-scripts\") pod \"cinder-api-0\" (UID: \"f8eeb421-2f6d-428f-93d5-5aca080785cc\") " pod="openstack/cinder-api-0" Oct 08 18:35:11 crc kubenswrapper[4859]: I1008 18:35:11.673216 4859 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/f8eeb421-2f6d-428f-93d5-5aca080785cc-internal-tls-certs\") pod \"cinder-api-0\" (UID: \"f8eeb421-2f6d-428f-93d5-5aca080785cc\") " pod="openstack/cinder-api-0" Oct 08 18:35:11 crc kubenswrapper[4859]: I1008 18:35:11.673241 4859 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/f8eeb421-2f6d-428f-93d5-5aca080785cc-config-data-custom\") pod \"cinder-api-0\" (UID: \"f8eeb421-2f6d-428f-93d5-5aca080785cc\") " pod="openstack/cinder-api-0" Oct 08 18:35:11 crc kubenswrapper[4859]: I1008 18:35:11.673282 4859 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/f8eeb421-2f6d-428f-93d5-5aca080785cc-logs\") pod \"cinder-api-0\" (UID: \"f8eeb421-2f6d-428f-93d5-5aca080785cc\") " pod="openstack/cinder-api-0" Oct 08 18:35:11 crc kubenswrapper[4859]: I1008 18:35:11.673345 4859 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/f8eeb421-2f6d-428f-93d5-5aca080785cc-etc-machine-id\") pod \"cinder-api-0\" (UID: \"f8eeb421-2f6d-428f-93d5-5aca080785cc\") " pod="openstack/cinder-api-0" Oct 08 18:35:11 crc kubenswrapper[4859]: I1008 18:35:11.673370 4859 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f8eeb421-2f6d-428f-93d5-5aca080785cc-combined-ca-bundle\") pod \"cinder-api-0\" (UID: \"f8eeb421-2f6d-428f-93d5-5aca080785cc\") " pod="openstack/cinder-api-0" Oct 08 18:35:11 crc kubenswrapper[4859]: I1008 18:35:11.673399 4859 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f8eeb421-2f6d-428f-93d5-5aca080785cc-config-data\") pod \"cinder-api-0\" (UID: \"f8eeb421-2f6d-428f-93d5-5aca080785cc\") " pod="openstack/cinder-api-0" Oct 08 18:35:11 crc kubenswrapper[4859]: I1008 18:35:11.673417 4859 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/f8eeb421-2f6d-428f-93d5-5aca080785cc-public-tls-certs\") pod \"cinder-api-0\" (UID: \"f8eeb421-2f6d-428f-93d5-5aca080785cc\") " pod="openstack/cinder-api-0" Oct 08 18:35:11 crc kubenswrapper[4859]: I1008 18:35:11.673439 4859 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-nchql\" (UniqueName: \"kubernetes.io/projected/f8eeb421-2f6d-428f-93d5-5aca080785cc-kube-api-access-nchql\") pod \"cinder-api-0\" (UID: \"f8eeb421-2f6d-428f-93d5-5aca080785cc\") " pod="openstack/cinder-api-0" Oct 08 18:35:11 crc kubenswrapper[4859]: I1008 18:35:11.673467 4859 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/f8eeb421-2f6d-428f-93d5-5aca080785cc-etc-machine-id\") pod \"cinder-api-0\" (UID: \"f8eeb421-2f6d-428f-93d5-5aca080785cc\") " pod="openstack/cinder-api-0" Oct 08 18:35:11 crc kubenswrapper[4859]: I1008 18:35:11.673978 4859 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/f8eeb421-2f6d-428f-93d5-5aca080785cc-logs\") pod \"cinder-api-0\" (UID: \"f8eeb421-2f6d-428f-93d5-5aca080785cc\") " pod="openstack/cinder-api-0" Oct 08 18:35:11 crc kubenswrapper[4859]: I1008 18:35:11.679043 4859 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/f8eeb421-2f6d-428f-93d5-5aca080785cc-public-tls-certs\") pod \"cinder-api-0\" (UID: \"f8eeb421-2f6d-428f-93d5-5aca080785cc\") " pod="openstack/cinder-api-0" Oct 08 18:35:11 crc kubenswrapper[4859]: I1008 18:35:11.680812 4859 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/f8eeb421-2f6d-428f-93d5-5aca080785cc-internal-tls-certs\") pod \"cinder-api-0\" (UID: \"f8eeb421-2f6d-428f-93d5-5aca080785cc\") " pod="openstack/cinder-api-0" Oct 08 18:35:11 crc kubenswrapper[4859]: I1008 18:35:11.681016 4859 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/f8eeb421-2f6d-428f-93d5-5aca080785cc-scripts\") pod \"cinder-api-0\" (UID: \"f8eeb421-2f6d-428f-93d5-5aca080785cc\") " pod="openstack/cinder-api-0" Oct 08 18:35:11 crc kubenswrapper[4859]: I1008 18:35:11.682171 4859 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f8eeb421-2f6d-428f-93d5-5aca080785cc-combined-ca-bundle\") pod \"cinder-api-0\" (UID: \"f8eeb421-2f6d-428f-93d5-5aca080785cc\") " pod="openstack/cinder-api-0" Oct 08 18:35:11 crc kubenswrapper[4859]: I1008 18:35:11.682328 4859 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/f8eeb421-2f6d-428f-93d5-5aca080785cc-config-data-custom\") pod \"cinder-api-0\" (UID: \"f8eeb421-2f6d-428f-93d5-5aca080785cc\") " pod="openstack/cinder-api-0" Oct 08 18:35:11 crc kubenswrapper[4859]: I1008 18:35:11.683090 4859 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f8eeb421-2f6d-428f-93d5-5aca080785cc-config-data\") pod \"cinder-api-0\" (UID: \"f8eeb421-2f6d-428f-93d5-5aca080785cc\") " pod="openstack/cinder-api-0" Oct 08 18:35:11 crc kubenswrapper[4859]: I1008 18:35:11.692299 4859 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-nchql\" (UniqueName: \"kubernetes.io/projected/f8eeb421-2f6d-428f-93d5-5aca080785cc-kube-api-access-nchql\") pod \"cinder-api-0\" (UID: \"f8eeb421-2f6d-428f-93d5-5aca080785cc\") " pod="openstack/cinder-api-0" Oct 08 18:35:11 crc kubenswrapper[4859]: I1008 18:35:11.790245 4859 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-api-0" Oct 08 18:35:11 crc kubenswrapper[4859]: I1008 18:35:11.975941 4859 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-default-external-api-0"] Oct 08 18:35:11 crc kubenswrapper[4859]: I1008 18:35:11.976455 4859 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/glance-default-external-api-0" podUID="b8433969-f079-49a2-95f3-ce4783d6a62c" containerName="glance-log" containerID="cri-o://e58a2a7e6d7cb1671e6f15be2eae9753bf9ba14ba510c49cc6d4d7121ba024ce" gracePeriod=30 Oct 08 18:35:11 crc kubenswrapper[4859]: I1008 18:35:11.976558 4859 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/glance-default-external-api-0" podUID="b8433969-f079-49a2-95f3-ce4783d6a62c" containerName="glance-httpd" containerID="cri-o://3d02d784fce2d216c2d98154bf821d302aeba5a184bb06406551e52cbd463ff4" gracePeriod=30 Oct 08 18:35:12 crc kubenswrapper[4859]: I1008 18:35:12.114728 4859 generic.go:334] "Generic (PLEG): container finished" podID="b8433969-f079-49a2-95f3-ce4783d6a62c" containerID="e58a2a7e6d7cb1671e6f15be2eae9753bf9ba14ba510c49cc6d4d7121ba024ce" exitCode=143 Oct 08 18:35:12 crc kubenswrapper[4859]: I1008 18:35:12.114817 4859 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"b8433969-f079-49a2-95f3-ce4783d6a62c","Type":"ContainerDied","Data":"e58a2a7e6d7cb1671e6f15be2eae9753bf9ba14ba510c49cc6d4d7121ba024ce"} Oct 08 18:35:12 crc kubenswrapper[4859]: I1008 18:35:12.123976 4859 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"d95266a3-07ae-4c7f-9cc7-af1a632290cc","Type":"ContainerStarted","Data":"11a1608a5dd4756100be705360e1e31235b194c38afb2ba257c21278b543ddc0"} Oct 08 18:35:12 crc kubenswrapper[4859]: I1008 18:35:12.125678 4859 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ceilometer-0" Oct 08 18:35:12 crc kubenswrapper[4859]: I1008 18:35:12.139883 4859 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-dd49-account-create-cnv2d" event={"ID":"4ee130b1-a5ee-428e-81f9-e4669ad95b11","Type":"ContainerDied","Data":"816f851c1db6e609c36fd3250ebf24fdca6dda096e47b77e1b14ca8a7d0487cf"} Oct 08 18:35:12 crc kubenswrapper[4859]: I1008 18:35:12.139933 4859 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="816f851c1db6e609c36fd3250ebf24fdca6dda096e47b77e1b14ca8a7d0487cf" Oct 08 18:35:12 crc kubenswrapper[4859]: I1008 18:35:12.139933 4859 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-dd49-account-create-cnv2d" Oct 08 18:35:12 crc kubenswrapper[4859]: I1008 18:35:12.158508 4859 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ceilometer-0" podStartSLOduration=2.076449638 podStartE2EDuration="6.158483845s" podCreationTimestamp="2025-10-08 18:35:06 +0000 UTC" firstStartedPulling="2025-10-08 18:35:07.122962874 +0000 UTC m=+1077.369802253" lastFinishedPulling="2025-10-08 18:35:11.204997081 +0000 UTC m=+1081.451836460" observedRunningTime="2025-10-08 18:35:12.150613492 +0000 UTC m=+1082.397452891" watchObservedRunningTime="2025-10-08 18:35:12.158483845 +0000 UTC m=+1082.405323234" Oct 08 18:35:12 crc kubenswrapper[4859]: I1008 18:35:12.326135 4859 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-api-0"] Oct 08 18:35:12 crc kubenswrapper[4859]: I1008 18:35:12.484609 4859 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="65e6bdd0-9ffe-48d7-9b02-444ce9b204dd" path="/var/lib/kubelet/pods/65e6bdd0-9ffe-48d7-9b02-444ce9b204dd/volumes" Oct 08 18:35:13 crc kubenswrapper[4859]: I1008 18:35:13.018066 4859 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-default-internal-api-0"] Oct 08 18:35:13 crc kubenswrapper[4859]: I1008 18:35:13.018307 4859 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/glance-default-internal-api-0" podUID="7bf3f855-41ae-4de9-ba6c-0fb4f51df141" containerName="glance-log" containerID="cri-o://28ec01a1442339cbd321103beebc7df267b3a533827492d256693d3f04972c5d" gracePeriod=30 Oct 08 18:35:13 crc kubenswrapper[4859]: I1008 18:35:13.018666 4859 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/glance-default-internal-api-0" podUID="7bf3f855-41ae-4de9-ba6c-0fb4f51df141" containerName="glance-httpd" containerID="cri-o://46ed96c026d937d863774320b18d16cbedb2414f358afc23e4af1c83103bb4e5" gracePeriod=30 Oct 08 18:35:13 crc kubenswrapper[4859]: I1008 18:35:13.162636 4859 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-api-0" event={"ID":"f8eeb421-2f6d-428f-93d5-5aca080785cc","Type":"ContainerStarted","Data":"ad54cb4de1b413f855ce990237edaeb2228047dc453d373cdcb59aa9df600961"} Oct 08 18:35:13 crc kubenswrapper[4859]: I1008 18:35:13.162971 4859 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-api-0" event={"ID":"f8eeb421-2f6d-428f-93d5-5aca080785cc","Type":"ContainerStarted","Data":"93df7080183af9e49f7b874abb5255deb7558cfe3cb63b62709644983c0561e1"} Oct 08 18:35:13 crc kubenswrapper[4859]: I1008 18:35:13.170077 4859 generic.go:334] "Generic (PLEG): container finished" podID="7bf3f855-41ae-4de9-ba6c-0fb4f51df141" containerID="28ec01a1442339cbd321103beebc7df267b3a533827492d256693d3f04972c5d" exitCode=143 Oct 08 18:35:13 crc kubenswrapper[4859]: I1008 18:35:13.170499 4859 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"7bf3f855-41ae-4de9-ba6c-0fb4f51df141","Type":"ContainerDied","Data":"28ec01a1442339cbd321103beebc7df267b3a533827492d256693d3f04972c5d"} Oct 08 18:35:14 crc kubenswrapper[4859]: I1008 18:35:14.183131 4859 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-api-0" event={"ID":"f8eeb421-2f6d-428f-93d5-5aca080785cc","Type":"ContainerStarted","Data":"9d9c818acb6771d0046cabbc193f10155595e3d2cb567ad743805bd042b6c56e"} Oct 08 18:35:14 crc kubenswrapper[4859]: I1008 18:35:14.210277 4859 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/cinder-api-0" podStartSLOduration=3.210257971 podStartE2EDuration="3.210257971s" podCreationTimestamp="2025-10-08 18:35:11 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-08 18:35:14.206097588 +0000 UTC m=+1084.452936977" watchObservedRunningTime="2025-10-08 18:35:14.210257971 +0000 UTC m=+1084.457097350" Oct 08 18:35:14 crc kubenswrapper[4859]: I1008 18:35:14.448877 4859 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Oct 08 18:35:15 crc kubenswrapper[4859]: I1008 18:35:15.196912 4859 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/cinder-api-0" Oct 08 18:35:15 crc kubenswrapper[4859]: I1008 18:35:15.197454 4859 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="d95266a3-07ae-4c7f-9cc7-af1a632290cc" containerName="ceilometer-central-agent" containerID="cri-o://9a699a097ec98f3331d90f2e9089ee5a2d6472715e420e0f0ac4c813b0f7c776" gracePeriod=30 Oct 08 18:35:15 crc kubenswrapper[4859]: I1008 18:35:15.197904 4859 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="d95266a3-07ae-4c7f-9cc7-af1a632290cc" containerName="proxy-httpd" containerID="cri-o://11a1608a5dd4756100be705360e1e31235b194c38afb2ba257c21278b543ddc0" gracePeriod=30 Oct 08 18:35:15 crc kubenswrapper[4859]: I1008 18:35:15.197920 4859 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="d95266a3-07ae-4c7f-9cc7-af1a632290cc" containerName="sg-core" containerID="cri-o://f4b804fb4923d970b1a669c891e974729f108bea4442714c5bb19f4176d284ba" gracePeriod=30 Oct 08 18:35:15 crc kubenswrapper[4859]: I1008 18:35:15.197937 4859 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="d95266a3-07ae-4c7f-9cc7-af1a632290cc" containerName="ceilometer-notification-agent" containerID="cri-o://83b10b00b3fde51de9b23a21cf635ce9b4b442e8ddab240ca122aee97845c3e0" gracePeriod=30 Oct 08 18:35:15 crc kubenswrapper[4859]: I1008 18:35:15.323712 4859 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/kube-state-metrics-0" Oct 08 18:35:15 crc kubenswrapper[4859]: I1008 18:35:15.766016 4859 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-external-api-0" Oct 08 18:35:15 crc kubenswrapper[4859]: I1008 18:35:15.864446 4859 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b8433969-f079-49a2-95f3-ce4783d6a62c-combined-ca-bundle\") pod \"b8433969-f079-49a2-95f3-ce4783d6a62c\" (UID: \"b8433969-f079-49a2-95f3-ce4783d6a62c\") " Oct 08 18:35:15 crc kubenswrapper[4859]: I1008 18:35:15.864495 4859 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"glance\" (UniqueName: \"kubernetes.io/local-volume/local-storage12-crc\") pod \"b8433969-f079-49a2-95f3-ce4783d6a62c\" (UID: \"b8433969-f079-49a2-95f3-ce4783d6a62c\") " Oct 08 18:35:15 crc kubenswrapper[4859]: I1008 18:35:15.864633 4859 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/b8433969-f079-49a2-95f3-ce4783d6a62c-public-tls-certs\") pod \"b8433969-f079-49a2-95f3-ce4783d6a62c\" (UID: \"b8433969-f079-49a2-95f3-ce4783d6a62c\") " Oct 08 18:35:15 crc kubenswrapper[4859]: I1008 18:35:15.864676 4859 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-xvctl\" (UniqueName: \"kubernetes.io/projected/b8433969-f079-49a2-95f3-ce4783d6a62c-kube-api-access-xvctl\") pod \"b8433969-f079-49a2-95f3-ce4783d6a62c\" (UID: \"b8433969-f079-49a2-95f3-ce4783d6a62c\") " Oct 08 18:35:15 crc kubenswrapper[4859]: I1008 18:35:15.864750 4859 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/b8433969-f079-49a2-95f3-ce4783d6a62c-logs\") pod \"b8433969-f079-49a2-95f3-ce4783d6a62c\" (UID: \"b8433969-f079-49a2-95f3-ce4783d6a62c\") " Oct 08 18:35:15 crc kubenswrapper[4859]: I1008 18:35:15.864789 4859 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/b8433969-f079-49a2-95f3-ce4783d6a62c-httpd-run\") pod \"b8433969-f079-49a2-95f3-ce4783d6a62c\" (UID: \"b8433969-f079-49a2-95f3-ce4783d6a62c\") " Oct 08 18:35:15 crc kubenswrapper[4859]: I1008 18:35:15.864848 4859 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/b8433969-f079-49a2-95f3-ce4783d6a62c-config-data\") pod \"b8433969-f079-49a2-95f3-ce4783d6a62c\" (UID: \"b8433969-f079-49a2-95f3-ce4783d6a62c\") " Oct 08 18:35:15 crc kubenswrapper[4859]: I1008 18:35:15.864871 4859 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/b8433969-f079-49a2-95f3-ce4783d6a62c-scripts\") pod \"b8433969-f079-49a2-95f3-ce4783d6a62c\" (UID: \"b8433969-f079-49a2-95f3-ce4783d6a62c\") " Oct 08 18:35:15 crc kubenswrapper[4859]: I1008 18:35:15.866059 4859 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/b8433969-f079-49a2-95f3-ce4783d6a62c-httpd-run" (OuterVolumeSpecName: "httpd-run") pod "b8433969-f079-49a2-95f3-ce4783d6a62c" (UID: "b8433969-f079-49a2-95f3-ce4783d6a62c"). InnerVolumeSpecName "httpd-run". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 08 18:35:15 crc kubenswrapper[4859]: I1008 18:35:15.866331 4859 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/b8433969-f079-49a2-95f3-ce4783d6a62c-logs" (OuterVolumeSpecName: "logs") pod "b8433969-f079-49a2-95f3-ce4783d6a62c" (UID: "b8433969-f079-49a2-95f3-ce4783d6a62c"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 08 18:35:15 crc kubenswrapper[4859]: I1008 18:35:15.872357 4859 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/local-volume/local-storage12-crc" (OuterVolumeSpecName: "glance") pod "b8433969-f079-49a2-95f3-ce4783d6a62c" (UID: "b8433969-f079-49a2-95f3-ce4783d6a62c"). InnerVolumeSpecName "local-storage12-crc". PluginName "kubernetes.io/local-volume", VolumeGidValue "" Oct 08 18:35:15 crc kubenswrapper[4859]: I1008 18:35:15.873832 4859 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b8433969-f079-49a2-95f3-ce4783d6a62c-kube-api-access-xvctl" (OuterVolumeSpecName: "kube-api-access-xvctl") pod "b8433969-f079-49a2-95f3-ce4783d6a62c" (UID: "b8433969-f079-49a2-95f3-ce4783d6a62c"). InnerVolumeSpecName "kube-api-access-xvctl". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 08 18:35:15 crc kubenswrapper[4859]: I1008 18:35:15.875543 4859 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b8433969-f079-49a2-95f3-ce4783d6a62c-scripts" (OuterVolumeSpecName: "scripts") pod "b8433969-f079-49a2-95f3-ce4783d6a62c" (UID: "b8433969-f079-49a2-95f3-ce4783d6a62c"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 08 18:35:15 crc kubenswrapper[4859]: I1008 18:35:15.906329 4859 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b8433969-f079-49a2-95f3-ce4783d6a62c-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "b8433969-f079-49a2-95f3-ce4783d6a62c" (UID: "b8433969-f079-49a2-95f3-ce4783d6a62c"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 08 18:35:15 crc kubenswrapper[4859]: I1008 18:35:15.947166 4859 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b8433969-f079-49a2-95f3-ce4783d6a62c-public-tls-certs" (OuterVolumeSpecName: "public-tls-certs") pod "b8433969-f079-49a2-95f3-ce4783d6a62c" (UID: "b8433969-f079-49a2-95f3-ce4783d6a62c"). InnerVolumeSpecName "public-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 08 18:35:15 crc kubenswrapper[4859]: I1008 18:35:15.956884 4859 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b8433969-f079-49a2-95f3-ce4783d6a62c-config-data" (OuterVolumeSpecName: "config-data") pod "b8433969-f079-49a2-95f3-ce4783d6a62c" (UID: "b8433969-f079-49a2-95f3-ce4783d6a62c"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 08 18:35:15 crc kubenswrapper[4859]: I1008 18:35:15.967044 4859 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-xvctl\" (UniqueName: \"kubernetes.io/projected/b8433969-f079-49a2-95f3-ce4783d6a62c-kube-api-access-xvctl\") on node \"crc\" DevicePath \"\"" Oct 08 18:35:15 crc kubenswrapper[4859]: I1008 18:35:15.967085 4859 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/b8433969-f079-49a2-95f3-ce4783d6a62c-logs\") on node \"crc\" DevicePath \"\"" Oct 08 18:35:15 crc kubenswrapper[4859]: I1008 18:35:15.967099 4859 reconciler_common.go:293] "Volume detached for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/b8433969-f079-49a2-95f3-ce4783d6a62c-httpd-run\") on node \"crc\" DevicePath \"\"" Oct 08 18:35:15 crc kubenswrapper[4859]: I1008 18:35:15.967111 4859 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/b8433969-f079-49a2-95f3-ce4783d6a62c-config-data\") on node \"crc\" DevicePath \"\"" Oct 08 18:35:15 crc kubenswrapper[4859]: I1008 18:35:15.967122 4859 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/b8433969-f079-49a2-95f3-ce4783d6a62c-scripts\") on node \"crc\" DevicePath \"\"" Oct 08 18:35:15 crc kubenswrapper[4859]: I1008 18:35:15.967132 4859 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b8433969-f079-49a2-95f3-ce4783d6a62c-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 08 18:35:15 crc kubenswrapper[4859]: I1008 18:35:15.967163 4859 reconciler_common.go:286] "operationExecutor.UnmountDevice started for volume \"local-storage12-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage12-crc\") on node \"crc\" " Oct 08 18:35:15 crc kubenswrapper[4859]: I1008 18:35:15.967174 4859 reconciler_common.go:293] "Volume detached for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/b8433969-f079-49a2-95f3-ce4783d6a62c-public-tls-certs\") on node \"crc\" DevicePath \"\"" Oct 08 18:35:15 crc kubenswrapper[4859]: I1008 18:35:15.985476 4859 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Oct 08 18:35:15 crc kubenswrapper[4859]: I1008 18:35:15.998481 4859 operation_generator.go:917] UnmountDevice succeeded for volume "local-storage12-crc" (UniqueName: "kubernetes.io/local-volume/local-storage12-crc") on node "crc" Oct 08 18:35:16 crc kubenswrapper[4859]: I1008 18:35:16.068544 4859 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/d95266a3-07ae-4c7f-9cc7-af1a632290cc-ceilometer-tls-certs\") pod \"d95266a3-07ae-4c7f-9cc7-af1a632290cc\" (UID: \"d95266a3-07ae-4c7f-9cc7-af1a632290cc\") " Oct 08 18:35:16 crc kubenswrapper[4859]: I1008 18:35:16.068641 4859 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/d95266a3-07ae-4c7f-9cc7-af1a632290cc-run-httpd\") pod \"d95266a3-07ae-4c7f-9cc7-af1a632290cc\" (UID: \"d95266a3-07ae-4c7f-9cc7-af1a632290cc\") " Oct 08 18:35:16 crc kubenswrapper[4859]: I1008 18:35:16.068727 4859 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-h2mfw\" (UniqueName: \"kubernetes.io/projected/d95266a3-07ae-4c7f-9cc7-af1a632290cc-kube-api-access-h2mfw\") pod \"d95266a3-07ae-4c7f-9cc7-af1a632290cc\" (UID: \"d95266a3-07ae-4c7f-9cc7-af1a632290cc\") " Oct 08 18:35:16 crc kubenswrapper[4859]: I1008 18:35:16.068765 4859 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/d95266a3-07ae-4c7f-9cc7-af1a632290cc-log-httpd\") pod \"d95266a3-07ae-4c7f-9cc7-af1a632290cc\" (UID: \"d95266a3-07ae-4c7f-9cc7-af1a632290cc\") " Oct 08 18:35:16 crc kubenswrapper[4859]: I1008 18:35:16.068811 4859 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/d95266a3-07ae-4c7f-9cc7-af1a632290cc-sg-core-conf-yaml\") pod \"d95266a3-07ae-4c7f-9cc7-af1a632290cc\" (UID: \"d95266a3-07ae-4c7f-9cc7-af1a632290cc\") " Oct 08 18:35:16 crc kubenswrapper[4859]: I1008 18:35:16.068870 4859 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d95266a3-07ae-4c7f-9cc7-af1a632290cc-combined-ca-bundle\") pod \"d95266a3-07ae-4c7f-9cc7-af1a632290cc\" (UID: \"d95266a3-07ae-4c7f-9cc7-af1a632290cc\") " Oct 08 18:35:16 crc kubenswrapper[4859]: I1008 18:35:16.068921 4859 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/d95266a3-07ae-4c7f-9cc7-af1a632290cc-scripts\") pod \"d95266a3-07ae-4c7f-9cc7-af1a632290cc\" (UID: \"d95266a3-07ae-4c7f-9cc7-af1a632290cc\") " Oct 08 18:35:16 crc kubenswrapper[4859]: I1008 18:35:16.068970 4859 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d95266a3-07ae-4c7f-9cc7-af1a632290cc-config-data\") pod \"d95266a3-07ae-4c7f-9cc7-af1a632290cc\" (UID: \"d95266a3-07ae-4c7f-9cc7-af1a632290cc\") " Oct 08 18:35:16 crc kubenswrapper[4859]: I1008 18:35:16.069352 4859 reconciler_common.go:293] "Volume detached for volume \"local-storage12-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage12-crc\") on node \"crc\" DevicePath \"\"" Oct 08 18:35:16 crc kubenswrapper[4859]: I1008 18:35:16.070079 4859 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/d95266a3-07ae-4c7f-9cc7-af1a632290cc-run-httpd" (OuterVolumeSpecName: "run-httpd") pod "d95266a3-07ae-4c7f-9cc7-af1a632290cc" (UID: "d95266a3-07ae-4c7f-9cc7-af1a632290cc"). InnerVolumeSpecName "run-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 08 18:35:16 crc kubenswrapper[4859]: I1008 18:35:16.071635 4859 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/d95266a3-07ae-4c7f-9cc7-af1a632290cc-log-httpd" (OuterVolumeSpecName: "log-httpd") pod "d95266a3-07ae-4c7f-9cc7-af1a632290cc" (UID: "d95266a3-07ae-4c7f-9cc7-af1a632290cc"). InnerVolumeSpecName "log-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 08 18:35:16 crc kubenswrapper[4859]: I1008 18:35:16.073102 4859 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d95266a3-07ae-4c7f-9cc7-af1a632290cc-scripts" (OuterVolumeSpecName: "scripts") pod "d95266a3-07ae-4c7f-9cc7-af1a632290cc" (UID: "d95266a3-07ae-4c7f-9cc7-af1a632290cc"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 08 18:35:16 crc kubenswrapper[4859]: I1008 18:35:16.074903 4859 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/d95266a3-07ae-4c7f-9cc7-af1a632290cc-kube-api-access-h2mfw" (OuterVolumeSpecName: "kube-api-access-h2mfw") pod "d95266a3-07ae-4c7f-9cc7-af1a632290cc" (UID: "d95266a3-07ae-4c7f-9cc7-af1a632290cc"). InnerVolumeSpecName "kube-api-access-h2mfw". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 08 18:35:16 crc kubenswrapper[4859]: I1008 18:35:16.097268 4859 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d95266a3-07ae-4c7f-9cc7-af1a632290cc-sg-core-conf-yaml" (OuterVolumeSpecName: "sg-core-conf-yaml") pod "d95266a3-07ae-4c7f-9cc7-af1a632290cc" (UID: "d95266a3-07ae-4c7f-9cc7-af1a632290cc"). InnerVolumeSpecName "sg-core-conf-yaml". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 08 18:35:16 crc kubenswrapper[4859]: I1008 18:35:16.120978 4859 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d95266a3-07ae-4c7f-9cc7-af1a632290cc-ceilometer-tls-certs" (OuterVolumeSpecName: "ceilometer-tls-certs") pod "d95266a3-07ae-4c7f-9cc7-af1a632290cc" (UID: "d95266a3-07ae-4c7f-9cc7-af1a632290cc"). InnerVolumeSpecName "ceilometer-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 08 18:35:16 crc kubenswrapper[4859]: I1008 18:35:16.166361 4859 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d95266a3-07ae-4c7f-9cc7-af1a632290cc-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "d95266a3-07ae-4c7f-9cc7-af1a632290cc" (UID: "d95266a3-07ae-4c7f-9cc7-af1a632290cc"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 08 18:35:16 crc kubenswrapper[4859]: I1008 18:35:16.171149 4859 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d95266a3-07ae-4c7f-9cc7-af1a632290cc-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 08 18:35:16 crc kubenswrapper[4859]: I1008 18:35:16.171185 4859 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/d95266a3-07ae-4c7f-9cc7-af1a632290cc-scripts\") on node \"crc\" DevicePath \"\"" Oct 08 18:35:16 crc kubenswrapper[4859]: I1008 18:35:16.171197 4859 reconciler_common.go:293] "Volume detached for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/d95266a3-07ae-4c7f-9cc7-af1a632290cc-ceilometer-tls-certs\") on node \"crc\" DevicePath \"\"" Oct 08 18:35:16 crc kubenswrapper[4859]: I1008 18:35:16.171210 4859 reconciler_common.go:293] "Volume detached for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/d95266a3-07ae-4c7f-9cc7-af1a632290cc-run-httpd\") on node \"crc\" DevicePath \"\"" Oct 08 18:35:16 crc kubenswrapper[4859]: I1008 18:35:16.171222 4859 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-h2mfw\" (UniqueName: \"kubernetes.io/projected/d95266a3-07ae-4c7f-9cc7-af1a632290cc-kube-api-access-h2mfw\") on node \"crc\" DevicePath \"\"" Oct 08 18:35:16 crc kubenswrapper[4859]: I1008 18:35:16.171237 4859 reconciler_common.go:293] "Volume detached for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/d95266a3-07ae-4c7f-9cc7-af1a632290cc-log-httpd\") on node \"crc\" DevicePath \"\"" Oct 08 18:35:16 crc kubenswrapper[4859]: I1008 18:35:16.171246 4859 reconciler_common.go:293] "Volume detached for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/d95266a3-07ae-4c7f-9cc7-af1a632290cc-sg-core-conf-yaml\") on node \"crc\" DevicePath \"\"" Oct 08 18:35:16 crc kubenswrapper[4859]: I1008 18:35:16.183193 4859 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d95266a3-07ae-4c7f-9cc7-af1a632290cc-config-data" (OuterVolumeSpecName: "config-data") pod "d95266a3-07ae-4c7f-9cc7-af1a632290cc" (UID: "d95266a3-07ae-4c7f-9cc7-af1a632290cc"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 08 18:35:16 crc kubenswrapper[4859]: I1008 18:35:16.208384 4859 generic.go:334] "Generic (PLEG): container finished" podID="d95266a3-07ae-4c7f-9cc7-af1a632290cc" containerID="11a1608a5dd4756100be705360e1e31235b194c38afb2ba257c21278b543ddc0" exitCode=0 Oct 08 18:35:16 crc kubenswrapper[4859]: I1008 18:35:16.208426 4859 generic.go:334] "Generic (PLEG): container finished" podID="d95266a3-07ae-4c7f-9cc7-af1a632290cc" containerID="f4b804fb4923d970b1a669c891e974729f108bea4442714c5bb19f4176d284ba" exitCode=2 Oct 08 18:35:16 crc kubenswrapper[4859]: I1008 18:35:16.208436 4859 generic.go:334] "Generic (PLEG): container finished" podID="d95266a3-07ae-4c7f-9cc7-af1a632290cc" containerID="83b10b00b3fde51de9b23a21cf635ce9b4b442e8ddab240ca122aee97845c3e0" exitCode=0 Oct 08 18:35:16 crc kubenswrapper[4859]: I1008 18:35:16.208444 4859 generic.go:334] "Generic (PLEG): container finished" podID="d95266a3-07ae-4c7f-9cc7-af1a632290cc" containerID="9a699a097ec98f3331d90f2e9089ee5a2d6472715e420e0f0ac4c813b0f7c776" exitCode=0 Oct 08 18:35:16 crc kubenswrapper[4859]: I1008 18:35:16.208490 4859 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"d95266a3-07ae-4c7f-9cc7-af1a632290cc","Type":"ContainerDied","Data":"11a1608a5dd4756100be705360e1e31235b194c38afb2ba257c21278b543ddc0"} Oct 08 18:35:16 crc kubenswrapper[4859]: I1008 18:35:16.208520 4859 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"d95266a3-07ae-4c7f-9cc7-af1a632290cc","Type":"ContainerDied","Data":"f4b804fb4923d970b1a669c891e974729f108bea4442714c5bb19f4176d284ba"} Oct 08 18:35:16 crc kubenswrapper[4859]: I1008 18:35:16.208533 4859 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"d95266a3-07ae-4c7f-9cc7-af1a632290cc","Type":"ContainerDied","Data":"83b10b00b3fde51de9b23a21cf635ce9b4b442e8ddab240ca122aee97845c3e0"} Oct 08 18:35:16 crc kubenswrapper[4859]: I1008 18:35:16.208543 4859 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"d95266a3-07ae-4c7f-9cc7-af1a632290cc","Type":"ContainerDied","Data":"9a699a097ec98f3331d90f2e9089ee5a2d6472715e420e0f0ac4c813b0f7c776"} Oct 08 18:35:16 crc kubenswrapper[4859]: I1008 18:35:16.208554 4859 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"d95266a3-07ae-4c7f-9cc7-af1a632290cc","Type":"ContainerDied","Data":"b77ebe586efb4734f6b2c8db2f0a9fc962bfe811a7a2985eb1ee4fe9dd8ed4b3"} Oct 08 18:35:16 crc kubenswrapper[4859]: I1008 18:35:16.208571 4859 scope.go:117] "RemoveContainer" containerID="11a1608a5dd4756100be705360e1e31235b194c38afb2ba257c21278b543ddc0" Oct 08 18:35:16 crc kubenswrapper[4859]: I1008 18:35:16.208792 4859 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Oct 08 18:35:16 crc kubenswrapper[4859]: I1008 18:35:16.211125 4859 generic.go:334] "Generic (PLEG): container finished" podID="b8433969-f079-49a2-95f3-ce4783d6a62c" containerID="3d02d784fce2d216c2d98154bf821d302aeba5a184bb06406551e52cbd463ff4" exitCode=0 Oct 08 18:35:16 crc kubenswrapper[4859]: I1008 18:35:16.211188 4859 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-external-api-0" Oct 08 18:35:16 crc kubenswrapper[4859]: I1008 18:35:16.211197 4859 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"b8433969-f079-49a2-95f3-ce4783d6a62c","Type":"ContainerDied","Data":"3d02d784fce2d216c2d98154bf821d302aeba5a184bb06406551e52cbd463ff4"} Oct 08 18:35:16 crc kubenswrapper[4859]: I1008 18:35:16.211252 4859 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"b8433969-f079-49a2-95f3-ce4783d6a62c","Type":"ContainerDied","Data":"4aa06945ba8e9a81faeb45633d2bc0119700cecd4455590a24af48f3a27a60db"} Oct 08 18:35:16 crc kubenswrapper[4859]: I1008 18:35:16.273193 4859 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d95266a3-07ae-4c7f-9cc7-af1a632290cc-config-data\") on node \"crc\" DevicePath \"\"" Oct 08 18:35:16 crc kubenswrapper[4859]: I1008 18:35:16.390417 4859 scope.go:117] "RemoveContainer" containerID="f4b804fb4923d970b1a669c891e974729f108bea4442714c5bb19f4176d284ba" Oct 08 18:35:16 crc kubenswrapper[4859]: I1008 18:35:16.400391 4859 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Oct 08 18:35:16 crc kubenswrapper[4859]: I1008 18:35:16.433102 4859 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/ceilometer-0"] Oct 08 18:35:16 crc kubenswrapper[4859]: I1008 18:35:16.438193 4859 scope.go:117] "RemoveContainer" containerID="83b10b00b3fde51de9b23a21cf635ce9b4b442e8ddab240ca122aee97845c3e0" Oct 08 18:35:16 crc kubenswrapper[4859]: I1008 18:35:16.455464 4859 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-default-external-api-0"] Oct 08 18:35:16 crc kubenswrapper[4859]: I1008 18:35:16.464334 4859 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/glance-default-external-api-0"] Oct 08 18:35:16 crc kubenswrapper[4859]: I1008 18:35:16.479943 4859 scope.go:117] "RemoveContainer" containerID="9a699a097ec98f3331d90f2e9089ee5a2d6472715e420e0f0ac4c813b0f7c776" Oct 08 18:35:16 crc kubenswrapper[4859]: I1008 18:35:16.496203 4859 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="b8433969-f079-49a2-95f3-ce4783d6a62c" path="/var/lib/kubelet/pods/b8433969-f079-49a2-95f3-ce4783d6a62c/volumes" Oct 08 18:35:16 crc kubenswrapper[4859]: I1008 18:35:16.497028 4859 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="d95266a3-07ae-4c7f-9cc7-af1a632290cc" path="/var/lib/kubelet/pods/d95266a3-07ae-4c7f-9cc7-af1a632290cc/volumes" Oct 08 18:35:16 crc kubenswrapper[4859]: I1008 18:35:16.497864 4859 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ceilometer-0"] Oct 08 18:35:16 crc kubenswrapper[4859]: E1008 18:35:16.498180 4859 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b8433969-f079-49a2-95f3-ce4783d6a62c" containerName="glance-log" Oct 08 18:35:16 crc kubenswrapper[4859]: I1008 18:35:16.498196 4859 state_mem.go:107] "Deleted CPUSet assignment" podUID="b8433969-f079-49a2-95f3-ce4783d6a62c" containerName="glance-log" Oct 08 18:35:16 crc kubenswrapper[4859]: E1008 18:35:16.498212 4859 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d95266a3-07ae-4c7f-9cc7-af1a632290cc" containerName="ceilometer-central-agent" Oct 08 18:35:16 crc kubenswrapper[4859]: I1008 18:35:16.498220 4859 state_mem.go:107] "Deleted CPUSet assignment" podUID="d95266a3-07ae-4c7f-9cc7-af1a632290cc" containerName="ceilometer-central-agent" Oct 08 18:35:16 crc kubenswrapper[4859]: E1008 18:35:16.498252 4859 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d95266a3-07ae-4c7f-9cc7-af1a632290cc" containerName="ceilometer-notification-agent" Oct 08 18:35:16 crc kubenswrapper[4859]: I1008 18:35:16.498261 4859 state_mem.go:107] "Deleted CPUSet assignment" podUID="d95266a3-07ae-4c7f-9cc7-af1a632290cc" containerName="ceilometer-notification-agent" Oct 08 18:35:16 crc kubenswrapper[4859]: E1008 18:35:16.498274 4859 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d95266a3-07ae-4c7f-9cc7-af1a632290cc" containerName="sg-core" Oct 08 18:35:16 crc kubenswrapper[4859]: I1008 18:35:16.498283 4859 state_mem.go:107] "Deleted CPUSet assignment" podUID="d95266a3-07ae-4c7f-9cc7-af1a632290cc" containerName="sg-core" Oct 08 18:35:16 crc kubenswrapper[4859]: E1008 18:35:16.498297 4859 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d95266a3-07ae-4c7f-9cc7-af1a632290cc" containerName="proxy-httpd" Oct 08 18:35:16 crc kubenswrapper[4859]: I1008 18:35:16.498305 4859 state_mem.go:107] "Deleted CPUSet assignment" podUID="d95266a3-07ae-4c7f-9cc7-af1a632290cc" containerName="proxy-httpd" Oct 08 18:35:16 crc kubenswrapper[4859]: E1008 18:35:16.498324 4859 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b8433969-f079-49a2-95f3-ce4783d6a62c" containerName="glance-httpd" Oct 08 18:35:16 crc kubenswrapper[4859]: I1008 18:35:16.498331 4859 state_mem.go:107] "Deleted CPUSet assignment" podUID="b8433969-f079-49a2-95f3-ce4783d6a62c" containerName="glance-httpd" Oct 08 18:35:16 crc kubenswrapper[4859]: I1008 18:35:16.498528 4859 memory_manager.go:354] "RemoveStaleState removing state" podUID="d95266a3-07ae-4c7f-9cc7-af1a632290cc" containerName="ceilometer-notification-agent" Oct 08 18:35:16 crc kubenswrapper[4859]: I1008 18:35:16.498544 4859 memory_manager.go:354] "RemoveStaleState removing state" podUID="b8433969-f079-49a2-95f3-ce4783d6a62c" containerName="glance-log" Oct 08 18:35:16 crc kubenswrapper[4859]: I1008 18:35:16.498554 4859 memory_manager.go:354] "RemoveStaleState removing state" podUID="d95266a3-07ae-4c7f-9cc7-af1a632290cc" containerName="sg-core" Oct 08 18:35:16 crc kubenswrapper[4859]: I1008 18:35:16.498565 4859 memory_manager.go:354] "RemoveStaleState removing state" podUID="d95266a3-07ae-4c7f-9cc7-af1a632290cc" containerName="ceilometer-central-agent" Oct 08 18:35:16 crc kubenswrapper[4859]: I1008 18:35:16.498581 4859 memory_manager.go:354] "RemoveStaleState removing state" podUID="b8433969-f079-49a2-95f3-ce4783d6a62c" containerName="glance-httpd" Oct 08 18:35:16 crc kubenswrapper[4859]: I1008 18:35:16.498599 4859 memory_manager.go:354] "RemoveStaleState removing state" podUID="d95266a3-07ae-4c7f-9cc7-af1a632290cc" containerName="proxy-httpd" Oct 08 18:35:16 crc kubenswrapper[4859]: I1008 18:35:16.507938 4859 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Oct 08 18:35:16 crc kubenswrapper[4859]: I1008 18:35:16.507987 4859 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/glance-default-external-api-0"] Oct 08 18:35:16 crc kubenswrapper[4859]: I1008 18:35:16.508988 4859 scope.go:117] "RemoveContainer" containerID="11a1608a5dd4756100be705360e1e31235b194c38afb2ba257c21278b543ddc0" Oct 08 18:35:16 crc kubenswrapper[4859]: I1008 18:35:16.509249 4859 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Oct 08 18:35:16 crc kubenswrapper[4859]: I1008 18:35:16.509897 4859 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-external-api-0" Oct 08 18:35:16 crc kubenswrapper[4859]: E1008 18:35:16.510242 4859 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"11a1608a5dd4756100be705360e1e31235b194c38afb2ba257c21278b543ddc0\": container with ID starting with 11a1608a5dd4756100be705360e1e31235b194c38afb2ba257c21278b543ddc0 not found: ID does not exist" containerID="11a1608a5dd4756100be705360e1e31235b194c38afb2ba257c21278b543ddc0" Oct 08 18:35:16 crc kubenswrapper[4859]: I1008 18:35:16.510308 4859 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"11a1608a5dd4756100be705360e1e31235b194c38afb2ba257c21278b543ddc0"} err="failed to get container status \"11a1608a5dd4756100be705360e1e31235b194c38afb2ba257c21278b543ddc0\": rpc error: code = NotFound desc = could not find container \"11a1608a5dd4756100be705360e1e31235b194c38afb2ba257c21278b543ddc0\": container with ID starting with 11a1608a5dd4756100be705360e1e31235b194c38afb2ba257c21278b543ddc0 not found: ID does not exist" Oct 08 18:35:16 crc kubenswrapper[4859]: I1008 18:35:16.510344 4859 scope.go:117] "RemoveContainer" containerID="f4b804fb4923d970b1a669c891e974729f108bea4442714c5bb19f4176d284ba" Oct 08 18:35:16 crc kubenswrapper[4859]: I1008 18:35:16.510655 4859 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-ceilometer-internal-svc" Oct 08 18:35:16 crc kubenswrapper[4859]: I1008 18:35:16.512415 4859 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-scripts" Oct 08 18:35:16 crc kubenswrapper[4859]: I1008 18:35:16.512580 4859 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-default-external-config-data" Oct 08 18:35:16 crc kubenswrapper[4859]: I1008 18:35:16.512648 4859 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-config-data" Oct 08 18:35:16 crc kubenswrapper[4859]: I1008 18:35:16.513863 4859 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-glance-default-public-svc" Oct 08 18:35:16 crc kubenswrapper[4859]: E1008 18:35:16.514905 4859 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"f4b804fb4923d970b1a669c891e974729f108bea4442714c5bb19f4176d284ba\": container with ID starting with f4b804fb4923d970b1a669c891e974729f108bea4442714c5bb19f4176d284ba not found: ID does not exist" containerID="f4b804fb4923d970b1a669c891e974729f108bea4442714c5bb19f4176d284ba" Oct 08 18:35:16 crc kubenswrapper[4859]: I1008 18:35:16.514953 4859 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"f4b804fb4923d970b1a669c891e974729f108bea4442714c5bb19f4176d284ba"} err="failed to get container status \"f4b804fb4923d970b1a669c891e974729f108bea4442714c5bb19f4176d284ba\": rpc error: code = NotFound desc = could not find container \"f4b804fb4923d970b1a669c891e974729f108bea4442714c5bb19f4176d284ba\": container with ID starting with f4b804fb4923d970b1a669c891e974729f108bea4442714c5bb19f4176d284ba not found: ID does not exist" Oct 08 18:35:16 crc kubenswrapper[4859]: I1008 18:35:16.514987 4859 scope.go:117] "RemoveContainer" containerID="83b10b00b3fde51de9b23a21cf635ce9b4b442e8ddab240ca122aee97845c3e0" Oct 08 18:35:16 crc kubenswrapper[4859]: I1008 18:35:16.515222 4859 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-external-api-0"] Oct 08 18:35:16 crc kubenswrapper[4859]: E1008 18:35:16.515478 4859 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"83b10b00b3fde51de9b23a21cf635ce9b4b442e8ddab240ca122aee97845c3e0\": container with ID starting with 83b10b00b3fde51de9b23a21cf635ce9b4b442e8ddab240ca122aee97845c3e0 not found: ID does not exist" containerID="83b10b00b3fde51de9b23a21cf635ce9b4b442e8ddab240ca122aee97845c3e0" Oct 08 18:35:16 crc kubenswrapper[4859]: I1008 18:35:16.515514 4859 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"83b10b00b3fde51de9b23a21cf635ce9b4b442e8ddab240ca122aee97845c3e0"} err="failed to get container status \"83b10b00b3fde51de9b23a21cf635ce9b4b442e8ddab240ca122aee97845c3e0\": rpc error: code = NotFound desc = could not find container \"83b10b00b3fde51de9b23a21cf635ce9b4b442e8ddab240ca122aee97845c3e0\": container with ID starting with 83b10b00b3fde51de9b23a21cf635ce9b4b442e8ddab240ca122aee97845c3e0 not found: ID does not exist" Oct 08 18:35:16 crc kubenswrapper[4859]: I1008 18:35:16.515535 4859 scope.go:117] "RemoveContainer" containerID="9a699a097ec98f3331d90f2e9089ee5a2d6472715e420e0f0ac4c813b0f7c776" Oct 08 18:35:16 crc kubenswrapper[4859]: E1008 18:35:16.515971 4859 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"9a699a097ec98f3331d90f2e9089ee5a2d6472715e420e0f0ac4c813b0f7c776\": container with ID starting with 9a699a097ec98f3331d90f2e9089ee5a2d6472715e420e0f0ac4c813b0f7c776 not found: ID does not exist" containerID="9a699a097ec98f3331d90f2e9089ee5a2d6472715e420e0f0ac4c813b0f7c776" Oct 08 18:35:16 crc kubenswrapper[4859]: I1008 18:35:16.515994 4859 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"9a699a097ec98f3331d90f2e9089ee5a2d6472715e420e0f0ac4c813b0f7c776"} err="failed to get container status \"9a699a097ec98f3331d90f2e9089ee5a2d6472715e420e0f0ac4c813b0f7c776\": rpc error: code = NotFound desc = could not find container \"9a699a097ec98f3331d90f2e9089ee5a2d6472715e420e0f0ac4c813b0f7c776\": container with ID starting with 9a699a097ec98f3331d90f2e9089ee5a2d6472715e420e0f0ac4c813b0f7c776 not found: ID does not exist" Oct 08 18:35:16 crc kubenswrapper[4859]: I1008 18:35:16.516007 4859 scope.go:117] "RemoveContainer" containerID="11a1608a5dd4756100be705360e1e31235b194c38afb2ba257c21278b543ddc0" Oct 08 18:35:16 crc kubenswrapper[4859]: I1008 18:35:16.518613 4859 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"11a1608a5dd4756100be705360e1e31235b194c38afb2ba257c21278b543ddc0"} err="failed to get container status \"11a1608a5dd4756100be705360e1e31235b194c38afb2ba257c21278b543ddc0\": rpc error: code = NotFound desc = could not find container \"11a1608a5dd4756100be705360e1e31235b194c38afb2ba257c21278b543ddc0\": container with ID starting with 11a1608a5dd4756100be705360e1e31235b194c38afb2ba257c21278b543ddc0 not found: ID does not exist" Oct 08 18:35:16 crc kubenswrapper[4859]: I1008 18:35:16.518643 4859 scope.go:117] "RemoveContainer" containerID="f4b804fb4923d970b1a669c891e974729f108bea4442714c5bb19f4176d284ba" Oct 08 18:35:16 crc kubenswrapper[4859]: I1008 18:35:16.521755 4859 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"f4b804fb4923d970b1a669c891e974729f108bea4442714c5bb19f4176d284ba"} err="failed to get container status \"f4b804fb4923d970b1a669c891e974729f108bea4442714c5bb19f4176d284ba\": rpc error: code = NotFound desc = could not find container \"f4b804fb4923d970b1a669c891e974729f108bea4442714c5bb19f4176d284ba\": container with ID starting with f4b804fb4923d970b1a669c891e974729f108bea4442714c5bb19f4176d284ba not found: ID does not exist" Oct 08 18:35:16 crc kubenswrapper[4859]: I1008 18:35:16.521784 4859 scope.go:117] "RemoveContainer" containerID="83b10b00b3fde51de9b23a21cf635ce9b4b442e8ddab240ca122aee97845c3e0" Oct 08 18:35:16 crc kubenswrapper[4859]: I1008 18:35:16.522801 4859 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"83b10b00b3fde51de9b23a21cf635ce9b4b442e8ddab240ca122aee97845c3e0"} err="failed to get container status \"83b10b00b3fde51de9b23a21cf635ce9b4b442e8ddab240ca122aee97845c3e0\": rpc error: code = NotFound desc = could not find container \"83b10b00b3fde51de9b23a21cf635ce9b4b442e8ddab240ca122aee97845c3e0\": container with ID starting with 83b10b00b3fde51de9b23a21cf635ce9b4b442e8ddab240ca122aee97845c3e0 not found: ID does not exist" Oct 08 18:35:16 crc kubenswrapper[4859]: I1008 18:35:16.522835 4859 scope.go:117] "RemoveContainer" containerID="9a699a097ec98f3331d90f2e9089ee5a2d6472715e420e0f0ac4c813b0f7c776" Oct 08 18:35:16 crc kubenswrapper[4859]: I1008 18:35:16.524807 4859 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"9a699a097ec98f3331d90f2e9089ee5a2d6472715e420e0f0ac4c813b0f7c776"} err="failed to get container status \"9a699a097ec98f3331d90f2e9089ee5a2d6472715e420e0f0ac4c813b0f7c776\": rpc error: code = NotFound desc = could not find container \"9a699a097ec98f3331d90f2e9089ee5a2d6472715e420e0f0ac4c813b0f7c776\": container with ID starting with 9a699a097ec98f3331d90f2e9089ee5a2d6472715e420e0f0ac4c813b0f7c776 not found: ID does not exist" Oct 08 18:35:16 crc kubenswrapper[4859]: I1008 18:35:16.524836 4859 scope.go:117] "RemoveContainer" containerID="11a1608a5dd4756100be705360e1e31235b194c38afb2ba257c21278b543ddc0" Oct 08 18:35:16 crc kubenswrapper[4859]: I1008 18:35:16.525309 4859 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"11a1608a5dd4756100be705360e1e31235b194c38afb2ba257c21278b543ddc0"} err="failed to get container status \"11a1608a5dd4756100be705360e1e31235b194c38afb2ba257c21278b543ddc0\": rpc error: code = NotFound desc = could not find container \"11a1608a5dd4756100be705360e1e31235b194c38afb2ba257c21278b543ddc0\": container with ID starting with 11a1608a5dd4756100be705360e1e31235b194c38afb2ba257c21278b543ddc0 not found: ID does not exist" Oct 08 18:35:16 crc kubenswrapper[4859]: I1008 18:35:16.525333 4859 scope.go:117] "RemoveContainer" containerID="f4b804fb4923d970b1a669c891e974729f108bea4442714c5bb19f4176d284ba" Oct 08 18:35:16 crc kubenswrapper[4859]: I1008 18:35:16.526769 4859 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"f4b804fb4923d970b1a669c891e974729f108bea4442714c5bb19f4176d284ba"} err="failed to get container status \"f4b804fb4923d970b1a669c891e974729f108bea4442714c5bb19f4176d284ba\": rpc error: code = NotFound desc = could not find container \"f4b804fb4923d970b1a669c891e974729f108bea4442714c5bb19f4176d284ba\": container with ID starting with f4b804fb4923d970b1a669c891e974729f108bea4442714c5bb19f4176d284ba not found: ID does not exist" Oct 08 18:35:16 crc kubenswrapper[4859]: I1008 18:35:16.526800 4859 scope.go:117] "RemoveContainer" containerID="83b10b00b3fde51de9b23a21cf635ce9b4b442e8ddab240ca122aee97845c3e0" Oct 08 18:35:16 crc kubenswrapper[4859]: I1008 18:35:16.528623 4859 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"83b10b00b3fde51de9b23a21cf635ce9b4b442e8ddab240ca122aee97845c3e0"} err="failed to get container status \"83b10b00b3fde51de9b23a21cf635ce9b4b442e8ddab240ca122aee97845c3e0\": rpc error: code = NotFound desc = could not find container \"83b10b00b3fde51de9b23a21cf635ce9b4b442e8ddab240ca122aee97845c3e0\": container with ID starting with 83b10b00b3fde51de9b23a21cf635ce9b4b442e8ddab240ca122aee97845c3e0 not found: ID does not exist" Oct 08 18:35:16 crc kubenswrapper[4859]: I1008 18:35:16.528655 4859 scope.go:117] "RemoveContainer" containerID="9a699a097ec98f3331d90f2e9089ee5a2d6472715e420e0f0ac4c813b0f7c776" Oct 08 18:35:16 crc kubenswrapper[4859]: I1008 18:35:16.532858 4859 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"9a699a097ec98f3331d90f2e9089ee5a2d6472715e420e0f0ac4c813b0f7c776"} err="failed to get container status \"9a699a097ec98f3331d90f2e9089ee5a2d6472715e420e0f0ac4c813b0f7c776\": rpc error: code = NotFound desc = could not find container \"9a699a097ec98f3331d90f2e9089ee5a2d6472715e420e0f0ac4c813b0f7c776\": container with ID starting with 9a699a097ec98f3331d90f2e9089ee5a2d6472715e420e0f0ac4c813b0f7c776 not found: ID does not exist" Oct 08 18:35:16 crc kubenswrapper[4859]: I1008 18:35:16.532909 4859 scope.go:117] "RemoveContainer" containerID="11a1608a5dd4756100be705360e1e31235b194c38afb2ba257c21278b543ddc0" Oct 08 18:35:16 crc kubenswrapper[4859]: I1008 18:35:16.533367 4859 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"11a1608a5dd4756100be705360e1e31235b194c38afb2ba257c21278b543ddc0"} err="failed to get container status \"11a1608a5dd4756100be705360e1e31235b194c38afb2ba257c21278b543ddc0\": rpc error: code = NotFound desc = could not find container \"11a1608a5dd4756100be705360e1e31235b194c38afb2ba257c21278b543ddc0\": container with ID starting with 11a1608a5dd4756100be705360e1e31235b194c38afb2ba257c21278b543ddc0 not found: ID does not exist" Oct 08 18:35:16 crc kubenswrapper[4859]: I1008 18:35:16.533464 4859 scope.go:117] "RemoveContainer" containerID="f4b804fb4923d970b1a669c891e974729f108bea4442714c5bb19f4176d284ba" Oct 08 18:35:16 crc kubenswrapper[4859]: I1008 18:35:16.534933 4859 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"f4b804fb4923d970b1a669c891e974729f108bea4442714c5bb19f4176d284ba"} err="failed to get container status \"f4b804fb4923d970b1a669c891e974729f108bea4442714c5bb19f4176d284ba\": rpc error: code = NotFound desc = could not find container \"f4b804fb4923d970b1a669c891e974729f108bea4442714c5bb19f4176d284ba\": container with ID starting with f4b804fb4923d970b1a669c891e974729f108bea4442714c5bb19f4176d284ba not found: ID does not exist" Oct 08 18:35:16 crc kubenswrapper[4859]: I1008 18:35:16.534980 4859 scope.go:117] "RemoveContainer" containerID="83b10b00b3fde51de9b23a21cf635ce9b4b442e8ddab240ca122aee97845c3e0" Oct 08 18:35:16 crc kubenswrapper[4859]: I1008 18:35:16.535673 4859 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"83b10b00b3fde51de9b23a21cf635ce9b4b442e8ddab240ca122aee97845c3e0"} err="failed to get container status \"83b10b00b3fde51de9b23a21cf635ce9b4b442e8ddab240ca122aee97845c3e0\": rpc error: code = NotFound desc = could not find container \"83b10b00b3fde51de9b23a21cf635ce9b4b442e8ddab240ca122aee97845c3e0\": container with ID starting with 83b10b00b3fde51de9b23a21cf635ce9b4b442e8ddab240ca122aee97845c3e0 not found: ID does not exist" Oct 08 18:35:16 crc kubenswrapper[4859]: I1008 18:35:16.535740 4859 scope.go:117] "RemoveContainer" containerID="9a699a097ec98f3331d90f2e9089ee5a2d6472715e420e0f0ac4c813b0f7c776" Oct 08 18:35:16 crc kubenswrapper[4859]: I1008 18:35:16.537919 4859 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"9a699a097ec98f3331d90f2e9089ee5a2d6472715e420e0f0ac4c813b0f7c776"} err="failed to get container status \"9a699a097ec98f3331d90f2e9089ee5a2d6472715e420e0f0ac4c813b0f7c776\": rpc error: code = NotFound desc = could not find container \"9a699a097ec98f3331d90f2e9089ee5a2d6472715e420e0f0ac4c813b0f7c776\": container with ID starting with 9a699a097ec98f3331d90f2e9089ee5a2d6472715e420e0f0ac4c813b0f7c776 not found: ID does not exist" Oct 08 18:35:16 crc kubenswrapper[4859]: I1008 18:35:16.537962 4859 scope.go:117] "RemoveContainer" containerID="3d02d784fce2d216c2d98154bf821d302aeba5a184bb06406551e52cbd463ff4" Oct 08 18:35:16 crc kubenswrapper[4859]: I1008 18:35:16.580816 4859 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/c3a48f64-6a96-4071-89cc-3f7bc1116694-ceilometer-tls-certs\") pod \"ceilometer-0\" (UID: \"c3a48f64-6a96-4071-89cc-3f7bc1116694\") " pod="openstack/ceilometer-0" Oct 08 18:35:16 crc kubenswrapper[4859]: I1008 18:35:16.580865 4859 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/21a00512-7854-48c3-a040-795eb7d79744-config-data\") pod \"glance-default-external-api-0\" (UID: \"21a00512-7854-48c3-a040-795eb7d79744\") " pod="openstack/glance-default-external-api-0" Oct 08 18:35:16 crc kubenswrapper[4859]: I1008 18:35:16.580917 4859 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/21a00512-7854-48c3-a040-795eb7d79744-logs\") pod \"glance-default-external-api-0\" (UID: \"21a00512-7854-48c3-a040-795eb7d79744\") " pod="openstack/glance-default-external-api-0" Oct 08 18:35:16 crc kubenswrapper[4859]: I1008 18:35:16.580943 4859 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/21a00512-7854-48c3-a040-795eb7d79744-scripts\") pod \"glance-default-external-api-0\" (UID: \"21a00512-7854-48c3-a040-795eb7d79744\") " pod="openstack/glance-default-external-api-0" Oct 08 18:35:16 crc kubenswrapper[4859]: I1008 18:35:16.581009 4859 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/c3a48f64-6a96-4071-89cc-3f7bc1116694-log-httpd\") pod \"ceilometer-0\" (UID: \"c3a48f64-6a96-4071-89cc-3f7bc1116694\") " pod="openstack/ceilometer-0" Oct 08 18:35:16 crc kubenswrapper[4859]: I1008 18:35:16.581031 4859 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/21a00512-7854-48c3-a040-795eb7d79744-public-tls-certs\") pod \"glance-default-external-api-0\" (UID: \"21a00512-7854-48c3-a040-795eb7d79744\") " pod="openstack/glance-default-external-api-0" Oct 08 18:35:16 crc kubenswrapper[4859]: I1008 18:35:16.581054 4859 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/21a00512-7854-48c3-a040-795eb7d79744-combined-ca-bundle\") pod \"glance-default-external-api-0\" (UID: \"21a00512-7854-48c3-a040-795eb7d79744\") " pod="openstack/glance-default-external-api-0" Oct 08 18:35:16 crc kubenswrapper[4859]: I1008 18:35:16.581079 4859 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-v4w8m\" (UniqueName: \"kubernetes.io/projected/c3a48f64-6a96-4071-89cc-3f7bc1116694-kube-api-access-v4w8m\") pod \"ceilometer-0\" (UID: \"c3a48f64-6a96-4071-89cc-3f7bc1116694\") " pod="openstack/ceilometer-0" Oct 08 18:35:16 crc kubenswrapper[4859]: I1008 18:35:16.581141 4859 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/c3a48f64-6a96-4071-89cc-3f7bc1116694-scripts\") pod \"ceilometer-0\" (UID: \"c3a48f64-6a96-4071-89cc-3f7bc1116694\") " pod="openstack/ceilometer-0" Oct 08 18:35:16 crc kubenswrapper[4859]: I1008 18:35:16.581163 4859 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/c3a48f64-6a96-4071-89cc-3f7bc1116694-run-httpd\") pod \"ceilometer-0\" (UID: \"c3a48f64-6a96-4071-89cc-3f7bc1116694\") " pod="openstack/ceilometer-0" Oct 08 18:35:16 crc kubenswrapper[4859]: I1008 18:35:16.581229 4859 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c3a48f64-6a96-4071-89cc-3f7bc1116694-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"c3a48f64-6a96-4071-89cc-3f7bc1116694\") " pod="openstack/ceilometer-0" Oct 08 18:35:16 crc kubenswrapper[4859]: I1008 18:35:16.581322 4859 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/21a00512-7854-48c3-a040-795eb7d79744-httpd-run\") pod \"glance-default-external-api-0\" (UID: \"21a00512-7854-48c3-a040-795eb7d79744\") " pod="openstack/glance-default-external-api-0" Oct 08 18:35:16 crc kubenswrapper[4859]: I1008 18:35:16.581347 4859 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/c3a48f64-6a96-4071-89cc-3f7bc1116694-config-data\") pod \"ceilometer-0\" (UID: \"c3a48f64-6a96-4071-89cc-3f7bc1116694\") " pod="openstack/ceilometer-0" Oct 08 18:35:16 crc kubenswrapper[4859]: I1008 18:35:16.581419 4859 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage12-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage12-crc\") pod \"glance-default-external-api-0\" (UID: \"21a00512-7854-48c3-a040-795eb7d79744\") " pod="openstack/glance-default-external-api-0" Oct 08 18:35:16 crc kubenswrapper[4859]: I1008 18:35:16.581480 4859 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-mxstz\" (UniqueName: \"kubernetes.io/projected/21a00512-7854-48c3-a040-795eb7d79744-kube-api-access-mxstz\") pod \"glance-default-external-api-0\" (UID: \"21a00512-7854-48c3-a040-795eb7d79744\") " pod="openstack/glance-default-external-api-0" Oct 08 18:35:16 crc kubenswrapper[4859]: I1008 18:35:16.581508 4859 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/c3a48f64-6a96-4071-89cc-3f7bc1116694-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"c3a48f64-6a96-4071-89cc-3f7bc1116694\") " pod="openstack/ceilometer-0" Oct 08 18:35:16 crc kubenswrapper[4859]: I1008 18:35:16.628155 4859 scope.go:117] "RemoveContainer" containerID="e58a2a7e6d7cb1671e6f15be2eae9753bf9ba14ba510c49cc6d4d7121ba024ce" Oct 08 18:35:16 crc kubenswrapper[4859]: I1008 18:35:16.655165 4859 scope.go:117] "RemoveContainer" containerID="3d02d784fce2d216c2d98154bf821d302aeba5a184bb06406551e52cbd463ff4" Oct 08 18:35:16 crc kubenswrapper[4859]: E1008 18:35:16.655641 4859 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"3d02d784fce2d216c2d98154bf821d302aeba5a184bb06406551e52cbd463ff4\": container with ID starting with 3d02d784fce2d216c2d98154bf821d302aeba5a184bb06406551e52cbd463ff4 not found: ID does not exist" containerID="3d02d784fce2d216c2d98154bf821d302aeba5a184bb06406551e52cbd463ff4" Oct 08 18:35:16 crc kubenswrapper[4859]: I1008 18:35:16.655675 4859 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"3d02d784fce2d216c2d98154bf821d302aeba5a184bb06406551e52cbd463ff4"} err="failed to get container status \"3d02d784fce2d216c2d98154bf821d302aeba5a184bb06406551e52cbd463ff4\": rpc error: code = NotFound desc = could not find container \"3d02d784fce2d216c2d98154bf821d302aeba5a184bb06406551e52cbd463ff4\": container with ID starting with 3d02d784fce2d216c2d98154bf821d302aeba5a184bb06406551e52cbd463ff4 not found: ID does not exist" Oct 08 18:35:16 crc kubenswrapper[4859]: I1008 18:35:16.655714 4859 scope.go:117] "RemoveContainer" containerID="e58a2a7e6d7cb1671e6f15be2eae9753bf9ba14ba510c49cc6d4d7121ba024ce" Oct 08 18:35:16 crc kubenswrapper[4859]: E1008 18:35:16.656113 4859 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"e58a2a7e6d7cb1671e6f15be2eae9753bf9ba14ba510c49cc6d4d7121ba024ce\": container with ID starting with e58a2a7e6d7cb1671e6f15be2eae9753bf9ba14ba510c49cc6d4d7121ba024ce not found: ID does not exist" containerID="e58a2a7e6d7cb1671e6f15be2eae9753bf9ba14ba510c49cc6d4d7121ba024ce" Oct 08 18:35:16 crc kubenswrapper[4859]: I1008 18:35:16.656138 4859 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"e58a2a7e6d7cb1671e6f15be2eae9753bf9ba14ba510c49cc6d4d7121ba024ce"} err="failed to get container status \"e58a2a7e6d7cb1671e6f15be2eae9753bf9ba14ba510c49cc6d4d7121ba024ce\": rpc error: code = NotFound desc = could not find container \"e58a2a7e6d7cb1671e6f15be2eae9753bf9ba14ba510c49cc6d4d7121ba024ce\": container with ID starting with e58a2a7e6d7cb1671e6f15be2eae9753bf9ba14ba510c49cc6d4d7121ba024ce not found: ID does not exist" Oct 08 18:35:16 crc kubenswrapper[4859]: I1008 18:35:16.683003 4859 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/21a00512-7854-48c3-a040-795eb7d79744-httpd-run\") pod \"glance-default-external-api-0\" (UID: \"21a00512-7854-48c3-a040-795eb7d79744\") " pod="openstack/glance-default-external-api-0" Oct 08 18:35:16 crc kubenswrapper[4859]: I1008 18:35:16.683056 4859 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/c3a48f64-6a96-4071-89cc-3f7bc1116694-config-data\") pod \"ceilometer-0\" (UID: \"c3a48f64-6a96-4071-89cc-3f7bc1116694\") " pod="openstack/ceilometer-0" Oct 08 18:35:16 crc kubenswrapper[4859]: I1008 18:35:16.683123 4859 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage12-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage12-crc\") pod \"glance-default-external-api-0\" (UID: \"21a00512-7854-48c3-a040-795eb7d79744\") " pod="openstack/glance-default-external-api-0" Oct 08 18:35:16 crc kubenswrapper[4859]: I1008 18:35:16.683161 4859 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-mxstz\" (UniqueName: \"kubernetes.io/projected/21a00512-7854-48c3-a040-795eb7d79744-kube-api-access-mxstz\") pod \"glance-default-external-api-0\" (UID: \"21a00512-7854-48c3-a040-795eb7d79744\") " pod="openstack/glance-default-external-api-0" Oct 08 18:35:16 crc kubenswrapper[4859]: I1008 18:35:16.683186 4859 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/c3a48f64-6a96-4071-89cc-3f7bc1116694-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"c3a48f64-6a96-4071-89cc-3f7bc1116694\") " pod="openstack/ceilometer-0" Oct 08 18:35:16 crc kubenswrapper[4859]: I1008 18:35:16.683210 4859 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/c3a48f64-6a96-4071-89cc-3f7bc1116694-ceilometer-tls-certs\") pod \"ceilometer-0\" (UID: \"c3a48f64-6a96-4071-89cc-3f7bc1116694\") " pod="openstack/ceilometer-0" Oct 08 18:35:16 crc kubenswrapper[4859]: I1008 18:35:16.683430 4859 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/21a00512-7854-48c3-a040-795eb7d79744-httpd-run\") pod \"glance-default-external-api-0\" (UID: \"21a00512-7854-48c3-a040-795eb7d79744\") " pod="openstack/glance-default-external-api-0" Oct 08 18:35:16 crc kubenswrapper[4859]: I1008 18:35:16.683750 4859 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage12-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage12-crc\") pod \"glance-default-external-api-0\" (UID: \"21a00512-7854-48c3-a040-795eb7d79744\") device mount path \"/mnt/openstack/pv12\"" pod="openstack/glance-default-external-api-0" Oct 08 18:35:16 crc kubenswrapper[4859]: I1008 18:35:16.683911 4859 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/21a00512-7854-48c3-a040-795eb7d79744-config-data\") pod \"glance-default-external-api-0\" (UID: \"21a00512-7854-48c3-a040-795eb7d79744\") " pod="openstack/glance-default-external-api-0" Oct 08 18:35:16 crc kubenswrapper[4859]: I1008 18:35:16.683994 4859 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/21a00512-7854-48c3-a040-795eb7d79744-logs\") pod \"glance-default-external-api-0\" (UID: \"21a00512-7854-48c3-a040-795eb7d79744\") " pod="openstack/glance-default-external-api-0" Oct 08 18:35:16 crc kubenswrapper[4859]: I1008 18:35:16.684025 4859 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/21a00512-7854-48c3-a040-795eb7d79744-scripts\") pod \"glance-default-external-api-0\" (UID: \"21a00512-7854-48c3-a040-795eb7d79744\") " pod="openstack/glance-default-external-api-0" Oct 08 18:35:16 crc kubenswrapper[4859]: I1008 18:35:16.684083 4859 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/c3a48f64-6a96-4071-89cc-3f7bc1116694-log-httpd\") pod \"ceilometer-0\" (UID: \"c3a48f64-6a96-4071-89cc-3f7bc1116694\") " pod="openstack/ceilometer-0" Oct 08 18:35:16 crc kubenswrapper[4859]: I1008 18:35:16.684107 4859 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/21a00512-7854-48c3-a040-795eb7d79744-public-tls-certs\") pod \"glance-default-external-api-0\" (UID: \"21a00512-7854-48c3-a040-795eb7d79744\") " pod="openstack/glance-default-external-api-0" Oct 08 18:35:16 crc kubenswrapper[4859]: I1008 18:35:16.684132 4859 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-v4w8m\" (UniqueName: \"kubernetes.io/projected/c3a48f64-6a96-4071-89cc-3f7bc1116694-kube-api-access-v4w8m\") pod \"ceilometer-0\" (UID: \"c3a48f64-6a96-4071-89cc-3f7bc1116694\") " pod="openstack/ceilometer-0" Oct 08 18:35:16 crc kubenswrapper[4859]: I1008 18:35:16.684155 4859 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/21a00512-7854-48c3-a040-795eb7d79744-combined-ca-bundle\") pod \"glance-default-external-api-0\" (UID: \"21a00512-7854-48c3-a040-795eb7d79744\") " pod="openstack/glance-default-external-api-0" Oct 08 18:35:16 crc kubenswrapper[4859]: I1008 18:35:16.684218 4859 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/c3a48f64-6a96-4071-89cc-3f7bc1116694-scripts\") pod \"ceilometer-0\" (UID: \"c3a48f64-6a96-4071-89cc-3f7bc1116694\") " pod="openstack/ceilometer-0" Oct 08 18:35:16 crc kubenswrapper[4859]: I1008 18:35:16.684243 4859 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/c3a48f64-6a96-4071-89cc-3f7bc1116694-run-httpd\") pod \"ceilometer-0\" (UID: \"c3a48f64-6a96-4071-89cc-3f7bc1116694\") " pod="openstack/ceilometer-0" Oct 08 18:35:16 crc kubenswrapper[4859]: I1008 18:35:16.684328 4859 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/21a00512-7854-48c3-a040-795eb7d79744-logs\") pod \"glance-default-external-api-0\" (UID: \"21a00512-7854-48c3-a040-795eb7d79744\") " pod="openstack/glance-default-external-api-0" Oct 08 18:35:16 crc kubenswrapper[4859]: I1008 18:35:16.684330 4859 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c3a48f64-6a96-4071-89cc-3f7bc1116694-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"c3a48f64-6a96-4071-89cc-3f7bc1116694\") " pod="openstack/ceilometer-0" Oct 08 18:35:16 crc kubenswrapper[4859]: I1008 18:35:16.685787 4859 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/c3a48f64-6a96-4071-89cc-3f7bc1116694-log-httpd\") pod \"ceilometer-0\" (UID: \"c3a48f64-6a96-4071-89cc-3f7bc1116694\") " pod="openstack/ceilometer-0" Oct 08 18:35:16 crc kubenswrapper[4859]: I1008 18:35:16.685985 4859 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/c3a48f64-6a96-4071-89cc-3f7bc1116694-run-httpd\") pod \"ceilometer-0\" (UID: \"c3a48f64-6a96-4071-89cc-3f7bc1116694\") " pod="openstack/ceilometer-0" Oct 08 18:35:16 crc kubenswrapper[4859]: I1008 18:35:16.692251 4859 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/21a00512-7854-48c3-a040-795eb7d79744-config-data\") pod \"glance-default-external-api-0\" (UID: \"21a00512-7854-48c3-a040-795eb7d79744\") " pod="openstack/glance-default-external-api-0" Oct 08 18:35:16 crc kubenswrapper[4859]: I1008 18:35:16.693242 4859 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/21a00512-7854-48c3-a040-795eb7d79744-scripts\") pod \"glance-default-external-api-0\" (UID: \"21a00512-7854-48c3-a040-795eb7d79744\") " pod="openstack/glance-default-external-api-0" Oct 08 18:35:16 crc kubenswrapper[4859]: I1008 18:35:16.693383 4859 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/c3a48f64-6a96-4071-89cc-3f7bc1116694-scripts\") pod \"ceilometer-0\" (UID: \"c3a48f64-6a96-4071-89cc-3f7bc1116694\") " pod="openstack/ceilometer-0" Oct 08 18:35:16 crc kubenswrapper[4859]: I1008 18:35:16.693524 4859 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/c3a48f64-6a96-4071-89cc-3f7bc1116694-ceilometer-tls-certs\") pod \"ceilometer-0\" (UID: \"c3a48f64-6a96-4071-89cc-3f7bc1116694\") " pod="openstack/ceilometer-0" Oct 08 18:35:16 crc kubenswrapper[4859]: I1008 18:35:16.693655 4859 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c3a48f64-6a96-4071-89cc-3f7bc1116694-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"c3a48f64-6a96-4071-89cc-3f7bc1116694\") " pod="openstack/ceilometer-0" Oct 08 18:35:16 crc kubenswrapper[4859]: I1008 18:35:16.700968 4859 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/c3a48f64-6a96-4071-89cc-3f7bc1116694-config-data\") pod \"ceilometer-0\" (UID: \"c3a48f64-6a96-4071-89cc-3f7bc1116694\") " pod="openstack/ceilometer-0" Oct 08 18:35:16 crc kubenswrapper[4859]: I1008 18:35:16.710482 4859 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/c3a48f64-6a96-4071-89cc-3f7bc1116694-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"c3a48f64-6a96-4071-89cc-3f7bc1116694\") " pod="openstack/ceilometer-0" Oct 08 18:35:16 crc kubenswrapper[4859]: I1008 18:35:16.715031 4859 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-v4w8m\" (UniqueName: \"kubernetes.io/projected/c3a48f64-6a96-4071-89cc-3f7bc1116694-kube-api-access-v4w8m\") pod \"ceilometer-0\" (UID: \"c3a48f64-6a96-4071-89cc-3f7bc1116694\") " pod="openstack/ceilometer-0" Oct 08 18:35:16 crc kubenswrapper[4859]: I1008 18:35:16.716379 4859 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/21a00512-7854-48c3-a040-795eb7d79744-combined-ca-bundle\") pod \"glance-default-external-api-0\" (UID: \"21a00512-7854-48c3-a040-795eb7d79744\") " pod="openstack/glance-default-external-api-0" Oct 08 18:35:16 crc kubenswrapper[4859]: I1008 18:35:16.720963 4859 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-mxstz\" (UniqueName: \"kubernetes.io/projected/21a00512-7854-48c3-a040-795eb7d79744-kube-api-access-mxstz\") pod \"glance-default-external-api-0\" (UID: \"21a00512-7854-48c3-a040-795eb7d79744\") " pod="openstack/glance-default-external-api-0" Oct 08 18:35:16 crc kubenswrapper[4859]: I1008 18:35:16.733928 4859 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/21a00512-7854-48c3-a040-795eb7d79744-public-tls-certs\") pod \"glance-default-external-api-0\" (UID: \"21a00512-7854-48c3-a040-795eb7d79744\") " pod="openstack/glance-default-external-api-0" Oct 08 18:35:16 crc kubenswrapper[4859]: I1008 18:35:16.740566 4859 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage12-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage12-crc\") pod \"glance-default-external-api-0\" (UID: \"21a00512-7854-48c3-a040-795eb7d79744\") " pod="openstack/glance-default-external-api-0" Oct 08 18:35:16 crc kubenswrapper[4859]: I1008 18:35:16.814238 4859 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-internal-api-0" Oct 08 18:35:16 crc kubenswrapper[4859]: I1008 18:35:16.830405 4859 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Oct 08 18:35:16 crc kubenswrapper[4859]: I1008 18:35:16.852669 4859 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-external-api-0" Oct 08 18:35:16 crc kubenswrapper[4859]: I1008 18:35:16.887080 4859 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/7bf3f855-41ae-4de9-ba6c-0fb4f51df141-logs\") pod \"7bf3f855-41ae-4de9-ba6c-0fb4f51df141\" (UID: \"7bf3f855-41ae-4de9-ba6c-0fb4f51df141\") " Oct 08 18:35:16 crc kubenswrapper[4859]: I1008 18:35:16.887137 4859 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"glance\" (UniqueName: \"kubernetes.io/local-volume/local-storage10-crc\") pod \"7bf3f855-41ae-4de9-ba6c-0fb4f51df141\" (UID: \"7bf3f855-41ae-4de9-ba6c-0fb4f51df141\") " Oct 08 18:35:16 crc kubenswrapper[4859]: I1008 18:35:16.887180 4859 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7bf3f855-41ae-4de9-ba6c-0fb4f51df141-combined-ca-bundle\") pod \"7bf3f855-41ae-4de9-ba6c-0fb4f51df141\" (UID: \"7bf3f855-41ae-4de9-ba6c-0fb4f51df141\") " Oct 08 18:35:16 crc kubenswrapper[4859]: I1008 18:35:16.887240 4859 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/7bf3f855-41ae-4de9-ba6c-0fb4f51df141-scripts\") pod \"7bf3f855-41ae-4de9-ba6c-0fb4f51df141\" (UID: \"7bf3f855-41ae-4de9-ba6c-0fb4f51df141\") " Oct 08 18:35:16 crc kubenswrapper[4859]: I1008 18:35:16.887277 4859 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-4ms4c\" (UniqueName: \"kubernetes.io/projected/7bf3f855-41ae-4de9-ba6c-0fb4f51df141-kube-api-access-4ms4c\") pod \"7bf3f855-41ae-4de9-ba6c-0fb4f51df141\" (UID: \"7bf3f855-41ae-4de9-ba6c-0fb4f51df141\") " Oct 08 18:35:16 crc kubenswrapper[4859]: I1008 18:35:16.887305 4859 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/7bf3f855-41ae-4de9-ba6c-0fb4f51df141-httpd-run\") pod \"7bf3f855-41ae-4de9-ba6c-0fb4f51df141\" (UID: \"7bf3f855-41ae-4de9-ba6c-0fb4f51df141\") " Oct 08 18:35:16 crc kubenswrapper[4859]: I1008 18:35:16.887362 4859 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/7bf3f855-41ae-4de9-ba6c-0fb4f51df141-internal-tls-certs\") pod \"7bf3f855-41ae-4de9-ba6c-0fb4f51df141\" (UID: \"7bf3f855-41ae-4de9-ba6c-0fb4f51df141\") " Oct 08 18:35:16 crc kubenswrapper[4859]: I1008 18:35:16.887437 4859 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/7bf3f855-41ae-4de9-ba6c-0fb4f51df141-config-data\") pod \"7bf3f855-41ae-4de9-ba6c-0fb4f51df141\" (UID: \"7bf3f855-41ae-4de9-ba6c-0fb4f51df141\") " Oct 08 18:35:16 crc kubenswrapper[4859]: I1008 18:35:16.893452 4859 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/7bf3f855-41ae-4de9-ba6c-0fb4f51df141-httpd-run" (OuterVolumeSpecName: "httpd-run") pod "7bf3f855-41ae-4de9-ba6c-0fb4f51df141" (UID: "7bf3f855-41ae-4de9-ba6c-0fb4f51df141"). InnerVolumeSpecName "httpd-run". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 08 18:35:16 crc kubenswrapper[4859]: I1008 18:35:16.895735 4859 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/7bf3f855-41ae-4de9-ba6c-0fb4f51df141-logs" (OuterVolumeSpecName: "logs") pod "7bf3f855-41ae-4de9-ba6c-0fb4f51df141" (UID: "7bf3f855-41ae-4de9-ba6c-0fb4f51df141"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 08 18:35:16 crc kubenswrapper[4859]: I1008 18:35:16.896041 4859 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/7bf3f855-41ae-4de9-ba6c-0fb4f51df141-scripts" (OuterVolumeSpecName: "scripts") pod "7bf3f855-41ae-4de9-ba6c-0fb4f51df141" (UID: "7bf3f855-41ae-4de9-ba6c-0fb4f51df141"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 08 18:35:16 crc kubenswrapper[4859]: I1008 18:35:16.902215 4859 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/local-volume/local-storage10-crc" (OuterVolumeSpecName: "glance") pod "7bf3f855-41ae-4de9-ba6c-0fb4f51df141" (UID: "7bf3f855-41ae-4de9-ba6c-0fb4f51df141"). InnerVolumeSpecName "local-storage10-crc". PluginName "kubernetes.io/local-volume", VolumeGidValue "" Oct 08 18:35:16 crc kubenswrapper[4859]: I1008 18:35:16.909132 4859 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/7bf3f855-41ae-4de9-ba6c-0fb4f51df141-kube-api-access-4ms4c" (OuterVolumeSpecName: "kube-api-access-4ms4c") pod "7bf3f855-41ae-4de9-ba6c-0fb4f51df141" (UID: "7bf3f855-41ae-4de9-ba6c-0fb4f51df141"). InnerVolumeSpecName "kube-api-access-4ms4c". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 08 18:35:16 crc kubenswrapper[4859]: I1008 18:35:16.947768 4859 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/7bf3f855-41ae-4de9-ba6c-0fb4f51df141-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "7bf3f855-41ae-4de9-ba6c-0fb4f51df141" (UID: "7bf3f855-41ae-4de9-ba6c-0fb4f51df141"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 08 18:35:16 crc kubenswrapper[4859]: I1008 18:35:16.960006 4859 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/7bf3f855-41ae-4de9-ba6c-0fb4f51df141-config-data" (OuterVolumeSpecName: "config-data") pod "7bf3f855-41ae-4de9-ba6c-0fb4f51df141" (UID: "7bf3f855-41ae-4de9-ba6c-0fb4f51df141"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 08 18:35:16 crc kubenswrapper[4859]: I1008 18:35:16.990613 4859 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/7bf3f855-41ae-4de9-ba6c-0fb4f51df141-config-data\") on node \"crc\" DevicePath \"\"" Oct 08 18:35:16 crc kubenswrapper[4859]: I1008 18:35:16.990648 4859 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/7bf3f855-41ae-4de9-ba6c-0fb4f51df141-logs\") on node \"crc\" DevicePath \"\"" Oct 08 18:35:16 crc kubenswrapper[4859]: I1008 18:35:16.990700 4859 reconciler_common.go:286] "operationExecutor.UnmountDevice started for volume \"local-storage10-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage10-crc\") on node \"crc\" " Oct 08 18:35:16 crc kubenswrapper[4859]: I1008 18:35:16.990717 4859 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7bf3f855-41ae-4de9-ba6c-0fb4f51df141-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 08 18:35:16 crc kubenswrapper[4859]: I1008 18:35:16.990729 4859 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/7bf3f855-41ae-4de9-ba6c-0fb4f51df141-scripts\") on node \"crc\" DevicePath \"\"" Oct 08 18:35:16 crc kubenswrapper[4859]: I1008 18:35:16.990739 4859 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-4ms4c\" (UniqueName: \"kubernetes.io/projected/7bf3f855-41ae-4de9-ba6c-0fb4f51df141-kube-api-access-4ms4c\") on node \"crc\" DevicePath \"\"" Oct 08 18:35:16 crc kubenswrapper[4859]: I1008 18:35:16.990750 4859 reconciler_common.go:293] "Volume detached for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/7bf3f855-41ae-4de9-ba6c-0fb4f51df141-httpd-run\") on node \"crc\" DevicePath \"\"" Oct 08 18:35:17 crc kubenswrapper[4859]: I1008 18:35:17.019503 4859 operation_generator.go:917] UnmountDevice succeeded for volume "local-storage10-crc" (UniqueName: "kubernetes.io/local-volume/local-storage10-crc") on node "crc" Oct 08 18:35:17 crc kubenswrapper[4859]: I1008 18:35:17.025471 4859 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/7bf3f855-41ae-4de9-ba6c-0fb4f51df141-internal-tls-certs" (OuterVolumeSpecName: "internal-tls-certs") pod "7bf3f855-41ae-4de9-ba6c-0fb4f51df141" (UID: "7bf3f855-41ae-4de9-ba6c-0fb4f51df141"). InnerVolumeSpecName "internal-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 08 18:35:17 crc kubenswrapper[4859]: I1008 18:35:17.092677 4859 reconciler_common.go:293] "Volume detached for volume \"local-storage10-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage10-crc\") on node \"crc\" DevicePath \"\"" Oct 08 18:35:17 crc kubenswrapper[4859]: I1008 18:35:17.092723 4859 reconciler_common.go:293] "Volume detached for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/7bf3f855-41ae-4de9-ba6c-0fb4f51df141-internal-tls-certs\") on node \"crc\" DevicePath \"\"" Oct 08 18:35:17 crc kubenswrapper[4859]: I1008 18:35:17.227445 4859 generic.go:334] "Generic (PLEG): container finished" podID="7bf3f855-41ae-4de9-ba6c-0fb4f51df141" containerID="46ed96c026d937d863774320b18d16cbedb2414f358afc23e4af1c83103bb4e5" exitCode=0 Oct 08 18:35:17 crc kubenswrapper[4859]: I1008 18:35:17.227513 4859 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-internal-api-0" Oct 08 18:35:17 crc kubenswrapper[4859]: I1008 18:35:17.227512 4859 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"7bf3f855-41ae-4de9-ba6c-0fb4f51df141","Type":"ContainerDied","Data":"46ed96c026d937d863774320b18d16cbedb2414f358afc23e4af1c83103bb4e5"} Oct 08 18:35:17 crc kubenswrapper[4859]: I1008 18:35:17.227586 4859 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"7bf3f855-41ae-4de9-ba6c-0fb4f51df141","Type":"ContainerDied","Data":"3eb8702bed8879b712d748acd343a4ab86e37d62f33dc654c6a14848f6eaccc5"} Oct 08 18:35:17 crc kubenswrapper[4859]: I1008 18:35:17.227606 4859 scope.go:117] "RemoveContainer" containerID="46ed96c026d937d863774320b18d16cbedb2414f358afc23e4af1c83103bb4e5" Oct 08 18:35:17 crc kubenswrapper[4859]: I1008 18:35:17.304068 4859 scope.go:117] "RemoveContainer" containerID="28ec01a1442339cbd321103beebc7df267b3a533827492d256693d3f04972c5d" Oct 08 18:35:17 crc kubenswrapper[4859]: I1008 18:35:17.306371 4859 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-default-internal-api-0"] Oct 08 18:35:17 crc kubenswrapper[4859]: I1008 18:35:17.318719 4859 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/glance-default-internal-api-0"] Oct 08 18:35:17 crc kubenswrapper[4859]: I1008 18:35:17.328002 4859 scope.go:117] "RemoveContainer" containerID="46ed96c026d937d863774320b18d16cbedb2414f358afc23e4af1c83103bb4e5" Oct 08 18:35:17 crc kubenswrapper[4859]: I1008 18:35:17.328063 4859 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/glance-default-internal-api-0"] Oct 08 18:35:17 crc kubenswrapper[4859]: E1008 18:35:17.328822 4859 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7bf3f855-41ae-4de9-ba6c-0fb4f51df141" containerName="glance-log" Oct 08 18:35:17 crc kubenswrapper[4859]: I1008 18:35:17.328934 4859 state_mem.go:107] "Deleted CPUSet assignment" podUID="7bf3f855-41ae-4de9-ba6c-0fb4f51df141" containerName="glance-log" Oct 08 18:35:17 crc kubenswrapper[4859]: E1008 18:35:17.329032 4859 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7bf3f855-41ae-4de9-ba6c-0fb4f51df141" containerName="glance-httpd" Oct 08 18:35:17 crc kubenswrapper[4859]: I1008 18:35:17.329089 4859 state_mem.go:107] "Deleted CPUSet assignment" podUID="7bf3f855-41ae-4de9-ba6c-0fb4f51df141" containerName="glance-httpd" Oct 08 18:35:17 crc kubenswrapper[4859]: I1008 18:35:17.329366 4859 memory_manager.go:354] "RemoveStaleState removing state" podUID="7bf3f855-41ae-4de9-ba6c-0fb4f51df141" containerName="glance-httpd" Oct 08 18:35:17 crc kubenswrapper[4859]: I1008 18:35:17.329472 4859 memory_manager.go:354] "RemoveStaleState removing state" podUID="7bf3f855-41ae-4de9-ba6c-0fb4f51df141" containerName="glance-log" Oct 08 18:35:17 crc kubenswrapper[4859]: E1008 18:35:17.329724 4859 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"46ed96c026d937d863774320b18d16cbedb2414f358afc23e4af1c83103bb4e5\": container with ID starting with 46ed96c026d937d863774320b18d16cbedb2414f358afc23e4af1c83103bb4e5 not found: ID does not exist" containerID="46ed96c026d937d863774320b18d16cbedb2414f358afc23e4af1c83103bb4e5" Oct 08 18:35:17 crc kubenswrapper[4859]: I1008 18:35:17.329763 4859 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"46ed96c026d937d863774320b18d16cbedb2414f358afc23e4af1c83103bb4e5"} err="failed to get container status \"46ed96c026d937d863774320b18d16cbedb2414f358afc23e4af1c83103bb4e5\": rpc error: code = NotFound desc = could not find container \"46ed96c026d937d863774320b18d16cbedb2414f358afc23e4af1c83103bb4e5\": container with ID starting with 46ed96c026d937d863774320b18d16cbedb2414f358afc23e4af1c83103bb4e5 not found: ID does not exist" Oct 08 18:35:17 crc kubenswrapper[4859]: I1008 18:35:17.329789 4859 scope.go:117] "RemoveContainer" containerID="28ec01a1442339cbd321103beebc7df267b3a533827492d256693d3f04972c5d" Oct 08 18:35:17 crc kubenswrapper[4859]: E1008 18:35:17.332141 4859 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"28ec01a1442339cbd321103beebc7df267b3a533827492d256693d3f04972c5d\": container with ID starting with 28ec01a1442339cbd321103beebc7df267b3a533827492d256693d3f04972c5d not found: ID does not exist" containerID="28ec01a1442339cbd321103beebc7df267b3a533827492d256693d3f04972c5d" Oct 08 18:35:17 crc kubenswrapper[4859]: I1008 18:35:17.332176 4859 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"28ec01a1442339cbd321103beebc7df267b3a533827492d256693d3f04972c5d"} err="failed to get container status \"28ec01a1442339cbd321103beebc7df267b3a533827492d256693d3f04972c5d\": rpc error: code = NotFound desc = could not find container \"28ec01a1442339cbd321103beebc7df267b3a533827492d256693d3f04972c5d\": container with ID starting with 28ec01a1442339cbd321103beebc7df267b3a533827492d256693d3f04972c5d not found: ID does not exist" Oct 08 18:35:17 crc kubenswrapper[4859]: I1008 18:35:17.334424 4859 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-internal-api-0" Oct 08 18:35:17 crc kubenswrapper[4859]: I1008 18:35:17.338823 4859 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-default-internal-config-data" Oct 08 18:35:17 crc kubenswrapper[4859]: I1008 18:35:17.339124 4859 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-glance-default-internal-svc" Oct 08 18:35:17 crc kubenswrapper[4859]: I1008 18:35:17.361675 4859 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-internal-api-0"] Oct 08 18:35:17 crc kubenswrapper[4859]: I1008 18:35:17.373446 4859 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Oct 08 18:35:17 crc kubenswrapper[4859]: I1008 18:35:17.399122 4859 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d14936ce-b30c-44e7-95bc-9a4d516e71b1-combined-ca-bundle\") pod \"glance-default-internal-api-0\" (UID: \"d14936ce-b30c-44e7-95bc-9a4d516e71b1\") " pod="openstack/glance-default-internal-api-0" Oct 08 18:35:17 crc kubenswrapper[4859]: I1008 18:35:17.399180 4859 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-pbrr8\" (UniqueName: \"kubernetes.io/projected/d14936ce-b30c-44e7-95bc-9a4d516e71b1-kube-api-access-pbrr8\") pod \"glance-default-internal-api-0\" (UID: \"d14936ce-b30c-44e7-95bc-9a4d516e71b1\") " pod="openstack/glance-default-internal-api-0" Oct 08 18:35:17 crc kubenswrapper[4859]: I1008 18:35:17.399309 4859 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/d14936ce-b30c-44e7-95bc-9a4d516e71b1-logs\") pod \"glance-default-internal-api-0\" (UID: \"d14936ce-b30c-44e7-95bc-9a4d516e71b1\") " pod="openstack/glance-default-internal-api-0" Oct 08 18:35:17 crc kubenswrapper[4859]: I1008 18:35:17.399412 4859 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage10-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage10-crc\") pod \"glance-default-internal-api-0\" (UID: \"d14936ce-b30c-44e7-95bc-9a4d516e71b1\") " pod="openstack/glance-default-internal-api-0" Oct 08 18:35:17 crc kubenswrapper[4859]: I1008 18:35:17.399538 4859 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/d14936ce-b30c-44e7-95bc-9a4d516e71b1-httpd-run\") pod \"glance-default-internal-api-0\" (UID: \"d14936ce-b30c-44e7-95bc-9a4d516e71b1\") " pod="openstack/glance-default-internal-api-0" Oct 08 18:35:17 crc kubenswrapper[4859]: I1008 18:35:17.399634 4859 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/d14936ce-b30c-44e7-95bc-9a4d516e71b1-scripts\") pod \"glance-default-internal-api-0\" (UID: \"d14936ce-b30c-44e7-95bc-9a4d516e71b1\") " pod="openstack/glance-default-internal-api-0" Oct 08 18:35:17 crc kubenswrapper[4859]: I1008 18:35:17.399682 4859 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d14936ce-b30c-44e7-95bc-9a4d516e71b1-config-data\") pod \"glance-default-internal-api-0\" (UID: \"d14936ce-b30c-44e7-95bc-9a4d516e71b1\") " pod="openstack/glance-default-internal-api-0" Oct 08 18:35:17 crc kubenswrapper[4859]: I1008 18:35:17.399722 4859 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/d14936ce-b30c-44e7-95bc-9a4d516e71b1-internal-tls-certs\") pod \"glance-default-internal-api-0\" (UID: \"d14936ce-b30c-44e7-95bc-9a4d516e71b1\") " pod="openstack/glance-default-internal-api-0" Oct 08 18:35:17 crc kubenswrapper[4859]: I1008 18:35:17.500941 4859 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/d14936ce-b30c-44e7-95bc-9a4d516e71b1-logs\") pod \"glance-default-internal-api-0\" (UID: \"d14936ce-b30c-44e7-95bc-9a4d516e71b1\") " pod="openstack/glance-default-internal-api-0" Oct 08 18:35:17 crc kubenswrapper[4859]: I1008 18:35:17.501027 4859 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage10-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage10-crc\") pod \"glance-default-internal-api-0\" (UID: \"d14936ce-b30c-44e7-95bc-9a4d516e71b1\") " pod="openstack/glance-default-internal-api-0" Oct 08 18:35:17 crc kubenswrapper[4859]: I1008 18:35:17.501135 4859 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/d14936ce-b30c-44e7-95bc-9a4d516e71b1-httpd-run\") pod \"glance-default-internal-api-0\" (UID: \"d14936ce-b30c-44e7-95bc-9a4d516e71b1\") " pod="openstack/glance-default-internal-api-0" Oct 08 18:35:17 crc kubenswrapper[4859]: I1008 18:35:17.501505 4859 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage10-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage10-crc\") pod \"glance-default-internal-api-0\" (UID: \"d14936ce-b30c-44e7-95bc-9a4d516e71b1\") device mount path \"/mnt/openstack/pv10\"" pod="openstack/glance-default-internal-api-0" Oct 08 18:35:17 crc kubenswrapper[4859]: I1008 18:35:17.501505 4859 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/d14936ce-b30c-44e7-95bc-9a4d516e71b1-logs\") pod \"glance-default-internal-api-0\" (UID: \"d14936ce-b30c-44e7-95bc-9a4d516e71b1\") " pod="openstack/glance-default-internal-api-0" Oct 08 18:35:17 crc kubenswrapper[4859]: I1008 18:35:17.501936 4859 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/d14936ce-b30c-44e7-95bc-9a4d516e71b1-scripts\") pod \"glance-default-internal-api-0\" (UID: \"d14936ce-b30c-44e7-95bc-9a4d516e71b1\") " pod="openstack/glance-default-internal-api-0" Oct 08 18:35:17 crc kubenswrapper[4859]: I1008 18:35:17.502204 4859 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d14936ce-b30c-44e7-95bc-9a4d516e71b1-config-data\") pod \"glance-default-internal-api-0\" (UID: \"d14936ce-b30c-44e7-95bc-9a4d516e71b1\") " pod="openstack/glance-default-internal-api-0" Oct 08 18:35:17 crc kubenswrapper[4859]: I1008 18:35:17.502238 4859 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/d14936ce-b30c-44e7-95bc-9a4d516e71b1-internal-tls-certs\") pod \"glance-default-internal-api-0\" (UID: \"d14936ce-b30c-44e7-95bc-9a4d516e71b1\") " pod="openstack/glance-default-internal-api-0" Oct 08 18:35:17 crc kubenswrapper[4859]: I1008 18:35:17.502457 4859 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d14936ce-b30c-44e7-95bc-9a4d516e71b1-combined-ca-bundle\") pod \"glance-default-internal-api-0\" (UID: \"d14936ce-b30c-44e7-95bc-9a4d516e71b1\") " pod="openstack/glance-default-internal-api-0" Oct 08 18:35:17 crc kubenswrapper[4859]: I1008 18:35:17.502510 4859 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-pbrr8\" (UniqueName: \"kubernetes.io/projected/d14936ce-b30c-44e7-95bc-9a4d516e71b1-kube-api-access-pbrr8\") pod \"glance-default-internal-api-0\" (UID: \"d14936ce-b30c-44e7-95bc-9a4d516e71b1\") " pod="openstack/glance-default-internal-api-0" Oct 08 18:35:17 crc kubenswrapper[4859]: I1008 18:35:17.502883 4859 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/d14936ce-b30c-44e7-95bc-9a4d516e71b1-httpd-run\") pod \"glance-default-internal-api-0\" (UID: \"d14936ce-b30c-44e7-95bc-9a4d516e71b1\") " pod="openstack/glance-default-internal-api-0" Oct 08 18:35:17 crc kubenswrapper[4859]: I1008 18:35:17.508500 4859 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/d14936ce-b30c-44e7-95bc-9a4d516e71b1-internal-tls-certs\") pod \"glance-default-internal-api-0\" (UID: \"d14936ce-b30c-44e7-95bc-9a4d516e71b1\") " pod="openstack/glance-default-internal-api-0" Oct 08 18:35:17 crc kubenswrapper[4859]: I1008 18:35:17.508821 4859 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d14936ce-b30c-44e7-95bc-9a4d516e71b1-config-data\") pod \"glance-default-internal-api-0\" (UID: \"d14936ce-b30c-44e7-95bc-9a4d516e71b1\") " pod="openstack/glance-default-internal-api-0" Oct 08 18:35:17 crc kubenswrapper[4859]: I1008 18:35:17.511632 4859 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/d14936ce-b30c-44e7-95bc-9a4d516e71b1-scripts\") pod \"glance-default-internal-api-0\" (UID: \"d14936ce-b30c-44e7-95bc-9a4d516e71b1\") " pod="openstack/glance-default-internal-api-0" Oct 08 18:35:17 crc kubenswrapper[4859]: I1008 18:35:17.518601 4859 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d14936ce-b30c-44e7-95bc-9a4d516e71b1-combined-ca-bundle\") pod \"glance-default-internal-api-0\" (UID: \"d14936ce-b30c-44e7-95bc-9a4d516e71b1\") " pod="openstack/glance-default-internal-api-0" Oct 08 18:35:17 crc kubenswrapper[4859]: I1008 18:35:17.538531 4859 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-external-api-0"] Oct 08 18:35:17 crc kubenswrapper[4859]: I1008 18:35:17.543096 4859 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-pbrr8\" (UniqueName: \"kubernetes.io/projected/d14936ce-b30c-44e7-95bc-9a4d516e71b1-kube-api-access-pbrr8\") pod \"glance-default-internal-api-0\" (UID: \"d14936ce-b30c-44e7-95bc-9a4d516e71b1\") " pod="openstack/glance-default-internal-api-0" Oct 08 18:35:17 crc kubenswrapper[4859]: I1008 18:35:17.606117 4859 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage10-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage10-crc\") pod \"glance-default-internal-api-0\" (UID: \"d14936ce-b30c-44e7-95bc-9a4d516e71b1\") " pod="openstack/glance-default-internal-api-0" Oct 08 18:35:17 crc kubenswrapper[4859]: I1008 18:35:17.702111 4859 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-internal-api-0" Oct 08 18:35:18 crc kubenswrapper[4859]: I1008 18:35:18.244489 4859 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"21a00512-7854-48c3-a040-795eb7d79744","Type":"ContainerStarted","Data":"aeb006a1ab39ccf579b860642fdee277420558a5d75644140f2c20db66ba48d3"} Oct 08 18:35:18 crc kubenswrapper[4859]: I1008 18:35:18.247320 4859 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"c3a48f64-6a96-4071-89cc-3f7bc1116694","Type":"ContainerStarted","Data":"fff6d6678da7c49d4e113d7c3d7de23ebfcc4fa849576243ee55d22c2a584db2"} Oct 08 18:35:18 crc kubenswrapper[4859]: I1008 18:35:18.247364 4859 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"c3a48f64-6a96-4071-89cc-3f7bc1116694","Type":"ContainerStarted","Data":"057ee2a69f865fcd963174397e456288f60ac2c44eb9d6e283a460bebd15b751"} Oct 08 18:35:18 crc kubenswrapper[4859]: I1008 18:35:18.325286 4859 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-internal-api-0"] Oct 08 18:35:18 crc kubenswrapper[4859]: W1008 18:35:18.331369 4859 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podd14936ce_b30c_44e7_95bc_9a4d516e71b1.slice/crio-293291c734e50153d83d3aa4fa6f148cea64225f439cffe762d6baf6100ccea2 WatchSource:0}: Error finding container 293291c734e50153d83d3aa4fa6f148cea64225f439cffe762d6baf6100ccea2: Status 404 returned error can't find the container with id 293291c734e50153d83d3aa4fa6f148cea64225f439cffe762d6baf6100ccea2 Oct 08 18:35:18 crc kubenswrapper[4859]: I1008 18:35:18.481629 4859 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="7bf3f855-41ae-4de9-ba6c-0fb4f51df141" path="/var/lib/kubelet/pods/7bf3f855-41ae-4de9-ba6c-0fb4f51df141/volumes" Oct 08 18:35:18 crc kubenswrapper[4859]: I1008 18:35:18.758095 4859 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell0-e838-account-create-6ndb4"] Oct 08 18:35:18 crc kubenswrapper[4859]: I1008 18:35:18.763287 4859 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-e838-account-create-6ndb4" Oct 08 18:35:18 crc kubenswrapper[4859]: I1008 18:35:18.766109 4859 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell0-db-secret" Oct 08 18:35:18 crc kubenswrapper[4859]: I1008 18:35:18.785654 4859 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-e838-account-create-6ndb4"] Oct 08 18:35:18 crc kubenswrapper[4859]: I1008 18:35:18.846896 4859 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-flnts\" (UniqueName: \"kubernetes.io/projected/4f171e6e-3d04-4e18-9b70-b183e9fb222d-kube-api-access-flnts\") pod \"nova-cell0-e838-account-create-6ndb4\" (UID: \"4f171e6e-3d04-4e18-9b70-b183e9fb222d\") " pod="openstack/nova-cell0-e838-account-create-6ndb4" Oct 08 18:35:18 crc kubenswrapper[4859]: I1008 18:35:18.947140 4859 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell1-f8cf-account-create-ftsj5"] Oct 08 18:35:18 crc kubenswrapper[4859]: I1008 18:35:18.948486 4859 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-f8cf-account-create-ftsj5" Oct 08 18:35:18 crc kubenswrapper[4859]: I1008 18:35:18.949215 4859 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-flnts\" (UniqueName: \"kubernetes.io/projected/4f171e6e-3d04-4e18-9b70-b183e9fb222d-kube-api-access-flnts\") pod \"nova-cell0-e838-account-create-6ndb4\" (UID: \"4f171e6e-3d04-4e18-9b70-b183e9fb222d\") " pod="openstack/nova-cell0-e838-account-create-6ndb4" Oct 08 18:35:18 crc kubenswrapper[4859]: I1008 18:35:18.956528 4859 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell1-db-secret" Oct 08 18:35:18 crc kubenswrapper[4859]: I1008 18:35:18.962935 4859 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-f8cf-account-create-ftsj5"] Oct 08 18:35:18 crc kubenswrapper[4859]: I1008 18:35:18.976662 4859 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-flnts\" (UniqueName: \"kubernetes.io/projected/4f171e6e-3d04-4e18-9b70-b183e9fb222d-kube-api-access-flnts\") pod \"nova-cell0-e838-account-create-6ndb4\" (UID: \"4f171e6e-3d04-4e18-9b70-b183e9fb222d\") " pod="openstack/nova-cell0-e838-account-create-6ndb4" Oct 08 18:35:19 crc kubenswrapper[4859]: I1008 18:35:19.051295 4859 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-8kr5t\" (UniqueName: \"kubernetes.io/projected/d68d0cff-c3b1-45bb-a00c-23febbb14aee-kube-api-access-8kr5t\") pod \"nova-cell1-f8cf-account-create-ftsj5\" (UID: \"d68d0cff-c3b1-45bb-a00c-23febbb14aee\") " pod="openstack/nova-cell1-f8cf-account-create-ftsj5" Oct 08 18:35:19 crc kubenswrapper[4859]: I1008 18:35:19.107984 4859 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-e838-account-create-6ndb4" Oct 08 18:35:19 crc kubenswrapper[4859]: I1008 18:35:19.152914 4859 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-8kr5t\" (UniqueName: \"kubernetes.io/projected/d68d0cff-c3b1-45bb-a00c-23febbb14aee-kube-api-access-8kr5t\") pod \"nova-cell1-f8cf-account-create-ftsj5\" (UID: \"d68d0cff-c3b1-45bb-a00c-23febbb14aee\") " pod="openstack/nova-cell1-f8cf-account-create-ftsj5" Oct 08 18:35:19 crc kubenswrapper[4859]: I1008 18:35:19.175402 4859 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-8kr5t\" (UniqueName: \"kubernetes.io/projected/d68d0cff-c3b1-45bb-a00c-23febbb14aee-kube-api-access-8kr5t\") pod \"nova-cell1-f8cf-account-create-ftsj5\" (UID: \"d68d0cff-c3b1-45bb-a00c-23febbb14aee\") " pod="openstack/nova-cell1-f8cf-account-create-ftsj5" Oct 08 18:35:19 crc kubenswrapper[4859]: I1008 18:35:19.288805 4859 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-f8cf-account-create-ftsj5" Oct 08 18:35:19 crc kubenswrapper[4859]: I1008 18:35:19.334199 4859 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"c3a48f64-6a96-4071-89cc-3f7bc1116694","Type":"ContainerStarted","Data":"b2db0d54d3ba08a9d3d383f5c23225939102d4c15f94e0ceabf208c65ab6a1e4"} Oct 08 18:35:19 crc kubenswrapper[4859]: I1008 18:35:19.355199 4859 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"21a00512-7854-48c3-a040-795eb7d79744","Type":"ContainerStarted","Data":"4ed7208c7d7b03dae625c1fe9e682f913155d27f4a82de2c6560e7af7b868370"} Oct 08 18:35:19 crc kubenswrapper[4859]: I1008 18:35:19.355266 4859 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"21a00512-7854-48c3-a040-795eb7d79744","Type":"ContainerStarted","Data":"15ea05047e40a09606a79d0c86cd0a77d48dd5e5a44bd7aa70931b203217a588"} Oct 08 18:35:19 crc kubenswrapper[4859]: I1008 18:35:19.361003 4859 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"d14936ce-b30c-44e7-95bc-9a4d516e71b1","Type":"ContainerStarted","Data":"d179120d776943079959cc1fe6b45c5cbb42bd1fa6fb09cf438ecb9ec7767420"} Oct 08 18:35:19 crc kubenswrapper[4859]: I1008 18:35:19.361046 4859 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"d14936ce-b30c-44e7-95bc-9a4d516e71b1","Type":"ContainerStarted","Data":"293291c734e50153d83d3aa4fa6f148cea64225f439cffe762d6baf6100ccea2"} Oct 08 18:35:19 crc kubenswrapper[4859]: I1008 18:35:19.385424 4859 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/glance-default-external-api-0" podStartSLOduration=3.385404849 podStartE2EDuration="3.385404849s" podCreationTimestamp="2025-10-08 18:35:16 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-08 18:35:19.381010399 +0000 UTC m=+1089.627849778" watchObservedRunningTime="2025-10-08 18:35:19.385404849 +0000 UTC m=+1089.632244218" Oct 08 18:35:19 crc kubenswrapper[4859]: I1008 18:35:19.641791 4859 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-e838-account-create-6ndb4"] Oct 08 18:35:19 crc kubenswrapper[4859]: I1008 18:35:19.829004 4859 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-f8cf-account-create-ftsj5"] Oct 08 18:35:19 crc kubenswrapper[4859]: W1008 18:35:19.835493 4859 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podd68d0cff_c3b1_45bb_a00c_23febbb14aee.slice/crio-d8414f6d45284e8e04b64655146eb710bc1bf9130537d6ea14fe43eb39840715 WatchSource:0}: Error finding container d8414f6d45284e8e04b64655146eb710bc1bf9130537d6ea14fe43eb39840715: Status 404 returned error can't find the container with id d8414f6d45284e8e04b64655146eb710bc1bf9130537d6ea14fe43eb39840715 Oct 08 18:35:19 crc kubenswrapper[4859]: I1008 18:35:19.973345 4859 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Oct 08 18:35:20 crc kubenswrapper[4859]: I1008 18:35:20.375169 4859 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"d14936ce-b30c-44e7-95bc-9a4d516e71b1","Type":"ContainerStarted","Data":"e746e00a86b507848c9a22b19ef1702eee502fcb8c1a5628df1266a4eaab4a7c"} Oct 08 18:35:20 crc kubenswrapper[4859]: I1008 18:35:20.377045 4859 generic.go:334] "Generic (PLEG): container finished" podID="4f171e6e-3d04-4e18-9b70-b183e9fb222d" containerID="8d304c8b88fe273732cb1eb09623aa4d5469a346872436266c01a7b9eec8dd66" exitCode=0 Oct 08 18:35:20 crc kubenswrapper[4859]: I1008 18:35:20.377111 4859 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-e838-account-create-6ndb4" event={"ID":"4f171e6e-3d04-4e18-9b70-b183e9fb222d","Type":"ContainerDied","Data":"8d304c8b88fe273732cb1eb09623aa4d5469a346872436266c01a7b9eec8dd66"} Oct 08 18:35:20 crc kubenswrapper[4859]: I1008 18:35:20.377135 4859 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-e838-account-create-6ndb4" event={"ID":"4f171e6e-3d04-4e18-9b70-b183e9fb222d","Type":"ContainerStarted","Data":"a1cdaa884c8ca64de007c7653133344e813344b13e2d15c8b2ac9e7dadcb6c9e"} Oct 08 18:35:20 crc kubenswrapper[4859]: I1008 18:35:20.379178 4859 generic.go:334] "Generic (PLEG): container finished" podID="d68d0cff-c3b1-45bb-a00c-23febbb14aee" containerID="07779e470ce52d0f4ab06a584cd0ab7f58267d01ef6cd06afd05e3def2f0a59b" exitCode=0 Oct 08 18:35:20 crc kubenswrapper[4859]: I1008 18:35:20.379224 4859 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-f8cf-account-create-ftsj5" event={"ID":"d68d0cff-c3b1-45bb-a00c-23febbb14aee","Type":"ContainerDied","Data":"07779e470ce52d0f4ab06a584cd0ab7f58267d01ef6cd06afd05e3def2f0a59b"} Oct 08 18:35:20 crc kubenswrapper[4859]: I1008 18:35:20.379242 4859 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-f8cf-account-create-ftsj5" event={"ID":"d68d0cff-c3b1-45bb-a00c-23febbb14aee","Type":"ContainerStarted","Data":"d8414f6d45284e8e04b64655146eb710bc1bf9130537d6ea14fe43eb39840715"} Oct 08 18:35:20 crc kubenswrapper[4859]: I1008 18:35:20.383810 4859 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"c3a48f64-6a96-4071-89cc-3f7bc1116694","Type":"ContainerStarted","Data":"4509cf6e879a4cf5f404205c0e2832d503607bfe89ccfbe5070b7ce8ba4acee6"} Oct 08 18:35:20 crc kubenswrapper[4859]: I1008 18:35:20.413972 4859 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/glance-default-internal-api-0" podStartSLOduration=3.413951431 podStartE2EDuration="3.413951431s" podCreationTimestamp="2025-10-08 18:35:17 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-08 18:35:20.401230735 +0000 UTC m=+1090.648070124" watchObservedRunningTime="2025-10-08 18:35:20.413951431 +0000 UTC m=+1090.660790810" Oct 08 18:35:21 crc kubenswrapper[4859]: I1008 18:35:21.400523 4859 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstackclient" event={"ID":"1a631d33-4c0c-46c9-a3d5-4d9dc9f6de84","Type":"ContainerStarted","Data":"609fbbf116fe423df155d8db401a219350e114d03b9a30ec4d530df5f1008826"} Oct 08 18:35:21 crc kubenswrapper[4859]: I1008 18:35:21.848319 4859 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-f8cf-account-create-ftsj5" Oct 08 18:35:21 crc kubenswrapper[4859]: I1008 18:35:21.853815 4859 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-e838-account-create-6ndb4" Oct 08 18:35:21 crc kubenswrapper[4859]: I1008 18:35:21.920090 4859 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-flnts\" (UniqueName: \"kubernetes.io/projected/4f171e6e-3d04-4e18-9b70-b183e9fb222d-kube-api-access-flnts\") pod \"4f171e6e-3d04-4e18-9b70-b183e9fb222d\" (UID: \"4f171e6e-3d04-4e18-9b70-b183e9fb222d\") " Oct 08 18:35:21 crc kubenswrapper[4859]: I1008 18:35:21.920348 4859 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-8kr5t\" (UniqueName: \"kubernetes.io/projected/d68d0cff-c3b1-45bb-a00c-23febbb14aee-kube-api-access-8kr5t\") pod \"d68d0cff-c3b1-45bb-a00c-23febbb14aee\" (UID: \"d68d0cff-c3b1-45bb-a00c-23febbb14aee\") " Oct 08 18:35:21 crc kubenswrapper[4859]: I1008 18:35:21.926986 4859 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/4f171e6e-3d04-4e18-9b70-b183e9fb222d-kube-api-access-flnts" (OuterVolumeSpecName: "kube-api-access-flnts") pod "4f171e6e-3d04-4e18-9b70-b183e9fb222d" (UID: "4f171e6e-3d04-4e18-9b70-b183e9fb222d"). InnerVolumeSpecName "kube-api-access-flnts". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 08 18:35:21 crc kubenswrapper[4859]: I1008 18:35:21.957988 4859 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/d68d0cff-c3b1-45bb-a00c-23febbb14aee-kube-api-access-8kr5t" (OuterVolumeSpecName: "kube-api-access-8kr5t") pod "d68d0cff-c3b1-45bb-a00c-23febbb14aee" (UID: "d68d0cff-c3b1-45bb-a00c-23febbb14aee"). InnerVolumeSpecName "kube-api-access-8kr5t". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 08 18:35:22 crc kubenswrapper[4859]: I1008 18:35:22.022851 4859 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-flnts\" (UniqueName: \"kubernetes.io/projected/4f171e6e-3d04-4e18-9b70-b183e9fb222d-kube-api-access-flnts\") on node \"crc\" DevicePath \"\"" Oct 08 18:35:22 crc kubenswrapper[4859]: I1008 18:35:22.022889 4859 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-8kr5t\" (UniqueName: \"kubernetes.io/projected/d68d0cff-c3b1-45bb-a00c-23febbb14aee-kube-api-access-8kr5t\") on node \"crc\" DevicePath \"\"" Oct 08 18:35:22 crc kubenswrapper[4859]: I1008 18:35:22.411290 4859 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-f8cf-account-create-ftsj5" event={"ID":"d68d0cff-c3b1-45bb-a00c-23febbb14aee","Type":"ContainerDied","Data":"d8414f6d45284e8e04b64655146eb710bc1bf9130537d6ea14fe43eb39840715"} Oct 08 18:35:22 crc kubenswrapper[4859]: I1008 18:35:22.411353 4859 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="d8414f6d45284e8e04b64655146eb710bc1bf9130537d6ea14fe43eb39840715" Oct 08 18:35:22 crc kubenswrapper[4859]: I1008 18:35:22.411438 4859 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-f8cf-account-create-ftsj5" Oct 08 18:35:22 crc kubenswrapper[4859]: I1008 18:35:22.415198 4859 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-e838-account-create-6ndb4" event={"ID":"4f171e6e-3d04-4e18-9b70-b183e9fb222d","Type":"ContainerDied","Data":"a1cdaa884c8ca64de007c7653133344e813344b13e2d15c8b2ac9e7dadcb6c9e"} Oct 08 18:35:22 crc kubenswrapper[4859]: I1008 18:35:22.415248 4859 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-e838-account-create-6ndb4" Oct 08 18:35:22 crc kubenswrapper[4859]: I1008 18:35:22.415266 4859 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="a1cdaa884c8ca64de007c7653133344e813344b13e2d15c8b2ac9e7dadcb6c9e" Oct 08 18:35:22 crc kubenswrapper[4859]: I1008 18:35:22.447499 4859 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/openstackclient" podStartSLOduration=3.153922448 podStartE2EDuration="33.447476008s" podCreationTimestamp="2025-10-08 18:34:49 +0000 UTC" firstStartedPulling="2025-10-08 18:34:50.82375341 +0000 UTC m=+1061.070592789" lastFinishedPulling="2025-10-08 18:35:21.11730695 +0000 UTC m=+1091.364146349" observedRunningTime="2025-10-08 18:35:22.443185821 +0000 UTC m=+1092.690025210" watchObservedRunningTime="2025-10-08 18:35:22.447476008 +0000 UTC m=+1092.694315387" Oct 08 18:35:23 crc kubenswrapper[4859]: I1008 18:35:23.425427 4859 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"c3a48f64-6a96-4071-89cc-3f7bc1116694","Type":"ContainerStarted","Data":"b6cd765532dc0caa882289f09ed55f099a165cdb9739f37148d51356f57f5994"} Oct 08 18:35:23 crc kubenswrapper[4859]: I1008 18:35:23.425869 4859 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ceilometer-0" Oct 08 18:35:23 crc kubenswrapper[4859]: I1008 18:35:23.425637 4859 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="c3a48f64-6a96-4071-89cc-3f7bc1116694" containerName="proxy-httpd" containerID="cri-o://b6cd765532dc0caa882289f09ed55f099a165cdb9739f37148d51356f57f5994" gracePeriod=30 Oct 08 18:35:23 crc kubenswrapper[4859]: I1008 18:35:23.425588 4859 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="c3a48f64-6a96-4071-89cc-3f7bc1116694" containerName="ceilometer-central-agent" containerID="cri-o://fff6d6678da7c49d4e113d7c3d7de23ebfcc4fa849576243ee55d22c2a584db2" gracePeriod=30 Oct 08 18:35:23 crc kubenswrapper[4859]: I1008 18:35:23.425705 4859 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="c3a48f64-6a96-4071-89cc-3f7bc1116694" containerName="ceilometer-notification-agent" containerID="cri-o://b2db0d54d3ba08a9d3d383f5c23225939102d4c15f94e0ceabf208c65ab6a1e4" gracePeriod=30 Oct 08 18:35:23 crc kubenswrapper[4859]: I1008 18:35:23.425642 4859 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="c3a48f64-6a96-4071-89cc-3f7bc1116694" containerName="sg-core" containerID="cri-o://4509cf6e879a4cf5f404205c0e2832d503607bfe89ccfbe5070b7ce8ba4acee6" gracePeriod=30 Oct 08 18:35:23 crc kubenswrapper[4859]: I1008 18:35:23.457601 4859 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ceilometer-0" podStartSLOduration=2.198214357 podStartE2EDuration="7.457578025s" podCreationTimestamp="2025-10-08 18:35:16 +0000 UTC" firstStartedPulling="2025-10-08 18:35:17.369653846 +0000 UTC m=+1087.616493225" lastFinishedPulling="2025-10-08 18:35:22.629017514 +0000 UTC m=+1092.875856893" observedRunningTime="2025-10-08 18:35:23.451419793 +0000 UTC m=+1093.698259192" watchObservedRunningTime="2025-10-08 18:35:23.457578025 +0000 UTC m=+1093.704417394" Oct 08 18:35:24 crc kubenswrapper[4859]: I1008 18:35:24.000039 4859 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/cinder-api-0" Oct 08 18:35:24 crc kubenswrapper[4859]: I1008 18:35:24.008522 4859 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell0-conductor-db-sync-9hb79"] Oct 08 18:35:24 crc kubenswrapper[4859]: E1008 18:35:24.008937 4859 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="4f171e6e-3d04-4e18-9b70-b183e9fb222d" containerName="mariadb-account-create" Oct 08 18:35:24 crc kubenswrapper[4859]: I1008 18:35:24.008955 4859 state_mem.go:107] "Deleted CPUSet assignment" podUID="4f171e6e-3d04-4e18-9b70-b183e9fb222d" containerName="mariadb-account-create" Oct 08 18:35:24 crc kubenswrapper[4859]: E1008 18:35:24.008989 4859 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d68d0cff-c3b1-45bb-a00c-23febbb14aee" containerName="mariadb-account-create" Oct 08 18:35:24 crc kubenswrapper[4859]: I1008 18:35:24.008995 4859 state_mem.go:107] "Deleted CPUSet assignment" podUID="d68d0cff-c3b1-45bb-a00c-23febbb14aee" containerName="mariadb-account-create" Oct 08 18:35:24 crc kubenswrapper[4859]: I1008 18:35:24.009141 4859 memory_manager.go:354] "RemoveStaleState removing state" podUID="d68d0cff-c3b1-45bb-a00c-23febbb14aee" containerName="mariadb-account-create" Oct 08 18:35:24 crc kubenswrapper[4859]: I1008 18:35:24.009170 4859 memory_manager.go:354] "RemoveStaleState removing state" podUID="4f171e6e-3d04-4e18-9b70-b183e9fb222d" containerName="mariadb-account-create" Oct 08 18:35:24 crc kubenswrapper[4859]: I1008 18:35:24.009741 4859 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-conductor-db-sync-9hb79" Oct 08 18:35:24 crc kubenswrapper[4859]: I1008 18:35:24.012203 4859 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell0-conductor-config-data" Oct 08 18:35:24 crc kubenswrapper[4859]: I1008 18:35:24.012619 4859 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-nova-dockercfg-hsn5b" Oct 08 18:35:24 crc kubenswrapper[4859]: I1008 18:35:24.012807 4859 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell0-conductor-scripts" Oct 08 18:35:24 crc kubenswrapper[4859]: I1008 18:35:24.022384 4859 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-conductor-db-sync-9hb79"] Oct 08 18:35:24 crc kubenswrapper[4859]: I1008 18:35:24.057916 4859 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-dhdhm\" (UniqueName: \"kubernetes.io/projected/b03ff9b0-98b5-4d0b-990a-253ada95b6d8-kube-api-access-dhdhm\") pod \"nova-cell0-conductor-db-sync-9hb79\" (UID: \"b03ff9b0-98b5-4d0b-990a-253ada95b6d8\") " pod="openstack/nova-cell0-conductor-db-sync-9hb79" Oct 08 18:35:24 crc kubenswrapper[4859]: I1008 18:35:24.057978 4859 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b03ff9b0-98b5-4d0b-990a-253ada95b6d8-combined-ca-bundle\") pod \"nova-cell0-conductor-db-sync-9hb79\" (UID: \"b03ff9b0-98b5-4d0b-990a-253ada95b6d8\") " pod="openstack/nova-cell0-conductor-db-sync-9hb79" Oct 08 18:35:24 crc kubenswrapper[4859]: I1008 18:35:24.058068 4859 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/b03ff9b0-98b5-4d0b-990a-253ada95b6d8-config-data\") pod \"nova-cell0-conductor-db-sync-9hb79\" (UID: \"b03ff9b0-98b5-4d0b-990a-253ada95b6d8\") " pod="openstack/nova-cell0-conductor-db-sync-9hb79" Oct 08 18:35:24 crc kubenswrapper[4859]: I1008 18:35:24.058089 4859 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/b03ff9b0-98b5-4d0b-990a-253ada95b6d8-scripts\") pod \"nova-cell0-conductor-db-sync-9hb79\" (UID: \"b03ff9b0-98b5-4d0b-990a-253ada95b6d8\") " pod="openstack/nova-cell0-conductor-db-sync-9hb79" Oct 08 18:35:24 crc kubenswrapper[4859]: I1008 18:35:24.160867 4859 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-dhdhm\" (UniqueName: \"kubernetes.io/projected/b03ff9b0-98b5-4d0b-990a-253ada95b6d8-kube-api-access-dhdhm\") pod \"nova-cell0-conductor-db-sync-9hb79\" (UID: \"b03ff9b0-98b5-4d0b-990a-253ada95b6d8\") " pod="openstack/nova-cell0-conductor-db-sync-9hb79" Oct 08 18:35:24 crc kubenswrapper[4859]: I1008 18:35:24.161274 4859 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b03ff9b0-98b5-4d0b-990a-253ada95b6d8-combined-ca-bundle\") pod \"nova-cell0-conductor-db-sync-9hb79\" (UID: \"b03ff9b0-98b5-4d0b-990a-253ada95b6d8\") " pod="openstack/nova-cell0-conductor-db-sync-9hb79" Oct 08 18:35:24 crc kubenswrapper[4859]: I1008 18:35:24.161364 4859 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/b03ff9b0-98b5-4d0b-990a-253ada95b6d8-config-data\") pod \"nova-cell0-conductor-db-sync-9hb79\" (UID: \"b03ff9b0-98b5-4d0b-990a-253ada95b6d8\") " pod="openstack/nova-cell0-conductor-db-sync-9hb79" Oct 08 18:35:24 crc kubenswrapper[4859]: I1008 18:35:24.161404 4859 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/b03ff9b0-98b5-4d0b-990a-253ada95b6d8-scripts\") pod \"nova-cell0-conductor-db-sync-9hb79\" (UID: \"b03ff9b0-98b5-4d0b-990a-253ada95b6d8\") " pod="openstack/nova-cell0-conductor-db-sync-9hb79" Oct 08 18:35:24 crc kubenswrapper[4859]: I1008 18:35:24.180195 4859 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/b03ff9b0-98b5-4d0b-990a-253ada95b6d8-config-data\") pod \"nova-cell0-conductor-db-sync-9hb79\" (UID: \"b03ff9b0-98b5-4d0b-990a-253ada95b6d8\") " pod="openstack/nova-cell0-conductor-db-sync-9hb79" Oct 08 18:35:24 crc kubenswrapper[4859]: I1008 18:35:24.180503 4859 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b03ff9b0-98b5-4d0b-990a-253ada95b6d8-combined-ca-bundle\") pod \"nova-cell0-conductor-db-sync-9hb79\" (UID: \"b03ff9b0-98b5-4d0b-990a-253ada95b6d8\") " pod="openstack/nova-cell0-conductor-db-sync-9hb79" Oct 08 18:35:24 crc kubenswrapper[4859]: I1008 18:35:24.183308 4859 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/b03ff9b0-98b5-4d0b-990a-253ada95b6d8-scripts\") pod \"nova-cell0-conductor-db-sync-9hb79\" (UID: \"b03ff9b0-98b5-4d0b-990a-253ada95b6d8\") " pod="openstack/nova-cell0-conductor-db-sync-9hb79" Oct 08 18:35:24 crc kubenswrapper[4859]: I1008 18:35:24.183462 4859 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-dhdhm\" (UniqueName: \"kubernetes.io/projected/b03ff9b0-98b5-4d0b-990a-253ada95b6d8-kube-api-access-dhdhm\") pod \"nova-cell0-conductor-db-sync-9hb79\" (UID: \"b03ff9b0-98b5-4d0b-990a-253ada95b6d8\") " pod="openstack/nova-cell0-conductor-db-sync-9hb79" Oct 08 18:35:24 crc kubenswrapper[4859]: I1008 18:35:24.288020 4859 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Oct 08 18:35:24 crc kubenswrapper[4859]: I1008 18:35:24.339330 4859 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-conductor-db-sync-9hb79" Oct 08 18:35:24 crc kubenswrapper[4859]: I1008 18:35:24.371128 4859 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/c3a48f64-6a96-4071-89cc-3f7bc1116694-log-httpd\") pod \"c3a48f64-6a96-4071-89cc-3f7bc1116694\" (UID: \"c3a48f64-6a96-4071-89cc-3f7bc1116694\") " Oct 08 18:35:24 crc kubenswrapper[4859]: I1008 18:35:24.371205 4859 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/c3a48f64-6a96-4071-89cc-3f7bc1116694-run-httpd\") pod \"c3a48f64-6a96-4071-89cc-3f7bc1116694\" (UID: \"c3a48f64-6a96-4071-89cc-3f7bc1116694\") " Oct 08 18:35:24 crc kubenswrapper[4859]: I1008 18:35:24.371237 4859 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-v4w8m\" (UniqueName: \"kubernetes.io/projected/c3a48f64-6a96-4071-89cc-3f7bc1116694-kube-api-access-v4w8m\") pod \"c3a48f64-6a96-4071-89cc-3f7bc1116694\" (UID: \"c3a48f64-6a96-4071-89cc-3f7bc1116694\") " Oct 08 18:35:24 crc kubenswrapper[4859]: I1008 18:35:24.371292 4859 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/c3a48f64-6a96-4071-89cc-3f7bc1116694-sg-core-conf-yaml\") pod \"c3a48f64-6a96-4071-89cc-3f7bc1116694\" (UID: \"c3a48f64-6a96-4071-89cc-3f7bc1116694\") " Oct 08 18:35:24 crc kubenswrapper[4859]: I1008 18:35:24.371343 4859 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/c3a48f64-6a96-4071-89cc-3f7bc1116694-scripts\") pod \"c3a48f64-6a96-4071-89cc-3f7bc1116694\" (UID: \"c3a48f64-6a96-4071-89cc-3f7bc1116694\") " Oct 08 18:35:24 crc kubenswrapper[4859]: I1008 18:35:24.371386 4859 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c3a48f64-6a96-4071-89cc-3f7bc1116694-combined-ca-bundle\") pod \"c3a48f64-6a96-4071-89cc-3f7bc1116694\" (UID: \"c3a48f64-6a96-4071-89cc-3f7bc1116694\") " Oct 08 18:35:24 crc kubenswrapper[4859]: I1008 18:35:24.371401 4859 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/c3a48f64-6a96-4071-89cc-3f7bc1116694-ceilometer-tls-certs\") pod \"c3a48f64-6a96-4071-89cc-3f7bc1116694\" (UID: \"c3a48f64-6a96-4071-89cc-3f7bc1116694\") " Oct 08 18:35:24 crc kubenswrapper[4859]: I1008 18:35:24.371422 4859 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/c3a48f64-6a96-4071-89cc-3f7bc1116694-config-data\") pod \"c3a48f64-6a96-4071-89cc-3f7bc1116694\" (UID: \"c3a48f64-6a96-4071-89cc-3f7bc1116694\") " Oct 08 18:35:24 crc kubenswrapper[4859]: I1008 18:35:24.372884 4859 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/c3a48f64-6a96-4071-89cc-3f7bc1116694-run-httpd" (OuterVolumeSpecName: "run-httpd") pod "c3a48f64-6a96-4071-89cc-3f7bc1116694" (UID: "c3a48f64-6a96-4071-89cc-3f7bc1116694"). InnerVolumeSpecName "run-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 08 18:35:24 crc kubenswrapper[4859]: I1008 18:35:24.373288 4859 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/c3a48f64-6a96-4071-89cc-3f7bc1116694-log-httpd" (OuterVolumeSpecName: "log-httpd") pod "c3a48f64-6a96-4071-89cc-3f7bc1116694" (UID: "c3a48f64-6a96-4071-89cc-3f7bc1116694"). InnerVolumeSpecName "log-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 08 18:35:24 crc kubenswrapper[4859]: I1008 18:35:24.386936 4859 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c3a48f64-6a96-4071-89cc-3f7bc1116694-scripts" (OuterVolumeSpecName: "scripts") pod "c3a48f64-6a96-4071-89cc-3f7bc1116694" (UID: "c3a48f64-6a96-4071-89cc-3f7bc1116694"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 08 18:35:24 crc kubenswrapper[4859]: I1008 18:35:24.395348 4859 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/c3a48f64-6a96-4071-89cc-3f7bc1116694-kube-api-access-v4w8m" (OuterVolumeSpecName: "kube-api-access-v4w8m") pod "c3a48f64-6a96-4071-89cc-3f7bc1116694" (UID: "c3a48f64-6a96-4071-89cc-3f7bc1116694"). InnerVolumeSpecName "kube-api-access-v4w8m". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 08 18:35:24 crc kubenswrapper[4859]: I1008 18:35:24.448109 4859 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c3a48f64-6a96-4071-89cc-3f7bc1116694-sg-core-conf-yaml" (OuterVolumeSpecName: "sg-core-conf-yaml") pod "c3a48f64-6a96-4071-89cc-3f7bc1116694" (UID: "c3a48f64-6a96-4071-89cc-3f7bc1116694"). InnerVolumeSpecName "sg-core-conf-yaml". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 08 18:35:24 crc kubenswrapper[4859]: I1008 18:35:24.450604 4859 generic.go:334] "Generic (PLEG): container finished" podID="c3a48f64-6a96-4071-89cc-3f7bc1116694" containerID="b6cd765532dc0caa882289f09ed55f099a165cdb9739f37148d51356f57f5994" exitCode=0 Oct 08 18:35:24 crc kubenswrapper[4859]: I1008 18:35:24.450731 4859 generic.go:334] "Generic (PLEG): container finished" podID="c3a48f64-6a96-4071-89cc-3f7bc1116694" containerID="4509cf6e879a4cf5f404205c0e2832d503607bfe89ccfbe5070b7ce8ba4acee6" exitCode=2 Oct 08 18:35:24 crc kubenswrapper[4859]: I1008 18:35:24.450789 4859 generic.go:334] "Generic (PLEG): container finished" podID="c3a48f64-6a96-4071-89cc-3f7bc1116694" containerID="b2db0d54d3ba08a9d3d383f5c23225939102d4c15f94e0ceabf208c65ab6a1e4" exitCode=0 Oct 08 18:35:24 crc kubenswrapper[4859]: I1008 18:35:24.450914 4859 generic.go:334] "Generic (PLEG): container finished" podID="c3a48f64-6a96-4071-89cc-3f7bc1116694" containerID="fff6d6678da7c49d4e113d7c3d7de23ebfcc4fa849576243ee55d22c2a584db2" exitCode=0 Oct 08 18:35:24 crc kubenswrapper[4859]: I1008 18:35:24.450985 4859 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"c3a48f64-6a96-4071-89cc-3f7bc1116694","Type":"ContainerDied","Data":"b6cd765532dc0caa882289f09ed55f099a165cdb9739f37148d51356f57f5994"} Oct 08 18:35:24 crc kubenswrapper[4859]: I1008 18:35:24.451068 4859 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"c3a48f64-6a96-4071-89cc-3f7bc1116694","Type":"ContainerDied","Data":"4509cf6e879a4cf5f404205c0e2832d503607bfe89ccfbe5070b7ce8ba4acee6"} Oct 08 18:35:24 crc kubenswrapper[4859]: I1008 18:35:24.451135 4859 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"c3a48f64-6a96-4071-89cc-3f7bc1116694","Type":"ContainerDied","Data":"b2db0d54d3ba08a9d3d383f5c23225939102d4c15f94e0ceabf208c65ab6a1e4"} Oct 08 18:35:24 crc kubenswrapper[4859]: I1008 18:35:24.451191 4859 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"c3a48f64-6a96-4071-89cc-3f7bc1116694","Type":"ContainerDied","Data":"fff6d6678da7c49d4e113d7c3d7de23ebfcc4fa849576243ee55d22c2a584db2"} Oct 08 18:35:24 crc kubenswrapper[4859]: I1008 18:35:24.451254 4859 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"c3a48f64-6a96-4071-89cc-3f7bc1116694","Type":"ContainerDied","Data":"057ee2a69f865fcd963174397e456288f60ac2c44eb9d6e283a460bebd15b751"} Oct 08 18:35:24 crc kubenswrapper[4859]: I1008 18:35:24.451316 4859 scope.go:117] "RemoveContainer" containerID="b6cd765532dc0caa882289f09ed55f099a165cdb9739f37148d51356f57f5994" Oct 08 18:35:24 crc kubenswrapper[4859]: I1008 18:35:24.451534 4859 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Oct 08 18:35:24 crc kubenswrapper[4859]: I1008 18:35:24.474097 4859 reconciler_common.go:293] "Volume detached for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/c3a48f64-6a96-4071-89cc-3f7bc1116694-sg-core-conf-yaml\") on node \"crc\" DevicePath \"\"" Oct 08 18:35:24 crc kubenswrapper[4859]: I1008 18:35:24.474451 4859 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/c3a48f64-6a96-4071-89cc-3f7bc1116694-scripts\") on node \"crc\" DevicePath \"\"" Oct 08 18:35:24 crc kubenswrapper[4859]: I1008 18:35:24.474465 4859 reconciler_common.go:293] "Volume detached for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/c3a48f64-6a96-4071-89cc-3f7bc1116694-log-httpd\") on node \"crc\" DevicePath \"\"" Oct 08 18:35:24 crc kubenswrapper[4859]: I1008 18:35:24.474476 4859 reconciler_common.go:293] "Volume detached for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/c3a48f64-6a96-4071-89cc-3f7bc1116694-run-httpd\") on node \"crc\" DevicePath \"\"" Oct 08 18:35:24 crc kubenswrapper[4859]: I1008 18:35:24.474488 4859 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-v4w8m\" (UniqueName: \"kubernetes.io/projected/c3a48f64-6a96-4071-89cc-3f7bc1116694-kube-api-access-v4w8m\") on node \"crc\" DevicePath \"\"" Oct 08 18:35:24 crc kubenswrapper[4859]: I1008 18:35:24.507982 4859 scope.go:117] "RemoveContainer" containerID="4509cf6e879a4cf5f404205c0e2832d503607bfe89ccfbe5070b7ce8ba4acee6" Oct 08 18:35:24 crc kubenswrapper[4859]: I1008 18:35:24.520469 4859 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c3a48f64-6a96-4071-89cc-3f7bc1116694-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "c3a48f64-6a96-4071-89cc-3f7bc1116694" (UID: "c3a48f64-6a96-4071-89cc-3f7bc1116694"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 08 18:35:24 crc kubenswrapper[4859]: I1008 18:35:24.525944 4859 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c3a48f64-6a96-4071-89cc-3f7bc1116694-ceilometer-tls-certs" (OuterVolumeSpecName: "ceilometer-tls-certs") pod "c3a48f64-6a96-4071-89cc-3f7bc1116694" (UID: "c3a48f64-6a96-4071-89cc-3f7bc1116694"). InnerVolumeSpecName "ceilometer-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 08 18:35:24 crc kubenswrapper[4859]: I1008 18:35:24.550672 4859 scope.go:117] "RemoveContainer" containerID="b2db0d54d3ba08a9d3d383f5c23225939102d4c15f94e0ceabf208c65ab6a1e4" Oct 08 18:35:24 crc kubenswrapper[4859]: I1008 18:35:24.579221 4859 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c3a48f64-6a96-4071-89cc-3f7bc1116694-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 08 18:35:24 crc kubenswrapper[4859]: I1008 18:35:24.579249 4859 reconciler_common.go:293] "Volume detached for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/c3a48f64-6a96-4071-89cc-3f7bc1116694-ceilometer-tls-certs\") on node \"crc\" DevicePath \"\"" Oct 08 18:35:24 crc kubenswrapper[4859]: I1008 18:35:24.600015 4859 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c3a48f64-6a96-4071-89cc-3f7bc1116694-config-data" (OuterVolumeSpecName: "config-data") pod "c3a48f64-6a96-4071-89cc-3f7bc1116694" (UID: "c3a48f64-6a96-4071-89cc-3f7bc1116694"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 08 18:35:24 crc kubenswrapper[4859]: I1008 18:35:24.672834 4859 scope.go:117] "RemoveContainer" containerID="fff6d6678da7c49d4e113d7c3d7de23ebfcc4fa849576243ee55d22c2a584db2" Oct 08 18:35:24 crc kubenswrapper[4859]: I1008 18:35:24.680594 4859 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/c3a48f64-6a96-4071-89cc-3f7bc1116694-config-data\") on node \"crc\" DevicePath \"\"" Oct 08 18:35:24 crc kubenswrapper[4859]: I1008 18:35:24.695348 4859 scope.go:117] "RemoveContainer" containerID="b6cd765532dc0caa882289f09ed55f099a165cdb9739f37148d51356f57f5994" Oct 08 18:35:24 crc kubenswrapper[4859]: E1008 18:35:24.695811 4859 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"b6cd765532dc0caa882289f09ed55f099a165cdb9739f37148d51356f57f5994\": container with ID starting with b6cd765532dc0caa882289f09ed55f099a165cdb9739f37148d51356f57f5994 not found: ID does not exist" containerID="b6cd765532dc0caa882289f09ed55f099a165cdb9739f37148d51356f57f5994" Oct 08 18:35:24 crc kubenswrapper[4859]: I1008 18:35:24.695844 4859 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"b6cd765532dc0caa882289f09ed55f099a165cdb9739f37148d51356f57f5994"} err="failed to get container status \"b6cd765532dc0caa882289f09ed55f099a165cdb9739f37148d51356f57f5994\": rpc error: code = NotFound desc = could not find container \"b6cd765532dc0caa882289f09ed55f099a165cdb9739f37148d51356f57f5994\": container with ID starting with b6cd765532dc0caa882289f09ed55f099a165cdb9739f37148d51356f57f5994 not found: ID does not exist" Oct 08 18:35:24 crc kubenswrapper[4859]: I1008 18:35:24.695867 4859 scope.go:117] "RemoveContainer" containerID="4509cf6e879a4cf5f404205c0e2832d503607bfe89ccfbe5070b7ce8ba4acee6" Oct 08 18:35:24 crc kubenswrapper[4859]: E1008 18:35:24.696137 4859 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"4509cf6e879a4cf5f404205c0e2832d503607bfe89ccfbe5070b7ce8ba4acee6\": container with ID starting with 4509cf6e879a4cf5f404205c0e2832d503607bfe89ccfbe5070b7ce8ba4acee6 not found: ID does not exist" containerID="4509cf6e879a4cf5f404205c0e2832d503607bfe89ccfbe5070b7ce8ba4acee6" Oct 08 18:35:24 crc kubenswrapper[4859]: I1008 18:35:24.696182 4859 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"4509cf6e879a4cf5f404205c0e2832d503607bfe89ccfbe5070b7ce8ba4acee6"} err="failed to get container status \"4509cf6e879a4cf5f404205c0e2832d503607bfe89ccfbe5070b7ce8ba4acee6\": rpc error: code = NotFound desc = could not find container \"4509cf6e879a4cf5f404205c0e2832d503607bfe89ccfbe5070b7ce8ba4acee6\": container with ID starting with 4509cf6e879a4cf5f404205c0e2832d503607bfe89ccfbe5070b7ce8ba4acee6 not found: ID does not exist" Oct 08 18:35:24 crc kubenswrapper[4859]: I1008 18:35:24.696212 4859 scope.go:117] "RemoveContainer" containerID="b2db0d54d3ba08a9d3d383f5c23225939102d4c15f94e0ceabf208c65ab6a1e4" Oct 08 18:35:24 crc kubenswrapper[4859]: E1008 18:35:24.696621 4859 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"b2db0d54d3ba08a9d3d383f5c23225939102d4c15f94e0ceabf208c65ab6a1e4\": container with ID starting with b2db0d54d3ba08a9d3d383f5c23225939102d4c15f94e0ceabf208c65ab6a1e4 not found: ID does not exist" containerID="b2db0d54d3ba08a9d3d383f5c23225939102d4c15f94e0ceabf208c65ab6a1e4" Oct 08 18:35:24 crc kubenswrapper[4859]: I1008 18:35:24.696647 4859 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"b2db0d54d3ba08a9d3d383f5c23225939102d4c15f94e0ceabf208c65ab6a1e4"} err="failed to get container status \"b2db0d54d3ba08a9d3d383f5c23225939102d4c15f94e0ceabf208c65ab6a1e4\": rpc error: code = NotFound desc = could not find container \"b2db0d54d3ba08a9d3d383f5c23225939102d4c15f94e0ceabf208c65ab6a1e4\": container with ID starting with b2db0d54d3ba08a9d3d383f5c23225939102d4c15f94e0ceabf208c65ab6a1e4 not found: ID does not exist" Oct 08 18:35:24 crc kubenswrapper[4859]: I1008 18:35:24.696661 4859 scope.go:117] "RemoveContainer" containerID="fff6d6678da7c49d4e113d7c3d7de23ebfcc4fa849576243ee55d22c2a584db2" Oct 08 18:35:24 crc kubenswrapper[4859]: E1008 18:35:24.697788 4859 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"fff6d6678da7c49d4e113d7c3d7de23ebfcc4fa849576243ee55d22c2a584db2\": container with ID starting with fff6d6678da7c49d4e113d7c3d7de23ebfcc4fa849576243ee55d22c2a584db2 not found: ID does not exist" containerID="fff6d6678da7c49d4e113d7c3d7de23ebfcc4fa849576243ee55d22c2a584db2" Oct 08 18:35:24 crc kubenswrapper[4859]: I1008 18:35:24.697813 4859 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"fff6d6678da7c49d4e113d7c3d7de23ebfcc4fa849576243ee55d22c2a584db2"} err="failed to get container status \"fff6d6678da7c49d4e113d7c3d7de23ebfcc4fa849576243ee55d22c2a584db2\": rpc error: code = NotFound desc = could not find container \"fff6d6678da7c49d4e113d7c3d7de23ebfcc4fa849576243ee55d22c2a584db2\": container with ID starting with fff6d6678da7c49d4e113d7c3d7de23ebfcc4fa849576243ee55d22c2a584db2 not found: ID does not exist" Oct 08 18:35:24 crc kubenswrapper[4859]: I1008 18:35:24.697828 4859 scope.go:117] "RemoveContainer" containerID="b6cd765532dc0caa882289f09ed55f099a165cdb9739f37148d51356f57f5994" Oct 08 18:35:24 crc kubenswrapper[4859]: I1008 18:35:24.698205 4859 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"b6cd765532dc0caa882289f09ed55f099a165cdb9739f37148d51356f57f5994"} err="failed to get container status \"b6cd765532dc0caa882289f09ed55f099a165cdb9739f37148d51356f57f5994\": rpc error: code = NotFound desc = could not find container \"b6cd765532dc0caa882289f09ed55f099a165cdb9739f37148d51356f57f5994\": container with ID starting with b6cd765532dc0caa882289f09ed55f099a165cdb9739f37148d51356f57f5994 not found: ID does not exist" Oct 08 18:35:24 crc kubenswrapper[4859]: I1008 18:35:24.698255 4859 scope.go:117] "RemoveContainer" containerID="4509cf6e879a4cf5f404205c0e2832d503607bfe89ccfbe5070b7ce8ba4acee6" Oct 08 18:35:24 crc kubenswrapper[4859]: I1008 18:35:24.698540 4859 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"4509cf6e879a4cf5f404205c0e2832d503607bfe89ccfbe5070b7ce8ba4acee6"} err="failed to get container status \"4509cf6e879a4cf5f404205c0e2832d503607bfe89ccfbe5070b7ce8ba4acee6\": rpc error: code = NotFound desc = could not find container \"4509cf6e879a4cf5f404205c0e2832d503607bfe89ccfbe5070b7ce8ba4acee6\": container with ID starting with 4509cf6e879a4cf5f404205c0e2832d503607bfe89ccfbe5070b7ce8ba4acee6 not found: ID does not exist" Oct 08 18:35:24 crc kubenswrapper[4859]: I1008 18:35:24.698560 4859 scope.go:117] "RemoveContainer" containerID="b2db0d54d3ba08a9d3d383f5c23225939102d4c15f94e0ceabf208c65ab6a1e4" Oct 08 18:35:24 crc kubenswrapper[4859]: I1008 18:35:24.698891 4859 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"b2db0d54d3ba08a9d3d383f5c23225939102d4c15f94e0ceabf208c65ab6a1e4"} err="failed to get container status \"b2db0d54d3ba08a9d3d383f5c23225939102d4c15f94e0ceabf208c65ab6a1e4\": rpc error: code = NotFound desc = could not find container \"b2db0d54d3ba08a9d3d383f5c23225939102d4c15f94e0ceabf208c65ab6a1e4\": container with ID starting with b2db0d54d3ba08a9d3d383f5c23225939102d4c15f94e0ceabf208c65ab6a1e4 not found: ID does not exist" Oct 08 18:35:24 crc kubenswrapper[4859]: I1008 18:35:24.698909 4859 scope.go:117] "RemoveContainer" containerID="fff6d6678da7c49d4e113d7c3d7de23ebfcc4fa849576243ee55d22c2a584db2" Oct 08 18:35:24 crc kubenswrapper[4859]: I1008 18:35:24.699259 4859 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"fff6d6678da7c49d4e113d7c3d7de23ebfcc4fa849576243ee55d22c2a584db2"} err="failed to get container status \"fff6d6678da7c49d4e113d7c3d7de23ebfcc4fa849576243ee55d22c2a584db2\": rpc error: code = NotFound desc = could not find container \"fff6d6678da7c49d4e113d7c3d7de23ebfcc4fa849576243ee55d22c2a584db2\": container with ID starting with fff6d6678da7c49d4e113d7c3d7de23ebfcc4fa849576243ee55d22c2a584db2 not found: ID does not exist" Oct 08 18:35:24 crc kubenswrapper[4859]: I1008 18:35:24.699276 4859 scope.go:117] "RemoveContainer" containerID="b6cd765532dc0caa882289f09ed55f099a165cdb9739f37148d51356f57f5994" Oct 08 18:35:24 crc kubenswrapper[4859]: I1008 18:35:24.699520 4859 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"b6cd765532dc0caa882289f09ed55f099a165cdb9739f37148d51356f57f5994"} err="failed to get container status \"b6cd765532dc0caa882289f09ed55f099a165cdb9739f37148d51356f57f5994\": rpc error: code = NotFound desc = could not find container \"b6cd765532dc0caa882289f09ed55f099a165cdb9739f37148d51356f57f5994\": container with ID starting with b6cd765532dc0caa882289f09ed55f099a165cdb9739f37148d51356f57f5994 not found: ID does not exist" Oct 08 18:35:24 crc kubenswrapper[4859]: I1008 18:35:24.699539 4859 scope.go:117] "RemoveContainer" containerID="4509cf6e879a4cf5f404205c0e2832d503607bfe89ccfbe5070b7ce8ba4acee6" Oct 08 18:35:24 crc kubenswrapper[4859]: I1008 18:35:24.699814 4859 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"4509cf6e879a4cf5f404205c0e2832d503607bfe89ccfbe5070b7ce8ba4acee6"} err="failed to get container status \"4509cf6e879a4cf5f404205c0e2832d503607bfe89ccfbe5070b7ce8ba4acee6\": rpc error: code = NotFound desc = could not find container \"4509cf6e879a4cf5f404205c0e2832d503607bfe89ccfbe5070b7ce8ba4acee6\": container with ID starting with 4509cf6e879a4cf5f404205c0e2832d503607bfe89ccfbe5070b7ce8ba4acee6 not found: ID does not exist" Oct 08 18:35:24 crc kubenswrapper[4859]: I1008 18:35:24.699852 4859 scope.go:117] "RemoveContainer" containerID="b2db0d54d3ba08a9d3d383f5c23225939102d4c15f94e0ceabf208c65ab6a1e4" Oct 08 18:35:24 crc kubenswrapper[4859]: I1008 18:35:24.703206 4859 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"b2db0d54d3ba08a9d3d383f5c23225939102d4c15f94e0ceabf208c65ab6a1e4"} err="failed to get container status \"b2db0d54d3ba08a9d3d383f5c23225939102d4c15f94e0ceabf208c65ab6a1e4\": rpc error: code = NotFound desc = could not find container \"b2db0d54d3ba08a9d3d383f5c23225939102d4c15f94e0ceabf208c65ab6a1e4\": container with ID starting with b2db0d54d3ba08a9d3d383f5c23225939102d4c15f94e0ceabf208c65ab6a1e4 not found: ID does not exist" Oct 08 18:35:24 crc kubenswrapper[4859]: I1008 18:35:24.703250 4859 scope.go:117] "RemoveContainer" containerID="fff6d6678da7c49d4e113d7c3d7de23ebfcc4fa849576243ee55d22c2a584db2" Oct 08 18:35:24 crc kubenswrapper[4859]: I1008 18:35:24.703586 4859 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"fff6d6678da7c49d4e113d7c3d7de23ebfcc4fa849576243ee55d22c2a584db2"} err="failed to get container status \"fff6d6678da7c49d4e113d7c3d7de23ebfcc4fa849576243ee55d22c2a584db2\": rpc error: code = NotFound desc = could not find container \"fff6d6678da7c49d4e113d7c3d7de23ebfcc4fa849576243ee55d22c2a584db2\": container with ID starting with fff6d6678da7c49d4e113d7c3d7de23ebfcc4fa849576243ee55d22c2a584db2 not found: ID does not exist" Oct 08 18:35:24 crc kubenswrapper[4859]: I1008 18:35:24.703627 4859 scope.go:117] "RemoveContainer" containerID="b6cd765532dc0caa882289f09ed55f099a165cdb9739f37148d51356f57f5994" Oct 08 18:35:24 crc kubenswrapper[4859]: I1008 18:35:24.703931 4859 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"b6cd765532dc0caa882289f09ed55f099a165cdb9739f37148d51356f57f5994"} err="failed to get container status \"b6cd765532dc0caa882289f09ed55f099a165cdb9739f37148d51356f57f5994\": rpc error: code = NotFound desc = could not find container \"b6cd765532dc0caa882289f09ed55f099a165cdb9739f37148d51356f57f5994\": container with ID starting with b6cd765532dc0caa882289f09ed55f099a165cdb9739f37148d51356f57f5994 not found: ID does not exist" Oct 08 18:35:24 crc kubenswrapper[4859]: I1008 18:35:24.703965 4859 scope.go:117] "RemoveContainer" containerID="4509cf6e879a4cf5f404205c0e2832d503607bfe89ccfbe5070b7ce8ba4acee6" Oct 08 18:35:24 crc kubenswrapper[4859]: I1008 18:35:24.704255 4859 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"4509cf6e879a4cf5f404205c0e2832d503607bfe89ccfbe5070b7ce8ba4acee6"} err="failed to get container status \"4509cf6e879a4cf5f404205c0e2832d503607bfe89ccfbe5070b7ce8ba4acee6\": rpc error: code = NotFound desc = could not find container \"4509cf6e879a4cf5f404205c0e2832d503607bfe89ccfbe5070b7ce8ba4acee6\": container with ID starting with 4509cf6e879a4cf5f404205c0e2832d503607bfe89ccfbe5070b7ce8ba4acee6 not found: ID does not exist" Oct 08 18:35:24 crc kubenswrapper[4859]: I1008 18:35:24.704279 4859 scope.go:117] "RemoveContainer" containerID="b2db0d54d3ba08a9d3d383f5c23225939102d4c15f94e0ceabf208c65ab6a1e4" Oct 08 18:35:24 crc kubenswrapper[4859]: I1008 18:35:24.704606 4859 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"b2db0d54d3ba08a9d3d383f5c23225939102d4c15f94e0ceabf208c65ab6a1e4"} err="failed to get container status \"b2db0d54d3ba08a9d3d383f5c23225939102d4c15f94e0ceabf208c65ab6a1e4\": rpc error: code = NotFound desc = could not find container \"b2db0d54d3ba08a9d3d383f5c23225939102d4c15f94e0ceabf208c65ab6a1e4\": container with ID starting with b2db0d54d3ba08a9d3d383f5c23225939102d4c15f94e0ceabf208c65ab6a1e4 not found: ID does not exist" Oct 08 18:35:24 crc kubenswrapper[4859]: I1008 18:35:24.704626 4859 scope.go:117] "RemoveContainer" containerID="fff6d6678da7c49d4e113d7c3d7de23ebfcc4fa849576243ee55d22c2a584db2" Oct 08 18:35:24 crc kubenswrapper[4859]: I1008 18:35:24.704899 4859 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"fff6d6678da7c49d4e113d7c3d7de23ebfcc4fa849576243ee55d22c2a584db2"} err="failed to get container status \"fff6d6678da7c49d4e113d7c3d7de23ebfcc4fa849576243ee55d22c2a584db2\": rpc error: code = NotFound desc = could not find container \"fff6d6678da7c49d4e113d7c3d7de23ebfcc4fa849576243ee55d22c2a584db2\": container with ID starting with fff6d6678da7c49d4e113d7c3d7de23ebfcc4fa849576243ee55d22c2a584db2 not found: ID does not exist" Oct 08 18:35:24 crc kubenswrapper[4859]: I1008 18:35:24.790668 4859 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Oct 08 18:35:24 crc kubenswrapper[4859]: I1008 18:35:24.799844 4859 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/ceilometer-0"] Oct 08 18:35:24 crc kubenswrapper[4859]: I1008 18:35:24.815049 4859 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ceilometer-0"] Oct 08 18:35:24 crc kubenswrapper[4859]: E1008 18:35:24.815434 4859 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c3a48f64-6a96-4071-89cc-3f7bc1116694" containerName="sg-core" Oct 08 18:35:24 crc kubenswrapper[4859]: I1008 18:35:24.815471 4859 state_mem.go:107] "Deleted CPUSet assignment" podUID="c3a48f64-6a96-4071-89cc-3f7bc1116694" containerName="sg-core" Oct 08 18:35:24 crc kubenswrapper[4859]: E1008 18:35:24.815496 4859 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c3a48f64-6a96-4071-89cc-3f7bc1116694" containerName="ceilometer-notification-agent" Oct 08 18:35:24 crc kubenswrapper[4859]: I1008 18:35:24.815506 4859 state_mem.go:107] "Deleted CPUSet assignment" podUID="c3a48f64-6a96-4071-89cc-3f7bc1116694" containerName="ceilometer-notification-agent" Oct 08 18:35:24 crc kubenswrapper[4859]: E1008 18:35:24.815533 4859 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c3a48f64-6a96-4071-89cc-3f7bc1116694" containerName="proxy-httpd" Oct 08 18:35:24 crc kubenswrapper[4859]: I1008 18:35:24.815539 4859 state_mem.go:107] "Deleted CPUSet assignment" podUID="c3a48f64-6a96-4071-89cc-3f7bc1116694" containerName="proxy-httpd" Oct 08 18:35:24 crc kubenswrapper[4859]: E1008 18:35:24.815551 4859 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c3a48f64-6a96-4071-89cc-3f7bc1116694" containerName="ceilometer-central-agent" Oct 08 18:35:24 crc kubenswrapper[4859]: I1008 18:35:24.815557 4859 state_mem.go:107] "Deleted CPUSet assignment" podUID="c3a48f64-6a96-4071-89cc-3f7bc1116694" containerName="ceilometer-central-agent" Oct 08 18:35:24 crc kubenswrapper[4859]: I1008 18:35:24.815729 4859 memory_manager.go:354] "RemoveStaleState removing state" podUID="c3a48f64-6a96-4071-89cc-3f7bc1116694" containerName="proxy-httpd" Oct 08 18:35:24 crc kubenswrapper[4859]: I1008 18:35:24.815740 4859 memory_manager.go:354] "RemoveStaleState removing state" podUID="c3a48f64-6a96-4071-89cc-3f7bc1116694" containerName="ceilometer-notification-agent" Oct 08 18:35:24 crc kubenswrapper[4859]: I1008 18:35:24.815760 4859 memory_manager.go:354] "RemoveStaleState removing state" podUID="c3a48f64-6a96-4071-89cc-3f7bc1116694" containerName="sg-core" Oct 08 18:35:24 crc kubenswrapper[4859]: I1008 18:35:24.815768 4859 memory_manager.go:354] "RemoveStaleState removing state" podUID="c3a48f64-6a96-4071-89cc-3f7bc1116694" containerName="ceilometer-central-agent" Oct 08 18:35:24 crc kubenswrapper[4859]: I1008 18:35:24.817406 4859 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Oct 08 18:35:24 crc kubenswrapper[4859]: I1008 18:35:24.819549 4859 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-ceilometer-internal-svc" Oct 08 18:35:24 crc kubenswrapper[4859]: I1008 18:35:24.820151 4859 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-config-data" Oct 08 18:35:24 crc kubenswrapper[4859]: I1008 18:35:24.820370 4859 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-scripts" Oct 08 18:35:24 crc kubenswrapper[4859]: I1008 18:35:24.838028 4859 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Oct 08 18:35:24 crc kubenswrapper[4859]: I1008 18:35:24.874745 4859 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-conductor-db-sync-9hb79"] Oct 08 18:35:24 crc kubenswrapper[4859]: I1008 18:35:24.884069 4859 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/135723dc-9767-4d2e-bf09-ad570aa5ee6d-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"135723dc-9767-4d2e-bf09-ad570aa5ee6d\") " pod="openstack/ceilometer-0" Oct 08 18:35:24 crc kubenswrapper[4859]: I1008 18:35:24.884342 4859 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/135723dc-9767-4d2e-bf09-ad570aa5ee6d-ceilometer-tls-certs\") pod \"ceilometer-0\" (UID: \"135723dc-9767-4d2e-bf09-ad570aa5ee6d\") " pod="openstack/ceilometer-0" Oct 08 18:35:24 crc kubenswrapper[4859]: I1008 18:35:24.884394 4859 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/135723dc-9767-4d2e-bf09-ad570aa5ee6d-scripts\") pod \"ceilometer-0\" (UID: \"135723dc-9767-4d2e-bf09-ad570aa5ee6d\") " pod="openstack/ceilometer-0" Oct 08 18:35:24 crc kubenswrapper[4859]: I1008 18:35:24.884555 4859 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/135723dc-9767-4d2e-bf09-ad570aa5ee6d-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"135723dc-9767-4d2e-bf09-ad570aa5ee6d\") " pod="openstack/ceilometer-0" Oct 08 18:35:24 crc kubenswrapper[4859]: I1008 18:35:24.884651 4859 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/135723dc-9767-4d2e-bf09-ad570aa5ee6d-run-httpd\") pod \"ceilometer-0\" (UID: \"135723dc-9767-4d2e-bf09-ad570aa5ee6d\") " pod="openstack/ceilometer-0" Oct 08 18:35:24 crc kubenswrapper[4859]: I1008 18:35:24.884761 4859 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-zpzk8\" (UniqueName: \"kubernetes.io/projected/135723dc-9767-4d2e-bf09-ad570aa5ee6d-kube-api-access-zpzk8\") pod \"ceilometer-0\" (UID: \"135723dc-9767-4d2e-bf09-ad570aa5ee6d\") " pod="openstack/ceilometer-0" Oct 08 18:35:24 crc kubenswrapper[4859]: I1008 18:35:24.884847 4859 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/135723dc-9767-4d2e-bf09-ad570aa5ee6d-log-httpd\") pod \"ceilometer-0\" (UID: \"135723dc-9767-4d2e-bf09-ad570aa5ee6d\") " pod="openstack/ceilometer-0" Oct 08 18:35:24 crc kubenswrapper[4859]: I1008 18:35:24.885013 4859 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/135723dc-9767-4d2e-bf09-ad570aa5ee6d-config-data\") pod \"ceilometer-0\" (UID: \"135723dc-9767-4d2e-bf09-ad570aa5ee6d\") " pod="openstack/ceilometer-0" Oct 08 18:35:24 crc kubenswrapper[4859]: I1008 18:35:24.988792 4859 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/135723dc-9767-4d2e-bf09-ad570aa5ee6d-scripts\") pod \"ceilometer-0\" (UID: \"135723dc-9767-4d2e-bf09-ad570aa5ee6d\") " pod="openstack/ceilometer-0" Oct 08 18:35:24 crc kubenswrapper[4859]: I1008 18:35:24.988862 4859 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/135723dc-9767-4d2e-bf09-ad570aa5ee6d-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"135723dc-9767-4d2e-bf09-ad570aa5ee6d\") " pod="openstack/ceilometer-0" Oct 08 18:35:24 crc kubenswrapper[4859]: I1008 18:35:24.988898 4859 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/135723dc-9767-4d2e-bf09-ad570aa5ee6d-run-httpd\") pod \"ceilometer-0\" (UID: \"135723dc-9767-4d2e-bf09-ad570aa5ee6d\") " pod="openstack/ceilometer-0" Oct 08 18:35:24 crc kubenswrapper[4859]: I1008 18:35:24.988939 4859 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-zpzk8\" (UniqueName: \"kubernetes.io/projected/135723dc-9767-4d2e-bf09-ad570aa5ee6d-kube-api-access-zpzk8\") pod \"ceilometer-0\" (UID: \"135723dc-9767-4d2e-bf09-ad570aa5ee6d\") " pod="openstack/ceilometer-0" Oct 08 18:35:24 crc kubenswrapper[4859]: I1008 18:35:24.988980 4859 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/135723dc-9767-4d2e-bf09-ad570aa5ee6d-log-httpd\") pod \"ceilometer-0\" (UID: \"135723dc-9767-4d2e-bf09-ad570aa5ee6d\") " pod="openstack/ceilometer-0" Oct 08 18:35:24 crc kubenswrapper[4859]: I1008 18:35:24.989016 4859 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/135723dc-9767-4d2e-bf09-ad570aa5ee6d-config-data\") pod \"ceilometer-0\" (UID: \"135723dc-9767-4d2e-bf09-ad570aa5ee6d\") " pod="openstack/ceilometer-0" Oct 08 18:35:24 crc kubenswrapper[4859]: I1008 18:35:24.989044 4859 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/135723dc-9767-4d2e-bf09-ad570aa5ee6d-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"135723dc-9767-4d2e-bf09-ad570aa5ee6d\") " pod="openstack/ceilometer-0" Oct 08 18:35:24 crc kubenswrapper[4859]: I1008 18:35:24.989062 4859 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/135723dc-9767-4d2e-bf09-ad570aa5ee6d-ceilometer-tls-certs\") pod \"ceilometer-0\" (UID: \"135723dc-9767-4d2e-bf09-ad570aa5ee6d\") " pod="openstack/ceilometer-0" Oct 08 18:35:24 crc kubenswrapper[4859]: I1008 18:35:24.989520 4859 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/135723dc-9767-4d2e-bf09-ad570aa5ee6d-run-httpd\") pod \"ceilometer-0\" (UID: \"135723dc-9767-4d2e-bf09-ad570aa5ee6d\") " pod="openstack/ceilometer-0" Oct 08 18:35:24 crc kubenswrapper[4859]: I1008 18:35:24.989556 4859 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/135723dc-9767-4d2e-bf09-ad570aa5ee6d-log-httpd\") pod \"ceilometer-0\" (UID: \"135723dc-9767-4d2e-bf09-ad570aa5ee6d\") " pod="openstack/ceilometer-0" Oct 08 18:35:24 crc kubenswrapper[4859]: I1008 18:35:24.994659 4859 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/135723dc-9767-4d2e-bf09-ad570aa5ee6d-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"135723dc-9767-4d2e-bf09-ad570aa5ee6d\") " pod="openstack/ceilometer-0" Oct 08 18:35:24 crc kubenswrapper[4859]: I1008 18:35:24.994717 4859 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/135723dc-9767-4d2e-bf09-ad570aa5ee6d-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"135723dc-9767-4d2e-bf09-ad570aa5ee6d\") " pod="openstack/ceilometer-0" Oct 08 18:35:24 crc kubenswrapper[4859]: I1008 18:35:24.996306 4859 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/135723dc-9767-4d2e-bf09-ad570aa5ee6d-config-data\") pod \"ceilometer-0\" (UID: \"135723dc-9767-4d2e-bf09-ad570aa5ee6d\") " pod="openstack/ceilometer-0" Oct 08 18:35:24 crc kubenswrapper[4859]: I1008 18:35:24.997663 4859 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/135723dc-9767-4d2e-bf09-ad570aa5ee6d-scripts\") pod \"ceilometer-0\" (UID: \"135723dc-9767-4d2e-bf09-ad570aa5ee6d\") " pod="openstack/ceilometer-0" Oct 08 18:35:24 crc kubenswrapper[4859]: I1008 18:35:24.998270 4859 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/135723dc-9767-4d2e-bf09-ad570aa5ee6d-ceilometer-tls-certs\") pod \"ceilometer-0\" (UID: \"135723dc-9767-4d2e-bf09-ad570aa5ee6d\") " pod="openstack/ceilometer-0" Oct 08 18:35:25 crc kubenswrapper[4859]: I1008 18:35:25.009394 4859 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-zpzk8\" (UniqueName: \"kubernetes.io/projected/135723dc-9767-4d2e-bf09-ad570aa5ee6d-kube-api-access-zpzk8\") pod \"ceilometer-0\" (UID: \"135723dc-9767-4d2e-bf09-ad570aa5ee6d\") " pod="openstack/ceilometer-0" Oct 08 18:35:25 crc kubenswrapper[4859]: I1008 18:35:25.137888 4859 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Oct 08 18:35:25 crc kubenswrapper[4859]: I1008 18:35:25.467654 4859 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-conductor-db-sync-9hb79" event={"ID":"b03ff9b0-98b5-4d0b-990a-253ada95b6d8","Type":"ContainerStarted","Data":"70c0958706465d13d2a058d84b3e16afeb9303d148a3ff84c164a7113efeb1c9"} Oct 08 18:35:25 crc kubenswrapper[4859]: I1008 18:35:25.597822 4859 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Oct 08 18:35:26 crc kubenswrapper[4859]: I1008 18:35:26.482443 4859 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="c3a48f64-6a96-4071-89cc-3f7bc1116694" path="/var/lib/kubelet/pods/c3a48f64-6a96-4071-89cc-3f7bc1116694/volumes" Oct 08 18:35:26 crc kubenswrapper[4859]: I1008 18:35:26.485256 4859 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"135723dc-9767-4d2e-bf09-ad570aa5ee6d","Type":"ContainerStarted","Data":"e8b8ab1aa42618805f0c5cb12f6ef9a1a2f4d979e05d79124e59dbd38ebec03f"} Oct 08 18:35:26 crc kubenswrapper[4859]: I1008 18:35:26.485294 4859 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"135723dc-9767-4d2e-bf09-ad570aa5ee6d","Type":"ContainerStarted","Data":"8c0c6f9e880bbe9e339983489e0b0f50589049373e11f98fc1645b2c18b63742"} Oct 08 18:35:26 crc kubenswrapper[4859]: I1008 18:35:26.853754 4859 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/glance-default-external-api-0" Oct 08 18:35:26 crc kubenswrapper[4859]: I1008 18:35:26.853806 4859 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/glance-default-external-api-0" Oct 08 18:35:26 crc kubenswrapper[4859]: I1008 18:35:26.892680 4859 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/glance-default-external-api-0" Oct 08 18:35:26 crc kubenswrapper[4859]: I1008 18:35:26.915723 4859 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/glance-default-external-api-0" Oct 08 18:35:27 crc kubenswrapper[4859]: I1008 18:35:27.495837 4859 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"135723dc-9767-4d2e-bf09-ad570aa5ee6d","Type":"ContainerStarted","Data":"55cd54864a5d4700d7ce2a2c53eb77e490165f5c160f17e3483c4203893fad70"} Oct 08 18:35:27 crc kubenswrapper[4859]: I1008 18:35:27.496188 4859 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/glance-default-external-api-0" Oct 08 18:35:27 crc kubenswrapper[4859]: I1008 18:35:27.496202 4859 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/glance-default-external-api-0" Oct 08 18:35:27 crc kubenswrapper[4859]: I1008 18:35:27.703342 4859 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/glance-default-internal-api-0" Oct 08 18:35:27 crc kubenswrapper[4859]: I1008 18:35:27.703418 4859 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/glance-default-internal-api-0" Oct 08 18:35:27 crc kubenswrapper[4859]: I1008 18:35:27.744325 4859 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/glance-default-internal-api-0" Oct 08 18:35:27 crc kubenswrapper[4859]: I1008 18:35:27.766093 4859 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/glance-default-internal-api-0" Oct 08 18:35:28 crc kubenswrapper[4859]: I1008 18:35:28.505749 4859 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/glance-default-internal-api-0" Oct 08 18:35:28 crc kubenswrapper[4859]: I1008 18:35:28.506121 4859 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/glance-default-internal-api-0" Oct 08 18:35:29 crc kubenswrapper[4859]: I1008 18:35:29.615669 4859 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/glance-default-external-api-0" Oct 08 18:35:29 crc kubenswrapper[4859]: I1008 18:35:29.616006 4859 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" Oct 08 18:35:29 crc kubenswrapper[4859]: I1008 18:35:29.696281 4859 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/glance-default-external-api-0" Oct 08 18:35:30 crc kubenswrapper[4859]: I1008 18:35:30.646208 4859 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/glance-default-internal-api-0" Oct 08 18:35:30 crc kubenswrapper[4859]: I1008 18:35:30.646700 4859 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" Oct 08 18:35:30 crc kubenswrapper[4859]: I1008 18:35:30.650378 4859 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/glance-default-internal-api-0" Oct 08 18:35:33 crc kubenswrapper[4859]: I1008 18:35:33.552224 4859 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"135723dc-9767-4d2e-bf09-ad570aa5ee6d","Type":"ContainerStarted","Data":"06338d13aaf10078c779ee7529acb2d54d17c0372c55baae66dd438f77d44c41"} Oct 08 18:35:33 crc kubenswrapper[4859]: I1008 18:35:33.555412 4859 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-conductor-db-sync-9hb79" event={"ID":"b03ff9b0-98b5-4d0b-990a-253ada95b6d8","Type":"ContainerStarted","Data":"dbccd234b8f52eb65169180bc0bf5c99fd5abfe26faafb1542803fd76eede24d"} Oct 08 18:35:33 crc kubenswrapper[4859]: I1008 18:35:33.578992 4859 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-cell0-conductor-db-sync-9hb79" podStartSLOduration=2.63687647 podStartE2EDuration="10.578974534s" podCreationTimestamp="2025-10-08 18:35:23 +0000 UTC" firstStartedPulling="2025-10-08 18:35:24.882537703 +0000 UTC m=+1095.129377082" lastFinishedPulling="2025-10-08 18:35:32.824635767 +0000 UTC m=+1103.071475146" observedRunningTime="2025-10-08 18:35:33.571483103 +0000 UTC m=+1103.818322512" watchObservedRunningTime="2025-10-08 18:35:33.578974534 +0000 UTC m=+1103.825813913" Oct 08 18:35:34 crc kubenswrapper[4859]: I1008 18:35:34.566646 4859 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"135723dc-9767-4d2e-bf09-ad570aa5ee6d","Type":"ContainerStarted","Data":"29d177c46e1eabafffe72b4a4a14e55bd3dafb7cc676aa4b7d776e4de32f5283"} Oct 08 18:35:34 crc kubenswrapper[4859]: I1008 18:35:34.567181 4859 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ceilometer-0" Oct 08 18:35:34 crc kubenswrapper[4859]: I1008 18:35:34.609381 4859 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ceilometer-0" podStartSLOduration=2.119865881 podStartE2EDuration="10.609359775s" podCreationTimestamp="2025-10-08 18:35:24 +0000 UTC" firstStartedPulling="2025-10-08 18:35:25.620176797 +0000 UTC m=+1095.867016176" lastFinishedPulling="2025-10-08 18:35:34.109670691 +0000 UTC m=+1104.356510070" observedRunningTime="2025-10-08 18:35:34.602327504 +0000 UTC m=+1104.849166903" watchObservedRunningTime="2025-10-08 18:35:34.609359775 +0000 UTC m=+1104.856199154" Oct 08 18:35:46 crc kubenswrapper[4859]: I1008 18:35:46.698205 4859 generic.go:334] "Generic (PLEG): container finished" podID="b03ff9b0-98b5-4d0b-990a-253ada95b6d8" containerID="dbccd234b8f52eb65169180bc0bf5c99fd5abfe26faafb1542803fd76eede24d" exitCode=0 Oct 08 18:35:46 crc kubenswrapper[4859]: I1008 18:35:46.698353 4859 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-conductor-db-sync-9hb79" event={"ID":"b03ff9b0-98b5-4d0b-990a-253ada95b6d8","Type":"ContainerDied","Data":"dbccd234b8f52eb65169180bc0bf5c99fd5abfe26faafb1542803fd76eede24d"} Oct 08 18:35:48 crc kubenswrapper[4859]: I1008 18:35:48.156224 4859 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-conductor-db-sync-9hb79" Oct 08 18:35:48 crc kubenswrapper[4859]: I1008 18:35:48.260055 4859 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/b03ff9b0-98b5-4d0b-990a-253ada95b6d8-config-data\") pod \"b03ff9b0-98b5-4d0b-990a-253ada95b6d8\" (UID: \"b03ff9b0-98b5-4d0b-990a-253ada95b6d8\") " Oct 08 18:35:48 crc kubenswrapper[4859]: I1008 18:35:48.260207 4859 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/b03ff9b0-98b5-4d0b-990a-253ada95b6d8-scripts\") pod \"b03ff9b0-98b5-4d0b-990a-253ada95b6d8\" (UID: \"b03ff9b0-98b5-4d0b-990a-253ada95b6d8\") " Oct 08 18:35:48 crc kubenswrapper[4859]: I1008 18:35:48.260248 4859 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b03ff9b0-98b5-4d0b-990a-253ada95b6d8-combined-ca-bundle\") pod \"b03ff9b0-98b5-4d0b-990a-253ada95b6d8\" (UID: \"b03ff9b0-98b5-4d0b-990a-253ada95b6d8\") " Oct 08 18:35:48 crc kubenswrapper[4859]: I1008 18:35:48.260328 4859 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-dhdhm\" (UniqueName: \"kubernetes.io/projected/b03ff9b0-98b5-4d0b-990a-253ada95b6d8-kube-api-access-dhdhm\") pod \"b03ff9b0-98b5-4d0b-990a-253ada95b6d8\" (UID: \"b03ff9b0-98b5-4d0b-990a-253ada95b6d8\") " Oct 08 18:35:48 crc kubenswrapper[4859]: I1008 18:35:48.270855 4859 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b03ff9b0-98b5-4d0b-990a-253ada95b6d8-scripts" (OuterVolumeSpecName: "scripts") pod "b03ff9b0-98b5-4d0b-990a-253ada95b6d8" (UID: "b03ff9b0-98b5-4d0b-990a-253ada95b6d8"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 08 18:35:48 crc kubenswrapper[4859]: I1008 18:35:48.274384 4859 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b03ff9b0-98b5-4d0b-990a-253ada95b6d8-kube-api-access-dhdhm" (OuterVolumeSpecName: "kube-api-access-dhdhm") pod "b03ff9b0-98b5-4d0b-990a-253ada95b6d8" (UID: "b03ff9b0-98b5-4d0b-990a-253ada95b6d8"). InnerVolumeSpecName "kube-api-access-dhdhm". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 08 18:35:48 crc kubenswrapper[4859]: I1008 18:35:48.290775 4859 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b03ff9b0-98b5-4d0b-990a-253ada95b6d8-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "b03ff9b0-98b5-4d0b-990a-253ada95b6d8" (UID: "b03ff9b0-98b5-4d0b-990a-253ada95b6d8"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 08 18:35:48 crc kubenswrapper[4859]: I1008 18:35:48.294823 4859 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b03ff9b0-98b5-4d0b-990a-253ada95b6d8-config-data" (OuterVolumeSpecName: "config-data") pod "b03ff9b0-98b5-4d0b-990a-253ada95b6d8" (UID: "b03ff9b0-98b5-4d0b-990a-253ada95b6d8"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 08 18:35:48 crc kubenswrapper[4859]: I1008 18:35:48.364653 4859 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b03ff9b0-98b5-4d0b-990a-253ada95b6d8-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 08 18:35:48 crc kubenswrapper[4859]: I1008 18:35:48.364715 4859 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-dhdhm\" (UniqueName: \"kubernetes.io/projected/b03ff9b0-98b5-4d0b-990a-253ada95b6d8-kube-api-access-dhdhm\") on node \"crc\" DevicePath \"\"" Oct 08 18:35:48 crc kubenswrapper[4859]: I1008 18:35:48.364750 4859 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/b03ff9b0-98b5-4d0b-990a-253ada95b6d8-config-data\") on node \"crc\" DevicePath \"\"" Oct 08 18:35:48 crc kubenswrapper[4859]: I1008 18:35:48.364759 4859 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/b03ff9b0-98b5-4d0b-990a-253ada95b6d8-scripts\") on node \"crc\" DevicePath \"\"" Oct 08 18:35:48 crc kubenswrapper[4859]: I1008 18:35:48.721253 4859 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-conductor-db-sync-9hb79" event={"ID":"b03ff9b0-98b5-4d0b-990a-253ada95b6d8","Type":"ContainerDied","Data":"70c0958706465d13d2a058d84b3e16afeb9303d148a3ff84c164a7113efeb1c9"} Oct 08 18:35:48 crc kubenswrapper[4859]: I1008 18:35:48.721645 4859 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="70c0958706465d13d2a058d84b3e16afeb9303d148a3ff84c164a7113efeb1c9" Oct 08 18:35:48 crc kubenswrapper[4859]: I1008 18:35:48.721747 4859 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-conductor-db-sync-9hb79" Oct 08 18:35:48 crc kubenswrapper[4859]: I1008 18:35:48.902909 4859 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell0-conductor-0"] Oct 08 18:35:48 crc kubenswrapper[4859]: E1008 18:35:48.903377 4859 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b03ff9b0-98b5-4d0b-990a-253ada95b6d8" containerName="nova-cell0-conductor-db-sync" Oct 08 18:35:48 crc kubenswrapper[4859]: I1008 18:35:48.903399 4859 state_mem.go:107] "Deleted CPUSet assignment" podUID="b03ff9b0-98b5-4d0b-990a-253ada95b6d8" containerName="nova-cell0-conductor-db-sync" Oct 08 18:35:48 crc kubenswrapper[4859]: I1008 18:35:48.903648 4859 memory_manager.go:354] "RemoveStaleState removing state" podUID="b03ff9b0-98b5-4d0b-990a-253ada95b6d8" containerName="nova-cell0-conductor-db-sync" Oct 08 18:35:48 crc kubenswrapper[4859]: I1008 18:35:48.904423 4859 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-conductor-0" Oct 08 18:35:48 crc kubenswrapper[4859]: I1008 18:35:48.907122 4859 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell0-conductor-config-data" Oct 08 18:35:48 crc kubenswrapper[4859]: I1008 18:35:48.907171 4859 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-nova-dockercfg-hsn5b" Oct 08 18:35:48 crc kubenswrapper[4859]: I1008 18:35:48.913609 4859 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-conductor-0"] Oct 08 18:35:48 crc kubenswrapper[4859]: I1008 18:35:48.977172 4859 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/8c9e1318-b507-4caa-832f-ea7c02f7c199-config-data\") pod \"nova-cell0-conductor-0\" (UID: \"8c9e1318-b507-4caa-832f-ea7c02f7c199\") " pod="openstack/nova-cell0-conductor-0" Oct 08 18:35:48 crc kubenswrapper[4859]: I1008 18:35:48.977229 4859 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-z47qp\" (UniqueName: \"kubernetes.io/projected/8c9e1318-b507-4caa-832f-ea7c02f7c199-kube-api-access-z47qp\") pod \"nova-cell0-conductor-0\" (UID: \"8c9e1318-b507-4caa-832f-ea7c02f7c199\") " pod="openstack/nova-cell0-conductor-0" Oct 08 18:35:48 crc kubenswrapper[4859]: I1008 18:35:48.977256 4859 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8c9e1318-b507-4caa-832f-ea7c02f7c199-combined-ca-bundle\") pod \"nova-cell0-conductor-0\" (UID: \"8c9e1318-b507-4caa-832f-ea7c02f7c199\") " pod="openstack/nova-cell0-conductor-0" Oct 08 18:35:49 crc kubenswrapper[4859]: I1008 18:35:49.079073 4859 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-z47qp\" (UniqueName: \"kubernetes.io/projected/8c9e1318-b507-4caa-832f-ea7c02f7c199-kube-api-access-z47qp\") pod \"nova-cell0-conductor-0\" (UID: \"8c9e1318-b507-4caa-832f-ea7c02f7c199\") " pod="openstack/nova-cell0-conductor-0" Oct 08 18:35:49 crc kubenswrapper[4859]: I1008 18:35:49.079150 4859 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8c9e1318-b507-4caa-832f-ea7c02f7c199-combined-ca-bundle\") pod \"nova-cell0-conductor-0\" (UID: \"8c9e1318-b507-4caa-832f-ea7c02f7c199\") " pod="openstack/nova-cell0-conductor-0" Oct 08 18:35:49 crc kubenswrapper[4859]: I1008 18:35:49.079293 4859 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/8c9e1318-b507-4caa-832f-ea7c02f7c199-config-data\") pod \"nova-cell0-conductor-0\" (UID: \"8c9e1318-b507-4caa-832f-ea7c02f7c199\") " pod="openstack/nova-cell0-conductor-0" Oct 08 18:35:49 crc kubenswrapper[4859]: I1008 18:35:49.087668 4859 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/8c9e1318-b507-4caa-832f-ea7c02f7c199-config-data\") pod \"nova-cell0-conductor-0\" (UID: \"8c9e1318-b507-4caa-832f-ea7c02f7c199\") " pod="openstack/nova-cell0-conductor-0" Oct 08 18:35:49 crc kubenswrapper[4859]: I1008 18:35:49.110407 4859 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8c9e1318-b507-4caa-832f-ea7c02f7c199-combined-ca-bundle\") pod \"nova-cell0-conductor-0\" (UID: \"8c9e1318-b507-4caa-832f-ea7c02f7c199\") " pod="openstack/nova-cell0-conductor-0" Oct 08 18:35:49 crc kubenswrapper[4859]: I1008 18:35:49.110408 4859 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-z47qp\" (UniqueName: \"kubernetes.io/projected/8c9e1318-b507-4caa-832f-ea7c02f7c199-kube-api-access-z47qp\") pod \"nova-cell0-conductor-0\" (UID: \"8c9e1318-b507-4caa-832f-ea7c02f7c199\") " pod="openstack/nova-cell0-conductor-0" Oct 08 18:35:49 crc kubenswrapper[4859]: I1008 18:35:49.222952 4859 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-conductor-0" Oct 08 18:35:49 crc kubenswrapper[4859]: I1008 18:35:49.728654 4859 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-conductor-0"] Oct 08 18:35:50 crc kubenswrapper[4859]: I1008 18:35:50.746156 4859 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-conductor-0" event={"ID":"8c9e1318-b507-4caa-832f-ea7c02f7c199","Type":"ContainerStarted","Data":"f717382cc600234775a31fd2ef62a2e2cc83423a6d4da37682083aacdf227273"} Oct 08 18:35:50 crc kubenswrapper[4859]: I1008 18:35:50.746503 4859 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-conductor-0" event={"ID":"8c9e1318-b507-4caa-832f-ea7c02f7c199","Type":"ContainerStarted","Data":"702d7f584c0ed45ed85adb8aa588a6c542dfbd7d2a19260f319536cc9dfa7c9d"} Oct 08 18:35:50 crc kubenswrapper[4859]: I1008 18:35:50.746737 4859 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-cell0-conductor-0" Oct 08 18:35:50 crc kubenswrapper[4859]: I1008 18:35:50.768767 4859 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-cell0-conductor-0" podStartSLOduration=2.768633951 podStartE2EDuration="2.768633951s" podCreationTimestamp="2025-10-08 18:35:48 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-08 18:35:50.764942325 +0000 UTC m=+1121.011781704" watchObservedRunningTime="2025-10-08 18:35:50.768633951 +0000 UTC m=+1121.015473370" Oct 08 18:35:54 crc kubenswrapper[4859]: I1008 18:35:54.255748 4859 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-cell0-conductor-0" Oct 08 18:35:54 crc kubenswrapper[4859]: I1008 18:35:54.733227 4859 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell0-cell-mapping-r2mmw"] Oct 08 18:35:54 crc kubenswrapper[4859]: I1008 18:35:54.735678 4859 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-cell-mapping-r2mmw" Oct 08 18:35:54 crc kubenswrapper[4859]: I1008 18:35:54.742141 4859 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell0-manage-scripts" Oct 08 18:35:54 crc kubenswrapper[4859]: I1008 18:35:54.743944 4859 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell0-manage-config-data" Oct 08 18:35:54 crc kubenswrapper[4859]: I1008 18:35:54.747927 4859 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-cell-mapping-r2mmw"] Oct 08 18:35:54 crc kubenswrapper[4859]: I1008 18:35:54.806986 4859 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-5h7hc\" (UniqueName: \"kubernetes.io/projected/76588317-83a8-42b4-8a4f-502d6889c7b5-kube-api-access-5h7hc\") pod \"nova-cell0-cell-mapping-r2mmw\" (UID: \"76588317-83a8-42b4-8a4f-502d6889c7b5\") " pod="openstack/nova-cell0-cell-mapping-r2mmw" Oct 08 18:35:54 crc kubenswrapper[4859]: I1008 18:35:54.807146 4859 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/76588317-83a8-42b4-8a4f-502d6889c7b5-combined-ca-bundle\") pod \"nova-cell0-cell-mapping-r2mmw\" (UID: \"76588317-83a8-42b4-8a4f-502d6889c7b5\") " pod="openstack/nova-cell0-cell-mapping-r2mmw" Oct 08 18:35:54 crc kubenswrapper[4859]: I1008 18:35:54.807181 4859 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/76588317-83a8-42b4-8a4f-502d6889c7b5-config-data\") pod \"nova-cell0-cell-mapping-r2mmw\" (UID: \"76588317-83a8-42b4-8a4f-502d6889c7b5\") " pod="openstack/nova-cell0-cell-mapping-r2mmw" Oct 08 18:35:54 crc kubenswrapper[4859]: I1008 18:35:54.807361 4859 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/76588317-83a8-42b4-8a4f-502d6889c7b5-scripts\") pod \"nova-cell0-cell-mapping-r2mmw\" (UID: \"76588317-83a8-42b4-8a4f-502d6889c7b5\") " pod="openstack/nova-cell0-cell-mapping-r2mmw" Oct 08 18:35:54 crc kubenswrapper[4859]: I1008 18:35:54.908678 4859 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/76588317-83a8-42b4-8a4f-502d6889c7b5-scripts\") pod \"nova-cell0-cell-mapping-r2mmw\" (UID: \"76588317-83a8-42b4-8a4f-502d6889c7b5\") " pod="openstack/nova-cell0-cell-mapping-r2mmw" Oct 08 18:35:54 crc kubenswrapper[4859]: I1008 18:35:54.908776 4859 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-5h7hc\" (UniqueName: \"kubernetes.io/projected/76588317-83a8-42b4-8a4f-502d6889c7b5-kube-api-access-5h7hc\") pod \"nova-cell0-cell-mapping-r2mmw\" (UID: \"76588317-83a8-42b4-8a4f-502d6889c7b5\") " pod="openstack/nova-cell0-cell-mapping-r2mmw" Oct 08 18:35:54 crc kubenswrapper[4859]: I1008 18:35:54.908870 4859 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/76588317-83a8-42b4-8a4f-502d6889c7b5-combined-ca-bundle\") pod \"nova-cell0-cell-mapping-r2mmw\" (UID: \"76588317-83a8-42b4-8a4f-502d6889c7b5\") " pod="openstack/nova-cell0-cell-mapping-r2mmw" Oct 08 18:35:54 crc kubenswrapper[4859]: I1008 18:35:54.908891 4859 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/76588317-83a8-42b4-8a4f-502d6889c7b5-config-data\") pod \"nova-cell0-cell-mapping-r2mmw\" (UID: \"76588317-83a8-42b4-8a4f-502d6889c7b5\") " pod="openstack/nova-cell0-cell-mapping-r2mmw" Oct 08 18:35:54 crc kubenswrapper[4859]: I1008 18:35:54.917341 4859 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/76588317-83a8-42b4-8a4f-502d6889c7b5-config-data\") pod \"nova-cell0-cell-mapping-r2mmw\" (UID: \"76588317-83a8-42b4-8a4f-502d6889c7b5\") " pod="openstack/nova-cell0-cell-mapping-r2mmw" Oct 08 18:35:54 crc kubenswrapper[4859]: I1008 18:35:54.920852 4859 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/76588317-83a8-42b4-8a4f-502d6889c7b5-scripts\") pod \"nova-cell0-cell-mapping-r2mmw\" (UID: \"76588317-83a8-42b4-8a4f-502d6889c7b5\") " pod="openstack/nova-cell0-cell-mapping-r2mmw" Oct 08 18:35:54 crc kubenswrapper[4859]: I1008 18:35:54.924410 4859 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/76588317-83a8-42b4-8a4f-502d6889c7b5-combined-ca-bundle\") pod \"nova-cell0-cell-mapping-r2mmw\" (UID: \"76588317-83a8-42b4-8a4f-502d6889c7b5\") " pod="openstack/nova-cell0-cell-mapping-r2mmw" Oct 08 18:35:54 crc kubenswrapper[4859]: I1008 18:35:54.924490 4859 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-api-0"] Oct 08 18:35:54 crc kubenswrapper[4859]: I1008 18:35:54.926312 4859 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Oct 08 18:35:54 crc kubenswrapper[4859]: I1008 18:35:54.934158 4859 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-api-config-data" Oct 08 18:35:54 crc kubenswrapper[4859]: I1008 18:35:54.944994 4859 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-5h7hc\" (UniqueName: \"kubernetes.io/projected/76588317-83a8-42b4-8a4f-502d6889c7b5-kube-api-access-5h7hc\") pod \"nova-cell0-cell-mapping-r2mmw\" (UID: \"76588317-83a8-42b4-8a4f-502d6889c7b5\") " pod="openstack/nova-cell0-cell-mapping-r2mmw" Oct 08 18:35:54 crc kubenswrapper[4859]: I1008 18:35:54.954495 4859 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-0"] Oct 08 18:35:55 crc kubenswrapper[4859]: I1008 18:35:55.012856 4859 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/2f61fa1e-ec4e-490e-91ef-90cf0887940d-logs\") pod \"nova-api-0\" (UID: \"2f61fa1e-ec4e-490e-91ef-90cf0887940d\") " pod="openstack/nova-api-0" Oct 08 18:35:55 crc kubenswrapper[4859]: I1008 18:35:55.012928 4859 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-vxk6g\" (UniqueName: \"kubernetes.io/projected/2f61fa1e-ec4e-490e-91ef-90cf0887940d-kube-api-access-vxk6g\") pod \"nova-api-0\" (UID: \"2f61fa1e-ec4e-490e-91ef-90cf0887940d\") " pod="openstack/nova-api-0" Oct 08 18:35:55 crc kubenswrapper[4859]: I1008 18:35:55.012985 4859 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2f61fa1e-ec4e-490e-91ef-90cf0887940d-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"2f61fa1e-ec4e-490e-91ef-90cf0887940d\") " pod="openstack/nova-api-0" Oct 08 18:35:55 crc kubenswrapper[4859]: I1008 18:35:55.013029 4859 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/2f61fa1e-ec4e-490e-91ef-90cf0887940d-config-data\") pod \"nova-api-0\" (UID: \"2f61fa1e-ec4e-490e-91ef-90cf0887940d\") " pod="openstack/nova-api-0" Oct 08 18:35:55 crc kubenswrapper[4859]: I1008 18:35:55.028630 4859 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-metadata-0"] Oct 08 18:35:55 crc kubenswrapper[4859]: I1008 18:35:55.030193 4859 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Oct 08 18:35:55 crc kubenswrapper[4859]: I1008 18:35:55.035195 4859 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-metadata-config-data" Oct 08 18:35:55 crc kubenswrapper[4859]: I1008 18:35:55.055372 4859 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-metadata-0"] Oct 08 18:35:55 crc kubenswrapper[4859]: I1008 18:35:55.065123 4859 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-cell-mapping-r2mmw" Oct 08 18:35:55 crc kubenswrapper[4859]: I1008 18:35:55.098835 4859 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell1-novncproxy-0"] Oct 08 18:35:55 crc kubenswrapper[4859]: I1008 18:35:55.100816 4859 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-novncproxy-0" Oct 08 18:35:55 crc kubenswrapper[4859]: I1008 18:35:55.106261 4859 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell1-novncproxy-config-data" Oct 08 18:35:55 crc kubenswrapper[4859]: I1008 18:35:55.116316 4859 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/8a8d7e8d-fea1-464e-b94b-e954ae1ee475-config-data\") pod \"nova-metadata-0\" (UID: \"8a8d7e8d-fea1-464e-b94b-e954ae1ee475\") " pod="openstack/nova-metadata-0" Oct 08 18:35:55 crc kubenswrapper[4859]: I1008 18:35:55.116423 4859 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/2f61fa1e-ec4e-490e-91ef-90cf0887940d-logs\") pod \"nova-api-0\" (UID: \"2f61fa1e-ec4e-490e-91ef-90cf0887940d\") " pod="openstack/nova-api-0" Oct 08 18:35:55 crc kubenswrapper[4859]: I1008 18:35:55.116458 4859 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-vxk6g\" (UniqueName: \"kubernetes.io/projected/2f61fa1e-ec4e-490e-91ef-90cf0887940d-kube-api-access-vxk6g\") pod \"nova-api-0\" (UID: \"2f61fa1e-ec4e-490e-91ef-90cf0887940d\") " pod="openstack/nova-api-0" Oct 08 18:35:55 crc kubenswrapper[4859]: I1008 18:35:55.116506 4859 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2f61fa1e-ec4e-490e-91ef-90cf0887940d-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"2f61fa1e-ec4e-490e-91ef-90cf0887940d\") " pod="openstack/nova-api-0" Oct 08 18:35:55 crc kubenswrapper[4859]: I1008 18:35:55.116525 4859 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/8a8d7e8d-fea1-464e-b94b-e954ae1ee475-logs\") pod \"nova-metadata-0\" (UID: \"8a8d7e8d-fea1-464e-b94b-e954ae1ee475\") " pod="openstack/nova-metadata-0" Oct 08 18:35:55 crc kubenswrapper[4859]: I1008 18:35:55.116567 4859 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/2f61fa1e-ec4e-490e-91ef-90cf0887940d-config-data\") pod \"nova-api-0\" (UID: \"2f61fa1e-ec4e-490e-91ef-90cf0887940d\") " pod="openstack/nova-api-0" Oct 08 18:35:55 crc kubenswrapper[4859]: I1008 18:35:55.116607 4859 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8a8d7e8d-fea1-464e-b94b-e954ae1ee475-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"8a8d7e8d-fea1-464e-b94b-e954ae1ee475\") " pod="openstack/nova-metadata-0" Oct 08 18:35:55 crc kubenswrapper[4859]: I1008 18:35:55.116626 4859 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-6xlmm\" (UniqueName: \"kubernetes.io/projected/8a8d7e8d-fea1-464e-b94b-e954ae1ee475-kube-api-access-6xlmm\") pod \"nova-metadata-0\" (UID: \"8a8d7e8d-fea1-464e-b94b-e954ae1ee475\") " pod="openstack/nova-metadata-0" Oct 08 18:35:55 crc kubenswrapper[4859]: I1008 18:35:55.119289 4859 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/2f61fa1e-ec4e-490e-91ef-90cf0887940d-logs\") pod \"nova-api-0\" (UID: \"2f61fa1e-ec4e-490e-91ef-90cf0887940d\") " pod="openstack/nova-api-0" Oct 08 18:35:55 crc kubenswrapper[4859]: I1008 18:35:55.140617 4859 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2f61fa1e-ec4e-490e-91ef-90cf0887940d-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"2f61fa1e-ec4e-490e-91ef-90cf0887940d\") " pod="openstack/nova-api-0" Oct 08 18:35:55 crc kubenswrapper[4859]: I1008 18:35:55.146852 4859 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/2f61fa1e-ec4e-490e-91ef-90cf0887940d-config-data\") pod \"nova-api-0\" (UID: \"2f61fa1e-ec4e-490e-91ef-90cf0887940d\") " pod="openstack/nova-api-0" Oct 08 18:35:55 crc kubenswrapper[4859]: I1008 18:35:55.166501 4859 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-novncproxy-0"] Oct 08 18:35:55 crc kubenswrapper[4859]: I1008 18:35:55.187496 4859 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-vxk6g\" (UniqueName: \"kubernetes.io/projected/2f61fa1e-ec4e-490e-91ef-90cf0887940d-kube-api-access-vxk6g\") pod \"nova-api-0\" (UID: \"2f61fa1e-ec4e-490e-91ef-90cf0887940d\") " pod="openstack/nova-api-0" Oct 08 18:35:55 crc kubenswrapper[4859]: I1008 18:35:55.200159 4859 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-694b6b9bcc-7zv88"] Oct 08 18:35:55 crc kubenswrapper[4859]: I1008 18:35:55.202624 4859 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-694b6b9bcc-7zv88" Oct 08 18:35:55 crc kubenswrapper[4859]: I1008 18:35:55.216334 4859 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/ceilometer-0" Oct 08 18:35:55 crc kubenswrapper[4859]: I1008 18:35:55.218059 4859 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/41c36921-eda9-432d-9d21-b478e4e49ee1-config-data\") pod \"nova-cell1-novncproxy-0\" (UID: \"41c36921-eda9-432d-9d21-b478e4e49ee1\") " pod="openstack/nova-cell1-novncproxy-0" Oct 08 18:35:55 crc kubenswrapper[4859]: I1008 18:35:55.218094 4859 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/8a8d7e8d-fea1-464e-b94b-e954ae1ee475-config-data\") pod \"nova-metadata-0\" (UID: \"8a8d7e8d-fea1-464e-b94b-e954ae1ee475\") " pod="openstack/nova-metadata-0" Oct 08 18:35:55 crc kubenswrapper[4859]: I1008 18:35:55.218118 4859 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/41c36921-eda9-432d-9d21-b478e4e49ee1-combined-ca-bundle\") pod \"nova-cell1-novncproxy-0\" (UID: \"41c36921-eda9-432d-9d21-b478e4e49ee1\") " pod="openstack/nova-cell1-novncproxy-0" Oct 08 18:35:55 crc kubenswrapper[4859]: I1008 18:35:55.218228 4859 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/8a8d7e8d-fea1-464e-b94b-e954ae1ee475-logs\") pod \"nova-metadata-0\" (UID: \"8a8d7e8d-fea1-464e-b94b-e954ae1ee475\") " pod="openstack/nova-metadata-0" Oct 08 18:35:55 crc kubenswrapper[4859]: I1008 18:35:55.218253 4859 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-ghzs4\" (UniqueName: \"kubernetes.io/projected/41c36921-eda9-432d-9d21-b478e4e49ee1-kube-api-access-ghzs4\") pod \"nova-cell1-novncproxy-0\" (UID: \"41c36921-eda9-432d-9d21-b478e4e49ee1\") " pod="openstack/nova-cell1-novncproxy-0" Oct 08 18:35:55 crc kubenswrapper[4859]: I1008 18:35:55.218295 4859 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8a8d7e8d-fea1-464e-b94b-e954ae1ee475-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"8a8d7e8d-fea1-464e-b94b-e954ae1ee475\") " pod="openstack/nova-metadata-0" Oct 08 18:35:55 crc kubenswrapper[4859]: I1008 18:35:55.218314 4859 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-6xlmm\" (UniqueName: \"kubernetes.io/projected/8a8d7e8d-fea1-464e-b94b-e954ae1ee475-kube-api-access-6xlmm\") pod \"nova-metadata-0\" (UID: \"8a8d7e8d-fea1-464e-b94b-e954ae1ee475\") " pod="openstack/nova-metadata-0" Oct 08 18:35:55 crc kubenswrapper[4859]: I1008 18:35:55.219173 4859 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/8a8d7e8d-fea1-464e-b94b-e954ae1ee475-logs\") pod \"nova-metadata-0\" (UID: \"8a8d7e8d-fea1-464e-b94b-e954ae1ee475\") " pod="openstack/nova-metadata-0" Oct 08 18:35:55 crc kubenswrapper[4859]: I1008 18:35:55.224703 4859 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-scheduler-0"] Oct 08 18:35:55 crc kubenswrapper[4859]: I1008 18:35:55.236059 4859 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Oct 08 18:35:55 crc kubenswrapper[4859]: I1008 18:35:55.238873 4859 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-694b6b9bcc-7zv88"] Oct 08 18:35:55 crc kubenswrapper[4859]: I1008 18:35:55.242728 4859 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/8a8d7e8d-fea1-464e-b94b-e954ae1ee475-config-data\") pod \"nova-metadata-0\" (UID: \"8a8d7e8d-fea1-464e-b94b-e954ae1ee475\") " pod="openstack/nova-metadata-0" Oct 08 18:35:55 crc kubenswrapper[4859]: I1008 18:35:55.243002 4859 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-scheduler-config-data" Oct 08 18:35:55 crc kubenswrapper[4859]: I1008 18:35:55.253422 4859 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8a8d7e8d-fea1-464e-b94b-e954ae1ee475-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"8a8d7e8d-fea1-464e-b94b-e954ae1ee475\") " pod="openstack/nova-metadata-0" Oct 08 18:35:55 crc kubenswrapper[4859]: I1008 18:35:55.257595 4859 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-scheduler-0"] Oct 08 18:35:55 crc kubenswrapper[4859]: I1008 18:35:55.273175 4859 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-6xlmm\" (UniqueName: \"kubernetes.io/projected/8a8d7e8d-fea1-464e-b94b-e954ae1ee475-kube-api-access-6xlmm\") pod \"nova-metadata-0\" (UID: \"8a8d7e8d-fea1-464e-b94b-e954ae1ee475\") " pod="openstack/nova-metadata-0" Oct 08 18:35:55 crc kubenswrapper[4859]: I1008 18:35:55.319457 4859 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/1ea66084-f5e7-4fcc-a439-c7806dde0197-dns-swift-storage-0\") pod \"dnsmasq-dns-694b6b9bcc-7zv88\" (UID: \"1ea66084-f5e7-4fcc-a439-c7806dde0197\") " pod="openstack/dnsmasq-dns-694b6b9bcc-7zv88" Oct 08 18:35:55 crc kubenswrapper[4859]: I1008 18:35:55.319501 4859 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-jd9kx\" (UniqueName: \"kubernetes.io/projected/1ea66084-f5e7-4fcc-a439-c7806dde0197-kube-api-access-jd9kx\") pod \"dnsmasq-dns-694b6b9bcc-7zv88\" (UID: \"1ea66084-f5e7-4fcc-a439-c7806dde0197\") " pod="openstack/dnsmasq-dns-694b6b9bcc-7zv88" Oct 08 18:35:55 crc kubenswrapper[4859]: I1008 18:35:55.319520 4859 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/1ea66084-f5e7-4fcc-a439-c7806dde0197-config\") pod \"dnsmasq-dns-694b6b9bcc-7zv88\" (UID: \"1ea66084-f5e7-4fcc-a439-c7806dde0197\") " pod="openstack/dnsmasq-dns-694b6b9bcc-7zv88" Oct 08 18:35:55 crc kubenswrapper[4859]: I1008 18:35:55.319554 4859 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/0178104e-34da-4178-b6e0-873404e282ef-config-data\") pod \"nova-scheduler-0\" (UID: \"0178104e-34da-4178-b6e0-873404e282ef\") " pod="openstack/nova-scheduler-0" Oct 08 18:35:55 crc kubenswrapper[4859]: I1008 18:35:55.319605 4859 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/41c36921-eda9-432d-9d21-b478e4e49ee1-config-data\") pod \"nova-cell1-novncproxy-0\" (UID: \"41c36921-eda9-432d-9d21-b478e4e49ee1\") " pod="openstack/nova-cell1-novncproxy-0" Oct 08 18:35:55 crc kubenswrapper[4859]: I1008 18:35:55.319633 4859 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/41c36921-eda9-432d-9d21-b478e4e49ee1-combined-ca-bundle\") pod \"nova-cell1-novncproxy-0\" (UID: \"41c36921-eda9-432d-9d21-b478e4e49ee1\") " pod="openstack/nova-cell1-novncproxy-0" Oct 08 18:35:55 crc kubenswrapper[4859]: I1008 18:35:55.319671 4859 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/1ea66084-f5e7-4fcc-a439-c7806dde0197-ovsdbserver-nb\") pod \"dnsmasq-dns-694b6b9bcc-7zv88\" (UID: \"1ea66084-f5e7-4fcc-a439-c7806dde0197\") " pod="openstack/dnsmasq-dns-694b6b9bcc-7zv88" Oct 08 18:35:55 crc kubenswrapper[4859]: I1008 18:35:55.319711 4859 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/1ea66084-f5e7-4fcc-a439-c7806dde0197-ovsdbserver-sb\") pod \"dnsmasq-dns-694b6b9bcc-7zv88\" (UID: \"1ea66084-f5e7-4fcc-a439-c7806dde0197\") " pod="openstack/dnsmasq-dns-694b6b9bcc-7zv88" Oct 08 18:35:55 crc kubenswrapper[4859]: I1008 18:35:55.319768 4859 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-s9nvk\" (UniqueName: \"kubernetes.io/projected/0178104e-34da-4178-b6e0-873404e282ef-kube-api-access-s9nvk\") pod \"nova-scheduler-0\" (UID: \"0178104e-34da-4178-b6e0-873404e282ef\") " pod="openstack/nova-scheduler-0" Oct 08 18:35:55 crc kubenswrapper[4859]: I1008 18:35:55.319786 4859 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/1ea66084-f5e7-4fcc-a439-c7806dde0197-dns-svc\") pod \"dnsmasq-dns-694b6b9bcc-7zv88\" (UID: \"1ea66084-f5e7-4fcc-a439-c7806dde0197\") " pod="openstack/dnsmasq-dns-694b6b9bcc-7zv88" Oct 08 18:35:55 crc kubenswrapper[4859]: I1008 18:35:55.319830 4859 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-ghzs4\" (UniqueName: \"kubernetes.io/projected/41c36921-eda9-432d-9d21-b478e4e49ee1-kube-api-access-ghzs4\") pod \"nova-cell1-novncproxy-0\" (UID: \"41c36921-eda9-432d-9d21-b478e4e49ee1\") " pod="openstack/nova-cell1-novncproxy-0" Oct 08 18:35:55 crc kubenswrapper[4859]: I1008 18:35:55.319864 4859 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0178104e-34da-4178-b6e0-873404e282ef-combined-ca-bundle\") pod \"nova-scheduler-0\" (UID: \"0178104e-34da-4178-b6e0-873404e282ef\") " pod="openstack/nova-scheduler-0" Oct 08 18:35:55 crc kubenswrapper[4859]: I1008 18:35:55.323544 4859 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/41c36921-eda9-432d-9d21-b478e4e49ee1-combined-ca-bundle\") pod \"nova-cell1-novncproxy-0\" (UID: \"41c36921-eda9-432d-9d21-b478e4e49ee1\") " pod="openstack/nova-cell1-novncproxy-0" Oct 08 18:35:55 crc kubenswrapper[4859]: I1008 18:35:55.327379 4859 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/41c36921-eda9-432d-9d21-b478e4e49ee1-config-data\") pod \"nova-cell1-novncproxy-0\" (UID: \"41c36921-eda9-432d-9d21-b478e4e49ee1\") " pod="openstack/nova-cell1-novncproxy-0" Oct 08 18:35:55 crc kubenswrapper[4859]: I1008 18:35:55.336064 4859 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Oct 08 18:35:55 crc kubenswrapper[4859]: I1008 18:35:55.368599 4859 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-ghzs4\" (UniqueName: \"kubernetes.io/projected/41c36921-eda9-432d-9d21-b478e4e49ee1-kube-api-access-ghzs4\") pod \"nova-cell1-novncproxy-0\" (UID: \"41c36921-eda9-432d-9d21-b478e4e49ee1\") " pod="openstack/nova-cell1-novncproxy-0" Oct 08 18:35:55 crc kubenswrapper[4859]: I1008 18:35:55.400657 4859 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Oct 08 18:35:55 crc kubenswrapper[4859]: I1008 18:35:55.421861 4859 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/1ea66084-f5e7-4fcc-a439-c7806dde0197-ovsdbserver-nb\") pod \"dnsmasq-dns-694b6b9bcc-7zv88\" (UID: \"1ea66084-f5e7-4fcc-a439-c7806dde0197\") " pod="openstack/dnsmasq-dns-694b6b9bcc-7zv88" Oct 08 18:35:55 crc kubenswrapper[4859]: I1008 18:35:55.421905 4859 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/1ea66084-f5e7-4fcc-a439-c7806dde0197-ovsdbserver-sb\") pod \"dnsmasq-dns-694b6b9bcc-7zv88\" (UID: \"1ea66084-f5e7-4fcc-a439-c7806dde0197\") " pod="openstack/dnsmasq-dns-694b6b9bcc-7zv88" Oct 08 18:35:55 crc kubenswrapper[4859]: I1008 18:35:55.421951 4859 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s9nvk\" (UniqueName: \"kubernetes.io/projected/0178104e-34da-4178-b6e0-873404e282ef-kube-api-access-s9nvk\") pod \"nova-scheduler-0\" (UID: \"0178104e-34da-4178-b6e0-873404e282ef\") " pod="openstack/nova-scheduler-0" Oct 08 18:35:55 crc kubenswrapper[4859]: I1008 18:35:55.421977 4859 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/1ea66084-f5e7-4fcc-a439-c7806dde0197-dns-svc\") pod \"dnsmasq-dns-694b6b9bcc-7zv88\" (UID: \"1ea66084-f5e7-4fcc-a439-c7806dde0197\") " pod="openstack/dnsmasq-dns-694b6b9bcc-7zv88" Oct 08 18:35:55 crc kubenswrapper[4859]: I1008 18:35:55.422030 4859 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0178104e-34da-4178-b6e0-873404e282ef-combined-ca-bundle\") pod \"nova-scheduler-0\" (UID: \"0178104e-34da-4178-b6e0-873404e282ef\") " pod="openstack/nova-scheduler-0" Oct 08 18:35:55 crc kubenswrapper[4859]: I1008 18:35:55.422076 4859 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/1ea66084-f5e7-4fcc-a439-c7806dde0197-dns-swift-storage-0\") pod \"dnsmasq-dns-694b6b9bcc-7zv88\" (UID: \"1ea66084-f5e7-4fcc-a439-c7806dde0197\") " pod="openstack/dnsmasq-dns-694b6b9bcc-7zv88" Oct 08 18:35:55 crc kubenswrapper[4859]: I1008 18:35:55.422096 4859 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-jd9kx\" (UniqueName: \"kubernetes.io/projected/1ea66084-f5e7-4fcc-a439-c7806dde0197-kube-api-access-jd9kx\") pod \"dnsmasq-dns-694b6b9bcc-7zv88\" (UID: \"1ea66084-f5e7-4fcc-a439-c7806dde0197\") " pod="openstack/dnsmasq-dns-694b6b9bcc-7zv88" Oct 08 18:35:55 crc kubenswrapper[4859]: I1008 18:35:55.422112 4859 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/1ea66084-f5e7-4fcc-a439-c7806dde0197-config\") pod \"dnsmasq-dns-694b6b9bcc-7zv88\" (UID: \"1ea66084-f5e7-4fcc-a439-c7806dde0197\") " pod="openstack/dnsmasq-dns-694b6b9bcc-7zv88" Oct 08 18:35:55 crc kubenswrapper[4859]: I1008 18:35:55.422130 4859 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/0178104e-34da-4178-b6e0-873404e282ef-config-data\") pod \"nova-scheduler-0\" (UID: \"0178104e-34da-4178-b6e0-873404e282ef\") " pod="openstack/nova-scheduler-0" Oct 08 18:35:55 crc kubenswrapper[4859]: I1008 18:35:55.424107 4859 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/1ea66084-f5e7-4fcc-a439-c7806dde0197-dns-svc\") pod \"dnsmasq-dns-694b6b9bcc-7zv88\" (UID: \"1ea66084-f5e7-4fcc-a439-c7806dde0197\") " pod="openstack/dnsmasq-dns-694b6b9bcc-7zv88" Oct 08 18:35:55 crc kubenswrapper[4859]: I1008 18:35:55.425197 4859 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/1ea66084-f5e7-4fcc-a439-c7806dde0197-ovsdbserver-nb\") pod \"dnsmasq-dns-694b6b9bcc-7zv88\" (UID: \"1ea66084-f5e7-4fcc-a439-c7806dde0197\") " pod="openstack/dnsmasq-dns-694b6b9bcc-7zv88" Oct 08 18:35:55 crc kubenswrapper[4859]: I1008 18:35:55.425458 4859 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/1ea66084-f5e7-4fcc-a439-c7806dde0197-dns-swift-storage-0\") pod \"dnsmasq-dns-694b6b9bcc-7zv88\" (UID: \"1ea66084-f5e7-4fcc-a439-c7806dde0197\") " pod="openstack/dnsmasq-dns-694b6b9bcc-7zv88" Oct 08 18:35:55 crc kubenswrapper[4859]: I1008 18:35:55.427026 4859 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/1ea66084-f5e7-4fcc-a439-c7806dde0197-ovsdbserver-sb\") pod \"dnsmasq-dns-694b6b9bcc-7zv88\" (UID: \"1ea66084-f5e7-4fcc-a439-c7806dde0197\") " pod="openstack/dnsmasq-dns-694b6b9bcc-7zv88" Oct 08 18:35:55 crc kubenswrapper[4859]: I1008 18:35:55.430060 4859 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/1ea66084-f5e7-4fcc-a439-c7806dde0197-config\") pod \"dnsmasq-dns-694b6b9bcc-7zv88\" (UID: \"1ea66084-f5e7-4fcc-a439-c7806dde0197\") " pod="openstack/dnsmasq-dns-694b6b9bcc-7zv88" Oct 08 18:35:55 crc kubenswrapper[4859]: I1008 18:35:55.430905 4859 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/0178104e-34da-4178-b6e0-873404e282ef-config-data\") pod \"nova-scheduler-0\" (UID: \"0178104e-34da-4178-b6e0-873404e282ef\") " pod="openstack/nova-scheduler-0" Oct 08 18:35:55 crc kubenswrapper[4859]: I1008 18:35:55.436344 4859 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0178104e-34da-4178-b6e0-873404e282ef-combined-ca-bundle\") pod \"nova-scheduler-0\" (UID: \"0178104e-34da-4178-b6e0-873404e282ef\") " pod="openstack/nova-scheduler-0" Oct 08 18:35:55 crc kubenswrapper[4859]: I1008 18:35:55.446800 4859 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-s9nvk\" (UniqueName: \"kubernetes.io/projected/0178104e-34da-4178-b6e0-873404e282ef-kube-api-access-s9nvk\") pod \"nova-scheduler-0\" (UID: \"0178104e-34da-4178-b6e0-873404e282ef\") " pod="openstack/nova-scheduler-0" Oct 08 18:35:55 crc kubenswrapper[4859]: I1008 18:35:55.453362 4859 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-jd9kx\" (UniqueName: \"kubernetes.io/projected/1ea66084-f5e7-4fcc-a439-c7806dde0197-kube-api-access-jd9kx\") pod \"dnsmasq-dns-694b6b9bcc-7zv88\" (UID: \"1ea66084-f5e7-4fcc-a439-c7806dde0197\") " pod="openstack/dnsmasq-dns-694b6b9bcc-7zv88" Oct 08 18:35:55 crc kubenswrapper[4859]: I1008 18:35:55.588069 4859 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-novncproxy-0" Oct 08 18:35:55 crc kubenswrapper[4859]: I1008 18:35:55.616181 4859 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-694b6b9bcc-7zv88" Oct 08 18:35:55 crc kubenswrapper[4859]: I1008 18:35:55.636912 4859 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Oct 08 18:35:55 crc kubenswrapper[4859]: I1008 18:35:55.758779 4859 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-cell-mapping-r2mmw"] Oct 08 18:35:55 crc kubenswrapper[4859]: I1008 18:35:55.934113 4859 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-0"] Oct 08 18:35:55 crc kubenswrapper[4859]: W1008 18:35:55.956842 4859 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod2f61fa1e_ec4e_490e_91ef_90cf0887940d.slice/crio-0d6bcfda0d0f0c376066ff120cd2fbb40dc18afff086c75338e5fa187fd3bf33 WatchSource:0}: Error finding container 0d6bcfda0d0f0c376066ff120cd2fbb40dc18afff086c75338e5fa187fd3bf33: Status 404 returned error can't find the container with id 0d6bcfda0d0f0c376066ff120cd2fbb40dc18afff086c75338e5fa187fd3bf33 Oct 08 18:35:56 crc kubenswrapper[4859]: I1008 18:35:56.012198 4859 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell1-conductor-db-sync-28hwn"] Oct 08 18:35:56 crc kubenswrapper[4859]: I1008 18:35:56.022794 4859 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-conductor-db-sync-28hwn" Oct 08 18:35:56 crc kubenswrapper[4859]: I1008 18:35:56.032508 4859 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell1-conductor-scripts" Oct 08 18:35:56 crc kubenswrapper[4859]: I1008 18:35:56.032860 4859 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell1-conductor-config-data" Oct 08 18:35:56 crc kubenswrapper[4859]: I1008 18:35:56.035523 4859 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-conductor-db-sync-28hwn"] Oct 08 18:35:56 crc kubenswrapper[4859]: I1008 18:35:56.072989 4859 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-metadata-0"] Oct 08 18:35:56 crc kubenswrapper[4859]: I1008 18:35:56.141603 4859 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/6b0220ac-1dc0-4ca9-a9ae-9335b57820fa-scripts\") pod \"nova-cell1-conductor-db-sync-28hwn\" (UID: \"6b0220ac-1dc0-4ca9-a9ae-9335b57820fa\") " pod="openstack/nova-cell1-conductor-db-sync-28hwn" Oct 08 18:35:56 crc kubenswrapper[4859]: I1008 18:35:56.141662 4859 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6b0220ac-1dc0-4ca9-a9ae-9335b57820fa-combined-ca-bundle\") pod \"nova-cell1-conductor-db-sync-28hwn\" (UID: \"6b0220ac-1dc0-4ca9-a9ae-9335b57820fa\") " pod="openstack/nova-cell1-conductor-db-sync-28hwn" Oct 08 18:35:56 crc kubenswrapper[4859]: I1008 18:35:56.141722 4859 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-9vt2x\" (UniqueName: \"kubernetes.io/projected/6b0220ac-1dc0-4ca9-a9ae-9335b57820fa-kube-api-access-9vt2x\") pod \"nova-cell1-conductor-db-sync-28hwn\" (UID: \"6b0220ac-1dc0-4ca9-a9ae-9335b57820fa\") " pod="openstack/nova-cell1-conductor-db-sync-28hwn" Oct 08 18:35:56 crc kubenswrapper[4859]: I1008 18:35:56.141805 4859 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/6b0220ac-1dc0-4ca9-a9ae-9335b57820fa-config-data\") pod \"nova-cell1-conductor-db-sync-28hwn\" (UID: \"6b0220ac-1dc0-4ca9-a9ae-9335b57820fa\") " pod="openstack/nova-cell1-conductor-db-sync-28hwn" Oct 08 18:35:56 crc kubenswrapper[4859]: W1008 18:35:56.219097 4859 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod41c36921_eda9_432d_9d21_b478e4e49ee1.slice/crio-00ee56302b363340528da79c2908dd8a1136515642bde815068e9117bb85782b WatchSource:0}: Error finding container 00ee56302b363340528da79c2908dd8a1136515642bde815068e9117bb85782b: Status 404 returned error can't find the container with id 00ee56302b363340528da79c2908dd8a1136515642bde815068e9117bb85782b Oct 08 18:35:56 crc kubenswrapper[4859]: I1008 18:35:56.245263 4859 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6b0220ac-1dc0-4ca9-a9ae-9335b57820fa-combined-ca-bundle\") pod \"nova-cell1-conductor-db-sync-28hwn\" (UID: \"6b0220ac-1dc0-4ca9-a9ae-9335b57820fa\") " pod="openstack/nova-cell1-conductor-db-sync-28hwn" Oct 08 18:35:56 crc kubenswrapper[4859]: I1008 18:35:56.245362 4859 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-9vt2x\" (UniqueName: \"kubernetes.io/projected/6b0220ac-1dc0-4ca9-a9ae-9335b57820fa-kube-api-access-9vt2x\") pod \"nova-cell1-conductor-db-sync-28hwn\" (UID: \"6b0220ac-1dc0-4ca9-a9ae-9335b57820fa\") " pod="openstack/nova-cell1-conductor-db-sync-28hwn" Oct 08 18:35:56 crc kubenswrapper[4859]: I1008 18:35:56.245515 4859 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/6b0220ac-1dc0-4ca9-a9ae-9335b57820fa-config-data\") pod \"nova-cell1-conductor-db-sync-28hwn\" (UID: \"6b0220ac-1dc0-4ca9-a9ae-9335b57820fa\") " pod="openstack/nova-cell1-conductor-db-sync-28hwn" Oct 08 18:35:56 crc kubenswrapper[4859]: I1008 18:35:56.245620 4859 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/6b0220ac-1dc0-4ca9-a9ae-9335b57820fa-scripts\") pod \"nova-cell1-conductor-db-sync-28hwn\" (UID: \"6b0220ac-1dc0-4ca9-a9ae-9335b57820fa\") " pod="openstack/nova-cell1-conductor-db-sync-28hwn" Oct 08 18:35:56 crc kubenswrapper[4859]: I1008 18:35:56.250442 4859 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-novncproxy-0"] Oct 08 18:35:56 crc kubenswrapper[4859]: I1008 18:35:56.253397 4859 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/6b0220ac-1dc0-4ca9-a9ae-9335b57820fa-scripts\") pod \"nova-cell1-conductor-db-sync-28hwn\" (UID: \"6b0220ac-1dc0-4ca9-a9ae-9335b57820fa\") " pod="openstack/nova-cell1-conductor-db-sync-28hwn" Oct 08 18:35:56 crc kubenswrapper[4859]: I1008 18:35:56.254069 4859 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/6b0220ac-1dc0-4ca9-a9ae-9335b57820fa-config-data\") pod \"nova-cell1-conductor-db-sync-28hwn\" (UID: \"6b0220ac-1dc0-4ca9-a9ae-9335b57820fa\") " pod="openstack/nova-cell1-conductor-db-sync-28hwn" Oct 08 18:35:56 crc kubenswrapper[4859]: I1008 18:35:56.260254 4859 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6b0220ac-1dc0-4ca9-a9ae-9335b57820fa-combined-ca-bundle\") pod \"nova-cell1-conductor-db-sync-28hwn\" (UID: \"6b0220ac-1dc0-4ca9-a9ae-9335b57820fa\") " pod="openstack/nova-cell1-conductor-db-sync-28hwn" Oct 08 18:35:56 crc kubenswrapper[4859]: I1008 18:35:56.266822 4859 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-9vt2x\" (UniqueName: \"kubernetes.io/projected/6b0220ac-1dc0-4ca9-a9ae-9335b57820fa-kube-api-access-9vt2x\") pod \"nova-cell1-conductor-db-sync-28hwn\" (UID: \"6b0220ac-1dc0-4ca9-a9ae-9335b57820fa\") " pod="openstack/nova-cell1-conductor-db-sync-28hwn" Oct 08 18:35:56 crc kubenswrapper[4859]: I1008 18:35:56.364715 4859 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-conductor-db-sync-28hwn" Oct 08 18:35:56 crc kubenswrapper[4859]: I1008 18:35:56.371394 4859 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-694b6b9bcc-7zv88"] Oct 08 18:35:56 crc kubenswrapper[4859]: I1008 18:35:56.389001 4859 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-scheduler-0"] Oct 08 18:35:56 crc kubenswrapper[4859]: I1008 18:35:56.829551 4859 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-cell-mapping-r2mmw" event={"ID":"76588317-83a8-42b4-8a4f-502d6889c7b5","Type":"ContainerStarted","Data":"38db145a1acab79ead1902ab5f38916c3ba433b24d3074e51ed13550d67ec7b4"} Oct 08 18:35:56 crc kubenswrapper[4859]: I1008 18:35:56.830131 4859 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-cell-mapping-r2mmw" event={"ID":"76588317-83a8-42b4-8a4f-502d6889c7b5","Type":"ContainerStarted","Data":"c64b0e8317dcbdf353293fe34270053e426611653787f744b079bb82f831ad49"} Oct 08 18:35:56 crc kubenswrapper[4859]: I1008 18:35:56.833441 4859 generic.go:334] "Generic (PLEG): container finished" podID="1ea66084-f5e7-4fcc-a439-c7806dde0197" containerID="02c74e0a8c552cfbca568550fb788525c9c2ce4b97fc9a1bcdfdf7756cc2a709" exitCode=0 Oct 08 18:35:56 crc kubenswrapper[4859]: I1008 18:35:56.833561 4859 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-694b6b9bcc-7zv88" event={"ID":"1ea66084-f5e7-4fcc-a439-c7806dde0197","Type":"ContainerDied","Data":"02c74e0a8c552cfbca568550fb788525c9c2ce4b97fc9a1bcdfdf7756cc2a709"} Oct 08 18:35:56 crc kubenswrapper[4859]: I1008 18:35:56.833622 4859 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-694b6b9bcc-7zv88" event={"ID":"1ea66084-f5e7-4fcc-a439-c7806dde0197","Type":"ContainerStarted","Data":"98ce65345d863f1e51428d1f5f987c53c4b454fb9c924dbd6bbd66f8c79621d7"} Oct 08 18:35:56 crc kubenswrapper[4859]: I1008 18:35:56.835707 4859 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"8a8d7e8d-fea1-464e-b94b-e954ae1ee475","Type":"ContainerStarted","Data":"59037ecb0d29a74aae0a85b65c900d4ee5e7dc0f9be715b0da96b423abd52be0"} Oct 08 18:35:56 crc kubenswrapper[4859]: I1008 18:35:56.837037 4859 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"2f61fa1e-ec4e-490e-91ef-90cf0887940d","Type":"ContainerStarted","Data":"0d6bcfda0d0f0c376066ff120cd2fbb40dc18afff086c75338e5fa187fd3bf33"} Oct 08 18:35:56 crc kubenswrapper[4859]: I1008 18:35:56.837992 4859 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"0178104e-34da-4178-b6e0-873404e282ef","Type":"ContainerStarted","Data":"163c22fdee9b9ac83ff176f22ea2757ca88f329656d4858aadf926430da91219"} Oct 08 18:35:56 crc kubenswrapper[4859]: I1008 18:35:56.839247 4859 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-conductor-db-sync-28hwn"] Oct 08 18:35:56 crc kubenswrapper[4859]: I1008 18:35:56.839894 4859 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-novncproxy-0" event={"ID":"41c36921-eda9-432d-9d21-b478e4e49ee1","Type":"ContainerStarted","Data":"00ee56302b363340528da79c2908dd8a1136515642bde815068e9117bb85782b"} Oct 08 18:35:56 crc kubenswrapper[4859]: I1008 18:35:56.855127 4859 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-cell0-cell-mapping-r2mmw" podStartSLOduration=2.855104968 podStartE2EDuration="2.855104968s" podCreationTimestamp="2025-10-08 18:35:54 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-08 18:35:56.846037219 +0000 UTC m=+1127.092876628" watchObservedRunningTime="2025-10-08 18:35:56.855104968 +0000 UTC m=+1127.101944357" Oct 08 18:35:57 crc kubenswrapper[4859]: I1008 18:35:57.851641 4859 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-conductor-db-sync-28hwn" event={"ID":"6b0220ac-1dc0-4ca9-a9ae-9335b57820fa","Type":"ContainerStarted","Data":"6fb9ae99d0df9addcb1423e79ad8723c8a3d60ad81028ad521e0f4c0280241b7"} Oct 08 18:35:57 crc kubenswrapper[4859]: I1008 18:35:57.851974 4859 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-conductor-db-sync-28hwn" event={"ID":"6b0220ac-1dc0-4ca9-a9ae-9335b57820fa","Type":"ContainerStarted","Data":"1015abdc89e469b6d2608efb2f3d45f112ac69e3c1e7af4973e5ac3aa3a4f65a"} Oct 08 18:35:57 crc kubenswrapper[4859]: I1008 18:35:57.854735 4859 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-694b6b9bcc-7zv88" event={"ID":"1ea66084-f5e7-4fcc-a439-c7806dde0197","Type":"ContainerStarted","Data":"46b13ba9c6e976285c67c445ae0c938d4f83886ca14e0c5e9153d7c03813fb0b"} Oct 08 18:35:57 crc kubenswrapper[4859]: I1008 18:35:57.879549 4859 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-cell1-conductor-db-sync-28hwn" podStartSLOduration=2.879527184 podStartE2EDuration="2.879527184s" podCreationTimestamp="2025-10-08 18:35:55 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-08 18:35:57.870950959 +0000 UTC m=+1128.117790338" watchObservedRunningTime="2025-10-08 18:35:57.879527184 +0000 UTC m=+1128.126366573" Oct 08 18:35:57 crc kubenswrapper[4859]: I1008 18:35:57.893362 4859 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-694b6b9bcc-7zv88" podStartSLOduration=2.893344569 podStartE2EDuration="2.893344569s" podCreationTimestamp="2025-10-08 18:35:55 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-08 18:35:57.890619131 +0000 UTC m=+1128.137458510" watchObservedRunningTime="2025-10-08 18:35:57.893344569 +0000 UTC m=+1128.140183948" Oct 08 18:35:58 crc kubenswrapper[4859]: I1008 18:35:58.810952 4859 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-metadata-0"] Oct 08 18:35:58 crc kubenswrapper[4859]: I1008 18:35:58.820388 4859 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell1-novncproxy-0"] Oct 08 18:35:58 crc kubenswrapper[4859]: I1008 18:35:58.872717 4859 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-694b6b9bcc-7zv88" Oct 08 18:35:59 crc kubenswrapper[4859]: I1008 18:35:59.891900 4859 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"8a8d7e8d-fea1-464e-b94b-e954ae1ee475","Type":"ContainerStarted","Data":"b6f9d87112cbd0d7a29f9e5c76a2db2677634ef786005d6541684a89d50bb62f"} Oct 08 18:35:59 crc kubenswrapper[4859]: I1008 18:35:59.896812 4859 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"2f61fa1e-ec4e-490e-91ef-90cf0887940d","Type":"ContainerStarted","Data":"fc645cdbe451b90429e2860adc57240bae13a8ad243741c1a12efdef93de92db"} Oct 08 18:35:59 crc kubenswrapper[4859]: I1008 18:35:59.899520 4859 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"0178104e-34da-4178-b6e0-873404e282ef","Type":"ContainerStarted","Data":"6060d204fee906aa340b4ed067014067f6a228091114803a60e21e70b3e3e192"} Oct 08 18:35:59 crc kubenswrapper[4859]: I1008 18:35:59.907751 4859 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-novncproxy-0" event={"ID":"41c36921-eda9-432d-9d21-b478e4e49ee1","Type":"ContainerStarted","Data":"e3197be4f4609dd18ce0c5f1b6a59c8574f21d7c1f4c6c037c5790afe9a61493"} Oct 08 18:35:59 crc kubenswrapper[4859]: I1008 18:35:59.907819 4859 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-cell1-novncproxy-0" podUID="41c36921-eda9-432d-9d21-b478e4e49ee1" containerName="nova-cell1-novncproxy-novncproxy" containerID="cri-o://e3197be4f4609dd18ce0c5f1b6a59c8574f21d7c1f4c6c037c5790afe9a61493" gracePeriod=30 Oct 08 18:35:59 crc kubenswrapper[4859]: I1008 18:35:59.922511 4859 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-scheduler-0" podStartSLOduration=2.125677033 podStartE2EDuration="4.922493438s" podCreationTimestamp="2025-10-08 18:35:55 +0000 UTC" firstStartedPulling="2025-10-08 18:35:56.401668626 +0000 UTC m=+1126.648508005" lastFinishedPulling="2025-10-08 18:35:59.198485031 +0000 UTC m=+1129.445324410" observedRunningTime="2025-10-08 18:35:59.915732645 +0000 UTC m=+1130.162572024" watchObservedRunningTime="2025-10-08 18:35:59.922493438 +0000 UTC m=+1130.169332817" Oct 08 18:35:59 crc kubenswrapper[4859]: I1008 18:35:59.939329 4859 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-cell1-novncproxy-0" podStartSLOduration=1.972898636 podStartE2EDuration="4.939305399s" podCreationTimestamp="2025-10-08 18:35:55 +0000 UTC" firstStartedPulling="2025-10-08 18:35:56.231417909 +0000 UTC m=+1126.478257288" lastFinishedPulling="2025-10-08 18:35:59.197824672 +0000 UTC m=+1129.444664051" observedRunningTime="2025-10-08 18:35:59.9334026 +0000 UTC m=+1130.180241989" watchObservedRunningTime="2025-10-08 18:35:59.939305399 +0000 UTC m=+1130.186144778" Oct 08 18:36:00 crc kubenswrapper[4859]: I1008 18:36:00.588782 4859 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-cell1-novncproxy-0" Oct 08 18:36:00 crc kubenswrapper[4859]: I1008 18:36:00.637942 4859 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-scheduler-0" Oct 08 18:36:00 crc kubenswrapper[4859]: I1008 18:36:00.918670 4859 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"8a8d7e8d-fea1-464e-b94b-e954ae1ee475","Type":"ContainerStarted","Data":"638df716932f73fd84e20153f3078d4a3c60f9f14faec6920c1a93c3de794dfb"} Oct 08 18:36:00 crc kubenswrapper[4859]: I1008 18:36:00.918783 4859 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-metadata-0" podUID="8a8d7e8d-fea1-464e-b94b-e954ae1ee475" containerName="nova-metadata-log" containerID="cri-o://b6f9d87112cbd0d7a29f9e5c76a2db2677634ef786005d6541684a89d50bb62f" gracePeriod=30 Oct 08 18:36:00 crc kubenswrapper[4859]: I1008 18:36:00.918856 4859 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-metadata-0" podUID="8a8d7e8d-fea1-464e-b94b-e954ae1ee475" containerName="nova-metadata-metadata" containerID="cri-o://638df716932f73fd84e20153f3078d4a3c60f9f14faec6920c1a93c3de794dfb" gracePeriod=30 Oct 08 18:36:00 crc kubenswrapper[4859]: I1008 18:36:00.923238 4859 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"2f61fa1e-ec4e-490e-91ef-90cf0887940d","Type":"ContainerStarted","Data":"41e75dcb065fad2bed5125c9c642da1af0bf7ef04626d2bc28bf09c54539e2fa"} Oct 08 18:36:00 crc kubenswrapper[4859]: I1008 18:36:00.947168 4859 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-metadata-0" podStartSLOduration=3.831485491 podStartE2EDuration="6.94715276s" podCreationTimestamp="2025-10-08 18:35:54 +0000 UTC" firstStartedPulling="2025-10-08 18:35:56.087559867 +0000 UTC m=+1126.334399246" lastFinishedPulling="2025-10-08 18:35:59.203227126 +0000 UTC m=+1129.450066515" observedRunningTime="2025-10-08 18:36:00.938872394 +0000 UTC m=+1131.185711773" watchObservedRunningTime="2025-10-08 18:36:00.94715276 +0000 UTC m=+1131.193992139" Oct 08 18:36:00 crc kubenswrapper[4859]: I1008 18:36:00.965073 4859 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-api-0" podStartSLOduration=3.723932276 podStartE2EDuration="6.965053562s" podCreationTimestamp="2025-10-08 18:35:54 +0000 UTC" firstStartedPulling="2025-10-08 18:35:55.958192788 +0000 UTC m=+1126.205032167" lastFinishedPulling="2025-10-08 18:35:59.199314064 +0000 UTC m=+1129.446153453" observedRunningTime="2025-10-08 18:36:00.958924547 +0000 UTC m=+1131.205763926" watchObservedRunningTime="2025-10-08 18:36:00.965053562 +0000 UTC m=+1131.211892941" Oct 08 18:36:01 crc kubenswrapper[4859]: I1008 18:36:01.549515 4859 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Oct 08 18:36:01 crc kubenswrapper[4859]: I1008 18:36:01.672017 4859 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-6xlmm\" (UniqueName: \"kubernetes.io/projected/8a8d7e8d-fea1-464e-b94b-e954ae1ee475-kube-api-access-6xlmm\") pod \"8a8d7e8d-fea1-464e-b94b-e954ae1ee475\" (UID: \"8a8d7e8d-fea1-464e-b94b-e954ae1ee475\") " Oct 08 18:36:01 crc kubenswrapper[4859]: I1008 18:36:01.672104 4859 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8a8d7e8d-fea1-464e-b94b-e954ae1ee475-combined-ca-bundle\") pod \"8a8d7e8d-fea1-464e-b94b-e954ae1ee475\" (UID: \"8a8d7e8d-fea1-464e-b94b-e954ae1ee475\") " Oct 08 18:36:01 crc kubenswrapper[4859]: I1008 18:36:01.672161 4859 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/8a8d7e8d-fea1-464e-b94b-e954ae1ee475-config-data\") pod \"8a8d7e8d-fea1-464e-b94b-e954ae1ee475\" (UID: \"8a8d7e8d-fea1-464e-b94b-e954ae1ee475\") " Oct 08 18:36:01 crc kubenswrapper[4859]: I1008 18:36:01.672266 4859 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/8a8d7e8d-fea1-464e-b94b-e954ae1ee475-logs\") pod \"8a8d7e8d-fea1-464e-b94b-e954ae1ee475\" (UID: \"8a8d7e8d-fea1-464e-b94b-e954ae1ee475\") " Oct 08 18:36:01 crc kubenswrapper[4859]: I1008 18:36:01.672606 4859 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/8a8d7e8d-fea1-464e-b94b-e954ae1ee475-logs" (OuterVolumeSpecName: "logs") pod "8a8d7e8d-fea1-464e-b94b-e954ae1ee475" (UID: "8a8d7e8d-fea1-464e-b94b-e954ae1ee475"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 08 18:36:01 crc kubenswrapper[4859]: I1008 18:36:01.678678 4859 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8a8d7e8d-fea1-464e-b94b-e954ae1ee475-kube-api-access-6xlmm" (OuterVolumeSpecName: "kube-api-access-6xlmm") pod "8a8d7e8d-fea1-464e-b94b-e954ae1ee475" (UID: "8a8d7e8d-fea1-464e-b94b-e954ae1ee475"). InnerVolumeSpecName "kube-api-access-6xlmm". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 08 18:36:01 crc kubenswrapper[4859]: I1008 18:36:01.701217 4859 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8a8d7e8d-fea1-464e-b94b-e954ae1ee475-config-data" (OuterVolumeSpecName: "config-data") pod "8a8d7e8d-fea1-464e-b94b-e954ae1ee475" (UID: "8a8d7e8d-fea1-464e-b94b-e954ae1ee475"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 08 18:36:01 crc kubenswrapper[4859]: I1008 18:36:01.704870 4859 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8a8d7e8d-fea1-464e-b94b-e954ae1ee475-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "8a8d7e8d-fea1-464e-b94b-e954ae1ee475" (UID: "8a8d7e8d-fea1-464e-b94b-e954ae1ee475"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 08 18:36:01 crc kubenswrapper[4859]: I1008 18:36:01.774862 4859 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-6xlmm\" (UniqueName: \"kubernetes.io/projected/8a8d7e8d-fea1-464e-b94b-e954ae1ee475-kube-api-access-6xlmm\") on node \"crc\" DevicePath \"\"" Oct 08 18:36:01 crc kubenswrapper[4859]: I1008 18:36:01.774914 4859 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8a8d7e8d-fea1-464e-b94b-e954ae1ee475-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 08 18:36:01 crc kubenswrapper[4859]: I1008 18:36:01.774928 4859 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/8a8d7e8d-fea1-464e-b94b-e954ae1ee475-config-data\") on node \"crc\" DevicePath \"\"" Oct 08 18:36:01 crc kubenswrapper[4859]: I1008 18:36:01.774936 4859 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/8a8d7e8d-fea1-464e-b94b-e954ae1ee475-logs\") on node \"crc\" DevicePath \"\"" Oct 08 18:36:01 crc kubenswrapper[4859]: I1008 18:36:01.934976 4859 generic.go:334] "Generic (PLEG): container finished" podID="8a8d7e8d-fea1-464e-b94b-e954ae1ee475" containerID="638df716932f73fd84e20153f3078d4a3c60f9f14faec6920c1a93c3de794dfb" exitCode=0 Oct 08 18:36:01 crc kubenswrapper[4859]: I1008 18:36:01.935036 4859 generic.go:334] "Generic (PLEG): container finished" podID="8a8d7e8d-fea1-464e-b94b-e954ae1ee475" containerID="b6f9d87112cbd0d7a29f9e5c76a2db2677634ef786005d6541684a89d50bb62f" exitCode=143 Oct 08 18:36:01 crc kubenswrapper[4859]: I1008 18:36:01.939854 4859 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Oct 08 18:36:01 crc kubenswrapper[4859]: I1008 18:36:01.939946 4859 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"8a8d7e8d-fea1-464e-b94b-e954ae1ee475","Type":"ContainerDied","Data":"638df716932f73fd84e20153f3078d4a3c60f9f14faec6920c1a93c3de794dfb"} Oct 08 18:36:01 crc kubenswrapper[4859]: I1008 18:36:01.940024 4859 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"8a8d7e8d-fea1-464e-b94b-e954ae1ee475","Type":"ContainerDied","Data":"b6f9d87112cbd0d7a29f9e5c76a2db2677634ef786005d6541684a89d50bb62f"} Oct 08 18:36:01 crc kubenswrapper[4859]: I1008 18:36:01.940038 4859 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"8a8d7e8d-fea1-464e-b94b-e954ae1ee475","Type":"ContainerDied","Data":"59037ecb0d29a74aae0a85b65c900d4ee5e7dc0f9be715b0da96b423abd52be0"} Oct 08 18:36:01 crc kubenswrapper[4859]: I1008 18:36:01.940080 4859 scope.go:117] "RemoveContainer" containerID="638df716932f73fd84e20153f3078d4a3c60f9f14faec6920c1a93c3de794dfb" Oct 08 18:36:01 crc kubenswrapper[4859]: I1008 18:36:01.981111 4859 scope.go:117] "RemoveContainer" containerID="b6f9d87112cbd0d7a29f9e5c76a2db2677634ef786005d6541684a89d50bb62f" Oct 08 18:36:02 crc kubenswrapper[4859]: I1008 18:36:02.023967 4859 scope.go:117] "RemoveContainer" containerID="638df716932f73fd84e20153f3078d4a3c60f9f14faec6920c1a93c3de794dfb" Oct 08 18:36:02 crc kubenswrapper[4859]: E1008 18:36:02.024655 4859 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"638df716932f73fd84e20153f3078d4a3c60f9f14faec6920c1a93c3de794dfb\": container with ID starting with 638df716932f73fd84e20153f3078d4a3c60f9f14faec6920c1a93c3de794dfb not found: ID does not exist" containerID="638df716932f73fd84e20153f3078d4a3c60f9f14faec6920c1a93c3de794dfb" Oct 08 18:36:02 crc kubenswrapper[4859]: I1008 18:36:02.024791 4859 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"638df716932f73fd84e20153f3078d4a3c60f9f14faec6920c1a93c3de794dfb"} err="failed to get container status \"638df716932f73fd84e20153f3078d4a3c60f9f14faec6920c1a93c3de794dfb\": rpc error: code = NotFound desc = could not find container \"638df716932f73fd84e20153f3078d4a3c60f9f14faec6920c1a93c3de794dfb\": container with ID starting with 638df716932f73fd84e20153f3078d4a3c60f9f14faec6920c1a93c3de794dfb not found: ID does not exist" Oct 08 18:36:02 crc kubenswrapper[4859]: I1008 18:36:02.024827 4859 scope.go:117] "RemoveContainer" containerID="b6f9d87112cbd0d7a29f9e5c76a2db2677634ef786005d6541684a89d50bb62f" Oct 08 18:36:02 crc kubenswrapper[4859]: E1008 18:36:02.025244 4859 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"b6f9d87112cbd0d7a29f9e5c76a2db2677634ef786005d6541684a89d50bb62f\": container with ID starting with b6f9d87112cbd0d7a29f9e5c76a2db2677634ef786005d6541684a89d50bb62f not found: ID does not exist" containerID="b6f9d87112cbd0d7a29f9e5c76a2db2677634ef786005d6541684a89d50bb62f" Oct 08 18:36:02 crc kubenswrapper[4859]: I1008 18:36:02.025268 4859 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"b6f9d87112cbd0d7a29f9e5c76a2db2677634ef786005d6541684a89d50bb62f"} err="failed to get container status \"b6f9d87112cbd0d7a29f9e5c76a2db2677634ef786005d6541684a89d50bb62f\": rpc error: code = NotFound desc = could not find container \"b6f9d87112cbd0d7a29f9e5c76a2db2677634ef786005d6541684a89d50bb62f\": container with ID starting with b6f9d87112cbd0d7a29f9e5c76a2db2677634ef786005d6541684a89d50bb62f not found: ID does not exist" Oct 08 18:36:02 crc kubenswrapper[4859]: I1008 18:36:02.025282 4859 scope.go:117] "RemoveContainer" containerID="638df716932f73fd84e20153f3078d4a3c60f9f14faec6920c1a93c3de794dfb" Oct 08 18:36:02 crc kubenswrapper[4859]: I1008 18:36:02.025516 4859 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"638df716932f73fd84e20153f3078d4a3c60f9f14faec6920c1a93c3de794dfb"} err="failed to get container status \"638df716932f73fd84e20153f3078d4a3c60f9f14faec6920c1a93c3de794dfb\": rpc error: code = NotFound desc = could not find container \"638df716932f73fd84e20153f3078d4a3c60f9f14faec6920c1a93c3de794dfb\": container with ID starting with 638df716932f73fd84e20153f3078d4a3c60f9f14faec6920c1a93c3de794dfb not found: ID does not exist" Oct 08 18:36:02 crc kubenswrapper[4859]: I1008 18:36:02.025533 4859 scope.go:117] "RemoveContainer" containerID="b6f9d87112cbd0d7a29f9e5c76a2db2677634ef786005d6541684a89d50bb62f" Oct 08 18:36:02 crc kubenswrapper[4859]: I1008 18:36:02.025795 4859 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"b6f9d87112cbd0d7a29f9e5c76a2db2677634ef786005d6541684a89d50bb62f"} err="failed to get container status \"b6f9d87112cbd0d7a29f9e5c76a2db2677634ef786005d6541684a89d50bb62f\": rpc error: code = NotFound desc = could not find container \"b6f9d87112cbd0d7a29f9e5c76a2db2677634ef786005d6541684a89d50bb62f\": container with ID starting with b6f9d87112cbd0d7a29f9e5c76a2db2677634ef786005d6541684a89d50bb62f not found: ID does not exist" Oct 08 18:36:02 crc kubenswrapper[4859]: I1008 18:36:02.028420 4859 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-metadata-0"] Oct 08 18:36:02 crc kubenswrapper[4859]: I1008 18:36:02.040559 4859 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-metadata-0"] Oct 08 18:36:02 crc kubenswrapper[4859]: I1008 18:36:02.057055 4859 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-metadata-0"] Oct 08 18:36:02 crc kubenswrapper[4859]: E1008 18:36:02.057836 4859 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8a8d7e8d-fea1-464e-b94b-e954ae1ee475" containerName="nova-metadata-log" Oct 08 18:36:02 crc kubenswrapper[4859]: I1008 18:36:02.057956 4859 state_mem.go:107] "Deleted CPUSet assignment" podUID="8a8d7e8d-fea1-464e-b94b-e954ae1ee475" containerName="nova-metadata-log" Oct 08 18:36:02 crc kubenswrapper[4859]: E1008 18:36:02.058044 4859 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8a8d7e8d-fea1-464e-b94b-e954ae1ee475" containerName="nova-metadata-metadata" Oct 08 18:36:02 crc kubenswrapper[4859]: I1008 18:36:02.058103 4859 state_mem.go:107] "Deleted CPUSet assignment" podUID="8a8d7e8d-fea1-464e-b94b-e954ae1ee475" containerName="nova-metadata-metadata" Oct 08 18:36:02 crc kubenswrapper[4859]: I1008 18:36:02.058395 4859 memory_manager.go:354] "RemoveStaleState removing state" podUID="8a8d7e8d-fea1-464e-b94b-e954ae1ee475" containerName="nova-metadata-metadata" Oct 08 18:36:02 crc kubenswrapper[4859]: I1008 18:36:02.058596 4859 memory_manager.go:354] "RemoveStaleState removing state" podUID="8a8d7e8d-fea1-464e-b94b-e954ae1ee475" containerName="nova-metadata-log" Oct 08 18:36:02 crc kubenswrapper[4859]: I1008 18:36:02.059717 4859 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Oct 08 18:36:02 crc kubenswrapper[4859]: I1008 18:36:02.064844 4859 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-nova-metadata-internal-svc" Oct 08 18:36:02 crc kubenswrapper[4859]: I1008 18:36:02.065302 4859 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-metadata-config-data" Oct 08 18:36:02 crc kubenswrapper[4859]: I1008 18:36:02.066482 4859 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-metadata-0"] Oct 08 18:36:02 crc kubenswrapper[4859]: I1008 18:36:02.190463 4859 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-6lzfp\" (UniqueName: \"kubernetes.io/projected/fd151d4e-0d70-4458-8ab8-ffabed5dddca-kube-api-access-6lzfp\") pod \"nova-metadata-0\" (UID: \"fd151d4e-0d70-4458-8ab8-ffabed5dddca\") " pod="openstack/nova-metadata-0" Oct 08 18:36:02 crc kubenswrapper[4859]: I1008 18:36:02.190867 4859 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/fd151d4e-0d70-4458-8ab8-ffabed5dddca-nova-metadata-tls-certs\") pod \"nova-metadata-0\" (UID: \"fd151d4e-0d70-4458-8ab8-ffabed5dddca\") " pod="openstack/nova-metadata-0" Oct 08 18:36:02 crc kubenswrapper[4859]: I1008 18:36:02.191133 4859 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/fd151d4e-0d70-4458-8ab8-ffabed5dddca-config-data\") pod \"nova-metadata-0\" (UID: \"fd151d4e-0d70-4458-8ab8-ffabed5dddca\") " pod="openstack/nova-metadata-0" Oct 08 18:36:02 crc kubenswrapper[4859]: I1008 18:36:02.191222 4859 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/fd151d4e-0d70-4458-8ab8-ffabed5dddca-logs\") pod \"nova-metadata-0\" (UID: \"fd151d4e-0d70-4458-8ab8-ffabed5dddca\") " pod="openstack/nova-metadata-0" Oct 08 18:36:02 crc kubenswrapper[4859]: I1008 18:36:02.191305 4859 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/fd151d4e-0d70-4458-8ab8-ffabed5dddca-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"fd151d4e-0d70-4458-8ab8-ffabed5dddca\") " pod="openstack/nova-metadata-0" Oct 08 18:36:02 crc kubenswrapper[4859]: I1008 18:36:02.293641 4859 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/fd151d4e-0d70-4458-8ab8-ffabed5dddca-config-data\") pod \"nova-metadata-0\" (UID: \"fd151d4e-0d70-4458-8ab8-ffabed5dddca\") " pod="openstack/nova-metadata-0" Oct 08 18:36:02 crc kubenswrapper[4859]: I1008 18:36:02.293760 4859 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/fd151d4e-0d70-4458-8ab8-ffabed5dddca-logs\") pod \"nova-metadata-0\" (UID: \"fd151d4e-0d70-4458-8ab8-ffabed5dddca\") " pod="openstack/nova-metadata-0" Oct 08 18:36:02 crc kubenswrapper[4859]: I1008 18:36:02.293838 4859 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/fd151d4e-0d70-4458-8ab8-ffabed5dddca-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"fd151d4e-0d70-4458-8ab8-ffabed5dddca\") " pod="openstack/nova-metadata-0" Oct 08 18:36:02 crc kubenswrapper[4859]: I1008 18:36:02.293933 4859 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-6lzfp\" (UniqueName: \"kubernetes.io/projected/fd151d4e-0d70-4458-8ab8-ffabed5dddca-kube-api-access-6lzfp\") pod \"nova-metadata-0\" (UID: \"fd151d4e-0d70-4458-8ab8-ffabed5dddca\") " pod="openstack/nova-metadata-0" Oct 08 18:36:02 crc kubenswrapper[4859]: I1008 18:36:02.293972 4859 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/fd151d4e-0d70-4458-8ab8-ffabed5dddca-nova-metadata-tls-certs\") pod \"nova-metadata-0\" (UID: \"fd151d4e-0d70-4458-8ab8-ffabed5dddca\") " pod="openstack/nova-metadata-0" Oct 08 18:36:02 crc kubenswrapper[4859]: I1008 18:36:02.295841 4859 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/fd151d4e-0d70-4458-8ab8-ffabed5dddca-logs\") pod \"nova-metadata-0\" (UID: \"fd151d4e-0d70-4458-8ab8-ffabed5dddca\") " pod="openstack/nova-metadata-0" Oct 08 18:36:02 crc kubenswrapper[4859]: I1008 18:36:02.299001 4859 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/fd151d4e-0d70-4458-8ab8-ffabed5dddca-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"fd151d4e-0d70-4458-8ab8-ffabed5dddca\") " pod="openstack/nova-metadata-0" Oct 08 18:36:02 crc kubenswrapper[4859]: I1008 18:36:02.299282 4859 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/fd151d4e-0d70-4458-8ab8-ffabed5dddca-config-data\") pod \"nova-metadata-0\" (UID: \"fd151d4e-0d70-4458-8ab8-ffabed5dddca\") " pod="openstack/nova-metadata-0" Oct 08 18:36:02 crc kubenswrapper[4859]: I1008 18:36:02.299515 4859 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/fd151d4e-0d70-4458-8ab8-ffabed5dddca-nova-metadata-tls-certs\") pod \"nova-metadata-0\" (UID: \"fd151d4e-0d70-4458-8ab8-ffabed5dddca\") " pod="openstack/nova-metadata-0" Oct 08 18:36:02 crc kubenswrapper[4859]: I1008 18:36:02.329076 4859 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-6lzfp\" (UniqueName: \"kubernetes.io/projected/fd151d4e-0d70-4458-8ab8-ffabed5dddca-kube-api-access-6lzfp\") pod \"nova-metadata-0\" (UID: \"fd151d4e-0d70-4458-8ab8-ffabed5dddca\") " pod="openstack/nova-metadata-0" Oct 08 18:36:02 crc kubenswrapper[4859]: I1008 18:36:02.380234 4859 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Oct 08 18:36:02 crc kubenswrapper[4859]: I1008 18:36:02.483615 4859 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="8a8d7e8d-fea1-464e-b94b-e954ae1ee475" path="/var/lib/kubelet/pods/8a8d7e8d-fea1-464e-b94b-e954ae1ee475/volumes" Oct 08 18:36:02 crc kubenswrapper[4859]: I1008 18:36:02.866798 4859 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-metadata-0"] Oct 08 18:36:02 crc kubenswrapper[4859]: I1008 18:36:02.946907 4859 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"fd151d4e-0d70-4458-8ab8-ffabed5dddca","Type":"ContainerStarted","Data":"ee47a246ebdff3be8ab7bdc0fd8ec71b5b70bf97a4f6503092cee86a44f7617d"} Oct 08 18:36:03 crc kubenswrapper[4859]: I1008 18:36:03.958348 4859 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"fd151d4e-0d70-4458-8ab8-ffabed5dddca","Type":"ContainerStarted","Data":"d2a2180ef27e62bc849417f71f26eec95810f30e12d337d28d89a70a6f32a434"} Oct 08 18:36:03 crc kubenswrapper[4859]: I1008 18:36:03.958673 4859 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"fd151d4e-0d70-4458-8ab8-ffabed5dddca","Type":"ContainerStarted","Data":"a1677fbe2b0ccc470b915d38b90540074c1f5035d3224a47de80d2518794d673"} Oct 08 18:36:03 crc kubenswrapper[4859]: I1008 18:36:03.977111 4859 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-metadata-0" podStartSLOduration=1.977089629 podStartE2EDuration="1.977089629s" podCreationTimestamp="2025-10-08 18:36:02 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-08 18:36:03.974947998 +0000 UTC m=+1134.221787387" watchObservedRunningTime="2025-10-08 18:36:03.977089629 +0000 UTC m=+1134.223929018" Oct 08 18:36:04 crc kubenswrapper[4859]: I1008 18:36:04.972413 4859 generic.go:334] "Generic (PLEG): container finished" podID="76588317-83a8-42b4-8a4f-502d6889c7b5" containerID="38db145a1acab79ead1902ab5f38916c3ba433b24d3074e51ed13550d67ec7b4" exitCode=0 Oct 08 18:36:04 crc kubenswrapper[4859]: I1008 18:36:04.972539 4859 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-cell-mapping-r2mmw" event={"ID":"76588317-83a8-42b4-8a4f-502d6889c7b5","Type":"ContainerDied","Data":"38db145a1acab79ead1902ab5f38916c3ba433b24d3074e51ed13550d67ec7b4"} Oct 08 18:36:05 crc kubenswrapper[4859]: I1008 18:36:05.337171 4859 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-api-0" Oct 08 18:36:05 crc kubenswrapper[4859]: I1008 18:36:05.337226 4859 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-api-0" Oct 08 18:36:05 crc kubenswrapper[4859]: I1008 18:36:05.619844 4859 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-694b6b9bcc-7zv88" Oct 08 18:36:05 crc kubenswrapper[4859]: I1008 18:36:05.638742 4859 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-scheduler-0" Oct 08 18:36:05 crc kubenswrapper[4859]: I1008 18:36:05.702590 4859 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-scheduler-0" Oct 08 18:36:05 crc kubenswrapper[4859]: I1008 18:36:05.730382 4859 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-5f74b5f5cc-tl7pk"] Oct 08 18:36:05 crc kubenswrapper[4859]: I1008 18:36:05.730745 4859 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-5f74b5f5cc-tl7pk" podUID="a03169fb-e503-427f-bafe-2f8767a23482" containerName="dnsmasq-dns" containerID="cri-o://a17c560ce5bf3618ca0c82f026671dacb150ab5c919db4a2fb2daccfe125f76e" gracePeriod=10 Oct 08 18:36:05 crc kubenswrapper[4859]: I1008 18:36:05.992154 4859 generic.go:334] "Generic (PLEG): container finished" podID="6b0220ac-1dc0-4ca9-a9ae-9335b57820fa" containerID="6fb9ae99d0df9addcb1423e79ad8723c8a3d60ad81028ad521e0f4c0280241b7" exitCode=0 Oct 08 18:36:05 crc kubenswrapper[4859]: I1008 18:36:05.992496 4859 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-conductor-db-sync-28hwn" event={"ID":"6b0220ac-1dc0-4ca9-a9ae-9335b57820fa","Type":"ContainerDied","Data":"6fb9ae99d0df9addcb1423e79ad8723c8a3d60ad81028ad521e0f4c0280241b7"} Oct 08 18:36:06 crc kubenswrapper[4859]: I1008 18:36:06.000531 4859 generic.go:334] "Generic (PLEG): container finished" podID="a03169fb-e503-427f-bafe-2f8767a23482" containerID="a17c560ce5bf3618ca0c82f026671dacb150ab5c919db4a2fb2daccfe125f76e" exitCode=0 Oct 08 18:36:06 crc kubenswrapper[4859]: I1008 18:36:06.000838 4859 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-5f74b5f5cc-tl7pk" event={"ID":"a03169fb-e503-427f-bafe-2f8767a23482","Type":"ContainerDied","Data":"a17c560ce5bf3618ca0c82f026671dacb150ab5c919db4a2fb2daccfe125f76e"} Oct 08 18:36:06 crc kubenswrapper[4859]: I1008 18:36:06.052104 4859 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-scheduler-0" Oct 08 18:36:06 crc kubenswrapper[4859]: I1008 18:36:06.373089 4859 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-5f74b5f5cc-tl7pk" Oct 08 18:36:06 crc kubenswrapper[4859]: I1008 18:36:06.424182 4859 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-api-0" podUID="2f61fa1e-ec4e-490e-91ef-90cf0887940d" containerName="nova-api-log" probeResult="failure" output="Get \"http://10.217.0.186:8774/\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Oct 08 18:36:06 crc kubenswrapper[4859]: I1008 18:36:06.424450 4859 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-api-0" podUID="2f61fa1e-ec4e-490e-91ef-90cf0887940d" containerName="nova-api-api" probeResult="failure" output="Get \"http://10.217.0.186:8774/\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Oct 08 18:36:06 crc kubenswrapper[4859]: I1008 18:36:06.510371 4859 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/a03169fb-e503-427f-bafe-2f8767a23482-ovsdbserver-sb\") pod \"a03169fb-e503-427f-bafe-2f8767a23482\" (UID: \"a03169fb-e503-427f-bafe-2f8767a23482\") " Oct 08 18:36:06 crc kubenswrapper[4859]: I1008 18:36:06.510423 4859 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/a03169fb-e503-427f-bafe-2f8767a23482-dns-swift-storage-0\") pod \"a03169fb-e503-427f-bafe-2f8767a23482\" (UID: \"a03169fb-e503-427f-bafe-2f8767a23482\") " Oct 08 18:36:06 crc kubenswrapper[4859]: I1008 18:36:06.510559 4859 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-77d9b\" (UniqueName: \"kubernetes.io/projected/a03169fb-e503-427f-bafe-2f8767a23482-kube-api-access-77d9b\") pod \"a03169fb-e503-427f-bafe-2f8767a23482\" (UID: \"a03169fb-e503-427f-bafe-2f8767a23482\") " Oct 08 18:36:06 crc kubenswrapper[4859]: I1008 18:36:06.510641 4859 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/a03169fb-e503-427f-bafe-2f8767a23482-config\") pod \"a03169fb-e503-427f-bafe-2f8767a23482\" (UID: \"a03169fb-e503-427f-bafe-2f8767a23482\") " Oct 08 18:36:06 crc kubenswrapper[4859]: I1008 18:36:06.510714 4859 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/a03169fb-e503-427f-bafe-2f8767a23482-dns-svc\") pod \"a03169fb-e503-427f-bafe-2f8767a23482\" (UID: \"a03169fb-e503-427f-bafe-2f8767a23482\") " Oct 08 18:36:06 crc kubenswrapper[4859]: I1008 18:36:06.510758 4859 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/a03169fb-e503-427f-bafe-2f8767a23482-ovsdbserver-nb\") pod \"a03169fb-e503-427f-bafe-2f8767a23482\" (UID: \"a03169fb-e503-427f-bafe-2f8767a23482\") " Oct 08 18:36:06 crc kubenswrapper[4859]: I1008 18:36:06.517897 4859 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a03169fb-e503-427f-bafe-2f8767a23482-kube-api-access-77d9b" (OuterVolumeSpecName: "kube-api-access-77d9b") pod "a03169fb-e503-427f-bafe-2f8767a23482" (UID: "a03169fb-e503-427f-bafe-2f8767a23482"). InnerVolumeSpecName "kube-api-access-77d9b". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 08 18:36:06 crc kubenswrapper[4859]: I1008 18:36:06.550558 4859 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-cell-mapping-r2mmw" Oct 08 18:36:06 crc kubenswrapper[4859]: I1008 18:36:06.603624 4859 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/a03169fb-e503-427f-bafe-2f8767a23482-dns-swift-storage-0" (OuterVolumeSpecName: "dns-swift-storage-0") pod "a03169fb-e503-427f-bafe-2f8767a23482" (UID: "a03169fb-e503-427f-bafe-2f8767a23482"). InnerVolumeSpecName "dns-swift-storage-0". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 08 18:36:06 crc kubenswrapper[4859]: I1008 18:36:06.609153 4859 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/a03169fb-e503-427f-bafe-2f8767a23482-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "a03169fb-e503-427f-bafe-2f8767a23482" (UID: "a03169fb-e503-427f-bafe-2f8767a23482"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 08 18:36:06 crc kubenswrapper[4859]: I1008 18:36:06.610828 4859 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/a03169fb-e503-427f-bafe-2f8767a23482-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "a03169fb-e503-427f-bafe-2f8767a23482" (UID: "a03169fb-e503-427f-bafe-2f8767a23482"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 08 18:36:06 crc kubenswrapper[4859]: I1008 18:36:06.612063 4859 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/76588317-83a8-42b4-8a4f-502d6889c7b5-combined-ca-bundle\") pod \"76588317-83a8-42b4-8a4f-502d6889c7b5\" (UID: \"76588317-83a8-42b4-8a4f-502d6889c7b5\") " Oct 08 18:36:06 crc kubenswrapper[4859]: I1008 18:36:06.612122 4859 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/76588317-83a8-42b4-8a4f-502d6889c7b5-scripts\") pod \"76588317-83a8-42b4-8a4f-502d6889c7b5\" (UID: \"76588317-83a8-42b4-8a4f-502d6889c7b5\") " Oct 08 18:36:06 crc kubenswrapper[4859]: I1008 18:36:06.612362 4859 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-5h7hc\" (UniqueName: \"kubernetes.io/projected/76588317-83a8-42b4-8a4f-502d6889c7b5-kube-api-access-5h7hc\") pod \"76588317-83a8-42b4-8a4f-502d6889c7b5\" (UID: \"76588317-83a8-42b4-8a4f-502d6889c7b5\") " Oct 08 18:36:06 crc kubenswrapper[4859]: I1008 18:36:06.612791 4859 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/76588317-83a8-42b4-8a4f-502d6889c7b5-config-data\") pod \"76588317-83a8-42b4-8a4f-502d6889c7b5\" (UID: \"76588317-83a8-42b4-8a4f-502d6889c7b5\") " Oct 08 18:36:06 crc kubenswrapper[4859]: I1008 18:36:06.613581 4859 reconciler_common.go:293] "Volume detached for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/a03169fb-e503-427f-bafe-2f8767a23482-dns-swift-storage-0\") on node \"crc\" DevicePath \"\"" Oct 08 18:36:06 crc kubenswrapper[4859]: I1008 18:36:06.613601 4859 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-77d9b\" (UniqueName: \"kubernetes.io/projected/a03169fb-e503-427f-bafe-2f8767a23482-kube-api-access-77d9b\") on node \"crc\" DevicePath \"\"" Oct 08 18:36:06 crc kubenswrapper[4859]: I1008 18:36:06.613615 4859 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/a03169fb-e503-427f-bafe-2f8767a23482-dns-svc\") on node \"crc\" DevicePath \"\"" Oct 08 18:36:06 crc kubenswrapper[4859]: I1008 18:36:06.613626 4859 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/a03169fb-e503-427f-bafe-2f8767a23482-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Oct 08 18:36:06 crc kubenswrapper[4859]: I1008 18:36:06.615393 4859 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/76588317-83a8-42b4-8a4f-502d6889c7b5-scripts" (OuterVolumeSpecName: "scripts") pod "76588317-83a8-42b4-8a4f-502d6889c7b5" (UID: "76588317-83a8-42b4-8a4f-502d6889c7b5"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 08 18:36:06 crc kubenswrapper[4859]: I1008 18:36:06.615595 4859 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/76588317-83a8-42b4-8a4f-502d6889c7b5-kube-api-access-5h7hc" (OuterVolumeSpecName: "kube-api-access-5h7hc") pod "76588317-83a8-42b4-8a4f-502d6889c7b5" (UID: "76588317-83a8-42b4-8a4f-502d6889c7b5"). InnerVolumeSpecName "kube-api-access-5h7hc". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 08 18:36:06 crc kubenswrapper[4859]: I1008 18:36:06.637021 4859 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/a03169fb-e503-427f-bafe-2f8767a23482-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "a03169fb-e503-427f-bafe-2f8767a23482" (UID: "a03169fb-e503-427f-bafe-2f8767a23482"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 08 18:36:06 crc kubenswrapper[4859]: I1008 18:36:06.637468 4859 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/a03169fb-e503-427f-bafe-2f8767a23482-config" (OuterVolumeSpecName: "config") pod "a03169fb-e503-427f-bafe-2f8767a23482" (UID: "a03169fb-e503-427f-bafe-2f8767a23482"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 08 18:36:06 crc kubenswrapper[4859]: I1008 18:36:06.644739 4859 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/76588317-83a8-42b4-8a4f-502d6889c7b5-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "76588317-83a8-42b4-8a4f-502d6889c7b5" (UID: "76588317-83a8-42b4-8a4f-502d6889c7b5"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 08 18:36:06 crc kubenswrapper[4859]: I1008 18:36:06.645801 4859 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/76588317-83a8-42b4-8a4f-502d6889c7b5-config-data" (OuterVolumeSpecName: "config-data") pod "76588317-83a8-42b4-8a4f-502d6889c7b5" (UID: "76588317-83a8-42b4-8a4f-502d6889c7b5"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 08 18:36:06 crc kubenswrapper[4859]: I1008 18:36:06.714840 4859 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/76588317-83a8-42b4-8a4f-502d6889c7b5-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 08 18:36:06 crc kubenswrapper[4859]: I1008 18:36:06.714871 4859 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/76588317-83a8-42b4-8a4f-502d6889c7b5-scripts\") on node \"crc\" DevicePath \"\"" Oct 08 18:36:06 crc kubenswrapper[4859]: I1008 18:36:06.714884 4859 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/a03169fb-e503-427f-bafe-2f8767a23482-config\") on node \"crc\" DevicePath \"\"" Oct 08 18:36:06 crc kubenswrapper[4859]: I1008 18:36:06.714892 4859 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-5h7hc\" (UniqueName: \"kubernetes.io/projected/76588317-83a8-42b4-8a4f-502d6889c7b5-kube-api-access-5h7hc\") on node \"crc\" DevicePath \"\"" Oct 08 18:36:06 crc kubenswrapper[4859]: I1008 18:36:06.714901 4859 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/76588317-83a8-42b4-8a4f-502d6889c7b5-config-data\") on node \"crc\" DevicePath \"\"" Oct 08 18:36:06 crc kubenswrapper[4859]: I1008 18:36:06.714910 4859 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/a03169fb-e503-427f-bafe-2f8767a23482-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Oct 08 18:36:07 crc kubenswrapper[4859]: I1008 18:36:07.016186 4859 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-5f74b5f5cc-tl7pk" event={"ID":"a03169fb-e503-427f-bafe-2f8767a23482","Type":"ContainerDied","Data":"e5814d1fa8bffbeb39ac72b0c31e8f6c976fa8b2b107e84302e99df288c6fc98"} Oct 08 18:36:07 crc kubenswrapper[4859]: I1008 18:36:07.016233 4859 scope.go:117] "RemoveContainer" containerID="a17c560ce5bf3618ca0c82f026671dacb150ab5c919db4a2fb2daccfe125f76e" Oct 08 18:36:07 crc kubenswrapper[4859]: I1008 18:36:07.016339 4859 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-5f74b5f5cc-tl7pk" Oct 08 18:36:07 crc kubenswrapper[4859]: I1008 18:36:07.025591 4859 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-cell-mapping-r2mmw" Oct 08 18:36:07 crc kubenswrapper[4859]: I1008 18:36:07.032244 4859 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-cell-mapping-r2mmw" event={"ID":"76588317-83a8-42b4-8a4f-502d6889c7b5","Type":"ContainerDied","Data":"c64b0e8317dcbdf353293fe34270053e426611653787f744b079bb82f831ad49"} Oct 08 18:36:07 crc kubenswrapper[4859]: I1008 18:36:07.032278 4859 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="c64b0e8317dcbdf353293fe34270053e426611653787f744b079bb82f831ad49" Oct 08 18:36:07 crc kubenswrapper[4859]: I1008 18:36:07.045908 4859 scope.go:117] "RemoveContainer" containerID="d5ffe6df38528c3cd0c30ee428245d2135e36d1f8e40e3dc7bf93bccb5962981" Oct 08 18:36:07 crc kubenswrapper[4859]: I1008 18:36:07.067596 4859 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-5f74b5f5cc-tl7pk"] Oct 08 18:36:07 crc kubenswrapper[4859]: I1008 18:36:07.079559 4859 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-5f74b5f5cc-tl7pk"] Oct 08 18:36:07 crc kubenswrapper[4859]: I1008 18:36:07.183609 4859 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-api-0"] Oct 08 18:36:07 crc kubenswrapper[4859]: I1008 18:36:07.183881 4859 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-api-0" podUID="2f61fa1e-ec4e-490e-91ef-90cf0887940d" containerName="nova-api-log" containerID="cri-o://fc645cdbe451b90429e2860adc57240bae13a8ad243741c1a12efdef93de92db" gracePeriod=30 Oct 08 18:36:07 crc kubenswrapper[4859]: I1008 18:36:07.184305 4859 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-api-0" podUID="2f61fa1e-ec4e-490e-91ef-90cf0887940d" containerName="nova-api-api" containerID="cri-o://41e75dcb065fad2bed5125c9c642da1af0bf7ef04626d2bc28bf09c54539e2fa" gracePeriod=30 Oct 08 18:36:07 crc kubenswrapper[4859]: I1008 18:36:07.191826 4859 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-scheduler-0"] Oct 08 18:36:07 crc kubenswrapper[4859]: I1008 18:36:07.250705 4859 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-metadata-0"] Oct 08 18:36:07 crc kubenswrapper[4859]: I1008 18:36:07.256987 4859 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-metadata-0" podUID="fd151d4e-0d70-4458-8ab8-ffabed5dddca" containerName="nova-metadata-log" containerID="cri-o://a1677fbe2b0ccc470b915d38b90540074c1f5035d3224a47de80d2518794d673" gracePeriod=30 Oct 08 18:36:07 crc kubenswrapper[4859]: I1008 18:36:07.257185 4859 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-metadata-0" podUID="fd151d4e-0d70-4458-8ab8-ffabed5dddca" containerName="nova-metadata-metadata" containerID="cri-o://d2a2180ef27e62bc849417f71f26eec95810f30e12d337d28d89a70a6f32a434" gracePeriod=30 Oct 08 18:36:07 crc kubenswrapper[4859]: I1008 18:36:07.381308 4859 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-metadata-0" Oct 08 18:36:07 crc kubenswrapper[4859]: I1008 18:36:07.381716 4859 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-metadata-0" Oct 08 18:36:07 crc kubenswrapper[4859]: I1008 18:36:07.498225 4859 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-conductor-db-sync-28hwn" Oct 08 18:36:07 crc kubenswrapper[4859]: I1008 18:36:07.630853 4859 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-9vt2x\" (UniqueName: \"kubernetes.io/projected/6b0220ac-1dc0-4ca9-a9ae-9335b57820fa-kube-api-access-9vt2x\") pod \"6b0220ac-1dc0-4ca9-a9ae-9335b57820fa\" (UID: \"6b0220ac-1dc0-4ca9-a9ae-9335b57820fa\") " Oct 08 18:36:07 crc kubenswrapper[4859]: I1008 18:36:07.631012 4859 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6b0220ac-1dc0-4ca9-a9ae-9335b57820fa-combined-ca-bundle\") pod \"6b0220ac-1dc0-4ca9-a9ae-9335b57820fa\" (UID: \"6b0220ac-1dc0-4ca9-a9ae-9335b57820fa\") " Oct 08 18:36:07 crc kubenswrapper[4859]: I1008 18:36:07.631091 4859 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/6b0220ac-1dc0-4ca9-a9ae-9335b57820fa-scripts\") pod \"6b0220ac-1dc0-4ca9-a9ae-9335b57820fa\" (UID: \"6b0220ac-1dc0-4ca9-a9ae-9335b57820fa\") " Oct 08 18:36:07 crc kubenswrapper[4859]: I1008 18:36:07.631166 4859 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/6b0220ac-1dc0-4ca9-a9ae-9335b57820fa-config-data\") pod \"6b0220ac-1dc0-4ca9-a9ae-9335b57820fa\" (UID: \"6b0220ac-1dc0-4ca9-a9ae-9335b57820fa\") " Oct 08 18:36:07 crc kubenswrapper[4859]: I1008 18:36:07.636256 4859 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6b0220ac-1dc0-4ca9-a9ae-9335b57820fa-scripts" (OuterVolumeSpecName: "scripts") pod "6b0220ac-1dc0-4ca9-a9ae-9335b57820fa" (UID: "6b0220ac-1dc0-4ca9-a9ae-9335b57820fa"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 08 18:36:07 crc kubenswrapper[4859]: I1008 18:36:07.636394 4859 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6b0220ac-1dc0-4ca9-a9ae-9335b57820fa-kube-api-access-9vt2x" (OuterVolumeSpecName: "kube-api-access-9vt2x") pod "6b0220ac-1dc0-4ca9-a9ae-9335b57820fa" (UID: "6b0220ac-1dc0-4ca9-a9ae-9335b57820fa"). InnerVolumeSpecName "kube-api-access-9vt2x". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 08 18:36:07 crc kubenswrapper[4859]: I1008 18:36:07.664998 4859 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6b0220ac-1dc0-4ca9-a9ae-9335b57820fa-config-data" (OuterVolumeSpecName: "config-data") pod "6b0220ac-1dc0-4ca9-a9ae-9335b57820fa" (UID: "6b0220ac-1dc0-4ca9-a9ae-9335b57820fa"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 08 18:36:07 crc kubenswrapper[4859]: I1008 18:36:07.690315 4859 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6b0220ac-1dc0-4ca9-a9ae-9335b57820fa-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "6b0220ac-1dc0-4ca9-a9ae-9335b57820fa" (UID: "6b0220ac-1dc0-4ca9-a9ae-9335b57820fa"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 08 18:36:07 crc kubenswrapper[4859]: I1008 18:36:07.739317 4859 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/6b0220ac-1dc0-4ca9-a9ae-9335b57820fa-scripts\") on node \"crc\" DevicePath \"\"" Oct 08 18:36:07 crc kubenswrapper[4859]: I1008 18:36:07.739357 4859 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/6b0220ac-1dc0-4ca9-a9ae-9335b57820fa-config-data\") on node \"crc\" DevicePath \"\"" Oct 08 18:36:07 crc kubenswrapper[4859]: I1008 18:36:07.739370 4859 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-9vt2x\" (UniqueName: \"kubernetes.io/projected/6b0220ac-1dc0-4ca9-a9ae-9335b57820fa-kube-api-access-9vt2x\") on node \"crc\" DevicePath \"\"" Oct 08 18:36:07 crc kubenswrapper[4859]: I1008 18:36:07.739586 4859 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6b0220ac-1dc0-4ca9-a9ae-9335b57820fa-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 08 18:36:07 crc kubenswrapper[4859]: I1008 18:36:07.771763 4859 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Oct 08 18:36:07 crc kubenswrapper[4859]: I1008 18:36:07.840428 4859 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-6lzfp\" (UniqueName: \"kubernetes.io/projected/fd151d4e-0d70-4458-8ab8-ffabed5dddca-kube-api-access-6lzfp\") pod \"fd151d4e-0d70-4458-8ab8-ffabed5dddca\" (UID: \"fd151d4e-0d70-4458-8ab8-ffabed5dddca\") " Oct 08 18:36:07 crc kubenswrapper[4859]: I1008 18:36:07.840508 4859 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/fd151d4e-0d70-4458-8ab8-ffabed5dddca-config-data\") pod \"fd151d4e-0d70-4458-8ab8-ffabed5dddca\" (UID: \"fd151d4e-0d70-4458-8ab8-ffabed5dddca\") " Oct 08 18:36:07 crc kubenswrapper[4859]: I1008 18:36:07.840544 4859 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/fd151d4e-0d70-4458-8ab8-ffabed5dddca-logs\") pod \"fd151d4e-0d70-4458-8ab8-ffabed5dddca\" (UID: \"fd151d4e-0d70-4458-8ab8-ffabed5dddca\") " Oct 08 18:36:07 crc kubenswrapper[4859]: I1008 18:36:07.840577 4859 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/fd151d4e-0d70-4458-8ab8-ffabed5dddca-combined-ca-bundle\") pod \"fd151d4e-0d70-4458-8ab8-ffabed5dddca\" (UID: \"fd151d4e-0d70-4458-8ab8-ffabed5dddca\") " Oct 08 18:36:07 crc kubenswrapper[4859]: I1008 18:36:07.840664 4859 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/fd151d4e-0d70-4458-8ab8-ffabed5dddca-nova-metadata-tls-certs\") pod \"fd151d4e-0d70-4458-8ab8-ffabed5dddca\" (UID: \"fd151d4e-0d70-4458-8ab8-ffabed5dddca\") " Oct 08 18:36:07 crc kubenswrapper[4859]: I1008 18:36:07.840926 4859 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/fd151d4e-0d70-4458-8ab8-ffabed5dddca-logs" (OuterVolumeSpecName: "logs") pod "fd151d4e-0d70-4458-8ab8-ffabed5dddca" (UID: "fd151d4e-0d70-4458-8ab8-ffabed5dddca"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 08 18:36:07 crc kubenswrapper[4859]: I1008 18:36:07.841233 4859 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/fd151d4e-0d70-4458-8ab8-ffabed5dddca-logs\") on node \"crc\" DevicePath \"\"" Oct 08 18:36:07 crc kubenswrapper[4859]: I1008 18:36:07.848375 4859 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/fd151d4e-0d70-4458-8ab8-ffabed5dddca-kube-api-access-6lzfp" (OuterVolumeSpecName: "kube-api-access-6lzfp") pod "fd151d4e-0d70-4458-8ab8-ffabed5dddca" (UID: "fd151d4e-0d70-4458-8ab8-ffabed5dddca"). InnerVolumeSpecName "kube-api-access-6lzfp". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 08 18:36:07 crc kubenswrapper[4859]: I1008 18:36:07.875199 4859 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/fd151d4e-0d70-4458-8ab8-ffabed5dddca-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "fd151d4e-0d70-4458-8ab8-ffabed5dddca" (UID: "fd151d4e-0d70-4458-8ab8-ffabed5dddca"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 08 18:36:07 crc kubenswrapper[4859]: I1008 18:36:07.879397 4859 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/fd151d4e-0d70-4458-8ab8-ffabed5dddca-config-data" (OuterVolumeSpecName: "config-data") pod "fd151d4e-0d70-4458-8ab8-ffabed5dddca" (UID: "fd151d4e-0d70-4458-8ab8-ffabed5dddca"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 08 18:36:07 crc kubenswrapper[4859]: I1008 18:36:07.911566 4859 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/fd151d4e-0d70-4458-8ab8-ffabed5dddca-nova-metadata-tls-certs" (OuterVolumeSpecName: "nova-metadata-tls-certs") pod "fd151d4e-0d70-4458-8ab8-ffabed5dddca" (UID: "fd151d4e-0d70-4458-8ab8-ffabed5dddca"). InnerVolumeSpecName "nova-metadata-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 08 18:36:07 crc kubenswrapper[4859]: I1008 18:36:07.943015 4859 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-6lzfp\" (UniqueName: \"kubernetes.io/projected/fd151d4e-0d70-4458-8ab8-ffabed5dddca-kube-api-access-6lzfp\") on node \"crc\" DevicePath \"\"" Oct 08 18:36:07 crc kubenswrapper[4859]: I1008 18:36:07.943082 4859 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/fd151d4e-0d70-4458-8ab8-ffabed5dddca-config-data\") on node \"crc\" DevicePath \"\"" Oct 08 18:36:07 crc kubenswrapper[4859]: I1008 18:36:07.943095 4859 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/fd151d4e-0d70-4458-8ab8-ffabed5dddca-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 08 18:36:07 crc kubenswrapper[4859]: I1008 18:36:07.943106 4859 reconciler_common.go:293] "Volume detached for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/fd151d4e-0d70-4458-8ab8-ffabed5dddca-nova-metadata-tls-certs\") on node \"crc\" DevicePath \"\"" Oct 08 18:36:08 crc kubenswrapper[4859]: I1008 18:36:08.062790 4859 generic.go:334] "Generic (PLEG): container finished" podID="fd151d4e-0d70-4458-8ab8-ffabed5dddca" containerID="d2a2180ef27e62bc849417f71f26eec95810f30e12d337d28d89a70a6f32a434" exitCode=0 Oct 08 18:36:08 crc kubenswrapper[4859]: I1008 18:36:08.062835 4859 generic.go:334] "Generic (PLEG): container finished" podID="fd151d4e-0d70-4458-8ab8-ffabed5dddca" containerID="a1677fbe2b0ccc470b915d38b90540074c1f5035d3224a47de80d2518794d673" exitCode=143 Oct 08 18:36:08 crc kubenswrapper[4859]: I1008 18:36:08.062836 4859 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Oct 08 18:36:08 crc kubenswrapper[4859]: I1008 18:36:08.062923 4859 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"fd151d4e-0d70-4458-8ab8-ffabed5dddca","Type":"ContainerDied","Data":"d2a2180ef27e62bc849417f71f26eec95810f30e12d337d28d89a70a6f32a434"} Oct 08 18:36:08 crc kubenswrapper[4859]: I1008 18:36:08.062979 4859 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"fd151d4e-0d70-4458-8ab8-ffabed5dddca","Type":"ContainerDied","Data":"a1677fbe2b0ccc470b915d38b90540074c1f5035d3224a47de80d2518794d673"} Oct 08 18:36:08 crc kubenswrapper[4859]: I1008 18:36:08.062990 4859 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"fd151d4e-0d70-4458-8ab8-ffabed5dddca","Type":"ContainerDied","Data":"ee47a246ebdff3be8ab7bdc0fd8ec71b5b70bf97a4f6503092cee86a44f7617d"} Oct 08 18:36:08 crc kubenswrapper[4859]: I1008 18:36:08.063012 4859 scope.go:117] "RemoveContainer" containerID="d2a2180ef27e62bc849417f71f26eec95810f30e12d337d28d89a70a6f32a434" Oct 08 18:36:08 crc kubenswrapper[4859]: I1008 18:36:08.071426 4859 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-conductor-db-sync-28hwn" event={"ID":"6b0220ac-1dc0-4ca9-a9ae-9335b57820fa","Type":"ContainerDied","Data":"1015abdc89e469b6d2608efb2f3d45f112ac69e3c1e7af4973e5ac3aa3a4f65a"} Oct 08 18:36:08 crc kubenswrapper[4859]: I1008 18:36:08.071471 4859 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="1015abdc89e469b6d2608efb2f3d45f112ac69e3c1e7af4973e5ac3aa3a4f65a" Oct 08 18:36:08 crc kubenswrapper[4859]: I1008 18:36:08.071468 4859 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-conductor-db-sync-28hwn" Oct 08 18:36:08 crc kubenswrapper[4859]: I1008 18:36:08.075986 4859 generic.go:334] "Generic (PLEG): container finished" podID="2f61fa1e-ec4e-490e-91ef-90cf0887940d" containerID="fc645cdbe451b90429e2860adc57240bae13a8ad243741c1a12efdef93de92db" exitCode=143 Oct 08 18:36:08 crc kubenswrapper[4859]: I1008 18:36:08.076045 4859 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"2f61fa1e-ec4e-490e-91ef-90cf0887940d","Type":"ContainerDied","Data":"fc645cdbe451b90429e2860adc57240bae13a8ad243741c1a12efdef93de92db"} Oct 08 18:36:08 crc kubenswrapper[4859]: I1008 18:36:08.078666 4859 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-scheduler-0" podUID="0178104e-34da-4178-b6e0-873404e282ef" containerName="nova-scheduler-scheduler" containerID="cri-o://6060d204fee906aa340b4ed067014067f6a228091114803a60e21e70b3e3e192" gracePeriod=30 Oct 08 18:36:08 crc kubenswrapper[4859]: I1008 18:36:08.099652 4859 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell1-conductor-0"] Oct 08 18:36:08 crc kubenswrapper[4859]: E1008 18:36:08.100380 4859 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a03169fb-e503-427f-bafe-2f8767a23482" containerName="dnsmasq-dns" Oct 08 18:36:08 crc kubenswrapper[4859]: I1008 18:36:08.100406 4859 state_mem.go:107] "Deleted CPUSet assignment" podUID="a03169fb-e503-427f-bafe-2f8767a23482" containerName="dnsmasq-dns" Oct 08 18:36:08 crc kubenswrapper[4859]: E1008 18:36:08.100419 4859 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="76588317-83a8-42b4-8a4f-502d6889c7b5" containerName="nova-manage" Oct 08 18:36:08 crc kubenswrapper[4859]: I1008 18:36:08.100428 4859 state_mem.go:107] "Deleted CPUSet assignment" podUID="76588317-83a8-42b4-8a4f-502d6889c7b5" containerName="nova-manage" Oct 08 18:36:08 crc kubenswrapper[4859]: E1008 18:36:08.100442 4859 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="fd151d4e-0d70-4458-8ab8-ffabed5dddca" containerName="nova-metadata-log" Oct 08 18:36:08 crc kubenswrapper[4859]: I1008 18:36:08.100452 4859 state_mem.go:107] "Deleted CPUSet assignment" podUID="fd151d4e-0d70-4458-8ab8-ffabed5dddca" containerName="nova-metadata-log" Oct 08 18:36:08 crc kubenswrapper[4859]: E1008 18:36:08.100480 4859 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a03169fb-e503-427f-bafe-2f8767a23482" containerName="init" Oct 08 18:36:08 crc kubenswrapper[4859]: I1008 18:36:08.100488 4859 state_mem.go:107] "Deleted CPUSet assignment" podUID="a03169fb-e503-427f-bafe-2f8767a23482" containerName="init" Oct 08 18:36:08 crc kubenswrapper[4859]: E1008 18:36:08.100512 4859 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="6b0220ac-1dc0-4ca9-a9ae-9335b57820fa" containerName="nova-cell1-conductor-db-sync" Oct 08 18:36:08 crc kubenswrapper[4859]: I1008 18:36:08.100519 4859 state_mem.go:107] "Deleted CPUSet assignment" podUID="6b0220ac-1dc0-4ca9-a9ae-9335b57820fa" containerName="nova-cell1-conductor-db-sync" Oct 08 18:36:08 crc kubenswrapper[4859]: E1008 18:36:08.100539 4859 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="fd151d4e-0d70-4458-8ab8-ffabed5dddca" containerName="nova-metadata-metadata" Oct 08 18:36:08 crc kubenswrapper[4859]: I1008 18:36:08.100547 4859 state_mem.go:107] "Deleted CPUSet assignment" podUID="fd151d4e-0d70-4458-8ab8-ffabed5dddca" containerName="nova-metadata-metadata" Oct 08 18:36:08 crc kubenswrapper[4859]: I1008 18:36:08.101609 4859 memory_manager.go:354] "RemoveStaleState removing state" podUID="a03169fb-e503-427f-bafe-2f8767a23482" containerName="dnsmasq-dns" Oct 08 18:36:08 crc kubenswrapper[4859]: I1008 18:36:08.101640 4859 memory_manager.go:354] "RemoveStaleState removing state" podUID="fd151d4e-0d70-4458-8ab8-ffabed5dddca" containerName="nova-metadata-metadata" Oct 08 18:36:08 crc kubenswrapper[4859]: I1008 18:36:08.101655 4859 memory_manager.go:354] "RemoveStaleState removing state" podUID="fd151d4e-0d70-4458-8ab8-ffabed5dddca" containerName="nova-metadata-log" Oct 08 18:36:08 crc kubenswrapper[4859]: I1008 18:36:08.101667 4859 memory_manager.go:354] "RemoveStaleState removing state" podUID="76588317-83a8-42b4-8a4f-502d6889c7b5" containerName="nova-manage" Oct 08 18:36:08 crc kubenswrapper[4859]: I1008 18:36:08.101696 4859 memory_manager.go:354] "RemoveStaleState removing state" podUID="6b0220ac-1dc0-4ca9-a9ae-9335b57820fa" containerName="nova-cell1-conductor-db-sync" Oct 08 18:36:08 crc kubenswrapper[4859]: I1008 18:36:08.102633 4859 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-conductor-0" Oct 08 18:36:08 crc kubenswrapper[4859]: I1008 18:36:08.106859 4859 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell1-conductor-config-data" Oct 08 18:36:08 crc kubenswrapper[4859]: I1008 18:36:08.119773 4859 scope.go:117] "RemoveContainer" containerID="a1677fbe2b0ccc470b915d38b90540074c1f5035d3224a47de80d2518794d673" Oct 08 18:36:08 crc kubenswrapper[4859]: I1008 18:36:08.119913 4859 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-conductor-0"] Oct 08 18:36:08 crc kubenswrapper[4859]: I1008 18:36:08.172729 4859 scope.go:117] "RemoveContainer" containerID="d2a2180ef27e62bc849417f71f26eec95810f30e12d337d28d89a70a6f32a434" Oct 08 18:36:08 crc kubenswrapper[4859]: E1008 18:36:08.173230 4859 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"d2a2180ef27e62bc849417f71f26eec95810f30e12d337d28d89a70a6f32a434\": container with ID starting with d2a2180ef27e62bc849417f71f26eec95810f30e12d337d28d89a70a6f32a434 not found: ID does not exist" containerID="d2a2180ef27e62bc849417f71f26eec95810f30e12d337d28d89a70a6f32a434" Oct 08 18:36:08 crc kubenswrapper[4859]: I1008 18:36:08.173261 4859 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"d2a2180ef27e62bc849417f71f26eec95810f30e12d337d28d89a70a6f32a434"} err="failed to get container status \"d2a2180ef27e62bc849417f71f26eec95810f30e12d337d28d89a70a6f32a434\": rpc error: code = NotFound desc = could not find container \"d2a2180ef27e62bc849417f71f26eec95810f30e12d337d28d89a70a6f32a434\": container with ID starting with d2a2180ef27e62bc849417f71f26eec95810f30e12d337d28d89a70a6f32a434 not found: ID does not exist" Oct 08 18:36:08 crc kubenswrapper[4859]: I1008 18:36:08.173288 4859 scope.go:117] "RemoveContainer" containerID="a1677fbe2b0ccc470b915d38b90540074c1f5035d3224a47de80d2518794d673" Oct 08 18:36:08 crc kubenswrapper[4859]: E1008 18:36:08.173659 4859 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"a1677fbe2b0ccc470b915d38b90540074c1f5035d3224a47de80d2518794d673\": container with ID starting with a1677fbe2b0ccc470b915d38b90540074c1f5035d3224a47de80d2518794d673 not found: ID does not exist" containerID="a1677fbe2b0ccc470b915d38b90540074c1f5035d3224a47de80d2518794d673" Oct 08 18:36:08 crc kubenswrapper[4859]: I1008 18:36:08.173704 4859 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"a1677fbe2b0ccc470b915d38b90540074c1f5035d3224a47de80d2518794d673"} err="failed to get container status \"a1677fbe2b0ccc470b915d38b90540074c1f5035d3224a47de80d2518794d673\": rpc error: code = NotFound desc = could not find container \"a1677fbe2b0ccc470b915d38b90540074c1f5035d3224a47de80d2518794d673\": container with ID starting with a1677fbe2b0ccc470b915d38b90540074c1f5035d3224a47de80d2518794d673 not found: ID does not exist" Oct 08 18:36:08 crc kubenswrapper[4859]: I1008 18:36:08.173722 4859 scope.go:117] "RemoveContainer" containerID="d2a2180ef27e62bc849417f71f26eec95810f30e12d337d28d89a70a6f32a434" Oct 08 18:36:08 crc kubenswrapper[4859]: I1008 18:36:08.173966 4859 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"d2a2180ef27e62bc849417f71f26eec95810f30e12d337d28d89a70a6f32a434"} err="failed to get container status \"d2a2180ef27e62bc849417f71f26eec95810f30e12d337d28d89a70a6f32a434\": rpc error: code = NotFound desc = could not find container \"d2a2180ef27e62bc849417f71f26eec95810f30e12d337d28d89a70a6f32a434\": container with ID starting with d2a2180ef27e62bc849417f71f26eec95810f30e12d337d28d89a70a6f32a434 not found: ID does not exist" Oct 08 18:36:08 crc kubenswrapper[4859]: I1008 18:36:08.173990 4859 scope.go:117] "RemoveContainer" containerID="a1677fbe2b0ccc470b915d38b90540074c1f5035d3224a47de80d2518794d673" Oct 08 18:36:08 crc kubenswrapper[4859]: I1008 18:36:08.174219 4859 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"a1677fbe2b0ccc470b915d38b90540074c1f5035d3224a47de80d2518794d673"} err="failed to get container status \"a1677fbe2b0ccc470b915d38b90540074c1f5035d3224a47de80d2518794d673\": rpc error: code = NotFound desc = could not find container \"a1677fbe2b0ccc470b915d38b90540074c1f5035d3224a47de80d2518794d673\": container with ID starting with a1677fbe2b0ccc470b915d38b90540074c1f5035d3224a47de80d2518794d673 not found: ID does not exist" Oct 08 18:36:08 crc kubenswrapper[4859]: I1008 18:36:08.184760 4859 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-metadata-0"] Oct 08 18:36:08 crc kubenswrapper[4859]: I1008 18:36:08.197543 4859 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-metadata-0"] Oct 08 18:36:08 crc kubenswrapper[4859]: I1008 18:36:08.236342 4859 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-metadata-0"] Oct 08 18:36:08 crc kubenswrapper[4859]: I1008 18:36:08.258247 4859 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Oct 08 18:36:08 crc kubenswrapper[4859]: I1008 18:36:08.259002 4859 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-metadata-0"] Oct 08 18:36:08 crc kubenswrapper[4859]: I1008 18:36:08.264888 4859 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-9klc5\" (UniqueName: \"kubernetes.io/projected/994d31f2-ad55-4817-a2e5-4e3743462cae-kube-api-access-9klc5\") pod \"nova-cell1-conductor-0\" (UID: \"994d31f2-ad55-4817-a2e5-4e3743462cae\") " pod="openstack/nova-cell1-conductor-0" Oct 08 18:36:08 crc kubenswrapper[4859]: I1008 18:36:08.264988 4859 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/994d31f2-ad55-4817-a2e5-4e3743462cae-combined-ca-bundle\") pod \"nova-cell1-conductor-0\" (UID: \"994d31f2-ad55-4817-a2e5-4e3743462cae\") " pod="openstack/nova-cell1-conductor-0" Oct 08 18:36:08 crc kubenswrapper[4859]: I1008 18:36:08.265065 4859 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/994d31f2-ad55-4817-a2e5-4e3743462cae-config-data\") pod \"nova-cell1-conductor-0\" (UID: \"994d31f2-ad55-4817-a2e5-4e3743462cae\") " pod="openstack/nova-cell1-conductor-0" Oct 08 18:36:08 crc kubenswrapper[4859]: I1008 18:36:08.266564 4859 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-nova-metadata-internal-svc" Oct 08 18:36:08 crc kubenswrapper[4859]: I1008 18:36:08.266770 4859 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-metadata-config-data" Oct 08 18:36:08 crc kubenswrapper[4859]: I1008 18:36:08.366436 4859 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f8de039d-85d9-444b-937e-d8d6f90bb77f-config-data\") pod \"nova-metadata-0\" (UID: \"f8de039d-85d9-444b-937e-d8d6f90bb77f\") " pod="openstack/nova-metadata-0" Oct 08 18:36:08 crc kubenswrapper[4859]: I1008 18:36:08.366490 4859 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/f8de039d-85d9-444b-937e-d8d6f90bb77f-logs\") pod \"nova-metadata-0\" (UID: \"f8de039d-85d9-444b-937e-d8d6f90bb77f\") " pod="openstack/nova-metadata-0" Oct 08 18:36:08 crc kubenswrapper[4859]: I1008 18:36:08.366951 4859 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-9klc5\" (UniqueName: \"kubernetes.io/projected/994d31f2-ad55-4817-a2e5-4e3743462cae-kube-api-access-9klc5\") pod \"nova-cell1-conductor-0\" (UID: \"994d31f2-ad55-4817-a2e5-4e3743462cae\") " pod="openstack/nova-cell1-conductor-0" Oct 08 18:36:08 crc kubenswrapper[4859]: I1008 18:36:08.367022 4859 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-jztkt\" (UniqueName: \"kubernetes.io/projected/f8de039d-85d9-444b-937e-d8d6f90bb77f-kube-api-access-jztkt\") pod \"nova-metadata-0\" (UID: \"f8de039d-85d9-444b-937e-d8d6f90bb77f\") " pod="openstack/nova-metadata-0" Oct 08 18:36:08 crc kubenswrapper[4859]: I1008 18:36:08.367044 4859 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f8de039d-85d9-444b-937e-d8d6f90bb77f-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"f8de039d-85d9-444b-937e-d8d6f90bb77f\") " pod="openstack/nova-metadata-0" Oct 08 18:36:08 crc kubenswrapper[4859]: I1008 18:36:08.367096 4859 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/994d31f2-ad55-4817-a2e5-4e3743462cae-combined-ca-bundle\") pod \"nova-cell1-conductor-0\" (UID: \"994d31f2-ad55-4817-a2e5-4e3743462cae\") " pod="openstack/nova-cell1-conductor-0" Oct 08 18:36:08 crc kubenswrapper[4859]: I1008 18:36:08.367913 4859 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/f8de039d-85d9-444b-937e-d8d6f90bb77f-nova-metadata-tls-certs\") pod \"nova-metadata-0\" (UID: \"f8de039d-85d9-444b-937e-d8d6f90bb77f\") " pod="openstack/nova-metadata-0" Oct 08 18:36:08 crc kubenswrapper[4859]: I1008 18:36:08.367954 4859 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/994d31f2-ad55-4817-a2e5-4e3743462cae-config-data\") pod \"nova-cell1-conductor-0\" (UID: \"994d31f2-ad55-4817-a2e5-4e3743462cae\") " pod="openstack/nova-cell1-conductor-0" Oct 08 18:36:08 crc kubenswrapper[4859]: I1008 18:36:08.372716 4859 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/994d31f2-ad55-4817-a2e5-4e3743462cae-combined-ca-bundle\") pod \"nova-cell1-conductor-0\" (UID: \"994d31f2-ad55-4817-a2e5-4e3743462cae\") " pod="openstack/nova-cell1-conductor-0" Oct 08 18:36:08 crc kubenswrapper[4859]: I1008 18:36:08.372775 4859 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/994d31f2-ad55-4817-a2e5-4e3743462cae-config-data\") pod \"nova-cell1-conductor-0\" (UID: \"994d31f2-ad55-4817-a2e5-4e3743462cae\") " pod="openstack/nova-cell1-conductor-0" Oct 08 18:36:08 crc kubenswrapper[4859]: I1008 18:36:08.385264 4859 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-9klc5\" (UniqueName: \"kubernetes.io/projected/994d31f2-ad55-4817-a2e5-4e3743462cae-kube-api-access-9klc5\") pod \"nova-cell1-conductor-0\" (UID: \"994d31f2-ad55-4817-a2e5-4e3743462cae\") " pod="openstack/nova-cell1-conductor-0" Oct 08 18:36:08 crc kubenswrapper[4859]: I1008 18:36:08.432194 4859 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-conductor-0" Oct 08 18:36:08 crc kubenswrapper[4859]: I1008 18:36:08.469173 4859 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f8de039d-85d9-444b-937e-d8d6f90bb77f-config-data\") pod \"nova-metadata-0\" (UID: \"f8de039d-85d9-444b-937e-d8d6f90bb77f\") " pod="openstack/nova-metadata-0" Oct 08 18:36:08 crc kubenswrapper[4859]: I1008 18:36:08.469217 4859 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/f8de039d-85d9-444b-937e-d8d6f90bb77f-logs\") pod \"nova-metadata-0\" (UID: \"f8de039d-85d9-444b-937e-d8d6f90bb77f\") " pod="openstack/nova-metadata-0" Oct 08 18:36:08 crc kubenswrapper[4859]: I1008 18:36:08.469288 4859 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-jztkt\" (UniqueName: \"kubernetes.io/projected/f8de039d-85d9-444b-937e-d8d6f90bb77f-kube-api-access-jztkt\") pod \"nova-metadata-0\" (UID: \"f8de039d-85d9-444b-937e-d8d6f90bb77f\") " pod="openstack/nova-metadata-0" Oct 08 18:36:08 crc kubenswrapper[4859]: I1008 18:36:08.469310 4859 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f8de039d-85d9-444b-937e-d8d6f90bb77f-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"f8de039d-85d9-444b-937e-d8d6f90bb77f\") " pod="openstack/nova-metadata-0" Oct 08 18:36:08 crc kubenswrapper[4859]: I1008 18:36:08.469344 4859 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/f8de039d-85d9-444b-937e-d8d6f90bb77f-nova-metadata-tls-certs\") pod \"nova-metadata-0\" (UID: \"f8de039d-85d9-444b-937e-d8d6f90bb77f\") " pod="openstack/nova-metadata-0" Oct 08 18:36:08 crc kubenswrapper[4859]: I1008 18:36:08.469899 4859 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/f8de039d-85d9-444b-937e-d8d6f90bb77f-logs\") pod \"nova-metadata-0\" (UID: \"f8de039d-85d9-444b-937e-d8d6f90bb77f\") " pod="openstack/nova-metadata-0" Oct 08 18:36:08 crc kubenswrapper[4859]: I1008 18:36:08.472905 4859 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/f8de039d-85d9-444b-937e-d8d6f90bb77f-nova-metadata-tls-certs\") pod \"nova-metadata-0\" (UID: \"f8de039d-85d9-444b-937e-d8d6f90bb77f\") " pod="openstack/nova-metadata-0" Oct 08 18:36:08 crc kubenswrapper[4859]: I1008 18:36:08.473280 4859 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f8de039d-85d9-444b-937e-d8d6f90bb77f-config-data\") pod \"nova-metadata-0\" (UID: \"f8de039d-85d9-444b-937e-d8d6f90bb77f\") " pod="openstack/nova-metadata-0" Oct 08 18:36:08 crc kubenswrapper[4859]: I1008 18:36:08.473519 4859 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f8de039d-85d9-444b-937e-d8d6f90bb77f-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"f8de039d-85d9-444b-937e-d8d6f90bb77f\") " pod="openstack/nova-metadata-0" Oct 08 18:36:08 crc kubenswrapper[4859]: I1008 18:36:08.493355 4859 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-jztkt\" (UniqueName: \"kubernetes.io/projected/f8de039d-85d9-444b-937e-d8d6f90bb77f-kube-api-access-jztkt\") pod \"nova-metadata-0\" (UID: \"f8de039d-85d9-444b-937e-d8d6f90bb77f\") " pod="openstack/nova-metadata-0" Oct 08 18:36:08 crc kubenswrapper[4859]: I1008 18:36:08.494352 4859 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="a03169fb-e503-427f-bafe-2f8767a23482" path="/var/lib/kubelet/pods/a03169fb-e503-427f-bafe-2f8767a23482/volumes" Oct 08 18:36:08 crc kubenswrapper[4859]: I1008 18:36:08.495520 4859 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="fd151d4e-0d70-4458-8ab8-ffabed5dddca" path="/var/lib/kubelet/pods/fd151d4e-0d70-4458-8ab8-ffabed5dddca/volumes" Oct 08 18:36:08 crc kubenswrapper[4859]: I1008 18:36:08.590549 4859 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Oct 08 18:36:08 crc kubenswrapper[4859]: I1008 18:36:08.912639 4859 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-conductor-0"] Oct 08 18:36:08 crc kubenswrapper[4859]: W1008 18:36:08.916577 4859 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod994d31f2_ad55_4817_a2e5_4e3743462cae.slice/crio-d2bc382eb59176b1cb61bbf106d9dd06f88cc35811c0b17669353acb65f7df60 WatchSource:0}: Error finding container d2bc382eb59176b1cb61bbf106d9dd06f88cc35811c0b17669353acb65f7df60: Status 404 returned error can't find the container with id d2bc382eb59176b1cb61bbf106d9dd06f88cc35811c0b17669353acb65f7df60 Oct 08 18:36:09 crc kubenswrapper[4859]: I1008 18:36:09.067057 4859 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-metadata-0"] Oct 08 18:36:09 crc kubenswrapper[4859]: I1008 18:36:09.089307 4859 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"f8de039d-85d9-444b-937e-d8d6f90bb77f","Type":"ContainerStarted","Data":"9345957d7543f2ca4931a21700a11bea59a58373aca68aa97f693fa971dc360b"} Oct 08 18:36:09 crc kubenswrapper[4859]: I1008 18:36:09.092051 4859 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-conductor-0" event={"ID":"994d31f2-ad55-4817-a2e5-4e3743462cae","Type":"ContainerStarted","Data":"d2bc382eb59176b1cb61bbf106d9dd06f88cc35811c0b17669353acb65f7df60"} Oct 08 18:36:10 crc kubenswrapper[4859]: I1008 18:36:10.113847 4859 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"f8de039d-85d9-444b-937e-d8d6f90bb77f","Type":"ContainerStarted","Data":"dce4317d8887820fbdd21c3fd9beefcc68a86c05505bb4dd820e11a9c745a137"} Oct 08 18:36:10 crc kubenswrapper[4859]: I1008 18:36:10.114543 4859 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"f8de039d-85d9-444b-937e-d8d6f90bb77f","Type":"ContainerStarted","Data":"719956fb0a5ddd763170476c1c204c6c8a78e5b6f8bdc405211df91e0e9eff5b"} Oct 08 18:36:10 crc kubenswrapper[4859]: I1008 18:36:10.119818 4859 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-conductor-0" event={"ID":"994d31f2-ad55-4817-a2e5-4e3743462cae","Type":"ContainerStarted","Data":"7a7f7145f443a770014b928c810d8208838c77a1c0c66b5b642c7c23067a639f"} Oct 08 18:36:10 crc kubenswrapper[4859]: I1008 18:36:10.120191 4859 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-cell1-conductor-0" Oct 08 18:36:10 crc kubenswrapper[4859]: I1008 18:36:10.144900 4859 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-metadata-0" podStartSLOduration=2.144878082 podStartE2EDuration="2.144878082s" podCreationTimestamp="2025-10-08 18:36:08 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-08 18:36:10.144602714 +0000 UTC m=+1140.391442123" watchObservedRunningTime="2025-10-08 18:36:10.144878082 +0000 UTC m=+1140.391717461" Oct 08 18:36:10 crc kubenswrapper[4859]: I1008 18:36:10.171080 4859 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-cell1-conductor-0" podStartSLOduration=2.171056721 podStartE2EDuration="2.171056721s" podCreationTimestamp="2025-10-08 18:36:08 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-08 18:36:10.164262477 +0000 UTC m=+1140.411101866" watchObservedRunningTime="2025-10-08 18:36:10.171056721 +0000 UTC m=+1140.417896100" Oct 08 18:36:10 crc kubenswrapper[4859]: E1008 18:36:10.640278 4859 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="6060d204fee906aa340b4ed067014067f6a228091114803a60e21e70b3e3e192" cmd=["/usr/bin/pgrep","-r","DRST","nova-scheduler"] Oct 08 18:36:10 crc kubenswrapper[4859]: E1008 18:36:10.642153 4859 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="6060d204fee906aa340b4ed067014067f6a228091114803a60e21e70b3e3e192" cmd=["/usr/bin/pgrep","-r","DRST","nova-scheduler"] Oct 08 18:36:10 crc kubenswrapper[4859]: E1008 18:36:10.644530 4859 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="6060d204fee906aa340b4ed067014067f6a228091114803a60e21e70b3e3e192" cmd=["/usr/bin/pgrep","-r","DRST","nova-scheduler"] Oct 08 18:36:10 crc kubenswrapper[4859]: E1008 18:36:10.644568 4859 prober.go:104] "Probe errored" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" probeType="Readiness" pod="openstack/nova-scheduler-0" podUID="0178104e-34da-4178-b6e0-873404e282ef" containerName="nova-scheduler-scheduler" Oct 08 18:36:11 crc kubenswrapper[4859]: I1008 18:36:11.778103 4859 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Oct 08 18:36:11 crc kubenswrapper[4859]: I1008 18:36:11.939681 4859 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0178104e-34da-4178-b6e0-873404e282ef-combined-ca-bundle\") pod \"0178104e-34da-4178-b6e0-873404e282ef\" (UID: \"0178104e-34da-4178-b6e0-873404e282ef\") " Oct 08 18:36:11 crc kubenswrapper[4859]: I1008 18:36:11.939795 4859 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-s9nvk\" (UniqueName: \"kubernetes.io/projected/0178104e-34da-4178-b6e0-873404e282ef-kube-api-access-s9nvk\") pod \"0178104e-34da-4178-b6e0-873404e282ef\" (UID: \"0178104e-34da-4178-b6e0-873404e282ef\") " Oct 08 18:36:11 crc kubenswrapper[4859]: I1008 18:36:11.939839 4859 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/0178104e-34da-4178-b6e0-873404e282ef-config-data\") pod \"0178104e-34da-4178-b6e0-873404e282ef\" (UID: \"0178104e-34da-4178-b6e0-873404e282ef\") " Oct 08 18:36:11 crc kubenswrapper[4859]: I1008 18:36:11.947114 4859 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/0178104e-34da-4178-b6e0-873404e282ef-kube-api-access-s9nvk" (OuterVolumeSpecName: "kube-api-access-s9nvk") pod "0178104e-34da-4178-b6e0-873404e282ef" (UID: "0178104e-34da-4178-b6e0-873404e282ef"). InnerVolumeSpecName "kube-api-access-s9nvk". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 08 18:36:11 crc kubenswrapper[4859]: I1008 18:36:11.972146 4859 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/0178104e-34da-4178-b6e0-873404e282ef-config-data" (OuterVolumeSpecName: "config-data") pod "0178104e-34da-4178-b6e0-873404e282ef" (UID: "0178104e-34da-4178-b6e0-873404e282ef"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 08 18:36:11 crc kubenswrapper[4859]: I1008 18:36:11.974073 4859 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/0178104e-34da-4178-b6e0-873404e282ef-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "0178104e-34da-4178-b6e0-873404e282ef" (UID: "0178104e-34da-4178-b6e0-873404e282ef"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 08 18:36:12 crc kubenswrapper[4859]: I1008 18:36:12.042159 4859 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-s9nvk\" (UniqueName: \"kubernetes.io/projected/0178104e-34da-4178-b6e0-873404e282ef-kube-api-access-s9nvk\") on node \"crc\" DevicePath \"\"" Oct 08 18:36:12 crc kubenswrapper[4859]: I1008 18:36:12.042201 4859 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/0178104e-34da-4178-b6e0-873404e282ef-config-data\") on node \"crc\" DevicePath \"\"" Oct 08 18:36:12 crc kubenswrapper[4859]: I1008 18:36:12.042211 4859 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0178104e-34da-4178-b6e0-873404e282ef-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 08 18:36:12 crc kubenswrapper[4859]: I1008 18:36:12.140145 4859 generic.go:334] "Generic (PLEG): container finished" podID="0178104e-34da-4178-b6e0-873404e282ef" containerID="6060d204fee906aa340b4ed067014067f6a228091114803a60e21e70b3e3e192" exitCode=0 Oct 08 18:36:12 crc kubenswrapper[4859]: I1008 18:36:12.140205 4859 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Oct 08 18:36:12 crc kubenswrapper[4859]: I1008 18:36:12.140211 4859 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"0178104e-34da-4178-b6e0-873404e282ef","Type":"ContainerDied","Data":"6060d204fee906aa340b4ed067014067f6a228091114803a60e21e70b3e3e192"} Oct 08 18:36:12 crc kubenswrapper[4859]: I1008 18:36:12.140351 4859 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"0178104e-34da-4178-b6e0-873404e282ef","Type":"ContainerDied","Data":"163c22fdee9b9ac83ff176f22ea2757ca88f329656d4858aadf926430da91219"} Oct 08 18:36:12 crc kubenswrapper[4859]: I1008 18:36:12.140390 4859 scope.go:117] "RemoveContainer" containerID="6060d204fee906aa340b4ed067014067f6a228091114803a60e21e70b3e3e192" Oct 08 18:36:12 crc kubenswrapper[4859]: I1008 18:36:12.168513 4859 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-scheduler-0"] Oct 08 18:36:12 crc kubenswrapper[4859]: I1008 18:36:12.172277 4859 scope.go:117] "RemoveContainer" containerID="6060d204fee906aa340b4ed067014067f6a228091114803a60e21e70b3e3e192" Oct 08 18:36:12 crc kubenswrapper[4859]: E1008 18:36:12.172868 4859 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"6060d204fee906aa340b4ed067014067f6a228091114803a60e21e70b3e3e192\": container with ID starting with 6060d204fee906aa340b4ed067014067f6a228091114803a60e21e70b3e3e192 not found: ID does not exist" containerID="6060d204fee906aa340b4ed067014067f6a228091114803a60e21e70b3e3e192" Oct 08 18:36:12 crc kubenswrapper[4859]: I1008 18:36:12.172926 4859 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"6060d204fee906aa340b4ed067014067f6a228091114803a60e21e70b3e3e192"} err="failed to get container status \"6060d204fee906aa340b4ed067014067f6a228091114803a60e21e70b3e3e192\": rpc error: code = NotFound desc = could not find container \"6060d204fee906aa340b4ed067014067f6a228091114803a60e21e70b3e3e192\": container with ID starting with 6060d204fee906aa340b4ed067014067f6a228091114803a60e21e70b3e3e192 not found: ID does not exist" Oct 08 18:36:12 crc kubenswrapper[4859]: I1008 18:36:12.177141 4859 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-scheduler-0"] Oct 08 18:36:12 crc kubenswrapper[4859]: I1008 18:36:12.192098 4859 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-scheduler-0"] Oct 08 18:36:12 crc kubenswrapper[4859]: E1008 18:36:12.193106 4859 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0178104e-34da-4178-b6e0-873404e282ef" containerName="nova-scheduler-scheduler" Oct 08 18:36:12 crc kubenswrapper[4859]: I1008 18:36:12.193140 4859 state_mem.go:107] "Deleted CPUSet assignment" podUID="0178104e-34da-4178-b6e0-873404e282ef" containerName="nova-scheduler-scheduler" Oct 08 18:36:12 crc kubenswrapper[4859]: I1008 18:36:12.193460 4859 memory_manager.go:354] "RemoveStaleState removing state" podUID="0178104e-34da-4178-b6e0-873404e282ef" containerName="nova-scheduler-scheduler" Oct 08 18:36:12 crc kubenswrapper[4859]: I1008 18:36:12.194499 4859 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Oct 08 18:36:12 crc kubenswrapper[4859]: I1008 18:36:12.198386 4859 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-scheduler-config-data" Oct 08 18:36:12 crc kubenswrapper[4859]: I1008 18:36:12.199767 4859 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-scheduler-0"] Oct 08 18:36:12 crc kubenswrapper[4859]: I1008 18:36:12.346444 4859 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e5f82275-e5ae-4d2d-8ab6-dc4afda848ba-combined-ca-bundle\") pod \"nova-scheduler-0\" (UID: \"e5f82275-e5ae-4d2d-8ab6-dc4afda848ba\") " pod="openstack/nova-scheduler-0" Oct 08 18:36:12 crc kubenswrapper[4859]: I1008 18:36:12.346918 4859 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/e5f82275-e5ae-4d2d-8ab6-dc4afda848ba-config-data\") pod \"nova-scheduler-0\" (UID: \"e5f82275-e5ae-4d2d-8ab6-dc4afda848ba\") " pod="openstack/nova-scheduler-0" Oct 08 18:36:12 crc kubenswrapper[4859]: I1008 18:36:12.347044 4859 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-xm54t\" (UniqueName: \"kubernetes.io/projected/e5f82275-e5ae-4d2d-8ab6-dc4afda848ba-kube-api-access-xm54t\") pod \"nova-scheduler-0\" (UID: \"e5f82275-e5ae-4d2d-8ab6-dc4afda848ba\") " pod="openstack/nova-scheduler-0" Oct 08 18:36:12 crc kubenswrapper[4859]: I1008 18:36:12.449177 4859 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e5f82275-e5ae-4d2d-8ab6-dc4afda848ba-combined-ca-bundle\") pod \"nova-scheduler-0\" (UID: \"e5f82275-e5ae-4d2d-8ab6-dc4afda848ba\") " pod="openstack/nova-scheduler-0" Oct 08 18:36:12 crc kubenswrapper[4859]: I1008 18:36:12.449358 4859 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/e5f82275-e5ae-4d2d-8ab6-dc4afda848ba-config-data\") pod \"nova-scheduler-0\" (UID: \"e5f82275-e5ae-4d2d-8ab6-dc4afda848ba\") " pod="openstack/nova-scheduler-0" Oct 08 18:36:12 crc kubenswrapper[4859]: I1008 18:36:12.449424 4859 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-xm54t\" (UniqueName: \"kubernetes.io/projected/e5f82275-e5ae-4d2d-8ab6-dc4afda848ba-kube-api-access-xm54t\") pod \"nova-scheduler-0\" (UID: \"e5f82275-e5ae-4d2d-8ab6-dc4afda848ba\") " pod="openstack/nova-scheduler-0" Oct 08 18:36:12 crc kubenswrapper[4859]: I1008 18:36:12.457779 4859 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/e5f82275-e5ae-4d2d-8ab6-dc4afda848ba-config-data\") pod \"nova-scheduler-0\" (UID: \"e5f82275-e5ae-4d2d-8ab6-dc4afda848ba\") " pod="openstack/nova-scheduler-0" Oct 08 18:36:12 crc kubenswrapper[4859]: I1008 18:36:12.462719 4859 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e5f82275-e5ae-4d2d-8ab6-dc4afda848ba-combined-ca-bundle\") pod \"nova-scheduler-0\" (UID: \"e5f82275-e5ae-4d2d-8ab6-dc4afda848ba\") " pod="openstack/nova-scheduler-0" Oct 08 18:36:12 crc kubenswrapper[4859]: I1008 18:36:12.467808 4859 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-xm54t\" (UniqueName: \"kubernetes.io/projected/e5f82275-e5ae-4d2d-8ab6-dc4afda848ba-kube-api-access-xm54t\") pod \"nova-scheduler-0\" (UID: \"e5f82275-e5ae-4d2d-8ab6-dc4afda848ba\") " pod="openstack/nova-scheduler-0" Oct 08 18:36:12 crc kubenswrapper[4859]: I1008 18:36:12.485422 4859 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="0178104e-34da-4178-b6e0-873404e282ef" path="/var/lib/kubelet/pods/0178104e-34da-4178-b6e0-873404e282ef/volumes" Oct 08 18:36:12 crc kubenswrapper[4859]: I1008 18:36:12.509262 4859 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Oct 08 18:36:12 crc kubenswrapper[4859]: I1008 18:36:12.954993 4859 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-scheduler-0"] Oct 08 18:36:13 crc kubenswrapper[4859]: I1008 18:36:13.143388 4859 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Oct 08 18:36:13 crc kubenswrapper[4859]: I1008 18:36:13.173937 4859 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"e5f82275-e5ae-4d2d-8ab6-dc4afda848ba","Type":"ContainerStarted","Data":"3f8cc4d2069435407bc2679facaa25edfb59b662f317d01fed8daa54ee38bff2"} Oct 08 18:36:13 crc kubenswrapper[4859]: I1008 18:36:13.195249 4859 generic.go:334] "Generic (PLEG): container finished" podID="2f61fa1e-ec4e-490e-91ef-90cf0887940d" containerID="41e75dcb065fad2bed5125c9c642da1af0bf7ef04626d2bc28bf09c54539e2fa" exitCode=0 Oct 08 18:36:13 crc kubenswrapper[4859]: I1008 18:36:13.195306 4859 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"2f61fa1e-ec4e-490e-91ef-90cf0887940d","Type":"ContainerDied","Data":"41e75dcb065fad2bed5125c9c642da1af0bf7ef04626d2bc28bf09c54539e2fa"} Oct 08 18:36:13 crc kubenswrapper[4859]: I1008 18:36:13.195339 4859 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"2f61fa1e-ec4e-490e-91ef-90cf0887940d","Type":"ContainerDied","Data":"0d6bcfda0d0f0c376066ff120cd2fbb40dc18afff086c75338e5fa187fd3bf33"} Oct 08 18:36:13 crc kubenswrapper[4859]: I1008 18:36:13.195359 4859 scope.go:117] "RemoveContainer" containerID="41e75dcb065fad2bed5125c9c642da1af0bf7ef04626d2bc28bf09c54539e2fa" Oct 08 18:36:13 crc kubenswrapper[4859]: I1008 18:36:13.195515 4859 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Oct 08 18:36:13 crc kubenswrapper[4859]: I1008 18:36:13.274124 4859 scope.go:117] "RemoveContainer" containerID="fc645cdbe451b90429e2860adc57240bae13a8ad243741c1a12efdef93de92db" Oct 08 18:36:13 crc kubenswrapper[4859]: I1008 18:36:13.274512 4859 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/2f61fa1e-ec4e-490e-91ef-90cf0887940d-logs\") pod \"2f61fa1e-ec4e-490e-91ef-90cf0887940d\" (UID: \"2f61fa1e-ec4e-490e-91ef-90cf0887940d\") " Oct 08 18:36:13 crc kubenswrapper[4859]: I1008 18:36:13.274567 4859 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-vxk6g\" (UniqueName: \"kubernetes.io/projected/2f61fa1e-ec4e-490e-91ef-90cf0887940d-kube-api-access-vxk6g\") pod \"2f61fa1e-ec4e-490e-91ef-90cf0887940d\" (UID: \"2f61fa1e-ec4e-490e-91ef-90cf0887940d\") " Oct 08 18:36:13 crc kubenswrapper[4859]: I1008 18:36:13.274756 4859 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2f61fa1e-ec4e-490e-91ef-90cf0887940d-combined-ca-bundle\") pod \"2f61fa1e-ec4e-490e-91ef-90cf0887940d\" (UID: \"2f61fa1e-ec4e-490e-91ef-90cf0887940d\") " Oct 08 18:36:13 crc kubenswrapper[4859]: I1008 18:36:13.274816 4859 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/2f61fa1e-ec4e-490e-91ef-90cf0887940d-config-data\") pod \"2f61fa1e-ec4e-490e-91ef-90cf0887940d\" (UID: \"2f61fa1e-ec4e-490e-91ef-90cf0887940d\") " Oct 08 18:36:13 crc kubenswrapper[4859]: I1008 18:36:13.275195 4859 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/2f61fa1e-ec4e-490e-91ef-90cf0887940d-logs" (OuterVolumeSpecName: "logs") pod "2f61fa1e-ec4e-490e-91ef-90cf0887940d" (UID: "2f61fa1e-ec4e-490e-91ef-90cf0887940d"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 08 18:36:13 crc kubenswrapper[4859]: I1008 18:36:13.275342 4859 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/2f61fa1e-ec4e-490e-91ef-90cf0887940d-logs\") on node \"crc\" DevicePath \"\"" Oct 08 18:36:13 crc kubenswrapper[4859]: I1008 18:36:13.288542 4859 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/2f61fa1e-ec4e-490e-91ef-90cf0887940d-kube-api-access-vxk6g" (OuterVolumeSpecName: "kube-api-access-vxk6g") pod "2f61fa1e-ec4e-490e-91ef-90cf0887940d" (UID: "2f61fa1e-ec4e-490e-91ef-90cf0887940d"). InnerVolumeSpecName "kube-api-access-vxk6g". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 08 18:36:13 crc kubenswrapper[4859]: I1008 18:36:13.320406 4859 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/2f61fa1e-ec4e-490e-91ef-90cf0887940d-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "2f61fa1e-ec4e-490e-91ef-90cf0887940d" (UID: "2f61fa1e-ec4e-490e-91ef-90cf0887940d"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 08 18:36:13 crc kubenswrapper[4859]: I1008 18:36:13.321522 4859 scope.go:117] "RemoveContainer" containerID="41e75dcb065fad2bed5125c9c642da1af0bf7ef04626d2bc28bf09c54539e2fa" Oct 08 18:36:13 crc kubenswrapper[4859]: E1008 18:36:13.322174 4859 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"41e75dcb065fad2bed5125c9c642da1af0bf7ef04626d2bc28bf09c54539e2fa\": container with ID starting with 41e75dcb065fad2bed5125c9c642da1af0bf7ef04626d2bc28bf09c54539e2fa not found: ID does not exist" containerID="41e75dcb065fad2bed5125c9c642da1af0bf7ef04626d2bc28bf09c54539e2fa" Oct 08 18:36:13 crc kubenswrapper[4859]: I1008 18:36:13.322269 4859 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"41e75dcb065fad2bed5125c9c642da1af0bf7ef04626d2bc28bf09c54539e2fa"} err="failed to get container status \"41e75dcb065fad2bed5125c9c642da1af0bf7ef04626d2bc28bf09c54539e2fa\": rpc error: code = NotFound desc = could not find container \"41e75dcb065fad2bed5125c9c642da1af0bf7ef04626d2bc28bf09c54539e2fa\": container with ID starting with 41e75dcb065fad2bed5125c9c642da1af0bf7ef04626d2bc28bf09c54539e2fa not found: ID does not exist" Oct 08 18:36:13 crc kubenswrapper[4859]: I1008 18:36:13.322363 4859 scope.go:117] "RemoveContainer" containerID="fc645cdbe451b90429e2860adc57240bae13a8ad243741c1a12efdef93de92db" Oct 08 18:36:13 crc kubenswrapper[4859]: E1008 18:36:13.323002 4859 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"fc645cdbe451b90429e2860adc57240bae13a8ad243741c1a12efdef93de92db\": container with ID starting with fc645cdbe451b90429e2860adc57240bae13a8ad243741c1a12efdef93de92db not found: ID does not exist" containerID="fc645cdbe451b90429e2860adc57240bae13a8ad243741c1a12efdef93de92db" Oct 08 18:36:13 crc kubenswrapper[4859]: I1008 18:36:13.323041 4859 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"fc645cdbe451b90429e2860adc57240bae13a8ad243741c1a12efdef93de92db"} err="failed to get container status \"fc645cdbe451b90429e2860adc57240bae13a8ad243741c1a12efdef93de92db\": rpc error: code = NotFound desc = could not find container \"fc645cdbe451b90429e2860adc57240bae13a8ad243741c1a12efdef93de92db\": container with ID starting with fc645cdbe451b90429e2860adc57240bae13a8ad243741c1a12efdef93de92db not found: ID does not exist" Oct 08 18:36:13 crc kubenswrapper[4859]: I1008 18:36:13.324258 4859 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/2f61fa1e-ec4e-490e-91ef-90cf0887940d-config-data" (OuterVolumeSpecName: "config-data") pod "2f61fa1e-ec4e-490e-91ef-90cf0887940d" (UID: "2f61fa1e-ec4e-490e-91ef-90cf0887940d"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 08 18:36:13 crc kubenswrapper[4859]: I1008 18:36:13.377573 4859 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2f61fa1e-ec4e-490e-91ef-90cf0887940d-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 08 18:36:13 crc kubenswrapper[4859]: I1008 18:36:13.377623 4859 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/2f61fa1e-ec4e-490e-91ef-90cf0887940d-config-data\") on node \"crc\" DevicePath \"\"" Oct 08 18:36:13 crc kubenswrapper[4859]: I1008 18:36:13.377638 4859 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-vxk6g\" (UniqueName: \"kubernetes.io/projected/2f61fa1e-ec4e-490e-91ef-90cf0887940d-kube-api-access-vxk6g\") on node \"crc\" DevicePath \"\"" Oct 08 18:36:13 crc kubenswrapper[4859]: I1008 18:36:13.537448 4859 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-api-0"] Oct 08 18:36:13 crc kubenswrapper[4859]: I1008 18:36:13.550858 4859 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-api-0"] Oct 08 18:36:13 crc kubenswrapper[4859]: I1008 18:36:13.595062 4859 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-metadata-0" Oct 08 18:36:13 crc kubenswrapper[4859]: I1008 18:36:13.595107 4859 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-metadata-0" Oct 08 18:36:13 crc kubenswrapper[4859]: I1008 18:36:13.598590 4859 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-api-0"] Oct 08 18:36:13 crc kubenswrapper[4859]: E1008 18:36:13.599633 4859 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2f61fa1e-ec4e-490e-91ef-90cf0887940d" containerName="nova-api-api" Oct 08 18:36:13 crc kubenswrapper[4859]: I1008 18:36:13.599654 4859 state_mem.go:107] "Deleted CPUSet assignment" podUID="2f61fa1e-ec4e-490e-91ef-90cf0887940d" containerName="nova-api-api" Oct 08 18:36:13 crc kubenswrapper[4859]: E1008 18:36:13.599699 4859 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2f61fa1e-ec4e-490e-91ef-90cf0887940d" containerName="nova-api-log" Oct 08 18:36:13 crc kubenswrapper[4859]: I1008 18:36:13.599707 4859 state_mem.go:107] "Deleted CPUSet assignment" podUID="2f61fa1e-ec4e-490e-91ef-90cf0887940d" containerName="nova-api-log" Oct 08 18:36:13 crc kubenswrapper[4859]: I1008 18:36:13.600276 4859 memory_manager.go:354] "RemoveStaleState removing state" podUID="2f61fa1e-ec4e-490e-91ef-90cf0887940d" containerName="nova-api-log" Oct 08 18:36:13 crc kubenswrapper[4859]: I1008 18:36:13.600317 4859 memory_manager.go:354] "RemoveStaleState removing state" podUID="2f61fa1e-ec4e-490e-91ef-90cf0887940d" containerName="nova-api-api" Oct 08 18:36:13 crc kubenswrapper[4859]: I1008 18:36:13.606182 4859 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Oct 08 18:36:13 crc kubenswrapper[4859]: I1008 18:36:13.614838 4859 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-0"] Oct 08 18:36:13 crc kubenswrapper[4859]: I1008 18:36:13.616365 4859 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-api-config-data" Oct 08 18:36:13 crc kubenswrapper[4859]: I1008 18:36:13.684821 4859 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/402e8a91-7238-4d32-800e-266a6c911802-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"402e8a91-7238-4d32-800e-266a6c911802\") " pod="openstack/nova-api-0" Oct 08 18:36:13 crc kubenswrapper[4859]: I1008 18:36:13.684869 4859 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/402e8a91-7238-4d32-800e-266a6c911802-logs\") pod \"nova-api-0\" (UID: \"402e8a91-7238-4d32-800e-266a6c911802\") " pod="openstack/nova-api-0" Oct 08 18:36:13 crc kubenswrapper[4859]: I1008 18:36:13.685130 4859 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/402e8a91-7238-4d32-800e-266a6c911802-config-data\") pod \"nova-api-0\" (UID: \"402e8a91-7238-4d32-800e-266a6c911802\") " pod="openstack/nova-api-0" Oct 08 18:36:13 crc kubenswrapper[4859]: I1008 18:36:13.685329 4859 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-gclhh\" (UniqueName: \"kubernetes.io/projected/402e8a91-7238-4d32-800e-266a6c911802-kube-api-access-gclhh\") pod \"nova-api-0\" (UID: \"402e8a91-7238-4d32-800e-266a6c911802\") " pod="openstack/nova-api-0" Oct 08 18:36:13 crc kubenswrapper[4859]: I1008 18:36:13.787260 4859 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/402e8a91-7238-4d32-800e-266a6c911802-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"402e8a91-7238-4d32-800e-266a6c911802\") " pod="openstack/nova-api-0" Oct 08 18:36:13 crc kubenswrapper[4859]: I1008 18:36:13.787618 4859 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/402e8a91-7238-4d32-800e-266a6c911802-logs\") pod \"nova-api-0\" (UID: \"402e8a91-7238-4d32-800e-266a6c911802\") " pod="openstack/nova-api-0" Oct 08 18:36:13 crc kubenswrapper[4859]: I1008 18:36:13.787754 4859 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/402e8a91-7238-4d32-800e-266a6c911802-config-data\") pod \"nova-api-0\" (UID: \"402e8a91-7238-4d32-800e-266a6c911802\") " pod="openstack/nova-api-0" Oct 08 18:36:13 crc kubenswrapper[4859]: I1008 18:36:13.788664 4859 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-gclhh\" (UniqueName: \"kubernetes.io/projected/402e8a91-7238-4d32-800e-266a6c911802-kube-api-access-gclhh\") pod \"nova-api-0\" (UID: \"402e8a91-7238-4d32-800e-266a6c911802\") " pod="openstack/nova-api-0" Oct 08 18:36:13 crc kubenswrapper[4859]: I1008 18:36:13.788388 4859 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/402e8a91-7238-4d32-800e-266a6c911802-logs\") pod \"nova-api-0\" (UID: \"402e8a91-7238-4d32-800e-266a6c911802\") " pod="openstack/nova-api-0" Oct 08 18:36:13 crc kubenswrapper[4859]: I1008 18:36:13.792664 4859 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/402e8a91-7238-4d32-800e-266a6c911802-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"402e8a91-7238-4d32-800e-266a6c911802\") " pod="openstack/nova-api-0" Oct 08 18:36:13 crc kubenswrapper[4859]: I1008 18:36:13.793224 4859 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/402e8a91-7238-4d32-800e-266a6c911802-config-data\") pod \"nova-api-0\" (UID: \"402e8a91-7238-4d32-800e-266a6c911802\") " pod="openstack/nova-api-0" Oct 08 18:36:13 crc kubenswrapper[4859]: I1008 18:36:13.806062 4859 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-gclhh\" (UniqueName: \"kubernetes.io/projected/402e8a91-7238-4d32-800e-266a6c911802-kube-api-access-gclhh\") pod \"nova-api-0\" (UID: \"402e8a91-7238-4d32-800e-266a6c911802\") " pod="openstack/nova-api-0" Oct 08 18:36:13 crc kubenswrapper[4859]: I1008 18:36:13.933058 4859 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Oct 08 18:36:14 crc kubenswrapper[4859]: I1008 18:36:14.207661 4859 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"e5f82275-e5ae-4d2d-8ab6-dc4afda848ba","Type":"ContainerStarted","Data":"4fe8f76dc74c296a52a35017a865d9caf429643b340fffe1bc9156fa36b2c04f"} Oct 08 18:36:14 crc kubenswrapper[4859]: I1008 18:36:14.227100 4859 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-scheduler-0" podStartSLOduration=2.227081593 podStartE2EDuration="2.227081593s" podCreationTimestamp="2025-10-08 18:36:12 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-08 18:36:14.222867133 +0000 UTC m=+1144.469706562" watchObservedRunningTime="2025-10-08 18:36:14.227081593 +0000 UTC m=+1144.473920972" Oct 08 18:36:14 crc kubenswrapper[4859]: I1008 18:36:14.426281 4859 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-0"] Oct 08 18:36:14 crc kubenswrapper[4859]: I1008 18:36:14.488533 4859 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="2f61fa1e-ec4e-490e-91ef-90cf0887940d" path="/var/lib/kubelet/pods/2f61fa1e-ec4e-490e-91ef-90cf0887940d/volumes" Oct 08 18:36:15 crc kubenswrapper[4859]: I1008 18:36:15.234940 4859 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"402e8a91-7238-4d32-800e-266a6c911802","Type":"ContainerStarted","Data":"45e4e70510b3cf6458e879dd30816ac1e0a0d242109940a427935dc1f27be0f1"} Oct 08 18:36:15 crc kubenswrapper[4859]: I1008 18:36:15.235941 4859 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"402e8a91-7238-4d32-800e-266a6c911802","Type":"ContainerStarted","Data":"2684d40d3cd07c780bb9fddbf1a08e723487ca1f6c6f95428d4d5792f5d94759"} Oct 08 18:36:15 crc kubenswrapper[4859]: I1008 18:36:15.236050 4859 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"402e8a91-7238-4d32-800e-266a6c911802","Type":"ContainerStarted","Data":"aab8da64925d89613990ff405f88562e44d85bf2325aa3e367b73d9d135a025e"} Oct 08 18:36:15 crc kubenswrapper[4859]: I1008 18:36:15.260255 4859 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-api-0" podStartSLOduration=2.260236108 podStartE2EDuration="2.260236108s" podCreationTimestamp="2025-10-08 18:36:13 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-08 18:36:15.250113839 +0000 UTC m=+1145.496953238" watchObservedRunningTime="2025-10-08 18:36:15.260236108 +0000 UTC m=+1145.507075487" Oct 08 18:36:17 crc kubenswrapper[4859]: I1008 18:36:17.510320 4859 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-scheduler-0" Oct 08 18:36:18 crc kubenswrapper[4859]: I1008 18:36:18.483954 4859 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-cell1-conductor-0" Oct 08 18:36:18 crc kubenswrapper[4859]: I1008 18:36:18.591246 4859 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-metadata-0" Oct 08 18:36:18 crc kubenswrapper[4859]: I1008 18:36:18.591314 4859 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-metadata-0" Oct 08 18:36:19 crc kubenswrapper[4859]: I1008 18:36:19.602827 4859 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-metadata-0" podUID="f8de039d-85d9-444b-937e-d8d6f90bb77f" containerName="nova-metadata-metadata" probeResult="failure" output="Get \"https://10.217.0.194:8775/\": net/http: request canceled (Client.Timeout exceeded while awaiting headers)" Oct 08 18:36:19 crc kubenswrapper[4859]: I1008 18:36:19.603039 4859 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-metadata-0" podUID="f8de039d-85d9-444b-937e-d8d6f90bb77f" containerName="nova-metadata-log" probeResult="failure" output="Get \"https://10.217.0.194:8775/\": net/http: request canceled (Client.Timeout exceeded while awaiting headers)" Oct 08 18:36:22 crc kubenswrapper[4859]: I1008 18:36:22.510585 4859 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-scheduler-0" Oct 08 18:36:22 crc kubenswrapper[4859]: I1008 18:36:22.542375 4859 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-scheduler-0" Oct 08 18:36:23 crc kubenswrapper[4859]: I1008 18:36:23.346011 4859 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-scheduler-0" Oct 08 18:36:23 crc kubenswrapper[4859]: I1008 18:36:23.933986 4859 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-api-0" Oct 08 18:36:23 crc kubenswrapper[4859]: I1008 18:36:23.934055 4859 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-api-0" Oct 08 18:36:25 crc kubenswrapper[4859]: I1008 18:36:25.017092 4859 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-api-0" podUID="402e8a91-7238-4d32-800e-266a6c911802" containerName="nova-api-log" probeResult="failure" output="Get \"http://10.217.0.196:8774/\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Oct 08 18:36:25 crc kubenswrapper[4859]: I1008 18:36:25.017201 4859 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-api-0" podUID="402e8a91-7238-4d32-800e-266a6c911802" containerName="nova-api-api" probeResult="failure" output="Get \"http://10.217.0.196:8774/\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Oct 08 18:36:28 crc kubenswrapper[4859]: I1008 18:36:28.598065 4859 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-metadata-0" Oct 08 18:36:28 crc kubenswrapper[4859]: I1008 18:36:28.598801 4859 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-metadata-0" Oct 08 18:36:28 crc kubenswrapper[4859]: I1008 18:36:28.611264 4859 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-metadata-0" Oct 08 18:36:28 crc kubenswrapper[4859]: I1008 18:36:28.612227 4859 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-metadata-0" Oct 08 18:36:29 crc kubenswrapper[4859]: E1008 18:36:29.956857 4859 manager.go:1116] Failed to create existing container: /kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod2f61fa1e_ec4e_490e_91ef_90cf0887940d.slice/crio-0d6bcfda0d0f0c376066ff120cd2fbb40dc18afff086c75338e5fa187fd3bf33: Error finding container 0d6bcfda0d0f0c376066ff120cd2fbb40dc18afff086c75338e5fa187fd3bf33: Status 404 returned error can't find the container with id 0d6bcfda0d0f0c376066ff120cd2fbb40dc18afff086c75338e5fa187fd3bf33 Oct 08 18:36:30 crc kubenswrapper[4859]: I1008 18:36:30.363961 4859 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-novncproxy-0" Oct 08 18:36:30 crc kubenswrapper[4859]: I1008 18:36:30.383983 4859 generic.go:334] "Generic (PLEG): container finished" podID="41c36921-eda9-432d-9d21-b478e4e49ee1" containerID="e3197be4f4609dd18ce0c5f1b6a59c8574f21d7c1f4c6c037c5790afe9a61493" exitCode=137 Oct 08 18:36:30 crc kubenswrapper[4859]: I1008 18:36:30.384054 4859 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-novncproxy-0" Oct 08 18:36:30 crc kubenswrapper[4859]: I1008 18:36:30.384043 4859 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-novncproxy-0" event={"ID":"41c36921-eda9-432d-9d21-b478e4e49ee1","Type":"ContainerDied","Data":"e3197be4f4609dd18ce0c5f1b6a59c8574f21d7c1f4c6c037c5790afe9a61493"} Oct 08 18:36:30 crc kubenswrapper[4859]: I1008 18:36:30.390222 4859 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-novncproxy-0" event={"ID":"41c36921-eda9-432d-9d21-b478e4e49ee1","Type":"ContainerDied","Data":"00ee56302b363340528da79c2908dd8a1136515642bde815068e9117bb85782b"} Oct 08 18:36:30 crc kubenswrapper[4859]: I1008 18:36:30.390430 4859 scope.go:117] "RemoveContainer" containerID="e3197be4f4609dd18ce0c5f1b6a59c8574f21d7c1f4c6c037c5790afe9a61493" Oct 08 18:36:30 crc kubenswrapper[4859]: I1008 18:36:30.423119 4859 scope.go:117] "RemoveContainer" containerID="e3197be4f4609dd18ce0c5f1b6a59c8574f21d7c1f4c6c037c5790afe9a61493" Oct 08 18:36:30 crc kubenswrapper[4859]: E1008 18:36:30.423513 4859 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"e3197be4f4609dd18ce0c5f1b6a59c8574f21d7c1f4c6c037c5790afe9a61493\": container with ID starting with e3197be4f4609dd18ce0c5f1b6a59c8574f21d7c1f4c6c037c5790afe9a61493 not found: ID does not exist" containerID="e3197be4f4609dd18ce0c5f1b6a59c8574f21d7c1f4c6c037c5790afe9a61493" Oct 08 18:36:30 crc kubenswrapper[4859]: I1008 18:36:30.423548 4859 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"e3197be4f4609dd18ce0c5f1b6a59c8574f21d7c1f4c6c037c5790afe9a61493"} err="failed to get container status \"e3197be4f4609dd18ce0c5f1b6a59c8574f21d7c1f4c6c037c5790afe9a61493\": rpc error: code = NotFound desc = could not find container \"e3197be4f4609dd18ce0c5f1b6a59c8574f21d7c1f4c6c037c5790afe9a61493\": container with ID starting with e3197be4f4609dd18ce0c5f1b6a59c8574f21d7c1f4c6c037c5790afe9a61493 not found: ID does not exist" Oct 08 18:36:30 crc kubenswrapper[4859]: I1008 18:36:30.508120 4859 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/41c36921-eda9-432d-9d21-b478e4e49ee1-config-data\") pod \"41c36921-eda9-432d-9d21-b478e4e49ee1\" (UID: \"41c36921-eda9-432d-9d21-b478e4e49ee1\") " Oct 08 18:36:30 crc kubenswrapper[4859]: I1008 18:36:30.508172 4859 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-ghzs4\" (UniqueName: \"kubernetes.io/projected/41c36921-eda9-432d-9d21-b478e4e49ee1-kube-api-access-ghzs4\") pod \"41c36921-eda9-432d-9d21-b478e4e49ee1\" (UID: \"41c36921-eda9-432d-9d21-b478e4e49ee1\") " Oct 08 18:36:30 crc kubenswrapper[4859]: I1008 18:36:30.508194 4859 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/41c36921-eda9-432d-9d21-b478e4e49ee1-combined-ca-bundle\") pod \"41c36921-eda9-432d-9d21-b478e4e49ee1\" (UID: \"41c36921-eda9-432d-9d21-b478e4e49ee1\") " Oct 08 18:36:30 crc kubenswrapper[4859]: I1008 18:36:30.513927 4859 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/41c36921-eda9-432d-9d21-b478e4e49ee1-kube-api-access-ghzs4" (OuterVolumeSpecName: "kube-api-access-ghzs4") pod "41c36921-eda9-432d-9d21-b478e4e49ee1" (UID: "41c36921-eda9-432d-9d21-b478e4e49ee1"). InnerVolumeSpecName "kube-api-access-ghzs4". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 08 18:36:30 crc kubenswrapper[4859]: I1008 18:36:30.543446 4859 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/41c36921-eda9-432d-9d21-b478e4e49ee1-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "41c36921-eda9-432d-9d21-b478e4e49ee1" (UID: "41c36921-eda9-432d-9d21-b478e4e49ee1"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 08 18:36:30 crc kubenswrapper[4859]: I1008 18:36:30.547956 4859 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/41c36921-eda9-432d-9d21-b478e4e49ee1-config-data" (OuterVolumeSpecName: "config-data") pod "41c36921-eda9-432d-9d21-b478e4e49ee1" (UID: "41c36921-eda9-432d-9d21-b478e4e49ee1"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 08 18:36:30 crc kubenswrapper[4859]: I1008 18:36:30.610127 4859 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/41c36921-eda9-432d-9d21-b478e4e49ee1-config-data\") on node \"crc\" DevicePath \"\"" Oct 08 18:36:30 crc kubenswrapper[4859]: I1008 18:36:30.610163 4859 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-ghzs4\" (UniqueName: \"kubernetes.io/projected/41c36921-eda9-432d-9d21-b478e4e49ee1-kube-api-access-ghzs4\") on node \"crc\" DevicePath \"\"" Oct 08 18:36:30 crc kubenswrapper[4859]: I1008 18:36:30.610175 4859 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/41c36921-eda9-432d-9d21-b478e4e49ee1-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 08 18:36:30 crc kubenswrapper[4859]: I1008 18:36:30.725820 4859 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell1-novncproxy-0"] Oct 08 18:36:30 crc kubenswrapper[4859]: I1008 18:36:30.735065 4859 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-cell1-novncproxy-0"] Oct 08 18:36:30 crc kubenswrapper[4859]: I1008 18:36:30.746492 4859 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell1-novncproxy-0"] Oct 08 18:36:30 crc kubenswrapper[4859]: E1008 18:36:30.747009 4859 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="41c36921-eda9-432d-9d21-b478e4e49ee1" containerName="nova-cell1-novncproxy-novncproxy" Oct 08 18:36:30 crc kubenswrapper[4859]: I1008 18:36:30.747034 4859 state_mem.go:107] "Deleted CPUSet assignment" podUID="41c36921-eda9-432d-9d21-b478e4e49ee1" containerName="nova-cell1-novncproxy-novncproxy" Oct 08 18:36:30 crc kubenswrapper[4859]: I1008 18:36:30.747316 4859 memory_manager.go:354] "RemoveStaleState removing state" podUID="41c36921-eda9-432d-9d21-b478e4e49ee1" containerName="nova-cell1-novncproxy-novncproxy" Oct 08 18:36:30 crc kubenswrapper[4859]: I1008 18:36:30.748125 4859 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-novncproxy-0" Oct 08 18:36:30 crc kubenswrapper[4859]: I1008 18:36:30.753583 4859 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell1-novncproxy-config-data" Oct 08 18:36:30 crc kubenswrapper[4859]: I1008 18:36:30.757287 4859 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-nova-novncproxy-cell1-vencrypt" Oct 08 18:36:30 crc kubenswrapper[4859]: I1008 18:36:30.763959 4859 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-novncproxy-0"] Oct 08 18:36:30 crc kubenswrapper[4859]: I1008 18:36:30.765519 4859 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-nova-novncproxy-cell1-public-svc" Oct 08 18:36:30 crc kubenswrapper[4859]: I1008 18:36:30.813236 4859 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-nmp6b\" (UniqueName: \"kubernetes.io/projected/f2fffadc-cd8a-4481-aafe-8f7e2dd338cf-kube-api-access-nmp6b\") pod \"nova-cell1-novncproxy-0\" (UID: \"f2fffadc-cd8a-4481-aafe-8f7e2dd338cf\") " pod="openstack/nova-cell1-novncproxy-0" Oct 08 18:36:30 crc kubenswrapper[4859]: I1008 18:36:30.813331 4859 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-novncproxy-tls-certs\" (UniqueName: \"kubernetes.io/secret/f2fffadc-cd8a-4481-aafe-8f7e2dd338cf-nova-novncproxy-tls-certs\") pod \"nova-cell1-novncproxy-0\" (UID: \"f2fffadc-cd8a-4481-aafe-8f7e2dd338cf\") " pod="openstack/nova-cell1-novncproxy-0" Oct 08 18:36:30 crc kubenswrapper[4859]: I1008 18:36:30.813466 4859 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"vencrypt-tls-certs\" (UniqueName: \"kubernetes.io/secret/f2fffadc-cd8a-4481-aafe-8f7e2dd338cf-vencrypt-tls-certs\") pod \"nova-cell1-novncproxy-0\" (UID: \"f2fffadc-cd8a-4481-aafe-8f7e2dd338cf\") " pod="openstack/nova-cell1-novncproxy-0" Oct 08 18:36:30 crc kubenswrapper[4859]: I1008 18:36:30.813509 4859 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f2fffadc-cd8a-4481-aafe-8f7e2dd338cf-config-data\") pod \"nova-cell1-novncproxy-0\" (UID: \"f2fffadc-cd8a-4481-aafe-8f7e2dd338cf\") " pod="openstack/nova-cell1-novncproxy-0" Oct 08 18:36:30 crc kubenswrapper[4859]: I1008 18:36:30.813570 4859 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f2fffadc-cd8a-4481-aafe-8f7e2dd338cf-combined-ca-bundle\") pod \"nova-cell1-novncproxy-0\" (UID: \"f2fffadc-cd8a-4481-aafe-8f7e2dd338cf\") " pod="openstack/nova-cell1-novncproxy-0" Oct 08 18:36:30 crc kubenswrapper[4859]: I1008 18:36:30.914986 4859 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"vencrypt-tls-certs\" (UniqueName: \"kubernetes.io/secret/f2fffadc-cd8a-4481-aafe-8f7e2dd338cf-vencrypt-tls-certs\") pod \"nova-cell1-novncproxy-0\" (UID: \"f2fffadc-cd8a-4481-aafe-8f7e2dd338cf\") " pod="openstack/nova-cell1-novncproxy-0" Oct 08 18:36:30 crc kubenswrapper[4859]: I1008 18:36:30.915033 4859 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f2fffadc-cd8a-4481-aafe-8f7e2dd338cf-config-data\") pod \"nova-cell1-novncproxy-0\" (UID: \"f2fffadc-cd8a-4481-aafe-8f7e2dd338cf\") " pod="openstack/nova-cell1-novncproxy-0" Oct 08 18:36:30 crc kubenswrapper[4859]: I1008 18:36:30.915078 4859 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f2fffadc-cd8a-4481-aafe-8f7e2dd338cf-combined-ca-bundle\") pod \"nova-cell1-novncproxy-0\" (UID: \"f2fffadc-cd8a-4481-aafe-8f7e2dd338cf\") " pod="openstack/nova-cell1-novncproxy-0" Oct 08 18:36:30 crc kubenswrapper[4859]: I1008 18:36:30.915111 4859 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-nmp6b\" (UniqueName: \"kubernetes.io/projected/f2fffadc-cd8a-4481-aafe-8f7e2dd338cf-kube-api-access-nmp6b\") pod \"nova-cell1-novncproxy-0\" (UID: \"f2fffadc-cd8a-4481-aafe-8f7e2dd338cf\") " pod="openstack/nova-cell1-novncproxy-0" Oct 08 18:36:30 crc kubenswrapper[4859]: I1008 18:36:30.915151 4859 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-novncproxy-tls-certs\" (UniqueName: \"kubernetes.io/secret/f2fffadc-cd8a-4481-aafe-8f7e2dd338cf-nova-novncproxy-tls-certs\") pod \"nova-cell1-novncproxy-0\" (UID: \"f2fffadc-cd8a-4481-aafe-8f7e2dd338cf\") " pod="openstack/nova-cell1-novncproxy-0" Oct 08 18:36:30 crc kubenswrapper[4859]: I1008 18:36:30.918824 4859 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-novncproxy-tls-certs\" (UniqueName: \"kubernetes.io/secret/f2fffadc-cd8a-4481-aafe-8f7e2dd338cf-nova-novncproxy-tls-certs\") pod \"nova-cell1-novncproxy-0\" (UID: \"f2fffadc-cd8a-4481-aafe-8f7e2dd338cf\") " pod="openstack/nova-cell1-novncproxy-0" Oct 08 18:36:30 crc kubenswrapper[4859]: I1008 18:36:30.918969 4859 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f2fffadc-cd8a-4481-aafe-8f7e2dd338cf-combined-ca-bundle\") pod \"nova-cell1-novncproxy-0\" (UID: \"f2fffadc-cd8a-4481-aafe-8f7e2dd338cf\") " pod="openstack/nova-cell1-novncproxy-0" Oct 08 18:36:30 crc kubenswrapper[4859]: I1008 18:36:30.920084 4859 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"vencrypt-tls-certs\" (UniqueName: \"kubernetes.io/secret/f2fffadc-cd8a-4481-aafe-8f7e2dd338cf-vencrypt-tls-certs\") pod \"nova-cell1-novncproxy-0\" (UID: \"f2fffadc-cd8a-4481-aafe-8f7e2dd338cf\") " pod="openstack/nova-cell1-novncproxy-0" Oct 08 18:36:30 crc kubenswrapper[4859]: I1008 18:36:30.920267 4859 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f2fffadc-cd8a-4481-aafe-8f7e2dd338cf-config-data\") pod \"nova-cell1-novncproxy-0\" (UID: \"f2fffadc-cd8a-4481-aafe-8f7e2dd338cf\") " pod="openstack/nova-cell1-novncproxy-0" Oct 08 18:36:30 crc kubenswrapper[4859]: I1008 18:36:30.935648 4859 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-nmp6b\" (UniqueName: \"kubernetes.io/projected/f2fffadc-cd8a-4481-aafe-8f7e2dd338cf-kube-api-access-nmp6b\") pod \"nova-cell1-novncproxy-0\" (UID: \"f2fffadc-cd8a-4481-aafe-8f7e2dd338cf\") " pod="openstack/nova-cell1-novncproxy-0" Oct 08 18:36:31 crc kubenswrapper[4859]: I1008 18:36:31.066954 4859 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-novncproxy-0" Oct 08 18:36:31 crc kubenswrapper[4859]: I1008 18:36:31.497260 4859 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-novncproxy-0"] Oct 08 18:36:32 crc kubenswrapper[4859]: I1008 18:36:32.414092 4859 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-novncproxy-0" event={"ID":"f2fffadc-cd8a-4481-aafe-8f7e2dd338cf","Type":"ContainerStarted","Data":"02b04483f2f24dacacc67ac3324584d0ae6306a7d88ec8022c9920d6f5586769"} Oct 08 18:36:32 crc kubenswrapper[4859]: I1008 18:36:32.414515 4859 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-novncproxy-0" event={"ID":"f2fffadc-cd8a-4481-aafe-8f7e2dd338cf","Type":"ContainerStarted","Data":"c1d50d3a7439a2981cbc8db65bd0b0102c5f35681d76847689c84d0f73b5b7e1"} Oct 08 18:36:32 crc kubenswrapper[4859]: I1008 18:36:32.446170 4859 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-cell1-novncproxy-0" podStartSLOduration=2.446151381 podStartE2EDuration="2.446151381s" podCreationTimestamp="2025-10-08 18:36:30 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-08 18:36:32.438978996 +0000 UTC m=+1162.685818385" watchObservedRunningTime="2025-10-08 18:36:32.446151381 +0000 UTC m=+1162.692990760" Oct 08 18:36:32 crc kubenswrapper[4859]: I1008 18:36:32.482573 4859 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="41c36921-eda9-432d-9d21-b478e4e49ee1" path="/var/lib/kubelet/pods/41c36921-eda9-432d-9d21-b478e4e49ee1/volumes" Oct 08 18:36:33 crc kubenswrapper[4859]: I1008 18:36:33.937377 4859 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-api-0" Oct 08 18:36:33 crc kubenswrapper[4859]: I1008 18:36:33.937720 4859 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-api-0" Oct 08 18:36:33 crc kubenswrapper[4859]: I1008 18:36:33.938125 4859 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-api-0" Oct 08 18:36:33 crc kubenswrapper[4859]: I1008 18:36:33.938165 4859 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-api-0" Oct 08 18:36:33 crc kubenswrapper[4859]: I1008 18:36:33.941240 4859 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-api-0" Oct 08 18:36:33 crc kubenswrapper[4859]: I1008 18:36:33.943976 4859 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-api-0" Oct 08 18:36:34 crc kubenswrapper[4859]: I1008 18:36:34.140085 4859 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-64986d45b9-sv5xb"] Oct 08 18:36:34 crc kubenswrapper[4859]: I1008 18:36:34.145862 4859 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-64986d45b9-sv5xb" Oct 08 18:36:34 crc kubenswrapper[4859]: I1008 18:36:34.157886 4859 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-64986d45b9-sv5xb"] Oct 08 18:36:34 crc kubenswrapper[4859]: I1008 18:36:34.291203 4859 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/93af9fa0-cdaa-4665-9120-870f91139ae6-ovsdbserver-nb\") pod \"dnsmasq-dns-64986d45b9-sv5xb\" (UID: \"93af9fa0-cdaa-4665-9120-870f91139ae6\") " pod="openstack/dnsmasq-dns-64986d45b9-sv5xb" Oct 08 18:36:34 crc kubenswrapper[4859]: I1008 18:36:34.291517 4859 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/93af9fa0-cdaa-4665-9120-870f91139ae6-dns-swift-storage-0\") pod \"dnsmasq-dns-64986d45b9-sv5xb\" (UID: \"93af9fa0-cdaa-4665-9120-870f91139ae6\") " pod="openstack/dnsmasq-dns-64986d45b9-sv5xb" Oct 08 18:36:34 crc kubenswrapper[4859]: I1008 18:36:34.291537 4859 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-x4hs7\" (UniqueName: \"kubernetes.io/projected/93af9fa0-cdaa-4665-9120-870f91139ae6-kube-api-access-x4hs7\") pod \"dnsmasq-dns-64986d45b9-sv5xb\" (UID: \"93af9fa0-cdaa-4665-9120-870f91139ae6\") " pod="openstack/dnsmasq-dns-64986d45b9-sv5xb" Oct 08 18:36:34 crc kubenswrapper[4859]: I1008 18:36:34.291559 4859 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/93af9fa0-cdaa-4665-9120-870f91139ae6-dns-svc\") pod \"dnsmasq-dns-64986d45b9-sv5xb\" (UID: \"93af9fa0-cdaa-4665-9120-870f91139ae6\") " pod="openstack/dnsmasq-dns-64986d45b9-sv5xb" Oct 08 18:36:34 crc kubenswrapper[4859]: I1008 18:36:34.291603 4859 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/93af9fa0-cdaa-4665-9120-870f91139ae6-ovsdbserver-sb\") pod \"dnsmasq-dns-64986d45b9-sv5xb\" (UID: \"93af9fa0-cdaa-4665-9120-870f91139ae6\") " pod="openstack/dnsmasq-dns-64986d45b9-sv5xb" Oct 08 18:36:34 crc kubenswrapper[4859]: I1008 18:36:34.291722 4859 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/93af9fa0-cdaa-4665-9120-870f91139ae6-config\") pod \"dnsmasq-dns-64986d45b9-sv5xb\" (UID: \"93af9fa0-cdaa-4665-9120-870f91139ae6\") " pod="openstack/dnsmasq-dns-64986d45b9-sv5xb" Oct 08 18:36:34 crc kubenswrapper[4859]: I1008 18:36:34.393836 4859 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/93af9fa0-cdaa-4665-9120-870f91139ae6-ovsdbserver-nb\") pod \"dnsmasq-dns-64986d45b9-sv5xb\" (UID: \"93af9fa0-cdaa-4665-9120-870f91139ae6\") " pod="openstack/dnsmasq-dns-64986d45b9-sv5xb" Oct 08 18:36:34 crc kubenswrapper[4859]: I1008 18:36:34.393917 4859 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/93af9fa0-cdaa-4665-9120-870f91139ae6-dns-swift-storage-0\") pod \"dnsmasq-dns-64986d45b9-sv5xb\" (UID: \"93af9fa0-cdaa-4665-9120-870f91139ae6\") " pod="openstack/dnsmasq-dns-64986d45b9-sv5xb" Oct 08 18:36:34 crc kubenswrapper[4859]: I1008 18:36:34.393940 4859 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-x4hs7\" (UniqueName: \"kubernetes.io/projected/93af9fa0-cdaa-4665-9120-870f91139ae6-kube-api-access-x4hs7\") pod \"dnsmasq-dns-64986d45b9-sv5xb\" (UID: \"93af9fa0-cdaa-4665-9120-870f91139ae6\") " pod="openstack/dnsmasq-dns-64986d45b9-sv5xb" Oct 08 18:36:34 crc kubenswrapper[4859]: I1008 18:36:34.393960 4859 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/93af9fa0-cdaa-4665-9120-870f91139ae6-dns-svc\") pod \"dnsmasq-dns-64986d45b9-sv5xb\" (UID: \"93af9fa0-cdaa-4665-9120-870f91139ae6\") " pod="openstack/dnsmasq-dns-64986d45b9-sv5xb" Oct 08 18:36:34 crc kubenswrapper[4859]: I1008 18:36:34.394002 4859 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/93af9fa0-cdaa-4665-9120-870f91139ae6-ovsdbserver-sb\") pod \"dnsmasq-dns-64986d45b9-sv5xb\" (UID: \"93af9fa0-cdaa-4665-9120-870f91139ae6\") " pod="openstack/dnsmasq-dns-64986d45b9-sv5xb" Oct 08 18:36:34 crc kubenswrapper[4859]: I1008 18:36:34.394019 4859 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/93af9fa0-cdaa-4665-9120-870f91139ae6-config\") pod \"dnsmasq-dns-64986d45b9-sv5xb\" (UID: \"93af9fa0-cdaa-4665-9120-870f91139ae6\") " pod="openstack/dnsmasq-dns-64986d45b9-sv5xb" Oct 08 18:36:34 crc kubenswrapper[4859]: I1008 18:36:34.394909 4859 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/93af9fa0-cdaa-4665-9120-870f91139ae6-config\") pod \"dnsmasq-dns-64986d45b9-sv5xb\" (UID: \"93af9fa0-cdaa-4665-9120-870f91139ae6\") " pod="openstack/dnsmasq-dns-64986d45b9-sv5xb" Oct 08 18:36:34 crc kubenswrapper[4859]: I1008 18:36:34.395444 4859 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/93af9fa0-cdaa-4665-9120-870f91139ae6-ovsdbserver-nb\") pod \"dnsmasq-dns-64986d45b9-sv5xb\" (UID: \"93af9fa0-cdaa-4665-9120-870f91139ae6\") " pod="openstack/dnsmasq-dns-64986d45b9-sv5xb" Oct 08 18:36:34 crc kubenswrapper[4859]: I1008 18:36:34.396020 4859 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/93af9fa0-cdaa-4665-9120-870f91139ae6-dns-swift-storage-0\") pod \"dnsmasq-dns-64986d45b9-sv5xb\" (UID: \"93af9fa0-cdaa-4665-9120-870f91139ae6\") " pod="openstack/dnsmasq-dns-64986d45b9-sv5xb" Oct 08 18:36:34 crc kubenswrapper[4859]: I1008 18:36:34.397187 4859 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/93af9fa0-cdaa-4665-9120-870f91139ae6-ovsdbserver-sb\") pod \"dnsmasq-dns-64986d45b9-sv5xb\" (UID: \"93af9fa0-cdaa-4665-9120-870f91139ae6\") " pod="openstack/dnsmasq-dns-64986d45b9-sv5xb" Oct 08 18:36:34 crc kubenswrapper[4859]: I1008 18:36:34.397304 4859 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/93af9fa0-cdaa-4665-9120-870f91139ae6-dns-svc\") pod \"dnsmasq-dns-64986d45b9-sv5xb\" (UID: \"93af9fa0-cdaa-4665-9120-870f91139ae6\") " pod="openstack/dnsmasq-dns-64986d45b9-sv5xb" Oct 08 18:36:34 crc kubenswrapper[4859]: I1008 18:36:34.430279 4859 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-x4hs7\" (UniqueName: \"kubernetes.io/projected/93af9fa0-cdaa-4665-9120-870f91139ae6-kube-api-access-x4hs7\") pod \"dnsmasq-dns-64986d45b9-sv5xb\" (UID: \"93af9fa0-cdaa-4665-9120-870f91139ae6\") " pod="openstack/dnsmasq-dns-64986d45b9-sv5xb" Oct 08 18:36:34 crc kubenswrapper[4859]: I1008 18:36:34.468779 4859 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-64986d45b9-sv5xb" Oct 08 18:36:34 crc kubenswrapper[4859]: I1008 18:36:34.964676 4859 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-64986d45b9-sv5xb"] Oct 08 18:36:35 crc kubenswrapper[4859]: I1008 18:36:35.445105 4859 generic.go:334] "Generic (PLEG): container finished" podID="93af9fa0-cdaa-4665-9120-870f91139ae6" containerID="c181b0f9e95d88022634d1baf22f6c42e474a38ddaee7084537b73bb541295e9" exitCode=0 Oct 08 18:36:35 crc kubenswrapper[4859]: I1008 18:36:35.445194 4859 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-64986d45b9-sv5xb" event={"ID":"93af9fa0-cdaa-4665-9120-870f91139ae6","Type":"ContainerDied","Data":"c181b0f9e95d88022634d1baf22f6c42e474a38ddaee7084537b73bb541295e9"} Oct 08 18:36:35 crc kubenswrapper[4859]: I1008 18:36:35.445818 4859 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-64986d45b9-sv5xb" event={"ID":"93af9fa0-cdaa-4665-9120-870f91139ae6","Type":"ContainerStarted","Data":"5012775aaa2f789a4fee328b2d9dbbed704068459b990b56c1bba3e39026678b"} Oct 08 18:36:36 crc kubenswrapper[4859]: I1008 18:36:36.067846 4859 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-cell1-novncproxy-0" Oct 08 18:36:36 crc kubenswrapper[4859]: I1008 18:36:36.128289 4859 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Oct 08 18:36:36 crc kubenswrapper[4859]: I1008 18:36:36.128600 4859 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="135723dc-9767-4d2e-bf09-ad570aa5ee6d" containerName="ceilometer-central-agent" containerID="cri-o://e8b8ab1aa42618805f0c5cb12f6ef9a1a2f4d979e05d79124e59dbd38ebec03f" gracePeriod=30 Oct 08 18:36:36 crc kubenswrapper[4859]: I1008 18:36:36.128624 4859 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="135723dc-9767-4d2e-bf09-ad570aa5ee6d" containerName="proxy-httpd" containerID="cri-o://29d177c46e1eabafffe72b4a4a14e55bd3dafb7cc676aa4b7d776e4de32f5283" gracePeriod=30 Oct 08 18:36:36 crc kubenswrapper[4859]: I1008 18:36:36.128706 4859 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="135723dc-9767-4d2e-bf09-ad570aa5ee6d" containerName="sg-core" containerID="cri-o://06338d13aaf10078c779ee7529acb2d54d17c0372c55baae66dd438f77d44c41" gracePeriod=30 Oct 08 18:36:36 crc kubenswrapper[4859]: I1008 18:36:36.128754 4859 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="135723dc-9767-4d2e-bf09-ad570aa5ee6d" containerName="ceilometer-notification-agent" containerID="cri-o://55cd54864a5d4700d7ce2a2c53eb77e490165f5c160f17e3483c4203893fad70" gracePeriod=30 Oct 08 18:36:36 crc kubenswrapper[4859]: I1008 18:36:36.458578 4859 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-64986d45b9-sv5xb" event={"ID":"93af9fa0-cdaa-4665-9120-870f91139ae6","Type":"ContainerStarted","Data":"2f3570ff9aac41b4211ea2b76be831b98d152e30595b7dface9313ee3be68aed"} Oct 08 18:36:36 crc kubenswrapper[4859]: I1008 18:36:36.459374 4859 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-64986d45b9-sv5xb" Oct 08 18:36:36 crc kubenswrapper[4859]: I1008 18:36:36.470290 4859 generic.go:334] "Generic (PLEG): container finished" podID="135723dc-9767-4d2e-bf09-ad570aa5ee6d" containerID="29d177c46e1eabafffe72b4a4a14e55bd3dafb7cc676aa4b7d776e4de32f5283" exitCode=0 Oct 08 18:36:36 crc kubenswrapper[4859]: I1008 18:36:36.470326 4859 generic.go:334] "Generic (PLEG): container finished" podID="135723dc-9767-4d2e-bf09-ad570aa5ee6d" containerID="06338d13aaf10078c779ee7529acb2d54d17c0372c55baae66dd438f77d44c41" exitCode=2 Oct 08 18:36:36 crc kubenswrapper[4859]: I1008 18:36:36.470333 4859 generic.go:334] "Generic (PLEG): container finished" podID="135723dc-9767-4d2e-bf09-ad570aa5ee6d" containerID="e8b8ab1aa42618805f0c5cb12f6ef9a1a2f4d979e05d79124e59dbd38ebec03f" exitCode=0 Oct 08 18:36:36 crc kubenswrapper[4859]: I1008 18:36:36.484378 4859 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-64986d45b9-sv5xb" podStartSLOduration=2.484360604 podStartE2EDuration="2.484360604s" podCreationTimestamp="2025-10-08 18:36:34 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-08 18:36:36.482405908 +0000 UTC m=+1166.729245297" watchObservedRunningTime="2025-10-08 18:36:36.484360604 +0000 UTC m=+1166.731199973" Oct 08 18:36:36 crc kubenswrapper[4859]: I1008 18:36:36.496611 4859 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"135723dc-9767-4d2e-bf09-ad570aa5ee6d","Type":"ContainerDied","Data":"29d177c46e1eabafffe72b4a4a14e55bd3dafb7cc676aa4b7d776e4de32f5283"} Oct 08 18:36:36 crc kubenswrapper[4859]: I1008 18:36:36.496649 4859 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"135723dc-9767-4d2e-bf09-ad570aa5ee6d","Type":"ContainerDied","Data":"06338d13aaf10078c779ee7529acb2d54d17c0372c55baae66dd438f77d44c41"} Oct 08 18:36:36 crc kubenswrapper[4859]: I1008 18:36:36.496658 4859 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"135723dc-9767-4d2e-bf09-ad570aa5ee6d","Type":"ContainerDied","Data":"e8b8ab1aa42618805f0c5cb12f6ef9a1a2f4d979e05d79124e59dbd38ebec03f"} Oct 08 18:36:36 crc kubenswrapper[4859]: I1008 18:36:36.548144 4859 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-api-0"] Oct 08 18:36:36 crc kubenswrapper[4859]: I1008 18:36:36.548347 4859 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-api-0" podUID="402e8a91-7238-4d32-800e-266a6c911802" containerName="nova-api-log" containerID="cri-o://2684d40d3cd07c780bb9fddbf1a08e723487ca1f6c6f95428d4d5792f5d94759" gracePeriod=30 Oct 08 18:36:36 crc kubenswrapper[4859]: I1008 18:36:36.548465 4859 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-api-0" podUID="402e8a91-7238-4d32-800e-266a6c911802" containerName="nova-api-api" containerID="cri-o://45e4e70510b3cf6458e879dd30816ac1e0a0d242109940a427935dc1f27be0f1" gracePeriod=30 Oct 08 18:36:37 crc kubenswrapper[4859]: I1008 18:36:37.481701 4859 generic.go:334] "Generic (PLEG): container finished" podID="402e8a91-7238-4d32-800e-266a6c911802" containerID="2684d40d3cd07c780bb9fddbf1a08e723487ca1f6c6f95428d4d5792f5d94759" exitCode=143 Oct 08 18:36:37 crc kubenswrapper[4859]: I1008 18:36:37.481785 4859 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"402e8a91-7238-4d32-800e-266a6c911802","Type":"ContainerDied","Data":"2684d40d3cd07c780bb9fddbf1a08e723487ca1f6c6f95428d4d5792f5d94759"} Oct 08 18:36:40 crc kubenswrapper[4859]: I1008 18:36:40.201194 4859 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Oct 08 18:36:40 crc kubenswrapper[4859]: I1008 18:36:40.283071 4859 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Oct 08 18:36:40 crc kubenswrapper[4859]: I1008 18:36:40.316571 4859 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/402e8a91-7238-4d32-800e-266a6c911802-combined-ca-bundle\") pod \"402e8a91-7238-4d32-800e-266a6c911802\" (UID: \"402e8a91-7238-4d32-800e-266a6c911802\") " Oct 08 18:36:40 crc kubenswrapper[4859]: I1008 18:36:40.316751 4859 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/402e8a91-7238-4d32-800e-266a6c911802-config-data\") pod \"402e8a91-7238-4d32-800e-266a6c911802\" (UID: \"402e8a91-7238-4d32-800e-266a6c911802\") " Oct 08 18:36:40 crc kubenswrapper[4859]: I1008 18:36:40.316827 4859 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/402e8a91-7238-4d32-800e-266a6c911802-logs\") pod \"402e8a91-7238-4d32-800e-266a6c911802\" (UID: \"402e8a91-7238-4d32-800e-266a6c911802\") " Oct 08 18:36:40 crc kubenswrapper[4859]: I1008 18:36:40.316872 4859 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-gclhh\" (UniqueName: \"kubernetes.io/projected/402e8a91-7238-4d32-800e-266a6c911802-kube-api-access-gclhh\") pod \"402e8a91-7238-4d32-800e-266a6c911802\" (UID: \"402e8a91-7238-4d32-800e-266a6c911802\") " Oct 08 18:36:40 crc kubenswrapper[4859]: I1008 18:36:40.319267 4859 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/402e8a91-7238-4d32-800e-266a6c911802-logs" (OuterVolumeSpecName: "logs") pod "402e8a91-7238-4d32-800e-266a6c911802" (UID: "402e8a91-7238-4d32-800e-266a6c911802"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 08 18:36:40 crc kubenswrapper[4859]: I1008 18:36:40.325835 4859 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/402e8a91-7238-4d32-800e-266a6c911802-kube-api-access-gclhh" (OuterVolumeSpecName: "kube-api-access-gclhh") pod "402e8a91-7238-4d32-800e-266a6c911802" (UID: "402e8a91-7238-4d32-800e-266a6c911802"). InnerVolumeSpecName "kube-api-access-gclhh". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 08 18:36:40 crc kubenswrapper[4859]: I1008 18:36:40.349043 4859 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/402e8a91-7238-4d32-800e-266a6c911802-config-data" (OuterVolumeSpecName: "config-data") pod "402e8a91-7238-4d32-800e-266a6c911802" (UID: "402e8a91-7238-4d32-800e-266a6c911802"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 08 18:36:40 crc kubenswrapper[4859]: I1008 18:36:40.350871 4859 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/402e8a91-7238-4d32-800e-266a6c911802-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "402e8a91-7238-4d32-800e-266a6c911802" (UID: "402e8a91-7238-4d32-800e-266a6c911802"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 08 18:36:40 crc kubenswrapper[4859]: I1008 18:36:40.418012 4859 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/135723dc-9767-4d2e-bf09-ad570aa5ee6d-log-httpd\") pod \"135723dc-9767-4d2e-bf09-ad570aa5ee6d\" (UID: \"135723dc-9767-4d2e-bf09-ad570aa5ee6d\") " Oct 08 18:36:40 crc kubenswrapper[4859]: I1008 18:36:40.418152 4859 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/135723dc-9767-4d2e-bf09-ad570aa5ee6d-run-httpd\") pod \"135723dc-9767-4d2e-bf09-ad570aa5ee6d\" (UID: \"135723dc-9767-4d2e-bf09-ad570aa5ee6d\") " Oct 08 18:36:40 crc kubenswrapper[4859]: I1008 18:36:40.418238 4859 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/135723dc-9767-4d2e-bf09-ad570aa5ee6d-scripts\") pod \"135723dc-9767-4d2e-bf09-ad570aa5ee6d\" (UID: \"135723dc-9767-4d2e-bf09-ad570aa5ee6d\") " Oct 08 18:36:40 crc kubenswrapper[4859]: I1008 18:36:40.418280 4859 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/135723dc-9767-4d2e-bf09-ad570aa5ee6d-ceilometer-tls-certs\") pod \"135723dc-9767-4d2e-bf09-ad570aa5ee6d\" (UID: \"135723dc-9767-4d2e-bf09-ad570aa5ee6d\") " Oct 08 18:36:40 crc kubenswrapper[4859]: I1008 18:36:40.418325 4859 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/135723dc-9767-4d2e-bf09-ad570aa5ee6d-sg-core-conf-yaml\") pod \"135723dc-9767-4d2e-bf09-ad570aa5ee6d\" (UID: \"135723dc-9767-4d2e-bf09-ad570aa5ee6d\") " Oct 08 18:36:40 crc kubenswrapper[4859]: I1008 18:36:40.418374 4859 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/135723dc-9767-4d2e-bf09-ad570aa5ee6d-combined-ca-bundle\") pod \"135723dc-9767-4d2e-bf09-ad570aa5ee6d\" (UID: \"135723dc-9767-4d2e-bf09-ad570aa5ee6d\") " Oct 08 18:36:40 crc kubenswrapper[4859]: I1008 18:36:40.418396 4859 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/135723dc-9767-4d2e-bf09-ad570aa5ee6d-run-httpd" (OuterVolumeSpecName: "run-httpd") pod "135723dc-9767-4d2e-bf09-ad570aa5ee6d" (UID: "135723dc-9767-4d2e-bf09-ad570aa5ee6d"). InnerVolumeSpecName "run-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 08 18:36:40 crc kubenswrapper[4859]: I1008 18:36:40.418414 4859 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-zpzk8\" (UniqueName: \"kubernetes.io/projected/135723dc-9767-4d2e-bf09-ad570aa5ee6d-kube-api-access-zpzk8\") pod \"135723dc-9767-4d2e-bf09-ad570aa5ee6d\" (UID: \"135723dc-9767-4d2e-bf09-ad570aa5ee6d\") " Oct 08 18:36:40 crc kubenswrapper[4859]: I1008 18:36:40.418449 4859 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/135723dc-9767-4d2e-bf09-ad570aa5ee6d-log-httpd" (OuterVolumeSpecName: "log-httpd") pod "135723dc-9767-4d2e-bf09-ad570aa5ee6d" (UID: "135723dc-9767-4d2e-bf09-ad570aa5ee6d"). InnerVolumeSpecName "log-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 08 18:36:40 crc kubenswrapper[4859]: I1008 18:36:40.418495 4859 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/135723dc-9767-4d2e-bf09-ad570aa5ee6d-config-data\") pod \"135723dc-9767-4d2e-bf09-ad570aa5ee6d\" (UID: \"135723dc-9767-4d2e-bf09-ad570aa5ee6d\") " Oct 08 18:36:40 crc kubenswrapper[4859]: I1008 18:36:40.419292 4859 reconciler_common.go:293] "Volume detached for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/135723dc-9767-4d2e-bf09-ad570aa5ee6d-log-httpd\") on node \"crc\" DevicePath \"\"" Oct 08 18:36:40 crc kubenswrapper[4859]: I1008 18:36:40.419306 4859 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/402e8a91-7238-4d32-800e-266a6c911802-logs\") on node \"crc\" DevicePath \"\"" Oct 08 18:36:40 crc kubenswrapper[4859]: I1008 18:36:40.419316 4859 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-gclhh\" (UniqueName: \"kubernetes.io/projected/402e8a91-7238-4d32-800e-266a6c911802-kube-api-access-gclhh\") on node \"crc\" DevicePath \"\"" Oct 08 18:36:40 crc kubenswrapper[4859]: I1008 18:36:40.419326 4859 reconciler_common.go:293] "Volume detached for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/135723dc-9767-4d2e-bf09-ad570aa5ee6d-run-httpd\") on node \"crc\" DevicePath \"\"" Oct 08 18:36:40 crc kubenswrapper[4859]: I1008 18:36:40.419333 4859 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/402e8a91-7238-4d32-800e-266a6c911802-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 08 18:36:40 crc kubenswrapper[4859]: I1008 18:36:40.419341 4859 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/402e8a91-7238-4d32-800e-266a6c911802-config-data\") on node \"crc\" DevicePath \"\"" Oct 08 18:36:40 crc kubenswrapper[4859]: I1008 18:36:40.423145 4859 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/135723dc-9767-4d2e-bf09-ad570aa5ee6d-scripts" (OuterVolumeSpecName: "scripts") pod "135723dc-9767-4d2e-bf09-ad570aa5ee6d" (UID: "135723dc-9767-4d2e-bf09-ad570aa5ee6d"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 08 18:36:40 crc kubenswrapper[4859]: I1008 18:36:40.423636 4859 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/135723dc-9767-4d2e-bf09-ad570aa5ee6d-kube-api-access-zpzk8" (OuterVolumeSpecName: "kube-api-access-zpzk8") pod "135723dc-9767-4d2e-bf09-ad570aa5ee6d" (UID: "135723dc-9767-4d2e-bf09-ad570aa5ee6d"). InnerVolumeSpecName "kube-api-access-zpzk8". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 08 18:36:40 crc kubenswrapper[4859]: I1008 18:36:40.483233 4859 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/135723dc-9767-4d2e-bf09-ad570aa5ee6d-sg-core-conf-yaml" (OuterVolumeSpecName: "sg-core-conf-yaml") pod "135723dc-9767-4d2e-bf09-ad570aa5ee6d" (UID: "135723dc-9767-4d2e-bf09-ad570aa5ee6d"). InnerVolumeSpecName "sg-core-conf-yaml". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 08 18:36:40 crc kubenswrapper[4859]: I1008 18:36:40.503539 4859 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/135723dc-9767-4d2e-bf09-ad570aa5ee6d-ceilometer-tls-certs" (OuterVolumeSpecName: "ceilometer-tls-certs") pod "135723dc-9767-4d2e-bf09-ad570aa5ee6d" (UID: "135723dc-9767-4d2e-bf09-ad570aa5ee6d"). InnerVolumeSpecName "ceilometer-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 08 18:36:40 crc kubenswrapper[4859]: I1008 18:36:40.514648 4859 generic.go:334] "Generic (PLEG): container finished" podID="135723dc-9767-4d2e-bf09-ad570aa5ee6d" containerID="55cd54864a5d4700d7ce2a2c53eb77e490165f5c160f17e3483c4203893fad70" exitCode=0 Oct 08 18:36:40 crc kubenswrapper[4859]: I1008 18:36:40.514720 4859 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Oct 08 18:36:40 crc kubenswrapper[4859]: I1008 18:36:40.514726 4859 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"135723dc-9767-4d2e-bf09-ad570aa5ee6d","Type":"ContainerDied","Data":"55cd54864a5d4700d7ce2a2c53eb77e490165f5c160f17e3483c4203893fad70"} Oct 08 18:36:40 crc kubenswrapper[4859]: I1008 18:36:40.515719 4859 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"135723dc-9767-4d2e-bf09-ad570aa5ee6d","Type":"ContainerDied","Data":"8c0c6f9e880bbe9e339983489e0b0f50589049373e11f98fc1645b2c18b63742"} Oct 08 18:36:40 crc kubenswrapper[4859]: I1008 18:36:40.515744 4859 scope.go:117] "RemoveContainer" containerID="29d177c46e1eabafffe72b4a4a14e55bd3dafb7cc676aa4b7d776e4de32f5283" Oct 08 18:36:40 crc kubenswrapper[4859]: I1008 18:36:40.520234 4859 generic.go:334] "Generic (PLEG): container finished" podID="402e8a91-7238-4d32-800e-266a6c911802" containerID="45e4e70510b3cf6458e879dd30816ac1e0a0d242109940a427935dc1f27be0f1" exitCode=0 Oct 08 18:36:40 crc kubenswrapper[4859]: I1008 18:36:40.520291 4859 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Oct 08 18:36:40 crc kubenswrapper[4859]: I1008 18:36:40.520278 4859 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"402e8a91-7238-4d32-800e-266a6c911802","Type":"ContainerDied","Data":"45e4e70510b3cf6458e879dd30816ac1e0a0d242109940a427935dc1f27be0f1"} Oct 08 18:36:40 crc kubenswrapper[4859]: I1008 18:36:40.520502 4859 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"402e8a91-7238-4d32-800e-266a6c911802","Type":"ContainerDied","Data":"aab8da64925d89613990ff405f88562e44d85bf2325aa3e367b73d9d135a025e"} Oct 08 18:36:40 crc kubenswrapper[4859]: I1008 18:36:40.521971 4859 reconciler_common.go:293] "Volume detached for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/135723dc-9767-4d2e-bf09-ad570aa5ee6d-ceilometer-tls-certs\") on node \"crc\" DevicePath \"\"" Oct 08 18:36:40 crc kubenswrapper[4859]: I1008 18:36:40.522130 4859 reconciler_common.go:293] "Volume detached for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/135723dc-9767-4d2e-bf09-ad570aa5ee6d-sg-core-conf-yaml\") on node \"crc\" DevicePath \"\"" Oct 08 18:36:40 crc kubenswrapper[4859]: I1008 18:36:40.522385 4859 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-zpzk8\" (UniqueName: \"kubernetes.io/projected/135723dc-9767-4d2e-bf09-ad570aa5ee6d-kube-api-access-zpzk8\") on node \"crc\" DevicePath \"\"" Oct 08 18:36:40 crc kubenswrapper[4859]: I1008 18:36:40.522727 4859 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/135723dc-9767-4d2e-bf09-ad570aa5ee6d-scripts\") on node \"crc\" DevicePath \"\"" Oct 08 18:36:40 crc kubenswrapper[4859]: E1008 18:36:40.544733 4859 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod402e8a91_7238_4d32_800e_266a6c911802.slice/crio-aab8da64925d89613990ff405f88562e44d85bf2325aa3e367b73d9d135a025e\": RecentStats: unable to find data in memory cache]" Oct 08 18:36:40 crc kubenswrapper[4859]: I1008 18:36:40.575612 4859 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/135723dc-9767-4d2e-bf09-ad570aa5ee6d-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "135723dc-9767-4d2e-bf09-ad570aa5ee6d" (UID: "135723dc-9767-4d2e-bf09-ad570aa5ee6d"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 08 18:36:40 crc kubenswrapper[4859]: I1008 18:36:40.589040 4859 scope.go:117] "RemoveContainer" containerID="06338d13aaf10078c779ee7529acb2d54d17c0372c55baae66dd438f77d44c41" Oct 08 18:36:40 crc kubenswrapper[4859]: I1008 18:36:40.622515 4859 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-api-0"] Oct 08 18:36:40 crc kubenswrapper[4859]: I1008 18:36:40.635874 4859 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-api-0"] Oct 08 18:36:40 crc kubenswrapper[4859]: I1008 18:36:40.639427 4859 scope.go:117] "RemoveContainer" containerID="55cd54864a5d4700d7ce2a2c53eb77e490165f5c160f17e3483c4203893fad70" Oct 08 18:36:40 crc kubenswrapper[4859]: I1008 18:36:40.639624 4859 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/135723dc-9767-4d2e-bf09-ad570aa5ee6d-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 08 18:36:40 crc kubenswrapper[4859]: I1008 18:36:40.649537 4859 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/135723dc-9767-4d2e-bf09-ad570aa5ee6d-config-data" (OuterVolumeSpecName: "config-data") pod "135723dc-9767-4d2e-bf09-ad570aa5ee6d" (UID: "135723dc-9767-4d2e-bf09-ad570aa5ee6d"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 08 18:36:40 crc kubenswrapper[4859]: I1008 18:36:40.661261 4859 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-api-0"] Oct 08 18:36:40 crc kubenswrapper[4859]: E1008 18:36:40.661696 4859 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="402e8a91-7238-4d32-800e-266a6c911802" containerName="nova-api-log" Oct 08 18:36:40 crc kubenswrapper[4859]: I1008 18:36:40.661717 4859 state_mem.go:107] "Deleted CPUSet assignment" podUID="402e8a91-7238-4d32-800e-266a6c911802" containerName="nova-api-log" Oct 08 18:36:40 crc kubenswrapper[4859]: E1008 18:36:40.661742 4859 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="402e8a91-7238-4d32-800e-266a6c911802" containerName="nova-api-api" Oct 08 18:36:40 crc kubenswrapper[4859]: I1008 18:36:40.661749 4859 state_mem.go:107] "Deleted CPUSet assignment" podUID="402e8a91-7238-4d32-800e-266a6c911802" containerName="nova-api-api" Oct 08 18:36:40 crc kubenswrapper[4859]: E1008 18:36:40.661763 4859 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="135723dc-9767-4d2e-bf09-ad570aa5ee6d" containerName="ceilometer-notification-agent" Oct 08 18:36:40 crc kubenswrapper[4859]: I1008 18:36:40.661769 4859 state_mem.go:107] "Deleted CPUSet assignment" podUID="135723dc-9767-4d2e-bf09-ad570aa5ee6d" containerName="ceilometer-notification-agent" Oct 08 18:36:40 crc kubenswrapper[4859]: E1008 18:36:40.661782 4859 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="135723dc-9767-4d2e-bf09-ad570aa5ee6d" containerName="ceilometer-central-agent" Oct 08 18:36:40 crc kubenswrapper[4859]: I1008 18:36:40.661788 4859 state_mem.go:107] "Deleted CPUSet assignment" podUID="135723dc-9767-4d2e-bf09-ad570aa5ee6d" containerName="ceilometer-central-agent" Oct 08 18:36:40 crc kubenswrapper[4859]: E1008 18:36:40.661798 4859 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="135723dc-9767-4d2e-bf09-ad570aa5ee6d" containerName="proxy-httpd" Oct 08 18:36:40 crc kubenswrapper[4859]: I1008 18:36:40.661804 4859 state_mem.go:107] "Deleted CPUSet assignment" podUID="135723dc-9767-4d2e-bf09-ad570aa5ee6d" containerName="proxy-httpd" Oct 08 18:36:40 crc kubenswrapper[4859]: E1008 18:36:40.661822 4859 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="135723dc-9767-4d2e-bf09-ad570aa5ee6d" containerName="sg-core" Oct 08 18:36:40 crc kubenswrapper[4859]: I1008 18:36:40.661828 4859 state_mem.go:107] "Deleted CPUSet assignment" podUID="135723dc-9767-4d2e-bf09-ad570aa5ee6d" containerName="sg-core" Oct 08 18:36:40 crc kubenswrapper[4859]: I1008 18:36:40.661988 4859 memory_manager.go:354] "RemoveStaleState removing state" podUID="402e8a91-7238-4d32-800e-266a6c911802" containerName="nova-api-api" Oct 08 18:36:40 crc kubenswrapper[4859]: I1008 18:36:40.661997 4859 memory_manager.go:354] "RemoveStaleState removing state" podUID="135723dc-9767-4d2e-bf09-ad570aa5ee6d" containerName="sg-core" Oct 08 18:36:40 crc kubenswrapper[4859]: I1008 18:36:40.662010 4859 memory_manager.go:354] "RemoveStaleState removing state" podUID="135723dc-9767-4d2e-bf09-ad570aa5ee6d" containerName="proxy-httpd" Oct 08 18:36:40 crc kubenswrapper[4859]: I1008 18:36:40.662020 4859 memory_manager.go:354] "RemoveStaleState removing state" podUID="135723dc-9767-4d2e-bf09-ad570aa5ee6d" containerName="ceilometer-central-agent" Oct 08 18:36:40 crc kubenswrapper[4859]: I1008 18:36:40.662032 4859 memory_manager.go:354] "RemoveStaleState removing state" podUID="402e8a91-7238-4d32-800e-266a6c911802" containerName="nova-api-log" Oct 08 18:36:40 crc kubenswrapper[4859]: I1008 18:36:40.662054 4859 memory_manager.go:354] "RemoveStaleState removing state" podUID="135723dc-9767-4d2e-bf09-ad570aa5ee6d" containerName="ceilometer-notification-agent" Oct 08 18:36:40 crc kubenswrapper[4859]: I1008 18:36:40.663132 4859 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Oct 08 18:36:40 crc kubenswrapper[4859]: I1008 18:36:40.665731 4859 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-nova-public-svc" Oct 08 18:36:40 crc kubenswrapper[4859]: I1008 18:36:40.666055 4859 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-nova-internal-svc" Oct 08 18:36:40 crc kubenswrapper[4859]: I1008 18:36:40.666266 4859 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-api-config-data" Oct 08 18:36:40 crc kubenswrapper[4859]: I1008 18:36:40.682060 4859 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-0"] Oct 08 18:36:40 crc kubenswrapper[4859]: I1008 18:36:40.689367 4859 scope.go:117] "RemoveContainer" containerID="e8b8ab1aa42618805f0c5cb12f6ef9a1a2f4d979e05d79124e59dbd38ebec03f" Oct 08 18:36:40 crc kubenswrapper[4859]: I1008 18:36:40.713648 4859 scope.go:117] "RemoveContainer" containerID="29d177c46e1eabafffe72b4a4a14e55bd3dafb7cc676aa4b7d776e4de32f5283" Oct 08 18:36:40 crc kubenswrapper[4859]: E1008 18:36:40.714236 4859 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"29d177c46e1eabafffe72b4a4a14e55bd3dafb7cc676aa4b7d776e4de32f5283\": container with ID starting with 29d177c46e1eabafffe72b4a4a14e55bd3dafb7cc676aa4b7d776e4de32f5283 not found: ID does not exist" containerID="29d177c46e1eabafffe72b4a4a14e55bd3dafb7cc676aa4b7d776e4de32f5283" Oct 08 18:36:40 crc kubenswrapper[4859]: I1008 18:36:40.714265 4859 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"29d177c46e1eabafffe72b4a4a14e55bd3dafb7cc676aa4b7d776e4de32f5283"} err="failed to get container status \"29d177c46e1eabafffe72b4a4a14e55bd3dafb7cc676aa4b7d776e4de32f5283\": rpc error: code = NotFound desc = could not find container \"29d177c46e1eabafffe72b4a4a14e55bd3dafb7cc676aa4b7d776e4de32f5283\": container with ID starting with 29d177c46e1eabafffe72b4a4a14e55bd3dafb7cc676aa4b7d776e4de32f5283 not found: ID does not exist" Oct 08 18:36:40 crc kubenswrapper[4859]: I1008 18:36:40.714303 4859 scope.go:117] "RemoveContainer" containerID="06338d13aaf10078c779ee7529acb2d54d17c0372c55baae66dd438f77d44c41" Oct 08 18:36:40 crc kubenswrapper[4859]: E1008 18:36:40.714565 4859 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"06338d13aaf10078c779ee7529acb2d54d17c0372c55baae66dd438f77d44c41\": container with ID starting with 06338d13aaf10078c779ee7529acb2d54d17c0372c55baae66dd438f77d44c41 not found: ID does not exist" containerID="06338d13aaf10078c779ee7529acb2d54d17c0372c55baae66dd438f77d44c41" Oct 08 18:36:40 crc kubenswrapper[4859]: I1008 18:36:40.714587 4859 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"06338d13aaf10078c779ee7529acb2d54d17c0372c55baae66dd438f77d44c41"} err="failed to get container status \"06338d13aaf10078c779ee7529acb2d54d17c0372c55baae66dd438f77d44c41\": rpc error: code = NotFound desc = could not find container \"06338d13aaf10078c779ee7529acb2d54d17c0372c55baae66dd438f77d44c41\": container with ID starting with 06338d13aaf10078c779ee7529acb2d54d17c0372c55baae66dd438f77d44c41 not found: ID does not exist" Oct 08 18:36:40 crc kubenswrapper[4859]: I1008 18:36:40.714605 4859 scope.go:117] "RemoveContainer" containerID="55cd54864a5d4700d7ce2a2c53eb77e490165f5c160f17e3483c4203893fad70" Oct 08 18:36:40 crc kubenswrapper[4859]: E1008 18:36:40.714811 4859 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"55cd54864a5d4700d7ce2a2c53eb77e490165f5c160f17e3483c4203893fad70\": container with ID starting with 55cd54864a5d4700d7ce2a2c53eb77e490165f5c160f17e3483c4203893fad70 not found: ID does not exist" containerID="55cd54864a5d4700d7ce2a2c53eb77e490165f5c160f17e3483c4203893fad70" Oct 08 18:36:40 crc kubenswrapper[4859]: I1008 18:36:40.714843 4859 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"55cd54864a5d4700d7ce2a2c53eb77e490165f5c160f17e3483c4203893fad70"} err="failed to get container status \"55cd54864a5d4700d7ce2a2c53eb77e490165f5c160f17e3483c4203893fad70\": rpc error: code = NotFound desc = could not find container \"55cd54864a5d4700d7ce2a2c53eb77e490165f5c160f17e3483c4203893fad70\": container with ID starting with 55cd54864a5d4700d7ce2a2c53eb77e490165f5c160f17e3483c4203893fad70 not found: ID does not exist" Oct 08 18:36:40 crc kubenswrapper[4859]: I1008 18:36:40.714861 4859 scope.go:117] "RemoveContainer" containerID="e8b8ab1aa42618805f0c5cb12f6ef9a1a2f4d979e05d79124e59dbd38ebec03f" Oct 08 18:36:40 crc kubenswrapper[4859]: E1008 18:36:40.715155 4859 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"e8b8ab1aa42618805f0c5cb12f6ef9a1a2f4d979e05d79124e59dbd38ebec03f\": container with ID starting with e8b8ab1aa42618805f0c5cb12f6ef9a1a2f4d979e05d79124e59dbd38ebec03f not found: ID does not exist" containerID="e8b8ab1aa42618805f0c5cb12f6ef9a1a2f4d979e05d79124e59dbd38ebec03f" Oct 08 18:36:40 crc kubenswrapper[4859]: I1008 18:36:40.715173 4859 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"e8b8ab1aa42618805f0c5cb12f6ef9a1a2f4d979e05d79124e59dbd38ebec03f"} err="failed to get container status \"e8b8ab1aa42618805f0c5cb12f6ef9a1a2f4d979e05d79124e59dbd38ebec03f\": rpc error: code = NotFound desc = could not find container \"e8b8ab1aa42618805f0c5cb12f6ef9a1a2f4d979e05d79124e59dbd38ebec03f\": container with ID starting with e8b8ab1aa42618805f0c5cb12f6ef9a1a2f4d979e05d79124e59dbd38ebec03f not found: ID does not exist" Oct 08 18:36:40 crc kubenswrapper[4859]: I1008 18:36:40.715185 4859 scope.go:117] "RemoveContainer" containerID="45e4e70510b3cf6458e879dd30816ac1e0a0d242109940a427935dc1f27be0f1" Oct 08 18:36:40 crc kubenswrapper[4859]: I1008 18:36:40.735460 4859 scope.go:117] "RemoveContainer" containerID="2684d40d3cd07c780bb9fddbf1a08e723487ca1f6c6f95428d4d5792f5d94759" Oct 08 18:36:40 crc kubenswrapper[4859]: I1008 18:36:40.741200 4859 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/01aea24c-9805-4a41-9843-c3366d597442-logs\") pod \"nova-api-0\" (UID: \"01aea24c-9805-4a41-9843-c3366d597442\") " pod="openstack/nova-api-0" Oct 08 18:36:40 crc kubenswrapper[4859]: I1008 18:36:40.741300 4859 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/01aea24c-9805-4a41-9843-c3366d597442-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"01aea24c-9805-4a41-9843-c3366d597442\") " pod="openstack/nova-api-0" Oct 08 18:36:40 crc kubenswrapper[4859]: I1008 18:36:40.741326 4859 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/01aea24c-9805-4a41-9843-c3366d597442-internal-tls-certs\") pod \"nova-api-0\" (UID: \"01aea24c-9805-4a41-9843-c3366d597442\") " pod="openstack/nova-api-0" Oct 08 18:36:40 crc kubenswrapper[4859]: I1008 18:36:40.741445 4859 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-hj89c\" (UniqueName: \"kubernetes.io/projected/01aea24c-9805-4a41-9843-c3366d597442-kube-api-access-hj89c\") pod \"nova-api-0\" (UID: \"01aea24c-9805-4a41-9843-c3366d597442\") " pod="openstack/nova-api-0" Oct 08 18:36:40 crc kubenswrapper[4859]: I1008 18:36:40.741505 4859 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/01aea24c-9805-4a41-9843-c3366d597442-config-data\") pod \"nova-api-0\" (UID: \"01aea24c-9805-4a41-9843-c3366d597442\") " pod="openstack/nova-api-0" Oct 08 18:36:40 crc kubenswrapper[4859]: I1008 18:36:40.741566 4859 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/01aea24c-9805-4a41-9843-c3366d597442-public-tls-certs\") pod \"nova-api-0\" (UID: \"01aea24c-9805-4a41-9843-c3366d597442\") " pod="openstack/nova-api-0" Oct 08 18:36:40 crc kubenswrapper[4859]: I1008 18:36:40.741657 4859 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/135723dc-9767-4d2e-bf09-ad570aa5ee6d-config-data\") on node \"crc\" DevicePath \"\"" Oct 08 18:36:40 crc kubenswrapper[4859]: I1008 18:36:40.755127 4859 scope.go:117] "RemoveContainer" containerID="45e4e70510b3cf6458e879dd30816ac1e0a0d242109940a427935dc1f27be0f1" Oct 08 18:36:40 crc kubenswrapper[4859]: E1008 18:36:40.755785 4859 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"45e4e70510b3cf6458e879dd30816ac1e0a0d242109940a427935dc1f27be0f1\": container with ID starting with 45e4e70510b3cf6458e879dd30816ac1e0a0d242109940a427935dc1f27be0f1 not found: ID does not exist" containerID="45e4e70510b3cf6458e879dd30816ac1e0a0d242109940a427935dc1f27be0f1" Oct 08 18:36:40 crc kubenswrapper[4859]: I1008 18:36:40.755832 4859 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"45e4e70510b3cf6458e879dd30816ac1e0a0d242109940a427935dc1f27be0f1"} err="failed to get container status \"45e4e70510b3cf6458e879dd30816ac1e0a0d242109940a427935dc1f27be0f1\": rpc error: code = NotFound desc = could not find container \"45e4e70510b3cf6458e879dd30816ac1e0a0d242109940a427935dc1f27be0f1\": container with ID starting with 45e4e70510b3cf6458e879dd30816ac1e0a0d242109940a427935dc1f27be0f1 not found: ID does not exist" Oct 08 18:36:40 crc kubenswrapper[4859]: I1008 18:36:40.755867 4859 scope.go:117] "RemoveContainer" containerID="2684d40d3cd07c780bb9fddbf1a08e723487ca1f6c6f95428d4d5792f5d94759" Oct 08 18:36:40 crc kubenswrapper[4859]: E1008 18:36:40.756198 4859 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"2684d40d3cd07c780bb9fddbf1a08e723487ca1f6c6f95428d4d5792f5d94759\": container with ID starting with 2684d40d3cd07c780bb9fddbf1a08e723487ca1f6c6f95428d4d5792f5d94759 not found: ID does not exist" containerID="2684d40d3cd07c780bb9fddbf1a08e723487ca1f6c6f95428d4d5792f5d94759" Oct 08 18:36:40 crc kubenswrapper[4859]: I1008 18:36:40.756232 4859 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"2684d40d3cd07c780bb9fddbf1a08e723487ca1f6c6f95428d4d5792f5d94759"} err="failed to get container status \"2684d40d3cd07c780bb9fddbf1a08e723487ca1f6c6f95428d4d5792f5d94759\": rpc error: code = NotFound desc = could not find container \"2684d40d3cd07c780bb9fddbf1a08e723487ca1f6c6f95428d4d5792f5d94759\": container with ID starting with 2684d40d3cd07c780bb9fddbf1a08e723487ca1f6c6f95428d4d5792f5d94759 not found: ID does not exist" Oct 08 18:36:40 crc kubenswrapper[4859]: I1008 18:36:40.844143 4859 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-hj89c\" (UniqueName: \"kubernetes.io/projected/01aea24c-9805-4a41-9843-c3366d597442-kube-api-access-hj89c\") pod \"nova-api-0\" (UID: \"01aea24c-9805-4a41-9843-c3366d597442\") " pod="openstack/nova-api-0" Oct 08 18:36:40 crc kubenswrapper[4859]: I1008 18:36:40.844207 4859 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/01aea24c-9805-4a41-9843-c3366d597442-config-data\") pod \"nova-api-0\" (UID: \"01aea24c-9805-4a41-9843-c3366d597442\") " pod="openstack/nova-api-0" Oct 08 18:36:40 crc kubenswrapper[4859]: I1008 18:36:40.844264 4859 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/01aea24c-9805-4a41-9843-c3366d597442-public-tls-certs\") pod \"nova-api-0\" (UID: \"01aea24c-9805-4a41-9843-c3366d597442\") " pod="openstack/nova-api-0" Oct 08 18:36:40 crc kubenswrapper[4859]: I1008 18:36:40.844310 4859 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/01aea24c-9805-4a41-9843-c3366d597442-logs\") pod \"nova-api-0\" (UID: \"01aea24c-9805-4a41-9843-c3366d597442\") " pod="openstack/nova-api-0" Oct 08 18:36:40 crc kubenswrapper[4859]: I1008 18:36:40.845040 4859 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/01aea24c-9805-4a41-9843-c3366d597442-logs\") pod \"nova-api-0\" (UID: \"01aea24c-9805-4a41-9843-c3366d597442\") " pod="openstack/nova-api-0" Oct 08 18:36:40 crc kubenswrapper[4859]: I1008 18:36:40.845714 4859 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/01aea24c-9805-4a41-9843-c3366d597442-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"01aea24c-9805-4a41-9843-c3366d597442\") " pod="openstack/nova-api-0" Oct 08 18:36:40 crc kubenswrapper[4859]: I1008 18:36:40.845790 4859 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/01aea24c-9805-4a41-9843-c3366d597442-internal-tls-certs\") pod \"nova-api-0\" (UID: \"01aea24c-9805-4a41-9843-c3366d597442\") " pod="openstack/nova-api-0" Oct 08 18:36:40 crc kubenswrapper[4859]: I1008 18:36:40.850360 4859 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/01aea24c-9805-4a41-9843-c3366d597442-config-data\") pod \"nova-api-0\" (UID: \"01aea24c-9805-4a41-9843-c3366d597442\") " pod="openstack/nova-api-0" Oct 08 18:36:40 crc kubenswrapper[4859]: I1008 18:36:40.850357 4859 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/01aea24c-9805-4a41-9843-c3366d597442-public-tls-certs\") pod \"nova-api-0\" (UID: \"01aea24c-9805-4a41-9843-c3366d597442\") " pod="openstack/nova-api-0" Oct 08 18:36:40 crc kubenswrapper[4859]: I1008 18:36:40.853529 4859 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/01aea24c-9805-4a41-9843-c3366d597442-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"01aea24c-9805-4a41-9843-c3366d597442\") " pod="openstack/nova-api-0" Oct 08 18:36:40 crc kubenswrapper[4859]: I1008 18:36:40.854329 4859 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/01aea24c-9805-4a41-9843-c3366d597442-internal-tls-certs\") pod \"nova-api-0\" (UID: \"01aea24c-9805-4a41-9843-c3366d597442\") " pod="openstack/nova-api-0" Oct 08 18:36:40 crc kubenswrapper[4859]: I1008 18:36:40.865581 4859 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-hj89c\" (UniqueName: \"kubernetes.io/projected/01aea24c-9805-4a41-9843-c3366d597442-kube-api-access-hj89c\") pod \"nova-api-0\" (UID: \"01aea24c-9805-4a41-9843-c3366d597442\") " pod="openstack/nova-api-0" Oct 08 18:36:40 crc kubenswrapper[4859]: I1008 18:36:40.883286 4859 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Oct 08 18:36:40 crc kubenswrapper[4859]: I1008 18:36:40.893960 4859 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/ceilometer-0"] Oct 08 18:36:40 crc kubenswrapper[4859]: I1008 18:36:40.905772 4859 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ceilometer-0"] Oct 08 18:36:40 crc kubenswrapper[4859]: I1008 18:36:40.908287 4859 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Oct 08 18:36:40 crc kubenswrapper[4859]: I1008 18:36:40.911592 4859 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-scripts" Oct 08 18:36:40 crc kubenswrapper[4859]: I1008 18:36:40.912179 4859 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-config-data" Oct 08 18:36:40 crc kubenswrapper[4859]: I1008 18:36:40.912386 4859 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-ceilometer-internal-svc" Oct 08 18:36:40 crc kubenswrapper[4859]: I1008 18:36:40.912897 4859 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Oct 08 18:36:40 crc kubenswrapper[4859]: I1008 18:36:40.996342 4859 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Oct 08 18:36:41 crc kubenswrapper[4859]: I1008 18:36:41.051498 4859 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-zl94n\" (UniqueName: \"kubernetes.io/projected/53819555-42d9-40cb-9897-94c5496c8da9-kube-api-access-zl94n\") pod \"ceilometer-0\" (UID: \"53819555-42d9-40cb-9897-94c5496c8da9\") " pod="openstack/ceilometer-0" Oct 08 18:36:41 crc kubenswrapper[4859]: I1008 18:36:41.051903 4859 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/53819555-42d9-40cb-9897-94c5496c8da9-scripts\") pod \"ceilometer-0\" (UID: \"53819555-42d9-40cb-9897-94c5496c8da9\") " pod="openstack/ceilometer-0" Oct 08 18:36:41 crc kubenswrapper[4859]: I1008 18:36:41.052005 4859 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/53819555-42d9-40cb-9897-94c5496c8da9-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"53819555-42d9-40cb-9897-94c5496c8da9\") " pod="openstack/ceilometer-0" Oct 08 18:36:41 crc kubenswrapper[4859]: I1008 18:36:41.052028 4859 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/53819555-42d9-40cb-9897-94c5496c8da9-config-data\") pod \"ceilometer-0\" (UID: \"53819555-42d9-40cb-9897-94c5496c8da9\") " pod="openstack/ceilometer-0" Oct 08 18:36:41 crc kubenswrapper[4859]: I1008 18:36:41.052120 4859 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/53819555-42d9-40cb-9897-94c5496c8da9-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"53819555-42d9-40cb-9897-94c5496c8da9\") " pod="openstack/ceilometer-0" Oct 08 18:36:41 crc kubenswrapper[4859]: I1008 18:36:41.052149 4859 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/53819555-42d9-40cb-9897-94c5496c8da9-ceilometer-tls-certs\") pod \"ceilometer-0\" (UID: \"53819555-42d9-40cb-9897-94c5496c8da9\") " pod="openstack/ceilometer-0" Oct 08 18:36:41 crc kubenswrapper[4859]: I1008 18:36:41.052312 4859 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/53819555-42d9-40cb-9897-94c5496c8da9-run-httpd\") pod \"ceilometer-0\" (UID: \"53819555-42d9-40cb-9897-94c5496c8da9\") " pod="openstack/ceilometer-0" Oct 08 18:36:41 crc kubenswrapper[4859]: I1008 18:36:41.052413 4859 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/53819555-42d9-40cb-9897-94c5496c8da9-log-httpd\") pod \"ceilometer-0\" (UID: \"53819555-42d9-40cb-9897-94c5496c8da9\") " pod="openstack/ceilometer-0" Oct 08 18:36:41 crc kubenswrapper[4859]: I1008 18:36:41.068132 4859 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-cell1-novncproxy-0" Oct 08 18:36:41 crc kubenswrapper[4859]: I1008 18:36:41.115749 4859 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-cell1-novncproxy-0" Oct 08 18:36:41 crc kubenswrapper[4859]: I1008 18:36:41.153924 4859 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-zl94n\" (UniqueName: \"kubernetes.io/projected/53819555-42d9-40cb-9897-94c5496c8da9-kube-api-access-zl94n\") pod \"ceilometer-0\" (UID: \"53819555-42d9-40cb-9897-94c5496c8da9\") " pod="openstack/ceilometer-0" Oct 08 18:36:41 crc kubenswrapper[4859]: I1008 18:36:41.153981 4859 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/53819555-42d9-40cb-9897-94c5496c8da9-scripts\") pod \"ceilometer-0\" (UID: \"53819555-42d9-40cb-9897-94c5496c8da9\") " pod="openstack/ceilometer-0" Oct 08 18:36:41 crc kubenswrapper[4859]: I1008 18:36:41.154058 4859 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/53819555-42d9-40cb-9897-94c5496c8da9-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"53819555-42d9-40cb-9897-94c5496c8da9\") " pod="openstack/ceilometer-0" Oct 08 18:36:41 crc kubenswrapper[4859]: I1008 18:36:41.154093 4859 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/53819555-42d9-40cb-9897-94c5496c8da9-config-data\") pod \"ceilometer-0\" (UID: \"53819555-42d9-40cb-9897-94c5496c8da9\") " pod="openstack/ceilometer-0" Oct 08 18:36:41 crc kubenswrapper[4859]: I1008 18:36:41.155167 4859 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/53819555-42d9-40cb-9897-94c5496c8da9-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"53819555-42d9-40cb-9897-94c5496c8da9\") " pod="openstack/ceilometer-0" Oct 08 18:36:41 crc kubenswrapper[4859]: I1008 18:36:41.155196 4859 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/53819555-42d9-40cb-9897-94c5496c8da9-ceilometer-tls-certs\") pod \"ceilometer-0\" (UID: \"53819555-42d9-40cb-9897-94c5496c8da9\") " pod="openstack/ceilometer-0" Oct 08 18:36:41 crc kubenswrapper[4859]: I1008 18:36:41.155246 4859 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/53819555-42d9-40cb-9897-94c5496c8da9-run-httpd\") pod \"ceilometer-0\" (UID: \"53819555-42d9-40cb-9897-94c5496c8da9\") " pod="openstack/ceilometer-0" Oct 08 18:36:41 crc kubenswrapper[4859]: I1008 18:36:41.155264 4859 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/53819555-42d9-40cb-9897-94c5496c8da9-log-httpd\") pod \"ceilometer-0\" (UID: \"53819555-42d9-40cb-9897-94c5496c8da9\") " pod="openstack/ceilometer-0" Oct 08 18:36:41 crc kubenswrapper[4859]: I1008 18:36:41.155949 4859 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/53819555-42d9-40cb-9897-94c5496c8da9-log-httpd\") pod \"ceilometer-0\" (UID: \"53819555-42d9-40cb-9897-94c5496c8da9\") " pod="openstack/ceilometer-0" Oct 08 18:36:41 crc kubenswrapper[4859]: I1008 18:36:41.157253 4859 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/53819555-42d9-40cb-9897-94c5496c8da9-run-httpd\") pod \"ceilometer-0\" (UID: \"53819555-42d9-40cb-9897-94c5496c8da9\") " pod="openstack/ceilometer-0" Oct 08 18:36:41 crc kubenswrapper[4859]: I1008 18:36:41.162066 4859 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/53819555-42d9-40cb-9897-94c5496c8da9-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"53819555-42d9-40cb-9897-94c5496c8da9\") " pod="openstack/ceilometer-0" Oct 08 18:36:41 crc kubenswrapper[4859]: I1008 18:36:41.162737 4859 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/53819555-42d9-40cb-9897-94c5496c8da9-config-data\") pod \"ceilometer-0\" (UID: \"53819555-42d9-40cb-9897-94c5496c8da9\") " pod="openstack/ceilometer-0" Oct 08 18:36:41 crc kubenswrapper[4859]: I1008 18:36:41.164998 4859 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/53819555-42d9-40cb-9897-94c5496c8da9-ceilometer-tls-certs\") pod \"ceilometer-0\" (UID: \"53819555-42d9-40cb-9897-94c5496c8da9\") " pod="openstack/ceilometer-0" Oct 08 18:36:41 crc kubenswrapper[4859]: I1008 18:36:41.168100 4859 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/53819555-42d9-40cb-9897-94c5496c8da9-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"53819555-42d9-40cb-9897-94c5496c8da9\") " pod="openstack/ceilometer-0" Oct 08 18:36:41 crc kubenswrapper[4859]: I1008 18:36:41.172237 4859 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/53819555-42d9-40cb-9897-94c5496c8da9-scripts\") pod \"ceilometer-0\" (UID: \"53819555-42d9-40cb-9897-94c5496c8da9\") " pod="openstack/ceilometer-0" Oct 08 18:36:41 crc kubenswrapper[4859]: I1008 18:36:41.175099 4859 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-zl94n\" (UniqueName: \"kubernetes.io/projected/53819555-42d9-40cb-9897-94c5496c8da9-kube-api-access-zl94n\") pod \"ceilometer-0\" (UID: \"53819555-42d9-40cb-9897-94c5496c8da9\") " pod="openstack/ceilometer-0" Oct 08 18:36:41 crc kubenswrapper[4859]: I1008 18:36:41.256672 4859 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Oct 08 18:36:41 crc kubenswrapper[4859]: I1008 18:36:41.456331 4859 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-0"] Oct 08 18:36:41 crc kubenswrapper[4859]: I1008 18:36:41.539713 4859 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"01aea24c-9805-4a41-9843-c3366d597442","Type":"ContainerStarted","Data":"33a274cc1a175459eb3c111f7ca40ff853a75cc34abb7325a37166cd63b59c7a"} Oct 08 18:36:41 crc kubenswrapper[4859]: I1008 18:36:41.557107 4859 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-cell1-novncproxy-0" Oct 08 18:36:41 crc kubenswrapper[4859]: I1008 18:36:41.761948 4859 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Oct 08 18:36:41 crc kubenswrapper[4859]: I1008 18:36:41.806537 4859 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell1-cell-mapping-9bz5d"] Oct 08 18:36:41 crc kubenswrapper[4859]: I1008 18:36:41.807980 4859 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-cell-mapping-9bz5d" Oct 08 18:36:41 crc kubenswrapper[4859]: I1008 18:36:41.810957 4859 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell1-manage-scripts" Oct 08 18:36:41 crc kubenswrapper[4859]: I1008 18:36:41.811056 4859 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell1-manage-config-data" Oct 08 18:36:41 crc kubenswrapper[4859]: I1008 18:36:41.832018 4859 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-cell-mapping-9bz5d"] Oct 08 18:36:41 crc kubenswrapper[4859]: I1008 18:36:41.867883 4859 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-9fkxw\" (UniqueName: \"kubernetes.io/projected/7a928d30-8b64-4daf-9731-abea3343acbc-kube-api-access-9fkxw\") pod \"nova-cell1-cell-mapping-9bz5d\" (UID: \"7a928d30-8b64-4daf-9731-abea3343acbc\") " pod="openstack/nova-cell1-cell-mapping-9bz5d" Oct 08 18:36:41 crc kubenswrapper[4859]: I1008 18:36:41.867937 4859 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/7a928d30-8b64-4daf-9731-abea3343acbc-config-data\") pod \"nova-cell1-cell-mapping-9bz5d\" (UID: \"7a928d30-8b64-4daf-9731-abea3343acbc\") " pod="openstack/nova-cell1-cell-mapping-9bz5d" Oct 08 18:36:41 crc kubenswrapper[4859]: I1008 18:36:41.868001 4859 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7a928d30-8b64-4daf-9731-abea3343acbc-combined-ca-bundle\") pod \"nova-cell1-cell-mapping-9bz5d\" (UID: \"7a928d30-8b64-4daf-9731-abea3343acbc\") " pod="openstack/nova-cell1-cell-mapping-9bz5d" Oct 08 18:36:41 crc kubenswrapper[4859]: I1008 18:36:41.868020 4859 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/7a928d30-8b64-4daf-9731-abea3343acbc-scripts\") pod \"nova-cell1-cell-mapping-9bz5d\" (UID: \"7a928d30-8b64-4daf-9731-abea3343acbc\") " pod="openstack/nova-cell1-cell-mapping-9bz5d" Oct 08 18:36:41 crc kubenswrapper[4859]: I1008 18:36:41.970562 4859 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-9fkxw\" (UniqueName: \"kubernetes.io/projected/7a928d30-8b64-4daf-9731-abea3343acbc-kube-api-access-9fkxw\") pod \"nova-cell1-cell-mapping-9bz5d\" (UID: \"7a928d30-8b64-4daf-9731-abea3343acbc\") " pod="openstack/nova-cell1-cell-mapping-9bz5d" Oct 08 18:36:41 crc kubenswrapper[4859]: I1008 18:36:41.970656 4859 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/7a928d30-8b64-4daf-9731-abea3343acbc-config-data\") pod \"nova-cell1-cell-mapping-9bz5d\" (UID: \"7a928d30-8b64-4daf-9731-abea3343acbc\") " pod="openstack/nova-cell1-cell-mapping-9bz5d" Oct 08 18:36:41 crc kubenswrapper[4859]: I1008 18:36:41.970745 4859 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7a928d30-8b64-4daf-9731-abea3343acbc-combined-ca-bundle\") pod \"nova-cell1-cell-mapping-9bz5d\" (UID: \"7a928d30-8b64-4daf-9731-abea3343acbc\") " pod="openstack/nova-cell1-cell-mapping-9bz5d" Oct 08 18:36:41 crc kubenswrapper[4859]: I1008 18:36:41.970768 4859 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/7a928d30-8b64-4daf-9731-abea3343acbc-scripts\") pod \"nova-cell1-cell-mapping-9bz5d\" (UID: \"7a928d30-8b64-4daf-9731-abea3343acbc\") " pod="openstack/nova-cell1-cell-mapping-9bz5d" Oct 08 18:36:41 crc kubenswrapper[4859]: I1008 18:36:41.976011 4859 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/7a928d30-8b64-4daf-9731-abea3343acbc-config-data\") pod \"nova-cell1-cell-mapping-9bz5d\" (UID: \"7a928d30-8b64-4daf-9731-abea3343acbc\") " pod="openstack/nova-cell1-cell-mapping-9bz5d" Oct 08 18:36:41 crc kubenswrapper[4859]: I1008 18:36:41.976391 4859 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7a928d30-8b64-4daf-9731-abea3343acbc-combined-ca-bundle\") pod \"nova-cell1-cell-mapping-9bz5d\" (UID: \"7a928d30-8b64-4daf-9731-abea3343acbc\") " pod="openstack/nova-cell1-cell-mapping-9bz5d" Oct 08 18:36:41 crc kubenswrapper[4859]: I1008 18:36:41.977110 4859 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/7a928d30-8b64-4daf-9731-abea3343acbc-scripts\") pod \"nova-cell1-cell-mapping-9bz5d\" (UID: \"7a928d30-8b64-4daf-9731-abea3343acbc\") " pod="openstack/nova-cell1-cell-mapping-9bz5d" Oct 08 18:36:41 crc kubenswrapper[4859]: I1008 18:36:41.989292 4859 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-9fkxw\" (UniqueName: \"kubernetes.io/projected/7a928d30-8b64-4daf-9731-abea3343acbc-kube-api-access-9fkxw\") pod \"nova-cell1-cell-mapping-9bz5d\" (UID: \"7a928d30-8b64-4daf-9731-abea3343acbc\") " pod="openstack/nova-cell1-cell-mapping-9bz5d" Oct 08 18:36:42 crc kubenswrapper[4859]: I1008 18:36:42.178188 4859 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-cell-mapping-9bz5d" Oct 08 18:36:42 crc kubenswrapper[4859]: I1008 18:36:42.482056 4859 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="135723dc-9767-4d2e-bf09-ad570aa5ee6d" path="/var/lib/kubelet/pods/135723dc-9767-4d2e-bf09-ad570aa5ee6d/volumes" Oct 08 18:36:42 crc kubenswrapper[4859]: I1008 18:36:42.483735 4859 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="402e8a91-7238-4d32-800e-266a6c911802" path="/var/lib/kubelet/pods/402e8a91-7238-4d32-800e-266a6c911802/volumes" Oct 08 18:36:42 crc kubenswrapper[4859]: I1008 18:36:42.552710 4859 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"53819555-42d9-40cb-9897-94c5496c8da9","Type":"ContainerStarted","Data":"02d02356c462635af406697e8bf10e3632da32d96aa7bd22d903057bff4cd3e8"} Oct 08 18:36:42 crc kubenswrapper[4859]: I1008 18:36:42.553013 4859 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"53819555-42d9-40cb-9897-94c5496c8da9","Type":"ContainerStarted","Data":"39dbd6775d6276772d7d0ee543c0fbee65b6999a12e91644fc909cf145c03f5a"} Oct 08 18:36:42 crc kubenswrapper[4859]: I1008 18:36:42.556663 4859 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"01aea24c-9805-4a41-9843-c3366d597442","Type":"ContainerStarted","Data":"229a8d291803695dc1c54a515b948e4483d245d8f32aae09cbcd61a368018a4d"} Oct 08 18:36:42 crc kubenswrapper[4859]: I1008 18:36:42.556725 4859 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"01aea24c-9805-4a41-9843-c3366d597442","Type":"ContainerStarted","Data":"9ff8822be5a79fd192a815cbffbe8666bbbb558d24c84188e5a97ffa080d975a"} Oct 08 18:36:42 crc kubenswrapper[4859]: I1008 18:36:42.585316 4859 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-api-0" podStartSLOduration=2.5852933460000003 podStartE2EDuration="2.585293346s" podCreationTimestamp="2025-10-08 18:36:40 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-08 18:36:42.575720713 +0000 UTC m=+1172.822560112" watchObservedRunningTime="2025-10-08 18:36:42.585293346 +0000 UTC m=+1172.832132755" Oct 08 18:36:42 crc kubenswrapper[4859]: W1008 18:36:42.676499 4859 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod7a928d30_8b64_4daf_9731_abea3343acbc.slice/crio-259d7b9b027af99448a2b0384912260c24f09e5a20df57aff64cc455fee3018c WatchSource:0}: Error finding container 259d7b9b027af99448a2b0384912260c24f09e5a20df57aff64cc455fee3018c: Status 404 returned error can't find the container with id 259d7b9b027af99448a2b0384912260c24f09e5a20df57aff64cc455fee3018c Oct 08 18:36:42 crc kubenswrapper[4859]: I1008 18:36:42.680010 4859 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-cell-mapping-9bz5d"] Oct 08 18:36:43 crc kubenswrapper[4859]: I1008 18:36:43.581860 4859 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-cell-mapping-9bz5d" event={"ID":"7a928d30-8b64-4daf-9731-abea3343acbc","Type":"ContainerStarted","Data":"7d3cd58642ece2fbfeef885a1db9ce1905a67a3195f6b96dfdd563340c9cc28b"} Oct 08 18:36:43 crc kubenswrapper[4859]: I1008 18:36:43.582269 4859 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-cell-mapping-9bz5d" event={"ID":"7a928d30-8b64-4daf-9731-abea3343acbc","Type":"ContainerStarted","Data":"259d7b9b027af99448a2b0384912260c24f09e5a20df57aff64cc455fee3018c"} Oct 08 18:36:43 crc kubenswrapper[4859]: I1008 18:36:43.589393 4859 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"53819555-42d9-40cb-9897-94c5496c8da9","Type":"ContainerStarted","Data":"ed898a8d668f2b60230819b6dce39b8d1b5bb7449597cc8a1ab0ff6de8a2b38a"} Oct 08 18:36:43 crc kubenswrapper[4859]: I1008 18:36:43.610617 4859 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-cell1-cell-mapping-9bz5d" podStartSLOduration=2.610598637 podStartE2EDuration="2.610598637s" podCreationTimestamp="2025-10-08 18:36:41 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-08 18:36:43.59846087 +0000 UTC m=+1173.845300279" watchObservedRunningTime="2025-10-08 18:36:43.610598637 +0000 UTC m=+1173.857438016" Oct 08 18:36:44 crc kubenswrapper[4859]: I1008 18:36:44.483619 4859 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-64986d45b9-sv5xb" Oct 08 18:36:44 crc kubenswrapper[4859]: I1008 18:36:44.550767 4859 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-694b6b9bcc-7zv88"] Oct 08 18:36:44 crc kubenswrapper[4859]: I1008 18:36:44.550992 4859 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-694b6b9bcc-7zv88" podUID="1ea66084-f5e7-4fcc-a439-c7806dde0197" containerName="dnsmasq-dns" containerID="cri-o://46b13ba9c6e976285c67c445ae0c938d4f83886ca14e0c5e9153d7c03813fb0b" gracePeriod=10 Oct 08 18:36:44 crc kubenswrapper[4859]: I1008 18:36:44.600942 4859 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"53819555-42d9-40cb-9897-94c5496c8da9","Type":"ContainerStarted","Data":"4109efc4602ca55f12c2b5e2e388e2641bb7b0a4018f6f39cc70209ab6901fec"} Oct 08 18:36:45 crc kubenswrapper[4859]: I1008 18:36:45.192605 4859 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-694b6b9bcc-7zv88" Oct 08 18:36:45 crc kubenswrapper[4859]: I1008 18:36:45.348446 4859 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/1ea66084-f5e7-4fcc-a439-c7806dde0197-ovsdbserver-sb\") pod \"1ea66084-f5e7-4fcc-a439-c7806dde0197\" (UID: \"1ea66084-f5e7-4fcc-a439-c7806dde0197\") " Oct 08 18:36:45 crc kubenswrapper[4859]: I1008 18:36:45.348583 4859 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/1ea66084-f5e7-4fcc-a439-c7806dde0197-dns-swift-storage-0\") pod \"1ea66084-f5e7-4fcc-a439-c7806dde0197\" (UID: \"1ea66084-f5e7-4fcc-a439-c7806dde0197\") " Oct 08 18:36:45 crc kubenswrapper[4859]: I1008 18:36:45.348666 4859 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-jd9kx\" (UniqueName: \"kubernetes.io/projected/1ea66084-f5e7-4fcc-a439-c7806dde0197-kube-api-access-jd9kx\") pod \"1ea66084-f5e7-4fcc-a439-c7806dde0197\" (UID: \"1ea66084-f5e7-4fcc-a439-c7806dde0197\") " Oct 08 18:36:45 crc kubenswrapper[4859]: I1008 18:36:45.348760 4859 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/1ea66084-f5e7-4fcc-a439-c7806dde0197-ovsdbserver-nb\") pod \"1ea66084-f5e7-4fcc-a439-c7806dde0197\" (UID: \"1ea66084-f5e7-4fcc-a439-c7806dde0197\") " Oct 08 18:36:45 crc kubenswrapper[4859]: I1008 18:36:45.348800 4859 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/1ea66084-f5e7-4fcc-a439-c7806dde0197-config\") pod \"1ea66084-f5e7-4fcc-a439-c7806dde0197\" (UID: \"1ea66084-f5e7-4fcc-a439-c7806dde0197\") " Oct 08 18:36:45 crc kubenswrapper[4859]: I1008 18:36:45.348854 4859 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/1ea66084-f5e7-4fcc-a439-c7806dde0197-dns-svc\") pod \"1ea66084-f5e7-4fcc-a439-c7806dde0197\" (UID: \"1ea66084-f5e7-4fcc-a439-c7806dde0197\") " Oct 08 18:36:45 crc kubenswrapper[4859]: I1008 18:36:45.354631 4859 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/1ea66084-f5e7-4fcc-a439-c7806dde0197-kube-api-access-jd9kx" (OuterVolumeSpecName: "kube-api-access-jd9kx") pod "1ea66084-f5e7-4fcc-a439-c7806dde0197" (UID: "1ea66084-f5e7-4fcc-a439-c7806dde0197"). InnerVolumeSpecName "kube-api-access-jd9kx". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 08 18:36:45 crc kubenswrapper[4859]: I1008 18:36:45.401318 4859 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1ea66084-f5e7-4fcc-a439-c7806dde0197-dns-swift-storage-0" (OuterVolumeSpecName: "dns-swift-storage-0") pod "1ea66084-f5e7-4fcc-a439-c7806dde0197" (UID: "1ea66084-f5e7-4fcc-a439-c7806dde0197"). InnerVolumeSpecName "dns-swift-storage-0". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 08 18:36:45 crc kubenswrapper[4859]: I1008 18:36:45.403612 4859 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1ea66084-f5e7-4fcc-a439-c7806dde0197-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "1ea66084-f5e7-4fcc-a439-c7806dde0197" (UID: "1ea66084-f5e7-4fcc-a439-c7806dde0197"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 08 18:36:45 crc kubenswrapper[4859]: I1008 18:36:45.410635 4859 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1ea66084-f5e7-4fcc-a439-c7806dde0197-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "1ea66084-f5e7-4fcc-a439-c7806dde0197" (UID: "1ea66084-f5e7-4fcc-a439-c7806dde0197"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 08 18:36:45 crc kubenswrapper[4859]: I1008 18:36:45.412246 4859 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1ea66084-f5e7-4fcc-a439-c7806dde0197-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "1ea66084-f5e7-4fcc-a439-c7806dde0197" (UID: "1ea66084-f5e7-4fcc-a439-c7806dde0197"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 08 18:36:45 crc kubenswrapper[4859]: I1008 18:36:45.416237 4859 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1ea66084-f5e7-4fcc-a439-c7806dde0197-config" (OuterVolumeSpecName: "config") pod "1ea66084-f5e7-4fcc-a439-c7806dde0197" (UID: "1ea66084-f5e7-4fcc-a439-c7806dde0197"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 08 18:36:45 crc kubenswrapper[4859]: I1008 18:36:45.451739 4859 reconciler_common.go:293] "Volume detached for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/1ea66084-f5e7-4fcc-a439-c7806dde0197-dns-swift-storage-0\") on node \"crc\" DevicePath \"\"" Oct 08 18:36:45 crc kubenswrapper[4859]: I1008 18:36:45.451776 4859 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-jd9kx\" (UniqueName: \"kubernetes.io/projected/1ea66084-f5e7-4fcc-a439-c7806dde0197-kube-api-access-jd9kx\") on node \"crc\" DevicePath \"\"" Oct 08 18:36:45 crc kubenswrapper[4859]: I1008 18:36:45.451788 4859 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/1ea66084-f5e7-4fcc-a439-c7806dde0197-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Oct 08 18:36:45 crc kubenswrapper[4859]: I1008 18:36:45.451803 4859 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/1ea66084-f5e7-4fcc-a439-c7806dde0197-config\") on node \"crc\" DevicePath \"\"" Oct 08 18:36:45 crc kubenswrapper[4859]: I1008 18:36:45.451811 4859 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/1ea66084-f5e7-4fcc-a439-c7806dde0197-dns-svc\") on node \"crc\" DevicePath \"\"" Oct 08 18:36:45 crc kubenswrapper[4859]: I1008 18:36:45.451820 4859 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/1ea66084-f5e7-4fcc-a439-c7806dde0197-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Oct 08 18:36:45 crc kubenswrapper[4859]: I1008 18:36:45.610264 4859 generic.go:334] "Generic (PLEG): container finished" podID="1ea66084-f5e7-4fcc-a439-c7806dde0197" containerID="46b13ba9c6e976285c67c445ae0c938d4f83886ca14e0c5e9153d7c03813fb0b" exitCode=0 Oct 08 18:36:45 crc kubenswrapper[4859]: I1008 18:36:45.610321 4859 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-694b6b9bcc-7zv88" Oct 08 18:36:45 crc kubenswrapper[4859]: I1008 18:36:45.610376 4859 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-694b6b9bcc-7zv88" event={"ID":"1ea66084-f5e7-4fcc-a439-c7806dde0197","Type":"ContainerDied","Data":"46b13ba9c6e976285c67c445ae0c938d4f83886ca14e0c5e9153d7c03813fb0b"} Oct 08 18:36:45 crc kubenswrapper[4859]: I1008 18:36:45.610434 4859 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-694b6b9bcc-7zv88" event={"ID":"1ea66084-f5e7-4fcc-a439-c7806dde0197","Type":"ContainerDied","Data":"98ce65345d863f1e51428d1f5f987c53c4b454fb9c924dbd6bbd66f8c79621d7"} Oct 08 18:36:45 crc kubenswrapper[4859]: I1008 18:36:45.610458 4859 scope.go:117] "RemoveContainer" containerID="46b13ba9c6e976285c67c445ae0c938d4f83886ca14e0c5e9153d7c03813fb0b" Oct 08 18:36:45 crc kubenswrapper[4859]: I1008 18:36:45.614753 4859 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"53819555-42d9-40cb-9897-94c5496c8da9","Type":"ContainerStarted","Data":"7b1ff3cbf184889bc5bf934a52abc99beb071721ad474c34a24dab1e5520680f"} Oct 08 18:36:45 crc kubenswrapper[4859]: I1008 18:36:45.615105 4859 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ceilometer-0" Oct 08 18:36:45 crc kubenswrapper[4859]: I1008 18:36:45.632044 4859 scope.go:117] "RemoveContainer" containerID="02c74e0a8c552cfbca568550fb788525c9c2ce4b97fc9a1bcdfdf7756cc2a709" Oct 08 18:36:45 crc kubenswrapper[4859]: I1008 18:36:45.644205 4859 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ceilometer-0" podStartSLOduration=2.544533101 podStartE2EDuration="5.644183973s" podCreationTimestamp="2025-10-08 18:36:40 +0000 UTC" firstStartedPulling="2025-10-08 18:36:41.758341875 +0000 UTC m=+1172.005181254" lastFinishedPulling="2025-10-08 18:36:44.857992747 +0000 UTC m=+1175.104832126" observedRunningTime="2025-10-08 18:36:45.635770512 +0000 UTC m=+1175.882609891" watchObservedRunningTime="2025-10-08 18:36:45.644183973 +0000 UTC m=+1175.891023352" Oct 08 18:36:45 crc kubenswrapper[4859]: I1008 18:36:45.665049 4859 scope.go:117] "RemoveContainer" containerID="46b13ba9c6e976285c67c445ae0c938d4f83886ca14e0c5e9153d7c03813fb0b" Oct 08 18:36:45 crc kubenswrapper[4859]: I1008 18:36:45.666900 4859 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-694b6b9bcc-7zv88"] Oct 08 18:36:45 crc kubenswrapper[4859]: E1008 18:36:45.667831 4859 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"46b13ba9c6e976285c67c445ae0c938d4f83886ca14e0c5e9153d7c03813fb0b\": container with ID starting with 46b13ba9c6e976285c67c445ae0c938d4f83886ca14e0c5e9153d7c03813fb0b not found: ID does not exist" containerID="46b13ba9c6e976285c67c445ae0c938d4f83886ca14e0c5e9153d7c03813fb0b" Oct 08 18:36:45 crc kubenswrapper[4859]: I1008 18:36:45.667926 4859 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"46b13ba9c6e976285c67c445ae0c938d4f83886ca14e0c5e9153d7c03813fb0b"} err="failed to get container status \"46b13ba9c6e976285c67c445ae0c938d4f83886ca14e0c5e9153d7c03813fb0b\": rpc error: code = NotFound desc = could not find container \"46b13ba9c6e976285c67c445ae0c938d4f83886ca14e0c5e9153d7c03813fb0b\": container with ID starting with 46b13ba9c6e976285c67c445ae0c938d4f83886ca14e0c5e9153d7c03813fb0b not found: ID does not exist" Oct 08 18:36:45 crc kubenswrapper[4859]: I1008 18:36:45.667966 4859 scope.go:117] "RemoveContainer" containerID="02c74e0a8c552cfbca568550fb788525c9c2ce4b97fc9a1bcdfdf7756cc2a709" Oct 08 18:36:45 crc kubenswrapper[4859]: E1008 18:36:45.669599 4859 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"02c74e0a8c552cfbca568550fb788525c9c2ce4b97fc9a1bcdfdf7756cc2a709\": container with ID starting with 02c74e0a8c552cfbca568550fb788525c9c2ce4b97fc9a1bcdfdf7756cc2a709 not found: ID does not exist" containerID="02c74e0a8c552cfbca568550fb788525c9c2ce4b97fc9a1bcdfdf7756cc2a709" Oct 08 18:36:45 crc kubenswrapper[4859]: I1008 18:36:45.669648 4859 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"02c74e0a8c552cfbca568550fb788525c9c2ce4b97fc9a1bcdfdf7756cc2a709"} err="failed to get container status \"02c74e0a8c552cfbca568550fb788525c9c2ce4b97fc9a1bcdfdf7756cc2a709\": rpc error: code = NotFound desc = could not find container \"02c74e0a8c552cfbca568550fb788525c9c2ce4b97fc9a1bcdfdf7756cc2a709\": container with ID starting with 02c74e0a8c552cfbca568550fb788525c9c2ce4b97fc9a1bcdfdf7756cc2a709 not found: ID does not exist" Oct 08 18:36:45 crc kubenswrapper[4859]: I1008 18:36:45.675253 4859 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-694b6b9bcc-7zv88"] Oct 08 18:36:46 crc kubenswrapper[4859]: I1008 18:36:46.483781 4859 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="1ea66084-f5e7-4fcc-a439-c7806dde0197" path="/var/lib/kubelet/pods/1ea66084-f5e7-4fcc-a439-c7806dde0197/volumes" Oct 08 18:36:48 crc kubenswrapper[4859]: I1008 18:36:48.647657 4859 generic.go:334] "Generic (PLEG): container finished" podID="7a928d30-8b64-4daf-9731-abea3343acbc" containerID="7d3cd58642ece2fbfeef885a1db9ce1905a67a3195f6b96dfdd563340c9cc28b" exitCode=0 Oct 08 18:36:48 crc kubenswrapper[4859]: I1008 18:36:48.648029 4859 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-cell-mapping-9bz5d" event={"ID":"7a928d30-8b64-4daf-9731-abea3343acbc","Type":"ContainerDied","Data":"7d3cd58642ece2fbfeef885a1db9ce1905a67a3195f6b96dfdd563340c9cc28b"} Oct 08 18:36:50 crc kubenswrapper[4859]: I1008 18:36:50.012183 4859 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-cell-mapping-9bz5d" Oct 08 18:36:50 crc kubenswrapper[4859]: I1008 18:36:50.141471 4859 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7a928d30-8b64-4daf-9731-abea3343acbc-combined-ca-bundle\") pod \"7a928d30-8b64-4daf-9731-abea3343acbc\" (UID: \"7a928d30-8b64-4daf-9731-abea3343acbc\") " Oct 08 18:36:50 crc kubenswrapper[4859]: I1008 18:36:50.141524 4859 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/7a928d30-8b64-4daf-9731-abea3343acbc-scripts\") pod \"7a928d30-8b64-4daf-9731-abea3343acbc\" (UID: \"7a928d30-8b64-4daf-9731-abea3343acbc\") " Oct 08 18:36:50 crc kubenswrapper[4859]: I1008 18:36:50.141565 4859 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/7a928d30-8b64-4daf-9731-abea3343acbc-config-data\") pod \"7a928d30-8b64-4daf-9731-abea3343acbc\" (UID: \"7a928d30-8b64-4daf-9731-abea3343acbc\") " Oct 08 18:36:50 crc kubenswrapper[4859]: I1008 18:36:50.141757 4859 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-9fkxw\" (UniqueName: \"kubernetes.io/projected/7a928d30-8b64-4daf-9731-abea3343acbc-kube-api-access-9fkxw\") pod \"7a928d30-8b64-4daf-9731-abea3343acbc\" (UID: \"7a928d30-8b64-4daf-9731-abea3343acbc\") " Oct 08 18:36:50 crc kubenswrapper[4859]: I1008 18:36:50.148125 4859 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/7a928d30-8b64-4daf-9731-abea3343acbc-kube-api-access-9fkxw" (OuterVolumeSpecName: "kube-api-access-9fkxw") pod "7a928d30-8b64-4daf-9731-abea3343acbc" (UID: "7a928d30-8b64-4daf-9731-abea3343acbc"). InnerVolumeSpecName "kube-api-access-9fkxw". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 08 18:36:50 crc kubenswrapper[4859]: I1008 18:36:50.164471 4859 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/7a928d30-8b64-4daf-9731-abea3343acbc-scripts" (OuterVolumeSpecName: "scripts") pod "7a928d30-8b64-4daf-9731-abea3343acbc" (UID: "7a928d30-8b64-4daf-9731-abea3343acbc"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 08 18:36:50 crc kubenswrapper[4859]: I1008 18:36:50.169756 4859 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/7a928d30-8b64-4daf-9731-abea3343acbc-config-data" (OuterVolumeSpecName: "config-data") pod "7a928d30-8b64-4daf-9731-abea3343acbc" (UID: "7a928d30-8b64-4daf-9731-abea3343acbc"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 08 18:36:50 crc kubenswrapper[4859]: I1008 18:36:50.175198 4859 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/7a928d30-8b64-4daf-9731-abea3343acbc-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "7a928d30-8b64-4daf-9731-abea3343acbc" (UID: "7a928d30-8b64-4daf-9731-abea3343acbc"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 08 18:36:50 crc kubenswrapper[4859]: I1008 18:36:50.243518 4859 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/7a928d30-8b64-4daf-9731-abea3343acbc-config-data\") on node \"crc\" DevicePath \"\"" Oct 08 18:36:50 crc kubenswrapper[4859]: I1008 18:36:50.243553 4859 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-9fkxw\" (UniqueName: \"kubernetes.io/projected/7a928d30-8b64-4daf-9731-abea3343acbc-kube-api-access-9fkxw\") on node \"crc\" DevicePath \"\"" Oct 08 18:36:50 crc kubenswrapper[4859]: I1008 18:36:50.243567 4859 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7a928d30-8b64-4daf-9731-abea3343acbc-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 08 18:36:50 crc kubenswrapper[4859]: I1008 18:36:50.243585 4859 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/7a928d30-8b64-4daf-9731-abea3343acbc-scripts\") on node \"crc\" DevicePath \"\"" Oct 08 18:36:50 crc kubenswrapper[4859]: I1008 18:36:50.684321 4859 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-cell-mapping-9bz5d" event={"ID":"7a928d30-8b64-4daf-9731-abea3343acbc","Type":"ContainerDied","Data":"259d7b9b027af99448a2b0384912260c24f09e5a20df57aff64cc455fee3018c"} Oct 08 18:36:50 crc kubenswrapper[4859]: I1008 18:36:50.684354 4859 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="259d7b9b027af99448a2b0384912260c24f09e5a20df57aff64cc455fee3018c" Oct 08 18:36:50 crc kubenswrapper[4859]: I1008 18:36:50.684407 4859 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-cell-mapping-9bz5d" Oct 08 18:36:50 crc kubenswrapper[4859]: I1008 18:36:50.847313 4859 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-scheduler-0"] Oct 08 18:36:50 crc kubenswrapper[4859]: I1008 18:36:50.847575 4859 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-scheduler-0" podUID="e5f82275-e5ae-4d2d-8ab6-dc4afda848ba" containerName="nova-scheduler-scheduler" containerID="cri-o://4fe8f76dc74c296a52a35017a865d9caf429643b340fffe1bc9156fa36b2c04f" gracePeriod=30 Oct 08 18:36:50 crc kubenswrapper[4859]: I1008 18:36:50.857161 4859 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-api-0"] Oct 08 18:36:50 crc kubenswrapper[4859]: I1008 18:36:50.857383 4859 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-api-0" podUID="01aea24c-9805-4a41-9843-c3366d597442" containerName="nova-api-log" containerID="cri-o://9ff8822be5a79fd192a815cbffbe8666bbbb558d24c84188e5a97ffa080d975a" gracePeriod=30 Oct 08 18:36:50 crc kubenswrapper[4859]: I1008 18:36:50.857542 4859 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-api-0" podUID="01aea24c-9805-4a41-9843-c3366d597442" containerName="nova-api-api" containerID="cri-o://229a8d291803695dc1c54a515b948e4483d245d8f32aae09cbcd61a368018a4d" gracePeriod=30 Oct 08 18:36:50 crc kubenswrapper[4859]: I1008 18:36:50.888986 4859 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-metadata-0"] Oct 08 18:36:50 crc kubenswrapper[4859]: I1008 18:36:50.889216 4859 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-metadata-0" podUID="f8de039d-85d9-444b-937e-d8d6f90bb77f" containerName="nova-metadata-log" containerID="cri-o://719956fb0a5ddd763170476c1c204c6c8a78e5b6f8bdc405211df91e0e9eff5b" gracePeriod=30 Oct 08 18:36:50 crc kubenswrapper[4859]: I1008 18:36:50.889633 4859 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-metadata-0" podUID="f8de039d-85d9-444b-937e-d8d6f90bb77f" containerName="nova-metadata-metadata" containerID="cri-o://dce4317d8887820fbdd21c3fd9beefcc68a86c05505bb4dd820e11a9c745a137" gracePeriod=30 Oct 08 18:36:51 crc kubenswrapper[4859]: I1008 18:36:51.570797 4859 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Oct 08 18:36:51 crc kubenswrapper[4859]: I1008 18:36:51.670093 4859 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/01aea24c-9805-4a41-9843-c3366d597442-internal-tls-certs\") pod \"01aea24c-9805-4a41-9843-c3366d597442\" (UID: \"01aea24c-9805-4a41-9843-c3366d597442\") " Oct 08 18:36:51 crc kubenswrapper[4859]: I1008 18:36:51.670637 4859 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/01aea24c-9805-4a41-9843-c3366d597442-public-tls-certs\") pod \"01aea24c-9805-4a41-9843-c3366d597442\" (UID: \"01aea24c-9805-4a41-9843-c3366d597442\") " Oct 08 18:36:51 crc kubenswrapper[4859]: I1008 18:36:51.670679 4859 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/01aea24c-9805-4a41-9843-c3366d597442-logs\") pod \"01aea24c-9805-4a41-9843-c3366d597442\" (UID: \"01aea24c-9805-4a41-9843-c3366d597442\") " Oct 08 18:36:51 crc kubenswrapper[4859]: I1008 18:36:51.670812 4859 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/01aea24c-9805-4a41-9843-c3366d597442-combined-ca-bundle\") pod \"01aea24c-9805-4a41-9843-c3366d597442\" (UID: \"01aea24c-9805-4a41-9843-c3366d597442\") " Oct 08 18:36:51 crc kubenswrapper[4859]: I1008 18:36:51.670874 4859 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-hj89c\" (UniqueName: \"kubernetes.io/projected/01aea24c-9805-4a41-9843-c3366d597442-kube-api-access-hj89c\") pod \"01aea24c-9805-4a41-9843-c3366d597442\" (UID: \"01aea24c-9805-4a41-9843-c3366d597442\") " Oct 08 18:36:51 crc kubenswrapper[4859]: I1008 18:36:51.670899 4859 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/01aea24c-9805-4a41-9843-c3366d597442-config-data\") pod \"01aea24c-9805-4a41-9843-c3366d597442\" (UID: \"01aea24c-9805-4a41-9843-c3366d597442\") " Oct 08 18:36:51 crc kubenswrapper[4859]: I1008 18:36:51.671630 4859 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/01aea24c-9805-4a41-9843-c3366d597442-logs" (OuterVolumeSpecName: "logs") pod "01aea24c-9805-4a41-9843-c3366d597442" (UID: "01aea24c-9805-4a41-9843-c3366d597442"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 08 18:36:51 crc kubenswrapper[4859]: I1008 18:36:51.680585 4859 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/01aea24c-9805-4a41-9843-c3366d597442-kube-api-access-hj89c" (OuterVolumeSpecName: "kube-api-access-hj89c") pod "01aea24c-9805-4a41-9843-c3366d597442" (UID: "01aea24c-9805-4a41-9843-c3366d597442"). InnerVolumeSpecName "kube-api-access-hj89c". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 08 18:36:51 crc kubenswrapper[4859]: I1008 18:36:51.695775 4859 generic.go:334] "Generic (PLEG): container finished" podID="01aea24c-9805-4a41-9843-c3366d597442" containerID="229a8d291803695dc1c54a515b948e4483d245d8f32aae09cbcd61a368018a4d" exitCode=0 Oct 08 18:36:51 crc kubenswrapper[4859]: I1008 18:36:51.695812 4859 generic.go:334] "Generic (PLEG): container finished" podID="01aea24c-9805-4a41-9843-c3366d597442" containerID="9ff8822be5a79fd192a815cbffbe8666bbbb558d24c84188e5a97ffa080d975a" exitCode=143 Oct 08 18:36:51 crc kubenswrapper[4859]: I1008 18:36:51.695874 4859 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"01aea24c-9805-4a41-9843-c3366d597442","Type":"ContainerDied","Data":"229a8d291803695dc1c54a515b948e4483d245d8f32aae09cbcd61a368018a4d"} Oct 08 18:36:51 crc kubenswrapper[4859]: I1008 18:36:51.695904 4859 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"01aea24c-9805-4a41-9843-c3366d597442","Type":"ContainerDied","Data":"9ff8822be5a79fd192a815cbffbe8666bbbb558d24c84188e5a97ffa080d975a"} Oct 08 18:36:51 crc kubenswrapper[4859]: I1008 18:36:51.695917 4859 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"01aea24c-9805-4a41-9843-c3366d597442","Type":"ContainerDied","Data":"33a274cc1a175459eb3c111f7ca40ff853a75cc34abb7325a37166cd63b59c7a"} Oct 08 18:36:51 crc kubenswrapper[4859]: I1008 18:36:51.695935 4859 scope.go:117] "RemoveContainer" containerID="229a8d291803695dc1c54a515b948e4483d245d8f32aae09cbcd61a368018a4d" Oct 08 18:36:51 crc kubenswrapper[4859]: I1008 18:36:51.696151 4859 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Oct 08 18:36:51 crc kubenswrapper[4859]: I1008 18:36:51.703042 4859 generic.go:334] "Generic (PLEG): container finished" podID="f8de039d-85d9-444b-937e-d8d6f90bb77f" containerID="719956fb0a5ddd763170476c1c204c6c8a78e5b6f8bdc405211df91e0e9eff5b" exitCode=143 Oct 08 18:36:51 crc kubenswrapper[4859]: I1008 18:36:51.703159 4859 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"f8de039d-85d9-444b-937e-d8d6f90bb77f","Type":"ContainerDied","Data":"719956fb0a5ddd763170476c1c204c6c8a78e5b6f8bdc405211df91e0e9eff5b"} Oct 08 18:36:51 crc kubenswrapper[4859]: I1008 18:36:51.711786 4859 generic.go:334] "Generic (PLEG): container finished" podID="e5f82275-e5ae-4d2d-8ab6-dc4afda848ba" containerID="4fe8f76dc74c296a52a35017a865d9caf429643b340fffe1bc9156fa36b2c04f" exitCode=0 Oct 08 18:36:51 crc kubenswrapper[4859]: I1008 18:36:51.712381 4859 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"e5f82275-e5ae-4d2d-8ab6-dc4afda848ba","Type":"ContainerDied","Data":"4fe8f76dc74c296a52a35017a865d9caf429643b340fffe1bc9156fa36b2c04f"} Oct 08 18:36:51 crc kubenswrapper[4859]: I1008 18:36:51.724304 4859 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/01aea24c-9805-4a41-9843-c3366d597442-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "01aea24c-9805-4a41-9843-c3366d597442" (UID: "01aea24c-9805-4a41-9843-c3366d597442"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 08 18:36:51 crc kubenswrapper[4859]: I1008 18:36:51.730603 4859 scope.go:117] "RemoveContainer" containerID="9ff8822be5a79fd192a815cbffbe8666bbbb558d24c84188e5a97ffa080d975a" Oct 08 18:36:51 crc kubenswrapper[4859]: I1008 18:36:51.730843 4859 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/01aea24c-9805-4a41-9843-c3366d597442-config-data" (OuterVolumeSpecName: "config-data") pod "01aea24c-9805-4a41-9843-c3366d597442" (UID: "01aea24c-9805-4a41-9843-c3366d597442"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 08 18:36:51 crc kubenswrapper[4859]: I1008 18:36:51.734420 4859 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/01aea24c-9805-4a41-9843-c3366d597442-internal-tls-certs" (OuterVolumeSpecName: "internal-tls-certs") pod "01aea24c-9805-4a41-9843-c3366d597442" (UID: "01aea24c-9805-4a41-9843-c3366d597442"). InnerVolumeSpecName "internal-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 08 18:36:51 crc kubenswrapper[4859]: I1008 18:36:51.759557 4859 scope.go:117] "RemoveContainer" containerID="229a8d291803695dc1c54a515b948e4483d245d8f32aae09cbcd61a368018a4d" Oct 08 18:36:51 crc kubenswrapper[4859]: E1008 18:36:51.760520 4859 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"229a8d291803695dc1c54a515b948e4483d245d8f32aae09cbcd61a368018a4d\": container with ID starting with 229a8d291803695dc1c54a515b948e4483d245d8f32aae09cbcd61a368018a4d not found: ID does not exist" containerID="229a8d291803695dc1c54a515b948e4483d245d8f32aae09cbcd61a368018a4d" Oct 08 18:36:51 crc kubenswrapper[4859]: I1008 18:36:51.760586 4859 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"229a8d291803695dc1c54a515b948e4483d245d8f32aae09cbcd61a368018a4d"} err="failed to get container status \"229a8d291803695dc1c54a515b948e4483d245d8f32aae09cbcd61a368018a4d\": rpc error: code = NotFound desc = could not find container \"229a8d291803695dc1c54a515b948e4483d245d8f32aae09cbcd61a368018a4d\": container with ID starting with 229a8d291803695dc1c54a515b948e4483d245d8f32aae09cbcd61a368018a4d not found: ID does not exist" Oct 08 18:36:51 crc kubenswrapper[4859]: I1008 18:36:51.760620 4859 scope.go:117] "RemoveContainer" containerID="9ff8822be5a79fd192a815cbffbe8666bbbb558d24c84188e5a97ffa080d975a" Oct 08 18:36:51 crc kubenswrapper[4859]: E1008 18:36:51.761092 4859 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"9ff8822be5a79fd192a815cbffbe8666bbbb558d24c84188e5a97ffa080d975a\": container with ID starting with 9ff8822be5a79fd192a815cbffbe8666bbbb558d24c84188e5a97ffa080d975a not found: ID does not exist" containerID="9ff8822be5a79fd192a815cbffbe8666bbbb558d24c84188e5a97ffa080d975a" Oct 08 18:36:51 crc kubenswrapper[4859]: I1008 18:36:51.761129 4859 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"9ff8822be5a79fd192a815cbffbe8666bbbb558d24c84188e5a97ffa080d975a"} err="failed to get container status \"9ff8822be5a79fd192a815cbffbe8666bbbb558d24c84188e5a97ffa080d975a\": rpc error: code = NotFound desc = could not find container \"9ff8822be5a79fd192a815cbffbe8666bbbb558d24c84188e5a97ffa080d975a\": container with ID starting with 9ff8822be5a79fd192a815cbffbe8666bbbb558d24c84188e5a97ffa080d975a not found: ID does not exist" Oct 08 18:36:51 crc kubenswrapper[4859]: I1008 18:36:51.761178 4859 scope.go:117] "RemoveContainer" containerID="229a8d291803695dc1c54a515b948e4483d245d8f32aae09cbcd61a368018a4d" Oct 08 18:36:51 crc kubenswrapper[4859]: I1008 18:36:51.761423 4859 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"229a8d291803695dc1c54a515b948e4483d245d8f32aae09cbcd61a368018a4d"} err="failed to get container status \"229a8d291803695dc1c54a515b948e4483d245d8f32aae09cbcd61a368018a4d\": rpc error: code = NotFound desc = could not find container \"229a8d291803695dc1c54a515b948e4483d245d8f32aae09cbcd61a368018a4d\": container with ID starting with 229a8d291803695dc1c54a515b948e4483d245d8f32aae09cbcd61a368018a4d not found: ID does not exist" Oct 08 18:36:51 crc kubenswrapper[4859]: I1008 18:36:51.761532 4859 scope.go:117] "RemoveContainer" containerID="9ff8822be5a79fd192a815cbffbe8666bbbb558d24c84188e5a97ffa080d975a" Oct 08 18:36:51 crc kubenswrapper[4859]: I1008 18:36:51.761870 4859 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"9ff8822be5a79fd192a815cbffbe8666bbbb558d24c84188e5a97ffa080d975a"} err="failed to get container status \"9ff8822be5a79fd192a815cbffbe8666bbbb558d24c84188e5a97ffa080d975a\": rpc error: code = NotFound desc = could not find container \"9ff8822be5a79fd192a815cbffbe8666bbbb558d24c84188e5a97ffa080d975a\": container with ID starting with 9ff8822be5a79fd192a815cbffbe8666bbbb558d24c84188e5a97ffa080d975a not found: ID does not exist" Oct 08 18:36:51 crc kubenswrapper[4859]: I1008 18:36:51.764096 4859 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/01aea24c-9805-4a41-9843-c3366d597442-public-tls-certs" (OuterVolumeSpecName: "public-tls-certs") pod "01aea24c-9805-4a41-9843-c3366d597442" (UID: "01aea24c-9805-4a41-9843-c3366d597442"). InnerVolumeSpecName "public-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 08 18:36:51 crc kubenswrapper[4859]: I1008 18:36:51.778000 4859 reconciler_common.go:293] "Volume detached for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/01aea24c-9805-4a41-9843-c3366d597442-internal-tls-certs\") on node \"crc\" DevicePath \"\"" Oct 08 18:36:51 crc kubenswrapper[4859]: I1008 18:36:51.778041 4859 reconciler_common.go:293] "Volume detached for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/01aea24c-9805-4a41-9843-c3366d597442-public-tls-certs\") on node \"crc\" DevicePath \"\"" Oct 08 18:36:51 crc kubenswrapper[4859]: I1008 18:36:51.778053 4859 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/01aea24c-9805-4a41-9843-c3366d597442-logs\") on node \"crc\" DevicePath \"\"" Oct 08 18:36:51 crc kubenswrapper[4859]: I1008 18:36:51.778069 4859 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/01aea24c-9805-4a41-9843-c3366d597442-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 08 18:36:51 crc kubenswrapper[4859]: I1008 18:36:51.778081 4859 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-hj89c\" (UniqueName: \"kubernetes.io/projected/01aea24c-9805-4a41-9843-c3366d597442-kube-api-access-hj89c\") on node \"crc\" DevicePath \"\"" Oct 08 18:36:51 crc kubenswrapper[4859]: I1008 18:36:51.778093 4859 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/01aea24c-9805-4a41-9843-c3366d597442-config-data\") on node \"crc\" DevicePath \"\"" Oct 08 18:36:51 crc kubenswrapper[4859]: I1008 18:36:51.886844 4859 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Oct 08 18:36:51 crc kubenswrapper[4859]: I1008 18:36:51.980766 4859 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-xm54t\" (UniqueName: \"kubernetes.io/projected/e5f82275-e5ae-4d2d-8ab6-dc4afda848ba-kube-api-access-xm54t\") pod \"e5f82275-e5ae-4d2d-8ab6-dc4afda848ba\" (UID: \"e5f82275-e5ae-4d2d-8ab6-dc4afda848ba\") " Oct 08 18:36:51 crc kubenswrapper[4859]: I1008 18:36:51.980866 4859 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/e5f82275-e5ae-4d2d-8ab6-dc4afda848ba-config-data\") pod \"e5f82275-e5ae-4d2d-8ab6-dc4afda848ba\" (UID: \"e5f82275-e5ae-4d2d-8ab6-dc4afda848ba\") " Oct 08 18:36:51 crc kubenswrapper[4859]: I1008 18:36:51.980967 4859 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e5f82275-e5ae-4d2d-8ab6-dc4afda848ba-combined-ca-bundle\") pod \"e5f82275-e5ae-4d2d-8ab6-dc4afda848ba\" (UID: \"e5f82275-e5ae-4d2d-8ab6-dc4afda848ba\") " Oct 08 18:36:51 crc kubenswrapper[4859]: I1008 18:36:51.996913 4859 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/e5f82275-e5ae-4d2d-8ab6-dc4afda848ba-kube-api-access-xm54t" (OuterVolumeSpecName: "kube-api-access-xm54t") pod "e5f82275-e5ae-4d2d-8ab6-dc4afda848ba" (UID: "e5f82275-e5ae-4d2d-8ab6-dc4afda848ba"). InnerVolumeSpecName "kube-api-access-xm54t". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 08 18:36:52 crc kubenswrapper[4859]: I1008 18:36:52.021119 4859 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e5f82275-e5ae-4d2d-8ab6-dc4afda848ba-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "e5f82275-e5ae-4d2d-8ab6-dc4afda848ba" (UID: "e5f82275-e5ae-4d2d-8ab6-dc4afda848ba"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 08 18:36:52 crc kubenswrapper[4859]: I1008 18:36:52.023381 4859 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e5f82275-e5ae-4d2d-8ab6-dc4afda848ba-config-data" (OuterVolumeSpecName: "config-data") pod "e5f82275-e5ae-4d2d-8ab6-dc4afda848ba" (UID: "e5f82275-e5ae-4d2d-8ab6-dc4afda848ba"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 08 18:36:52 crc kubenswrapper[4859]: I1008 18:36:52.082895 4859 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-xm54t\" (UniqueName: \"kubernetes.io/projected/e5f82275-e5ae-4d2d-8ab6-dc4afda848ba-kube-api-access-xm54t\") on node \"crc\" DevicePath \"\"" Oct 08 18:36:52 crc kubenswrapper[4859]: I1008 18:36:52.082933 4859 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/e5f82275-e5ae-4d2d-8ab6-dc4afda848ba-config-data\") on node \"crc\" DevicePath \"\"" Oct 08 18:36:52 crc kubenswrapper[4859]: I1008 18:36:52.082944 4859 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e5f82275-e5ae-4d2d-8ab6-dc4afda848ba-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 08 18:36:52 crc kubenswrapper[4859]: I1008 18:36:52.099128 4859 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-api-0"] Oct 08 18:36:52 crc kubenswrapper[4859]: I1008 18:36:52.106490 4859 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-api-0"] Oct 08 18:36:52 crc kubenswrapper[4859]: I1008 18:36:52.119983 4859 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-api-0"] Oct 08 18:36:52 crc kubenswrapper[4859]: E1008 18:36:52.120441 4859 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e5f82275-e5ae-4d2d-8ab6-dc4afda848ba" containerName="nova-scheduler-scheduler" Oct 08 18:36:52 crc kubenswrapper[4859]: I1008 18:36:52.120467 4859 state_mem.go:107] "Deleted CPUSet assignment" podUID="e5f82275-e5ae-4d2d-8ab6-dc4afda848ba" containerName="nova-scheduler-scheduler" Oct 08 18:36:52 crc kubenswrapper[4859]: E1008 18:36:52.120486 4859 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="01aea24c-9805-4a41-9843-c3366d597442" containerName="nova-api-log" Oct 08 18:36:52 crc kubenswrapper[4859]: I1008 18:36:52.120494 4859 state_mem.go:107] "Deleted CPUSet assignment" podUID="01aea24c-9805-4a41-9843-c3366d597442" containerName="nova-api-log" Oct 08 18:36:52 crc kubenswrapper[4859]: E1008 18:36:52.120516 4859 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7a928d30-8b64-4daf-9731-abea3343acbc" containerName="nova-manage" Oct 08 18:36:52 crc kubenswrapper[4859]: I1008 18:36:52.120523 4859 state_mem.go:107] "Deleted CPUSet assignment" podUID="7a928d30-8b64-4daf-9731-abea3343acbc" containerName="nova-manage" Oct 08 18:36:52 crc kubenswrapper[4859]: E1008 18:36:52.120540 4859 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="01aea24c-9805-4a41-9843-c3366d597442" containerName="nova-api-api" Oct 08 18:36:52 crc kubenswrapper[4859]: I1008 18:36:52.120549 4859 state_mem.go:107] "Deleted CPUSet assignment" podUID="01aea24c-9805-4a41-9843-c3366d597442" containerName="nova-api-api" Oct 08 18:36:52 crc kubenswrapper[4859]: E1008 18:36:52.120581 4859 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1ea66084-f5e7-4fcc-a439-c7806dde0197" containerName="dnsmasq-dns" Oct 08 18:36:52 crc kubenswrapper[4859]: I1008 18:36:52.120591 4859 state_mem.go:107] "Deleted CPUSet assignment" podUID="1ea66084-f5e7-4fcc-a439-c7806dde0197" containerName="dnsmasq-dns" Oct 08 18:36:52 crc kubenswrapper[4859]: E1008 18:36:52.120609 4859 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1ea66084-f5e7-4fcc-a439-c7806dde0197" containerName="init" Oct 08 18:36:52 crc kubenswrapper[4859]: I1008 18:36:52.120617 4859 state_mem.go:107] "Deleted CPUSet assignment" podUID="1ea66084-f5e7-4fcc-a439-c7806dde0197" containerName="init" Oct 08 18:36:52 crc kubenswrapper[4859]: I1008 18:36:52.120849 4859 memory_manager.go:354] "RemoveStaleState removing state" podUID="1ea66084-f5e7-4fcc-a439-c7806dde0197" containerName="dnsmasq-dns" Oct 08 18:36:52 crc kubenswrapper[4859]: I1008 18:36:52.120880 4859 memory_manager.go:354] "RemoveStaleState removing state" podUID="7a928d30-8b64-4daf-9731-abea3343acbc" containerName="nova-manage" Oct 08 18:36:52 crc kubenswrapper[4859]: I1008 18:36:52.120891 4859 memory_manager.go:354] "RemoveStaleState removing state" podUID="e5f82275-e5ae-4d2d-8ab6-dc4afda848ba" containerName="nova-scheduler-scheduler" Oct 08 18:36:52 crc kubenswrapper[4859]: I1008 18:36:52.120905 4859 memory_manager.go:354] "RemoveStaleState removing state" podUID="01aea24c-9805-4a41-9843-c3366d597442" containerName="nova-api-log" Oct 08 18:36:52 crc kubenswrapper[4859]: I1008 18:36:52.120930 4859 memory_manager.go:354] "RemoveStaleState removing state" podUID="01aea24c-9805-4a41-9843-c3366d597442" containerName="nova-api-api" Oct 08 18:36:52 crc kubenswrapper[4859]: I1008 18:36:52.122126 4859 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Oct 08 18:36:52 crc kubenswrapper[4859]: I1008 18:36:52.124615 4859 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-nova-public-svc" Oct 08 18:36:52 crc kubenswrapper[4859]: I1008 18:36:52.125965 4859 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-api-config-data" Oct 08 18:36:52 crc kubenswrapper[4859]: I1008 18:36:52.126301 4859 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-nova-internal-svc" Oct 08 18:36:52 crc kubenswrapper[4859]: I1008 18:36:52.142280 4859 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-0"] Oct 08 18:36:52 crc kubenswrapper[4859]: I1008 18:36:52.285645 4859 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/e56f2ed9-f7a3-4157-bb50-88fbd86d9c60-logs\") pod \"nova-api-0\" (UID: \"e56f2ed9-f7a3-4157-bb50-88fbd86d9c60\") " pod="openstack/nova-api-0" Oct 08 18:36:52 crc kubenswrapper[4859]: I1008 18:36:52.285772 4859 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/e56f2ed9-f7a3-4157-bb50-88fbd86d9c60-config-data\") pod \"nova-api-0\" (UID: \"e56f2ed9-f7a3-4157-bb50-88fbd86d9c60\") " pod="openstack/nova-api-0" Oct 08 18:36:52 crc kubenswrapper[4859]: I1008 18:36:52.285806 4859 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/e56f2ed9-f7a3-4157-bb50-88fbd86d9c60-public-tls-certs\") pod \"nova-api-0\" (UID: \"e56f2ed9-f7a3-4157-bb50-88fbd86d9c60\") " pod="openstack/nova-api-0" Oct 08 18:36:52 crc kubenswrapper[4859]: I1008 18:36:52.285871 4859 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-sq9tx\" (UniqueName: \"kubernetes.io/projected/e56f2ed9-f7a3-4157-bb50-88fbd86d9c60-kube-api-access-sq9tx\") pod \"nova-api-0\" (UID: \"e56f2ed9-f7a3-4157-bb50-88fbd86d9c60\") " pod="openstack/nova-api-0" Oct 08 18:36:52 crc kubenswrapper[4859]: I1008 18:36:52.286075 4859 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/e56f2ed9-f7a3-4157-bb50-88fbd86d9c60-internal-tls-certs\") pod \"nova-api-0\" (UID: \"e56f2ed9-f7a3-4157-bb50-88fbd86d9c60\") " pod="openstack/nova-api-0" Oct 08 18:36:52 crc kubenswrapper[4859]: I1008 18:36:52.286165 4859 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e56f2ed9-f7a3-4157-bb50-88fbd86d9c60-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"e56f2ed9-f7a3-4157-bb50-88fbd86d9c60\") " pod="openstack/nova-api-0" Oct 08 18:36:52 crc kubenswrapper[4859]: I1008 18:36:52.388068 4859 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/e56f2ed9-f7a3-4157-bb50-88fbd86d9c60-logs\") pod \"nova-api-0\" (UID: \"e56f2ed9-f7a3-4157-bb50-88fbd86d9c60\") " pod="openstack/nova-api-0" Oct 08 18:36:52 crc kubenswrapper[4859]: I1008 18:36:52.388152 4859 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/e56f2ed9-f7a3-4157-bb50-88fbd86d9c60-config-data\") pod \"nova-api-0\" (UID: \"e56f2ed9-f7a3-4157-bb50-88fbd86d9c60\") " pod="openstack/nova-api-0" Oct 08 18:36:52 crc kubenswrapper[4859]: I1008 18:36:52.388175 4859 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/e56f2ed9-f7a3-4157-bb50-88fbd86d9c60-public-tls-certs\") pod \"nova-api-0\" (UID: \"e56f2ed9-f7a3-4157-bb50-88fbd86d9c60\") " pod="openstack/nova-api-0" Oct 08 18:36:52 crc kubenswrapper[4859]: I1008 18:36:52.388197 4859 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-sq9tx\" (UniqueName: \"kubernetes.io/projected/e56f2ed9-f7a3-4157-bb50-88fbd86d9c60-kube-api-access-sq9tx\") pod \"nova-api-0\" (UID: \"e56f2ed9-f7a3-4157-bb50-88fbd86d9c60\") " pod="openstack/nova-api-0" Oct 08 18:36:52 crc kubenswrapper[4859]: I1008 18:36:52.388228 4859 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/e56f2ed9-f7a3-4157-bb50-88fbd86d9c60-internal-tls-certs\") pod \"nova-api-0\" (UID: \"e56f2ed9-f7a3-4157-bb50-88fbd86d9c60\") " pod="openstack/nova-api-0" Oct 08 18:36:52 crc kubenswrapper[4859]: I1008 18:36:52.388249 4859 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e56f2ed9-f7a3-4157-bb50-88fbd86d9c60-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"e56f2ed9-f7a3-4157-bb50-88fbd86d9c60\") " pod="openstack/nova-api-0" Oct 08 18:36:52 crc kubenswrapper[4859]: I1008 18:36:52.389415 4859 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/e56f2ed9-f7a3-4157-bb50-88fbd86d9c60-logs\") pod \"nova-api-0\" (UID: \"e56f2ed9-f7a3-4157-bb50-88fbd86d9c60\") " pod="openstack/nova-api-0" Oct 08 18:36:52 crc kubenswrapper[4859]: I1008 18:36:52.393110 4859 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/e56f2ed9-f7a3-4157-bb50-88fbd86d9c60-internal-tls-certs\") pod \"nova-api-0\" (UID: \"e56f2ed9-f7a3-4157-bb50-88fbd86d9c60\") " pod="openstack/nova-api-0" Oct 08 18:36:52 crc kubenswrapper[4859]: I1008 18:36:52.393277 4859 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/e56f2ed9-f7a3-4157-bb50-88fbd86d9c60-config-data\") pod \"nova-api-0\" (UID: \"e56f2ed9-f7a3-4157-bb50-88fbd86d9c60\") " pod="openstack/nova-api-0" Oct 08 18:36:52 crc kubenswrapper[4859]: I1008 18:36:52.396461 4859 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e56f2ed9-f7a3-4157-bb50-88fbd86d9c60-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"e56f2ed9-f7a3-4157-bb50-88fbd86d9c60\") " pod="openstack/nova-api-0" Oct 08 18:36:52 crc kubenswrapper[4859]: I1008 18:36:52.398217 4859 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/e56f2ed9-f7a3-4157-bb50-88fbd86d9c60-public-tls-certs\") pod \"nova-api-0\" (UID: \"e56f2ed9-f7a3-4157-bb50-88fbd86d9c60\") " pod="openstack/nova-api-0" Oct 08 18:36:52 crc kubenswrapper[4859]: I1008 18:36:52.405471 4859 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-sq9tx\" (UniqueName: \"kubernetes.io/projected/e56f2ed9-f7a3-4157-bb50-88fbd86d9c60-kube-api-access-sq9tx\") pod \"nova-api-0\" (UID: \"e56f2ed9-f7a3-4157-bb50-88fbd86d9c60\") " pod="openstack/nova-api-0" Oct 08 18:36:52 crc kubenswrapper[4859]: I1008 18:36:52.443282 4859 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Oct 08 18:36:52 crc kubenswrapper[4859]: I1008 18:36:52.487524 4859 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="01aea24c-9805-4a41-9843-c3366d597442" path="/var/lib/kubelet/pods/01aea24c-9805-4a41-9843-c3366d597442/volumes" Oct 08 18:36:52 crc kubenswrapper[4859]: I1008 18:36:52.724544 4859 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"e5f82275-e5ae-4d2d-8ab6-dc4afda848ba","Type":"ContainerDied","Data":"3f8cc4d2069435407bc2679facaa25edfb59b662f317d01fed8daa54ee38bff2"} Oct 08 18:36:52 crc kubenswrapper[4859]: I1008 18:36:52.724925 4859 scope.go:117] "RemoveContainer" containerID="4fe8f76dc74c296a52a35017a865d9caf429643b340fffe1bc9156fa36b2c04f" Oct 08 18:36:52 crc kubenswrapper[4859]: I1008 18:36:52.724883 4859 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Oct 08 18:36:52 crc kubenswrapper[4859]: I1008 18:36:52.746291 4859 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-scheduler-0"] Oct 08 18:36:52 crc kubenswrapper[4859]: I1008 18:36:52.760242 4859 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-scheduler-0"] Oct 08 18:36:52 crc kubenswrapper[4859]: I1008 18:36:52.769915 4859 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-scheduler-0"] Oct 08 18:36:52 crc kubenswrapper[4859]: I1008 18:36:52.771355 4859 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Oct 08 18:36:52 crc kubenswrapper[4859]: I1008 18:36:52.774467 4859 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-scheduler-config-data" Oct 08 18:36:52 crc kubenswrapper[4859]: I1008 18:36:52.780994 4859 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-scheduler-0"] Oct 08 18:36:52 crc kubenswrapper[4859]: I1008 18:36:52.865845 4859 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-0"] Oct 08 18:36:52 crc kubenswrapper[4859]: W1008 18:36:52.870170 4859 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pode56f2ed9_f7a3_4157_bb50_88fbd86d9c60.slice/crio-7e21164ef1c1846213a59ab3002c756bdbedb3ca855063ed38dc90e09e01163d WatchSource:0}: Error finding container 7e21164ef1c1846213a59ab3002c756bdbedb3ca855063ed38dc90e09e01163d: Status 404 returned error can't find the container with id 7e21164ef1c1846213a59ab3002c756bdbedb3ca855063ed38dc90e09e01163d Oct 08 18:36:52 crc kubenswrapper[4859]: I1008 18:36:52.900430 4859 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-79f4t\" (UniqueName: \"kubernetes.io/projected/70906bc9-3b52-4436-8d68-cdb37f0d2478-kube-api-access-79f4t\") pod \"nova-scheduler-0\" (UID: \"70906bc9-3b52-4436-8d68-cdb37f0d2478\") " pod="openstack/nova-scheduler-0" Oct 08 18:36:52 crc kubenswrapper[4859]: I1008 18:36:52.900536 4859 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/70906bc9-3b52-4436-8d68-cdb37f0d2478-config-data\") pod \"nova-scheduler-0\" (UID: \"70906bc9-3b52-4436-8d68-cdb37f0d2478\") " pod="openstack/nova-scheduler-0" Oct 08 18:36:52 crc kubenswrapper[4859]: I1008 18:36:52.900630 4859 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/70906bc9-3b52-4436-8d68-cdb37f0d2478-combined-ca-bundle\") pod \"nova-scheduler-0\" (UID: \"70906bc9-3b52-4436-8d68-cdb37f0d2478\") " pod="openstack/nova-scheduler-0" Oct 08 18:36:53 crc kubenswrapper[4859]: I1008 18:36:53.002840 4859 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-79f4t\" (UniqueName: \"kubernetes.io/projected/70906bc9-3b52-4436-8d68-cdb37f0d2478-kube-api-access-79f4t\") pod \"nova-scheduler-0\" (UID: \"70906bc9-3b52-4436-8d68-cdb37f0d2478\") " pod="openstack/nova-scheduler-0" Oct 08 18:36:53 crc kubenswrapper[4859]: I1008 18:36:53.002935 4859 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/70906bc9-3b52-4436-8d68-cdb37f0d2478-config-data\") pod \"nova-scheduler-0\" (UID: \"70906bc9-3b52-4436-8d68-cdb37f0d2478\") " pod="openstack/nova-scheduler-0" Oct 08 18:36:53 crc kubenswrapper[4859]: I1008 18:36:53.003032 4859 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/70906bc9-3b52-4436-8d68-cdb37f0d2478-combined-ca-bundle\") pod \"nova-scheduler-0\" (UID: \"70906bc9-3b52-4436-8d68-cdb37f0d2478\") " pod="openstack/nova-scheduler-0" Oct 08 18:36:53 crc kubenswrapper[4859]: I1008 18:36:53.009097 4859 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/70906bc9-3b52-4436-8d68-cdb37f0d2478-config-data\") pod \"nova-scheduler-0\" (UID: \"70906bc9-3b52-4436-8d68-cdb37f0d2478\") " pod="openstack/nova-scheduler-0" Oct 08 18:36:53 crc kubenswrapper[4859]: I1008 18:36:53.009377 4859 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/70906bc9-3b52-4436-8d68-cdb37f0d2478-combined-ca-bundle\") pod \"nova-scheduler-0\" (UID: \"70906bc9-3b52-4436-8d68-cdb37f0d2478\") " pod="openstack/nova-scheduler-0" Oct 08 18:36:53 crc kubenswrapper[4859]: I1008 18:36:53.021431 4859 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-79f4t\" (UniqueName: \"kubernetes.io/projected/70906bc9-3b52-4436-8d68-cdb37f0d2478-kube-api-access-79f4t\") pod \"nova-scheduler-0\" (UID: \"70906bc9-3b52-4436-8d68-cdb37f0d2478\") " pod="openstack/nova-scheduler-0" Oct 08 18:36:53 crc kubenswrapper[4859]: I1008 18:36:53.109596 4859 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Oct 08 18:36:53 crc kubenswrapper[4859]: I1008 18:36:53.547764 4859 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-scheduler-0"] Oct 08 18:36:53 crc kubenswrapper[4859]: I1008 18:36:53.737459 4859 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"e56f2ed9-f7a3-4157-bb50-88fbd86d9c60","Type":"ContainerStarted","Data":"cae112af652d6012b04b5914ca0b51c4f43ae7bb5b7ca2bb31b1769b0100762e"} Oct 08 18:36:53 crc kubenswrapper[4859]: I1008 18:36:53.737895 4859 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"e56f2ed9-f7a3-4157-bb50-88fbd86d9c60","Type":"ContainerStarted","Data":"e98e9d395b545aa48e86a522f095f97c479dbf3b918f49666e5749148f6f689d"} Oct 08 18:36:53 crc kubenswrapper[4859]: I1008 18:36:53.737913 4859 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"e56f2ed9-f7a3-4157-bb50-88fbd86d9c60","Type":"ContainerStarted","Data":"7e21164ef1c1846213a59ab3002c756bdbedb3ca855063ed38dc90e09e01163d"} Oct 08 18:36:53 crc kubenswrapper[4859]: I1008 18:36:53.744883 4859 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"70906bc9-3b52-4436-8d68-cdb37f0d2478","Type":"ContainerStarted","Data":"6b952f5849a0d0cb268d5a9ef7b249a9a56d23cbecc607af5dc2c40e33b2cb2d"} Oct 08 18:36:53 crc kubenswrapper[4859]: I1008 18:36:53.744943 4859 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"70906bc9-3b52-4436-8d68-cdb37f0d2478","Type":"ContainerStarted","Data":"5f490c5c789ce3712bada8bb9bbc183c9a63d40ac3aed9c35d55b9fa7a54b3e1"} Oct 08 18:36:53 crc kubenswrapper[4859]: I1008 18:36:53.782523 4859 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-scheduler-0" podStartSLOduration=1.782501748 podStartE2EDuration="1.782501748s" podCreationTimestamp="2025-10-08 18:36:52 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-08 18:36:53.781372916 +0000 UTC m=+1184.028212325" watchObservedRunningTime="2025-10-08 18:36:53.782501748 +0000 UTC m=+1184.029341137" Oct 08 18:36:53 crc kubenswrapper[4859]: I1008 18:36:53.783791 4859 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-api-0" podStartSLOduration=1.783784265 podStartE2EDuration="1.783784265s" podCreationTimestamp="2025-10-08 18:36:52 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-08 18:36:53.760461888 +0000 UTC m=+1184.007301297" watchObservedRunningTime="2025-10-08 18:36:53.783784265 +0000 UTC m=+1184.030623654" Oct 08 18:36:54 crc kubenswrapper[4859]: I1008 18:36:54.035126 4859 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/nova-metadata-0" podUID="f8de039d-85d9-444b-937e-d8d6f90bb77f" containerName="nova-metadata-log" probeResult="failure" output="Get \"https://10.217.0.194:8775/\": read tcp 10.217.0.2:60466->10.217.0.194:8775: read: connection reset by peer" Oct 08 18:36:54 crc kubenswrapper[4859]: I1008 18:36:54.036036 4859 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/nova-metadata-0" podUID="f8de039d-85d9-444b-937e-d8d6f90bb77f" containerName="nova-metadata-metadata" probeResult="failure" output="Get \"https://10.217.0.194:8775/\": read tcp 10.217.0.2:60474->10.217.0.194:8775: read: connection reset by peer" Oct 08 18:36:54 crc kubenswrapper[4859]: I1008 18:36:54.473769 4859 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Oct 08 18:36:54 crc kubenswrapper[4859]: I1008 18:36:54.482751 4859 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="e5f82275-e5ae-4d2d-8ab6-dc4afda848ba" path="/var/lib/kubelet/pods/e5f82275-e5ae-4d2d-8ab6-dc4afda848ba/volumes" Oct 08 18:36:54 crc kubenswrapper[4859]: I1008 18:36:54.633089 4859 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f8de039d-85d9-444b-937e-d8d6f90bb77f-config-data\") pod \"f8de039d-85d9-444b-937e-d8d6f90bb77f\" (UID: \"f8de039d-85d9-444b-937e-d8d6f90bb77f\") " Oct 08 18:36:54 crc kubenswrapper[4859]: I1008 18:36:54.633142 4859 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/f8de039d-85d9-444b-937e-d8d6f90bb77f-nova-metadata-tls-certs\") pod \"f8de039d-85d9-444b-937e-d8d6f90bb77f\" (UID: \"f8de039d-85d9-444b-937e-d8d6f90bb77f\") " Oct 08 18:36:54 crc kubenswrapper[4859]: I1008 18:36:54.633200 4859 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f8de039d-85d9-444b-937e-d8d6f90bb77f-combined-ca-bundle\") pod \"f8de039d-85d9-444b-937e-d8d6f90bb77f\" (UID: \"f8de039d-85d9-444b-937e-d8d6f90bb77f\") " Oct 08 18:36:54 crc kubenswrapper[4859]: I1008 18:36:54.633399 4859 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-jztkt\" (UniqueName: \"kubernetes.io/projected/f8de039d-85d9-444b-937e-d8d6f90bb77f-kube-api-access-jztkt\") pod \"f8de039d-85d9-444b-937e-d8d6f90bb77f\" (UID: \"f8de039d-85d9-444b-937e-d8d6f90bb77f\") " Oct 08 18:36:54 crc kubenswrapper[4859]: I1008 18:36:54.633450 4859 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/f8de039d-85d9-444b-937e-d8d6f90bb77f-logs\") pod \"f8de039d-85d9-444b-937e-d8d6f90bb77f\" (UID: \"f8de039d-85d9-444b-937e-d8d6f90bb77f\") " Oct 08 18:36:54 crc kubenswrapper[4859]: I1008 18:36:54.634195 4859 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/f8de039d-85d9-444b-937e-d8d6f90bb77f-logs" (OuterVolumeSpecName: "logs") pod "f8de039d-85d9-444b-937e-d8d6f90bb77f" (UID: "f8de039d-85d9-444b-937e-d8d6f90bb77f"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 08 18:36:54 crc kubenswrapper[4859]: I1008 18:36:54.638955 4859 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/f8de039d-85d9-444b-937e-d8d6f90bb77f-kube-api-access-jztkt" (OuterVolumeSpecName: "kube-api-access-jztkt") pod "f8de039d-85d9-444b-937e-d8d6f90bb77f" (UID: "f8de039d-85d9-444b-937e-d8d6f90bb77f"). InnerVolumeSpecName "kube-api-access-jztkt". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 08 18:36:54 crc kubenswrapper[4859]: I1008 18:36:54.663415 4859 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f8de039d-85d9-444b-937e-d8d6f90bb77f-config-data" (OuterVolumeSpecName: "config-data") pod "f8de039d-85d9-444b-937e-d8d6f90bb77f" (UID: "f8de039d-85d9-444b-937e-d8d6f90bb77f"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 08 18:36:54 crc kubenswrapper[4859]: I1008 18:36:54.671295 4859 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f8de039d-85d9-444b-937e-d8d6f90bb77f-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "f8de039d-85d9-444b-937e-d8d6f90bb77f" (UID: "f8de039d-85d9-444b-937e-d8d6f90bb77f"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 08 18:36:54 crc kubenswrapper[4859]: I1008 18:36:54.715403 4859 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f8de039d-85d9-444b-937e-d8d6f90bb77f-nova-metadata-tls-certs" (OuterVolumeSpecName: "nova-metadata-tls-certs") pod "f8de039d-85d9-444b-937e-d8d6f90bb77f" (UID: "f8de039d-85d9-444b-937e-d8d6f90bb77f"). InnerVolumeSpecName "nova-metadata-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 08 18:36:54 crc kubenswrapper[4859]: I1008 18:36:54.734994 4859 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f8de039d-85d9-444b-937e-d8d6f90bb77f-config-data\") on node \"crc\" DevicePath \"\"" Oct 08 18:36:54 crc kubenswrapper[4859]: I1008 18:36:54.735042 4859 reconciler_common.go:293] "Volume detached for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/f8de039d-85d9-444b-937e-d8d6f90bb77f-nova-metadata-tls-certs\") on node \"crc\" DevicePath \"\"" Oct 08 18:36:54 crc kubenswrapper[4859]: I1008 18:36:54.735061 4859 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f8de039d-85d9-444b-937e-d8d6f90bb77f-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 08 18:36:54 crc kubenswrapper[4859]: I1008 18:36:54.735079 4859 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-jztkt\" (UniqueName: \"kubernetes.io/projected/f8de039d-85d9-444b-937e-d8d6f90bb77f-kube-api-access-jztkt\") on node \"crc\" DevicePath \"\"" Oct 08 18:36:54 crc kubenswrapper[4859]: I1008 18:36:54.735094 4859 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/f8de039d-85d9-444b-937e-d8d6f90bb77f-logs\") on node \"crc\" DevicePath \"\"" Oct 08 18:36:54 crc kubenswrapper[4859]: I1008 18:36:54.755885 4859 generic.go:334] "Generic (PLEG): container finished" podID="f8de039d-85d9-444b-937e-d8d6f90bb77f" containerID="dce4317d8887820fbdd21c3fd9beefcc68a86c05505bb4dd820e11a9c745a137" exitCode=0 Oct 08 18:36:54 crc kubenswrapper[4859]: I1008 18:36:54.755965 4859 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Oct 08 18:36:54 crc kubenswrapper[4859]: I1008 18:36:54.755995 4859 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"f8de039d-85d9-444b-937e-d8d6f90bb77f","Type":"ContainerDied","Data":"dce4317d8887820fbdd21c3fd9beefcc68a86c05505bb4dd820e11a9c745a137"} Oct 08 18:36:54 crc kubenswrapper[4859]: I1008 18:36:54.756058 4859 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"f8de039d-85d9-444b-937e-d8d6f90bb77f","Type":"ContainerDied","Data":"9345957d7543f2ca4931a21700a11bea59a58373aca68aa97f693fa971dc360b"} Oct 08 18:36:54 crc kubenswrapper[4859]: I1008 18:36:54.756088 4859 scope.go:117] "RemoveContainer" containerID="dce4317d8887820fbdd21c3fd9beefcc68a86c05505bb4dd820e11a9c745a137" Oct 08 18:36:54 crc kubenswrapper[4859]: I1008 18:36:54.806521 4859 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-metadata-0"] Oct 08 18:36:54 crc kubenswrapper[4859]: I1008 18:36:54.813676 4859 scope.go:117] "RemoveContainer" containerID="719956fb0a5ddd763170476c1c204c6c8a78e5b6f8bdc405211df91e0e9eff5b" Oct 08 18:36:54 crc kubenswrapper[4859]: I1008 18:36:54.814948 4859 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-metadata-0"] Oct 08 18:36:54 crc kubenswrapper[4859]: I1008 18:36:54.830554 4859 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-metadata-0"] Oct 08 18:36:54 crc kubenswrapper[4859]: E1008 18:36:54.831061 4859 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f8de039d-85d9-444b-937e-d8d6f90bb77f" containerName="nova-metadata-log" Oct 08 18:36:54 crc kubenswrapper[4859]: I1008 18:36:54.831088 4859 state_mem.go:107] "Deleted CPUSet assignment" podUID="f8de039d-85d9-444b-937e-d8d6f90bb77f" containerName="nova-metadata-log" Oct 08 18:36:54 crc kubenswrapper[4859]: E1008 18:36:54.831105 4859 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f8de039d-85d9-444b-937e-d8d6f90bb77f" containerName="nova-metadata-metadata" Oct 08 18:36:54 crc kubenswrapper[4859]: I1008 18:36:54.831112 4859 state_mem.go:107] "Deleted CPUSet assignment" podUID="f8de039d-85d9-444b-937e-d8d6f90bb77f" containerName="nova-metadata-metadata" Oct 08 18:36:54 crc kubenswrapper[4859]: I1008 18:36:54.831321 4859 memory_manager.go:354] "RemoveStaleState removing state" podUID="f8de039d-85d9-444b-937e-d8d6f90bb77f" containerName="nova-metadata-log" Oct 08 18:36:54 crc kubenswrapper[4859]: I1008 18:36:54.831344 4859 memory_manager.go:354] "RemoveStaleState removing state" podUID="f8de039d-85d9-444b-937e-d8d6f90bb77f" containerName="nova-metadata-metadata" Oct 08 18:36:54 crc kubenswrapper[4859]: I1008 18:36:54.832499 4859 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Oct 08 18:36:54 crc kubenswrapper[4859]: I1008 18:36:54.835650 4859 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-nova-metadata-internal-svc" Oct 08 18:36:54 crc kubenswrapper[4859]: I1008 18:36:54.836006 4859 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-metadata-config-data" Oct 08 18:36:54 crc kubenswrapper[4859]: I1008 18:36:54.842147 4859 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-metadata-0"] Oct 08 18:36:54 crc kubenswrapper[4859]: I1008 18:36:54.867501 4859 scope.go:117] "RemoveContainer" containerID="dce4317d8887820fbdd21c3fd9beefcc68a86c05505bb4dd820e11a9c745a137" Oct 08 18:36:54 crc kubenswrapper[4859]: E1008 18:36:54.868006 4859 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"dce4317d8887820fbdd21c3fd9beefcc68a86c05505bb4dd820e11a9c745a137\": container with ID starting with dce4317d8887820fbdd21c3fd9beefcc68a86c05505bb4dd820e11a9c745a137 not found: ID does not exist" containerID="dce4317d8887820fbdd21c3fd9beefcc68a86c05505bb4dd820e11a9c745a137" Oct 08 18:36:54 crc kubenswrapper[4859]: I1008 18:36:54.868044 4859 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"dce4317d8887820fbdd21c3fd9beefcc68a86c05505bb4dd820e11a9c745a137"} err="failed to get container status \"dce4317d8887820fbdd21c3fd9beefcc68a86c05505bb4dd820e11a9c745a137\": rpc error: code = NotFound desc = could not find container \"dce4317d8887820fbdd21c3fd9beefcc68a86c05505bb4dd820e11a9c745a137\": container with ID starting with dce4317d8887820fbdd21c3fd9beefcc68a86c05505bb4dd820e11a9c745a137 not found: ID does not exist" Oct 08 18:36:54 crc kubenswrapper[4859]: I1008 18:36:54.868065 4859 scope.go:117] "RemoveContainer" containerID="719956fb0a5ddd763170476c1c204c6c8a78e5b6f8bdc405211df91e0e9eff5b" Oct 08 18:36:54 crc kubenswrapper[4859]: E1008 18:36:54.868364 4859 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"719956fb0a5ddd763170476c1c204c6c8a78e5b6f8bdc405211df91e0e9eff5b\": container with ID starting with 719956fb0a5ddd763170476c1c204c6c8a78e5b6f8bdc405211df91e0e9eff5b not found: ID does not exist" containerID="719956fb0a5ddd763170476c1c204c6c8a78e5b6f8bdc405211df91e0e9eff5b" Oct 08 18:36:54 crc kubenswrapper[4859]: I1008 18:36:54.868390 4859 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"719956fb0a5ddd763170476c1c204c6c8a78e5b6f8bdc405211df91e0e9eff5b"} err="failed to get container status \"719956fb0a5ddd763170476c1c204c6c8a78e5b6f8bdc405211df91e0e9eff5b\": rpc error: code = NotFound desc = could not find container \"719956fb0a5ddd763170476c1c204c6c8a78e5b6f8bdc405211df91e0e9eff5b\": container with ID starting with 719956fb0a5ddd763170476c1c204c6c8a78e5b6f8bdc405211df91e0e9eff5b not found: ID does not exist" Oct 08 18:36:54 crc kubenswrapper[4859]: I1008 18:36:54.937860 4859 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-5qhtv\" (UniqueName: \"kubernetes.io/projected/6188ecfc-2b44-443d-a966-8b91ae535533-kube-api-access-5qhtv\") pod \"nova-metadata-0\" (UID: \"6188ecfc-2b44-443d-a966-8b91ae535533\") " pod="openstack/nova-metadata-0" Oct 08 18:36:54 crc kubenswrapper[4859]: I1008 18:36:54.937914 4859 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6188ecfc-2b44-443d-a966-8b91ae535533-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"6188ecfc-2b44-443d-a966-8b91ae535533\") " pod="openstack/nova-metadata-0" Oct 08 18:36:54 crc kubenswrapper[4859]: I1008 18:36:54.937958 4859 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/6188ecfc-2b44-443d-a966-8b91ae535533-nova-metadata-tls-certs\") pod \"nova-metadata-0\" (UID: \"6188ecfc-2b44-443d-a966-8b91ae535533\") " pod="openstack/nova-metadata-0" Oct 08 18:36:54 crc kubenswrapper[4859]: I1008 18:36:54.938017 4859 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/6188ecfc-2b44-443d-a966-8b91ae535533-config-data\") pod \"nova-metadata-0\" (UID: \"6188ecfc-2b44-443d-a966-8b91ae535533\") " pod="openstack/nova-metadata-0" Oct 08 18:36:54 crc kubenswrapper[4859]: I1008 18:36:54.938045 4859 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/6188ecfc-2b44-443d-a966-8b91ae535533-logs\") pod \"nova-metadata-0\" (UID: \"6188ecfc-2b44-443d-a966-8b91ae535533\") " pod="openstack/nova-metadata-0" Oct 08 18:36:55 crc kubenswrapper[4859]: I1008 18:36:55.042977 4859 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-5qhtv\" (UniqueName: \"kubernetes.io/projected/6188ecfc-2b44-443d-a966-8b91ae535533-kube-api-access-5qhtv\") pod \"nova-metadata-0\" (UID: \"6188ecfc-2b44-443d-a966-8b91ae535533\") " pod="openstack/nova-metadata-0" Oct 08 18:36:55 crc kubenswrapper[4859]: I1008 18:36:55.043589 4859 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6188ecfc-2b44-443d-a966-8b91ae535533-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"6188ecfc-2b44-443d-a966-8b91ae535533\") " pod="openstack/nova-metadata-0" Oct 08 18:36:55 crc kubenswrapper[4859]: I1008 18:36:55.043812 4859 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/6188ecfc-2b44-443d-a966-8b91ae535533-nova-metadata-tls-certs\") pod \"nova-metadata-0\" (UID: \"6188ecfc-2b44-443d-a966-8b91ae535533\") " pod="openstack/nova-metadata-0" Oct 08 18:36:55 crc kubenswrapper[4859]: I1008 18:36:55.044069 4859 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/6188ecfc-2b44-443d-a966-8b91ae535533-config-data\") pod \"nova-metadata-0\" (UID: \"6188ecfc-2b44-443d-a966-8b91ae535533\") " pod="openstack/nova-metadata-0" Oct 08 18:36:55 crc kubenswrapper[4859]: I1008 18:36:55.044199 4859 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/6188ecfc-2b44-443d-a966-8b91ae535533-logs\") pod \"nova-metadata-0\" (UID: \"6188ecfc-2b44-443d-a966-8b91ae535533\") " pod="openstack/nova-metadata-0" Oct 08 18:36:55 crc kubenswrapper[4859]: I1008 18:36:55.044758 4859 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/6188ecfc-2b44-443d-a966-8b91ae535533-logs\") pod \"nova-metadata-0\" (UID: \"6188ecfc-2b44-443d-a966-8b91ae535533\") " pod="openstack/nova-metadata-0" Oct 08 18:36:55 crc kubenswrapper[4859]: I1008 18:36:55.047524 4859 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/6188ecfc-2b44-443d-a966-8b91ae535533-nova-metadata-tls-certs\") pod \"nova-metadata-0\" (UID: \"6188ecfc-2b44-443d-a966-8b91ae535533\") " pod="openstack/nova-metadata-0" Oct 08 18:36:55 crc kubenswrapper[4859]: I1008 18:36:55.048104 4859 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/6188ecfc-2b44-443d-a966-8b91ae535533-config-data\") pod \"nova-metadata-0\" (UID: \"6188ecfc-2b44-443d-a966-8b91ae535533\") " pod="openstack/nova-metadata-0" Oct 08 18:36:55 crc kubenswrapper[4859]: I1008 18:36:55.050717 4859 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6188ecfc-2b44-443d-a966-8b91ae535533-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"6188ecfc-2b44-443d-a966-8b91ae535533\") " pod="openstack/nova-metadata-0" Oct 08 18:36:55 crc kubenswrapper[4859]: I1008 18:36:55.066568 4859 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-5qhtv\" (UniqueName: \"kubernetes.io/projected/6188ecfc-2b44-443d-a966-8b91ae535533-kube-api-access-5qhtv\") pod \"nova-metadata-0\" (UID: \"6188ecfc-2b44-443d-a966-8b91ae535533\") " pod="openstack/nova-metadata-0" Oct 08 18:36:55 crc kubenswrapper[4859]: I1008 18:36:55.163834 4859 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Oct 08 18:36:55 crc kubenswrapper[4859]: I1008 18:36:55.637450 4859 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-metadata-0"] Oct 08 18:36:55 crc kubenswrapper[4859]: W1008 18:36:55.642872 4859 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod6188ecfc_2b44_443d_a966_8b91ae535533.slice/crio-d9cdd81a942149baa62e3e9717efddf9369f3a56922a477e154484d4db2b59a0 WatchSource:0}: Error finding container d9cdd81a942149baa62e3e9717efddf9369f3a56922a477e154484d4db2b59a0: Status 404 returned error can't find the container with id d9cdd81a942149baa62e3e9717efddf9369f3a56922a477e154484d4db2b59a0 Oct 08 18:36:55 crc kubenswrapper[4859]: I1008 18:36:55.766036 4859 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"6188ecfc-2b44-443d-a966-8b91ae535533","Type":"ContainerStarted","Data":"d9cdd81a942149baa62e3e9717efddf9369f3a56922a477e154484d4db2b59a0"} Oct 08 18:36:56 crc kubenswrapper[4859]: I1008 18:36:56.493601 4859 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="f8de039d-85d9-444b-937e-d8d6f90bb77f" path="/var/lib/kubelet/pods/f8de039d-85d9-444b-937e-d8d6f90bb77f/volumes" Oct 08 18:36:56 crc kubenswrapper[4859]: I1008 18:36:56.778621 4859 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"6188ecfc-2b44-443d-a966-8b91ae535533","Type":"ContainerStarted","Data":"c983313fd5c0961ab0dca0e7d92ff3b0d7c886c3ef04bc15d77432d3d42d8141"} Oct 08 18:36:56 crc kubenswrapper[4859]: I1008 18:36:56.778664 4859 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"6188ecfc-2b44-443d-a966-8b91ae535533","Type":"ContainerStarted","Data":"1869ffaddad4dbad939f0237faf927be4e3a64fb17d2afb319d5c81452a44c0a"} Oct 08 18:36:56 crc kubenswrapper[4859]: I1008 18:36:56.806317 4859 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-metadata-0" podStartSLOduration=2.806297651 podStartE2EDuration="2.806297651s" podCreationTimestamp="2025-10-08 18:36:54 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-08 18:36:56.796352357 +0000 UTC m=+1187.043191736" watchObservedRunningTime="2025-10-08 18:36:56.806297651 +0000 UTC m=+1187.053137040" Oct 08 18:36:58 crc kubenswrapper[4859]: I1008 18:36:58.110704 4859 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-scheduler-0" Oct 08 18:37:00 crc kubenswrapper[4859]: I1008 18:37:00.164841 4859 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-metadata-0" Oct 08 18:37:00 crc kubenswrapper[4859]: I1008 18:37:00.165303 4859 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-metadata-0" Oct 08 18:37:02 crc kubenswrapper[4859]: I1008 18:37:02.444307 4859 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-api-0" Oct 08 18:37:02 crc kubenswrapper[4859]: I1008 18:37:02.445846 4859 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-api-0" Oct 08 18:37:03 crc kubenswrapper[4859]: I1008 18:37:03.110740 4859 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-scheduler-0" Oct 08 18:37:03 crc kubenswrapper[4859]: I1008 18:37:03.160994 4859 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-scheduler-0" Oct 08 18:37:03 crc kubenswrapper[4859]: I1008 18:37:03.459979 4859 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-api-0" podUID="e56f2ed9-f7a3-4157-bb50-88fbd86d9c60" containerName="nova-api-api" probeResult="failure" output="Get \"https://10.217.0.202:8774/\": net/http: request canceled (Client.Timeout exceeded while awaiting headers)" Oct 08 18:37:03 crc kubenswrapper[4859]: I1008 18:37:03.459984 4859 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-api-0" podUID="e56f2ed9-f7a3-4157-bb50-88fbd86d9c60" containerName="nova-api-log" probeResult="failure" output="Get \"https://10.217.0.202:8774/\": net/http: request canceled (Client.Timeout exceeded while awaiting headers)" Oct 08 18:37:03 crc kubenswrapper[4859]: I1008 18:37:03.887157 4859 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-scheduler-0" Oct 08 18:37:05 crc kubenswrapper[4859]: I1008 18:37:05.165037 4859 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-metadata-0" Oct 08 18:37:05 crc kubenswrapper[4859]: I1008 18:37:05.165489 4859 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-metadata-0" Oct 08 18:37:06 crc kubenswrapper[4859]: I1008 18:37:06.180001 4859 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-metadata-0" podUID="6188ecfc-2b44-443d-a966-8b91ae535533" containerName="nova-metadata-log" probeResult="failure" output="Get \"https://10.217.0.204:8775/\": net/http: request canceled (Client.Timeout exceeded while awaiting headers)" Oct 08 18:37:06 crc kubenswrapper[4859]: I1008 18:37:06.180043 4859 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-metadata-0" podUID="6188ecfc-2b44-443d-a966-8b91ae535533" containerName="nova-metadata-metadata" probeResult="failure" output="Get \"https://10.217.0.204:8775/\": net/http: request canceled (Client.Timeout exceeded while awaiting headers)" Oct 08 18:37:11 crc kubenswrapper[4859]: I1008 18:37:11.263433 4859 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/ceilometer-0" Oct 08 18:37:12 crc kubenswrapper[4859]: I1008 18:37:12.468364 4859 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-api-0" Oct 08 18:37:12 crc kubenswrapper[4859]: I1008 18:37:12.489080 4859 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-api-0" Oct 08 18:37:12 crc kubenswrapper[4859]: I1008 18:37:12.489183 4859 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-api-0" Oct 08 18:37:12 crc kubenswrapper[4859]: I1008 18:37:12.489257 4859 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-api-0" Oct 08 18:37:12 crc kubenswrapper[4859]: I1008 18:37:12.954174 4859 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-api-0" Oct 08 18:37:12 crc kubenswrapper[4859]: I1008 18:37:12.960429 4859 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-api-0" Oct 08 18:37:15 crc kubenswrapper[4859]: I1008 18:37:15.172222 4859 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-metadata-0" Oct 08 18:37:15 crc kubenswrapper[4859]: I1008 18:37:15.176216 4859 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-metadata-0" Oct 08 18:37:15 crc kubenswrapper[4859]: I1008 18:37:15.180247 4859 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-metadata-0" Oct 08 18:37:15 crc kubenswrapper[4859]: I1008 18:37:15.986558 4859 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-metadata-0" Oct 08 18:37:17 crc kubenswrapper[4859]: I1008 18:37:17.924821 4859 patch_prober.go:28] interesting pod/machine-config-daemon-82s52 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 08 18:37:17 crc kubenswrapper[4859]: I1008 18:37:17.924956 4859 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-82s52" podUID="b23a6a6c-9d92-4e7b-840e-55cfda873a2d" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 08 18:37:19 crc kubenswrapper[4859]: I1008 18:37:19.182857 4859 scope.go:117] "RemoveContainer" containerID="d933baaac4cf8d90574c5c0f78b0557752e9fdb5a092f5f3890fbe21c3263e1d" Oct 08 18:37:19 crc kubenswrapper[4859]: I1008 18:37:19.208041 4859 scope.go:117] "RemoveContainer" containerID="2570425de29f3db779ddcaded0192a9d6e247e3e1514413fba87a1bad871db4f" Oct 08 18:37:19 crc kubenswrapper[4859]: I1008 18:37:19.278248 4859 scope.go:117] "RemoveContainer" containerID="97c3d33e64a8245acdfb332d454850d2a9d7d87cafb6bbb784f824a373d3ca04" Oct 08 18:37:24 crc kubenswrapper[4859]: I1008 18:37:24.278997 4859 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/rabbitmq-server-0"] Oct 08 18:37:25 crc kubenswrapper[4859]: I1008 18:37:25.036528 4859 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/rabbitmq-cell1-server-0"] Oct 08 18:37:28 crc kubenswrapper[4859]: I1008 18:37:28.942423 4859 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/rabbitmq-server-0" podUID="ba5a78be-e7b3-401a-bdb9-ba31071d7075" containerName="rabbitmq" containerID="cri-o://1ff56b52d2d31d1f67ef0e984e2373890024eeaf7a31c66677cb07168c88da23" gracePeriod=604796 Oct 08 18:37:29 crc kubenswrapper[4859]: I1008 18:37:29.153976 4859 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/rabbitmq-cell1-server-0" podUID="3075d5c2-3e74-4a4c-829f-248e87f45f3b" containerName="rabbitmq" containerID="cri-o://56c96709e82513e92c3d9367fe8c571724d57b12903fa91ea473fc776eee2ec6" gracePeriod=604796 Oct 08 18:37:34 crc kubenswrapper[4859]: I1008 18:37:34.906054 4859 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/rabbitmq-server-0" podUID="ba5a78be-e7b3-401a-bdb9-ba31071d7075" containerName="rabbitmq" probeResult="failure" output="dial tcp 10.217.0.100:5671: connect: connection refused" Oct 08 18:37:35 crc kubenswrapper[4859]: I1008 18:37:35.176064 4859 generic.go:334] "Generic (PLEG): container finished" podID="ba5a78be-e7b3-401a-bdb9-ba31071d7075" containerID="1ff56b52d2d31d1f67ef0e984e2373890024eeaf7a31c66677cb07168c88da23" exitCode=0 Oct 08 18:37:35 crc kubenswrapper[4859]: I1008 18:37:35.176222 4859 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-server-0" event={"ID":"ba5a78be-e7b3-401a-bdb9-ba31071d7075","Type":"ContainerDied","Data":"1ff56b52d2d31d1f67ef0e984e2373890024eeaf7a31c66677cb07168c88da23"} Oct 08 18:37:35 crc kubenswrapper[4859]: I1008 18:37:35.189974 4859 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/rabbitmq-cell1-server-0" podUID="3075d5c2-3e74-4a4c-829f-248e87f45f3b" containerName="rabbitmq" probeResult="failure" output="dial tcp 10.217.0.101:5671: connect: connection refused" Oct 08 18:37:35 crc kubenswrapper[4859]: I1008 18:37:35.518345 4859 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-server-0" Oct 08 18:37:35 crc kubenswrapper[4859]: I1008 18:37:35.649362 4859 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/ba5a78be-e7b3-401a-bdb9-ba31071d7075-rabbitmq-tls\") pod \"ba5a78be-e7b3-401a-bdb9-ba31071d7075\" (UID: \"ba5a78be-e7b3-401a-bdb9-ba31071d7075\") " Oct 08 18:37:35 crc kubenswrapper[4859]: I1008 18:37:35.649737 4859 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/ba5a78be-e7b3-401a-bdb9-ba31071d7075-erlang-cookie-secret\") pod \"ba5a78be-e7b3-401a-bdb9-ba31071d7075\" (UID: \"ba5a78be-e7b3-401a-bdb9-ba31071d7075\") " Oct 08 18:37:35 crc kubenswrapper[4859]: I1008 18:37:35.649805 4859 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/ba5a78be-e7b3-401a-bdb9-ba31071d7075-server-conf\") pod \"ba5a78be-e7b3-401a-bdb9-ba31071d7075\" (UID: \"ba5a78be-e7b3-401a-bdb9-ba31071d7075\") " Oct 08 18:37:35 crc kubenswrapper[4859]: I1008 18:37:35.649836 4859 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/ba5a78be-e7b3-401a-bdb9-ba31071d7075-rabbitmq-plugins\") pod \"ba5a78be-e7b3-401a-bdb9-ba31071d7075\" (UID: \"ba5a78be-e7b3-401a-bdb9-ba31071d7075\") " Oct 08 18:37:35 crc kubenswrapper[4859]: I1008 18:37:35.649850 4859 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/ba5a78be-e7b3-401a-bdb9-ba31071d7075-config-data\") pod \"ba5a78be-e7b3-401a-bdb9-ba31071d7075\" (UID: \"ba5a78be-e7b3-401a-bdb9-ba31071d7075\") " Oct 08 18:37:35 crc kubenswrapper[4859]: I1008 18:37:35.649888 4859 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"persistence\" (UniqueName: \"kubernetes.io/local-volume/local-storage08-crc\") pod \"ba5a78be-e7b3-401a-bdb9-ba31071d7075\" (UID: \"ba5a78be-e7b3-401a-bdb9-ba31071d7075\") " Oct 08 18:37:35 crc kubenswrapper[4859]: I1008 18:37:35.649931 4859 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/ba5a78be-e7b3-401a-bdb9-ba31071d7075-pod-info\") pod \"ba5a78be-e7b3-401a-bdb9-ba31071d7075\" (UID: \"ba5a78be-e7b3-401a-bdb9-ba31071d7075\") " Oct 08 18:37:35 crc kubenswrapper[4859]: I1008 18:37:35.649982 4859 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/ba5a78be-e7b3-401a-bdb9-ba31071d7075-rabbitmq-erlang-cookie\") pod \"ba5a78be-e7b3-401a-bdb9-ba31071d7075\" (UID: \"ba5a78be-e7b3-401a-bdb9-ba31071d7075\") " Oct 08 18:37:35 crc kubenswrapper[4859]: I1008 18:37:35.650026 4859 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/ba5a78be-e7b3-401a-bdb9-ba31071d7075-rabbitmq-confd\") pod \"ba5a78be-e7b3-401a-bdb9-ba31071d7075\" (UID: \"ba5a78be-e7b3-401a-bdb9-ba31071d7075\") " Oct 08 18:37:35 crc kubenswrapper[4859]: I1008 18:37:35.650054 4859 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-sbb5z\" (UniqueName: \"kubernetes.io/projected/ba5a78be-e7b3-401a-bdb9-ba31071d7075-kube-api-access-sbb5z\") pod \"ba5a78be-e7b3-401a-bdb9-ba31071d7075\" (UID: \"ba5a78be-e7b3-401a-bdb9-ba31071d7075\") " Oct 08 18:37:35 crc kubenswrapper[4859]: I1008 18:37:35.650073 4859 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/ba5a78be-e7b3-401a-bdb9-ba31071d7075-plugins-conf\") pod \"ba5a78be-e7b3-401a-bdb9-ba31071d7075\" (UID: \"ba5a78be-e7b3-401a-bdb9-ba31071d7075\") " Oct 08 18:37:35 crc kubenswrapper[4859]: I1008 18:37:35.653772 4859 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/ba5a78be-e7b3-401a-bdb9-ba31071d7075-rabbitmq-plugins" (OuterVolumeSpecName: "rabbitmq-plugins") pod "ba5a78be-e7b3-401a-bdb9-ba31071d7075" (UID: "ba5a78be-e7b3-401a-bdb9-ba31071d7075"). InnerVolumeSpecName "rabbitmq-plugins". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 08 18:37:35 crc kubenswrapper[4859]: I1008 18:37:35.654504 4859 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/ba5a78be-e7b3-401a-bdb9-ba31071d7075-plugins-conf" (OuterVolumeSpecName: "plugins-conf") pod "ba5a78be-e7b3-401a-bdb9-ba31071d7075" (UID: "ba5a78be-e7b3-401a-bdb9-ba31071d7075"). InnerVolumeSpecName "plugins-conf". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 08 18:37:35 crc kubenswrapper[4859]: I1008 18:37:35.656203 4859 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/ba5a78be-e7b3-401a-bdb9-ba31071d7075-rabbitmq-erlang-cookie" (OuterVolumeSpecName: "rabbitmq-erlang-cookie") pod "ba5a78be-e7b3-401a-bdb9-ba31071d7075" (UID: "ba5a78be-e7b3-401a-bdb9-ba31071d7075"). InnerVolumeSpecName "rabbitmq-erlang-cookie". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 08 18:37:35 crc kubenswrapper[4859]: I1008 18:37:35.657818 4859 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/ba5a78be-e7b3-401a-bdb9-ba31071d7075-erlang-cookie-secret" (OuterVolumeSpecName: "erlang-cookie-secret") pod "ba5a78be-e7b3-401a-bdb9-ba31071d7075" (UID: "ba5a78be-e7b3-401a-bdb9-ba31071d7075"). InnerVolumeSpecName "erlang-cookie-secret". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 08 18:37:35 crc kubenswrapper[4859]: I1008 18:37:35.659586 4859 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/ba5a78be-e7b3-401a-bdb9-ba31071d7075-kube-api-access-sbb5z" (OuterVolumeSpecName: "kube-api-access-sbb5z") pod "ba5a78be-e7b3-401a-bdb9-ba31071d7075" (UID: "ba5a78be-e7b3-401a-bdb9-ba31071d7075"). InnerVolumeSpecName "kube-api-access-sbb5z". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 08 18:37:35 crc kubenswrapper[4859]: I1008 18:37:35.660657 4859 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/local-volume/local-storage08-crc" (OuterVolumeSpecName: "persistence") pod "ba5a78be-e7b3-401a-bdb9-ba31071d7075" (UID: "ba5a78be-e7b3-401a-bdb9-ba31071d7075"). InnerVolumeSpecName "local-storage08-crc". PluginName "kubernetes.io/local-volume", VolumeGidValue "" Oct 08 18:37:35 crc kubenswrapper[4859]: I1008 18:37:35.680573 4859 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/ba5a78be-e7b3-401a-bdb9-ba31071d7075-rabbitmq-tls" (OuterVolumeSpecName: "rabbitmq-tls") pod "ba5a78be-e7b3-401a-bdb9-ba31071d7075" (UID: "ba5a78be-e7b3-401a-bdb9-ba31071d7075"). InnerVolumeSpecName "rabbitmq-tls". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 08 18:37:35 crc kubenswrapper[4859]: I1008 18:37:35.682326 4859 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/ba5a78be-e7b3-401a-bdb9-ba31071d7075-config-data" (OuterVolumeSpecName: "config-data") pod "ba5a78be-e7b3-401a-bdb9-ba31071d7075" (UID: "ba5a78be-e7b3-401a-bdb9-ba31071d7075"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 08 18:37:35 crc kubenswrapper[4859]: I1008 18:37:35.685022 4859 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/downward-api/ba5a78be-e7b3-401a-bdb9-ba31071d7075-pod-info" (OuterVolumeSpecName: "pod-info") pod "ba5a78be-e7b3-401a-bdb9-ba31071d7075" (UID: "ba5a78be-e7b3-401a-bdb9-ba31071d7075"). InnerVolumeSpecName "pod-info". PluginName "kubernetes.io/downward-api", VolumeGidValue "" Oct 08 18:37:35 crc kubenswrapper[4859]: I1008 18:37:35.722826 4859 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-cell1-server-0" Oct 08 18:37:35 crc kubenswrapper[4859]: I1008 18:37:35.746557 4859 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/ba5a78be-e7b3-401a-bdb9-ba31071d7075-server-conf" (OuterVolumeSpecName: "server-conf") pod "ba5a78be-e7b3-401a-bdb9-ba31071d7075" (UID: "ba5a78be-e7b3-401a-bdb9-ba31071d7075"). InnerVolumeSpecName "server-conf". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 08 18:37:35 crc kubenswrapper[4859]: I1008 18:37:35.752013 4859 reconciler_common.go:293] "Volume detached for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/ba5a78be-e7b3-401a-bdb9-ba31071d7075-rabbitmq-plugins\") on node \"crc\" DevicePath \"\"" Oct 08 18:37:35 crc kubenswrapper[4859]: I1008 18:37:35.752044 4859 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/ba5a78be-e7b3-401a-bdb9-ba31071d7075-config-data\") on node \"crc\" DevicePath \"\"" Oct 08 18:37:35 crc kubenswrapper[4859]: I1008 18:37:35.752070 4859 reconciler_common.go:286] "operationExecutor.UnmountDevice started for volume \"local-storage08-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage08-crc\") on node \"crc\" " Oct 08 18:37:35 crc kubenswrapper[4859]: I1008 18:37:35.752082 4859 reconciler_common.go:293] "Volume detached for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/ba5a78be-e7b3-401a-bdb9-ba31071d7075-pod-info\") on node \"crc\" DevicePath \"\"" Oct 08 18:37:35 crc kubenswrapper[4859]: I1008 18:37:35.752095 4859 reconciler_common.go:293] "Volume detached for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/ba5a78be-e7b3-401a-bdb9-ba31071d7075-rabbitmq-erlang-cookie\") on node \"crc\" DevicePath \"\"" Oct 08 18:37:35 crc kubenswrapper[4859]: I1008 18:37:35.752109 4859 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-sbb5z\" (UniqueName: \"kubernetes.io/projected/ba5a78be-e7b3-401a-bdb9-ba31071d7075-kube-api-access-sbb5z\") on node \"crc\" DevicePath \"\"" Oct 08 18:37:35 crc kubenswrapper[4859]: I1008 18:37:35.752118 4859 reconciler_common.go:293] "Volume detached for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/ba5a78be-e7b3-401a-bdb9-ba31071d7075-plugins-conf\") on node \"crc\" DevicePath \"\"" Oct 08 18:37:35 crc kubenswrapper[4859]: I1008 18:37:35.752126 4859 reconciler_common.go:293] "Volume detached for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/ba5a78be-e7b3-401a-bdb9-ba31071d7075-rabbitmq-tls\") on node \"crc\" DevicePath \"\"" Oct 08 18:37:35 crc kubenswrapper[4859]: I1008 18:37:35.752134 4859 reconciler_common.go:293] "Volume detached for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/ba5a78be-e7b3-401a-bdb9-ba31071d7075-erlang-cookie-secret\") on node \"crc\" DevicePath \"\"" Oct 08 18:37:35 crc kubenswrapper[4859]: I1008 18:37:35.752144 4859 reconciler_common.go:293] "Volume detached for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/ba5a78be-e7b3-401a-bdb9-ba31071d7075-server-conf\") on node \"crc\" DevicePath \"\"" Oct 08 18:37:35 crc kubenswrapper[4859]: I1008 18:37:35.772507 4859 operation_generator.go:917] UnmountDevice succeeded for volume "local-storage08-crc" (UniqueName: "kubernetes.io/local-volume/local-storage08-crc") on node "crc" Oct 08 18:37:35 crc kubenswrapper[4859]: I1008 18:37:35.792224 4859 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/ba5a78be-e7b3-401a-bdb9-ba31071d7075-rabbitmq-confd" (OuterVolumeSpecName: "rabbitmq-confd") pod "ba5a78be-e7b3-401a-bdb9-ba31071d7075" (UID: "ba5a78be-e7b3-401a-bdb9-ba31071d7075"). InnerVolumeSpecName "rabbitmq-confd". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 08 18:37:35 crc kubenswrapper[4859]: I1008 18:37:35.853440 4859 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/3075d5c2-3e74-4a4c-829f-248e87f45f3b-plugins-conf\") pod \"3075d5c2-3e74-4a4c-829f-248e87f45f3b\" (UID: \"3075d5c2-3e74-4a4c-829f-248e87f45f3b\") " Oct 08 18:37:35 crc kubenswrapper[4859]: I1008 18:37:35.853526 4859 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-xvbcd\" (UniqueName: \"kubernetes.io/projected/3075d5c2-3e74-4a4c-829f-248e87f45f3b-kube-api-access-xvbcd\") pod \"3075d5c2-3e74-4a4c-829f-248e87f45f3b\" (UID: \"3075d5c2-3e74-4a4c-829f-248e87f45f3b\") " Oct 08 18:37:35 crc kubenswrapper[4859]: I1008 18:37:35.853646 4859 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/3075d5c2-3e74-4a4c-829f-248e87f45f3b-server-conf\") pod \"3075d5c2-3e74-4a4c-829f-248e87f45f3b\" (UID: \"3075d5c2-3e74-4a4c-829f-248e87f45f3b\") " Oct 08 18:37:35 crc kubenswrapper[4859]: I1008 18:37:35.853732 4859 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/3075d5c2-3e74-4a4c-829f-248e87f45f3b-rabbitmq-erlang-cookie\") pod \"3075d5c2-3e74-4a4c-829f-248e87f45f3b\" (UID: \"3075d5c2-3e74-4a4c-829f-248e87f45f3b\") " Oct 08 18:37:35 crc kubenswrapper[4859]: I1008 18:37:35.853809 4859 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/3075d5c2-3e74-4a4c-829f-248e87f45f3b-pod-info\") pod \"3075d5c2-3e74-4a4c-829f-248e87f45f3b\" (UID: \"3075d5c2-3e74-4a4c-829f-248e87f45f3b\") " Oct 08 18:37:35 crc kubenswrapper[4859]: I1008 18:37:35.853829 4859 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/3075d5c2-3e74-4a4c-829f-248e87f45f3b-rabbitmq-plugins\") pod \"3075d5c2-3e74-4a4c-829f-248e87f45f3b\" (UID: \"3075d5c2-3e74-4a4c-829f-248e87f45f3b\") " Oct 08 18:37:35 crc kubenswrapper[4859]: I1008 18:37:35.853913 4859 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/3075d5c2-3e74-4a4c-829f-248e87f45f3b-rabbitmq-tls\") pod \"3075d5c2-3e74-4a4c-829f-248e87f45f3b\" (UID: \"3075d5c2-3e74-4a4c-829f-248e87f45f3b\") " Oct 08 18:37:35 crc kubenswrapper[4859]: I1008 18:37:35.853961 4859 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/3075d5c2-3e74-4a4c-829f-248e87f45f3b-config-data\") pod \"3075d5c2-3e74-4a4c-829f-248e87f45f3b\" (UID: \"3075d5c2-3e74-4a4c-829f-248e87f45f3b\") " Oct 08 18:37:35 crc kubenswrapper[4859]: I1008 18:37:35.853999 4859 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"persistence\" (UniqueName: \"kubernetes.io/local-volume/local-storage09-crc\") pod \"3075d5c2-3e74-4a4c-829f-248e87f45f3b\" (UID: \"3075d5c2-3e74-4a4c-829f-248e87f45f3b\") " Oct 08 18:37:35 crc kubenswrapper[4859]: I1008 18:37:35.854054 4859 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/3075d5c2-3e74-4a4c-829f-248e87f45f3b-rabbitmq-confd\") pod \"3075d5c2-3e74-4a4c-829f-248e87f45f3b\" (UID: \"3075d5c2-3e74-4a4c-829f-248e87f45f3b\") " Oct 08 18:37:35 crc kubenswrapper[4859]: I1008 18:37:35.854076 4859 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/3075d5c2-3e74-4a4c-829f-248e87f45f3b-erlang-cookie-secret\") pod \"3075d5c2-3e74-4a4c-829f-248e87f45f3b\" (UID: \"3075d5c2-3e74-4a4c-829f-248e87f45f3b\") " Oct 08 18:37:35 crc kubenswrapper[4859]: I1008 18:37:35.854888 4859 reconciler_common.go:293] "Volume detached for volume \"local-storage08-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage08-crc\") on node \"crc\" DevicePath \"\"" Oct 08 18:37:35 crc kubenswrapper[4859]: I1008 18:37:35.854930 4859 reconciler_common.go:293] "Volume detached for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/ba5a78be-e7b3-401a-bdb9-ba31071d7075-rabbitmq-confd\") on node \"crc\" DevicePath \"\"" Oct 08 18:37:35 crc kubenswrapper[4859]: I1008 18:37:35.856351 4859 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/3075d5c2-3e74-4a4c-829f-248e87f45f3b-rabbitmq-erlang-cookie" (OuterVolumeSpecName: "rabbitmq-erlang-cookie") pod "3075d5c2-3e74-4a4c-829f-248e87f45f3b" (UID: "3075d5c2-3e74-4a4c-829f-248e87f45f3b"). InnerVolumeSpecName "rabbitmq-erlang-cookie". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 08 18:37:35 crc kubenswrapper[4859]: I1008 18:37:35.856369 4859 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/3075d5c2-3e74-4a4c-829f-248e87f45f3b-rabbitmq-plugins" (OuterVolumeSpecName: "rabbitmq-plugins") pod "3075d5c2-3e74-4a4c-829f-248e87f45f3b" (UID: "3075d5c2-3e74-4a4c-829f-248e87f45f3b"). InnerVolumeSpecName "rabbitmq-plugins". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 08 18:37:35 crc kubenswrapper[4859]: I1008 18:37:35.857014 4859 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/3075d5c2-3e74-4a4c-829f-248e87f45f3b-plugins-conf" (OuterVolumeSpecName: "plugins-conf") pod "3075d5c2-3e74-4a4c-829f-248e87f45f3b" (UID: "3075d5c2-3e74-4a4c-829f-248e87f45f3b"). InnerVolumeSpecName "plugins-conf". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 08 18:37:35 crc kubenswrapper[4859]: I1008 18:37:35.858402 4859 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/local-volume/local-storage09-crc" (OuterVolumeSpecName: "persistence") pod "3075d5c2-3e74-4a4c-829f-248e87f45f3b" (UID: "3075d5c2-3e74-4a4c-829f-248e87f45f3b"). InnerVolumeSpecName "local-storage09-crc". PluginName "kubernetes.io/local-volume", VolumeGidValue "" Oct 08 18:37:35 crc kubenswrapper[4859]: I1008 18:37:35.860237 4859 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/3075d5c2-3e74-4a4c-829f-248e87f45f3b-rabbitmq-tls" (OuterVolumeSpecName: "rabbitmq-tls") pod "3075d5c2-3e74-4a4c-829f-248e87f45f3b" (UID: "3075d5c2-3e74-4a4c-829f-248e87f45f3b"). InnerVolumeSpecName "rabbitmq-tls". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 08 18:37:35 crc kubenswrapper[4859]: I1008 18:37:35.860282 4859 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/3075d5c2-3e74-4a4c-829f-248e87f45f3b-kube-api-access-xvbcd" (OuterVolumeSpecName: "kube-api-access-xvbcd") pod "3075d5c2-3e74-4a4c-829f-248e87f45f3b" (UID: "3075d5c2-3e74-4a4c-829f-248e87f45f3b"). InnerVolumeSpecName "kube-api-access-xvbcd". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 08 18:37:35 crc kubenswrapper[4859]: I1008 18:37:35.861034 4859 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/3075d5c2-3e74-4a4c-829f-248e87f45f3b-erlang-cookie-secret" (OuterVolumeSpecName: "erlang-cookie-secret") pod "3075d5c2-3e74-4a4c-829f-248e87f45f3b" (UID: "3075d5c2-3e74-4a4c-829f-248e87f45f3b"). InnerVolumeSpecName "erlang-cookie-secret". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 08 18:37:35 crc kubenswrapper[4859]: I1008 18:37:35.862854 4859 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/downward-api/3075d5c2-3e74-4a4c-829f-248e87f45f3b-pod-info" (OuterVolumeSpecName: "pod-info") pod "3075d5c2-3e74-4a4c-829f-248e87f45f3b" (UID: "3075d5c2-3e74-4a4c-829f-248e87f45f3b"). InnerVolumeSpecName "pod-info". PluginName "kubernetes.io/downward-api", VolumeGidValue "" Oct 08 18:37:35 crc kubenswrapper[4859]: I1008 18:37:35.881004 4859 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/3075d5c2-3e74-4a4c-829f-248e87f45f3b-config-data" (OuterVolumeSpecName: "config-data") pod "3075d5c2-3e74-4a4c-829f-248e87f45f3b" (UID: "3075d5c2-3e74-4a4c-829f-248e87f45f3b"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 08 18:37:35 crc kubenswrapper[4859]: I1008 18:37:35.906088 4859 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/3075d5c2-3e74-4a4c-829f-248e87f45f3b-server-conf" (OuterVolumeSpecName: "server-conf") pod "3075d5c2-3e74-4a4c-829f-248e87f45f3b" (UID: "3075d5c2-3e74-4a4c-829f-248e87f45f3b"). InnerVolumeSpecName "server-conf". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 08 18:37:35 crc kubenswrapper[4859]: I1008 18:37:35.953031 4859 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/3075d5c2-3e74-4a4c-829f-248e87f45f3b-rabbitmq-confd" (OuterVolumeSpecName: "rabbitmq-confd") pod "3075d5c2-3e74-4a4c-829f-248e87f45f3b" (UID: "3075d5c2-3e74-4a4c-829f-248e87f45f3b"). InnerVolumeSpecName "rabbitmq-confd". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 08 18:37:35 crc kubenswrapper[4859]: I1008 18:37:35.956494 4859 reconciler_common.go:293] "Volume detached for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/3075d5c2-3e74-4a4c-829f-248e87f45f3b-plugins-conf\") on node \"crc\" DevicePath \"\"" Oct 08 18:37:35 crc kubenswrapper[4859]: I1008 18:37:35.956533 4859 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-xvbcd\" (UniqueName: \"kubernetes.io/projected/3075d5c2-3e74-4a4c-829f-248e87f45f3b-kube-api-access-xvbcd\") on node \"crc\" DevicePath \"\"" Oct 08 18:37:35 crc kubenswrapper[4859]: I1008 18:37:35.956550 4859 reconciler_common.go:293] "Volume detached for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/3075d5c2-3e74-4a4c-829f-248e87f45f3b-server-conf\") on node \"crc\" DevicePath \"\"" Oct 08 18:37:35 crc kubenswrapper[4859]: I1008 18:37:35.956563 4859 reconciler_common.go:293] "Volume detached for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/3075d5c2-3e74-4a4c-829f-248e87f45f3b-rabbitmq-erlang-cookie\") on node \"crc\" DevicePath \"\"" Oct 08 18:37:35 crc kubenswrapper[4859]: I1008 18:37:35.956576 4859 reconciler_common.go:293] "Volume detached for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/3075d5c2-3e74-4a4c-829f-248e87f45f3b-rabbitmq-plugins\") on node \"crc\" DevicePath \"\"" Oct 08 18:37:35 crc kubenswrapper[4859]: I1008 18:37:35.956589 4859 reconciler_common.go:293] "Volume detached for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/3075d5c2-3e74-4a4c-829f-248e87f45f3b-pod-info\") on node \"crc\" DevicePath \"\"" Oct 08 18:37:35 crc kubenswrapper[4859]: I1008 18:37:35.956600 4859 reconciler_common.go:293] "Volume detached for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/3075d5c2-3e74-4a4c-829f-248e87f45f3b-rabbitmq-tls\") on node \"crc\" DevicePath \"\"" Oct 08 18:37:35 crc kubenswrapper[4859]: I1008 18:37:35.956611 4859 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/3075d5c2-3e74-4a4c-829f-248e87f45f3b-config-data\") on node \"crc\" DevicePath \"\"" Oct 08 18:37:35 crc kubenswrapper[4859]: I1008 18:37:35.956646 4859 reconciler_common.go:286] "operationExecutor.UnmountDevice started for volume \"local-storage09-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage09-crc\") on node \"crc\" " Oct 08 18:37:35 crc kubenswrapper[4859]: I1008 18:37:35.956660 4859 reconciler_common.go:293] "Volume detached for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/3075d5c2-3e74-4a4c-829f-248e87f45f3b-rabbitmq-confd\") on node \"crc\" DevicePath \"\"" Oct 08 18:37:35 crc kubenswrapper[4859]: I1008 18:37:35.956672 4859 reconciler_common.go:293] "Volume detached for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/3075d5c2-3e74-4a4c-829f-248e87f45f3b-erlang-cookie-secret\") on node \"crc\" DevicePath \"\"" Oct 08 18:37:35 crc kubenswrapper[4859]: I1008 18:37:35.979322 4859 operation_generator.go:917] UnmountDevice succeeded for volume "local-storage09-crc" (UniqueName: "kubernetes.io/local-volume/local-storage09-crc") on node "crc" Oct 08 18:37:36 crc kubenswrapper[4859]: I1008 18:37:36.058023 4859 reconciler_common.go:293] "Volume detached for volume \"local-storage09-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage09-crc\") on node \"crc\" DevicePath \"\"" Oct 08 18:37:36 crc kubenswrapper[4859]: I1008 18:37:36.190874 4859 generic.go:334] "Generic (PLEG): container finished" podID="3075d5c2-3e74-4a4c-829f-248e87f45f3b" containerID="56c96709e82513e92c3d9367fe8c571724d57b12903fa91ea473fc776eee2ec6" exitCode=0 Oct 08 18:37:36 crc kubenswrapper[4859]: I1008 18:37:36.190945 4859 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-cell1-server-0" event={"ID":"3075d5c2-3e74-4a4c-829f-248e87f45f3b","Type":"ContainerDied","Data":"56c96709e82513e92c3d9367fe8c571724d57b12903fa91ea473fc776eee2ec6"} Oct 08 18:37:36 crc kubenswrapper[4859]: I1008 18:37:36.190981 4859 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-cell1-server-0" Oct 08 18:37:36 crc kubenswrapper[4859]: I1008 18:37:36.191181 4859 scope.go:117] "RemoveContainer" containerID="56c96709e82513e92c3d9367fe8c571724d57b12903fa91ea473fc776eee2ec6" Oct 08 18:37:36 crc kubenswrapper[4859]: I1008 18:37:36.191165 4859 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-cell1-server-0" event={"ID":"3075d5c2-3e74-4a4c-829f-248e87f45f3b","Type":"ContainerDied","Data":"2a8542e26bacc34037bb28e50535483d385fada430925298e7c6d4b368c6d9c1"} Oct 08 18:37:36 crc kubenswrapper[4859]: I1008 18:37:36.193588 4859 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-server-0" event={"ID":"ba5a78be-e7b3-401a-bdb9-ba31071d7075","Type":"ContainerDied","Data":"b85c8c35dbecff3d9510cdce48f70961ca8f93abedbf900186ddf7c96916969a"} Oct 08 18:37:36 crc kubenswrapper[4859]: I1008 18:37:36.193668 4859 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-server-0" Oct 08 18:37:36 crc kubenswrapper[4859]: I1008 18:37:36.249447 4859 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/rabbitmq-server-0"] Oct 08 18:37:36 crc kubenswrapper[4859]: I1008 18:37:36.252084 4859 scope.go:117] "RemoveContainer" containerID="fecccb4aab6e4695efa36ed5ba960994a10727b9aac69914dd84867213d10810" Oct 08 18:37:36 crc kubenswrapper[4859]: I1008 18:37:36.260433 4859 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/rabbitmq-server-0"] Oct 08 18:37:36 crc kubenswrapper[4859]: I1008 18:37:36.275987 4859 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/rabbitmq-cell1-server-0"] Oct 08 18:37:36 crc kubenswrapper[4859]: I1008 18:37:36.296865 4859 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/rabbitmq-cell1-server-0"] Oct 08 18:37:36 crc kubenswrapper[4859]: I1008 18:37:36.298223 4859 scope.go:117] "RemoveContainer" containerID="56c96709e82513e92c3d9367fe8c571724d57b12903fa91ea473fc776eee2ec6" Oct 08 18:37:36 crc kubenswrapper[4859]: E1008 18:37:36.311125 4859 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"56c96709e82513e92c3d9367fe8c571724d57b12903fa91ea473fc776eee2ec6\": container with ID starting with 56c96709e82513e92c3d9367fe8c571724d57b12903fa91ea473fc776eee2ec6 not found: ID does not exist" containerID="56c96709e82513e92c3d9367fe8c571724d57b12903fa91ea473fc776eee2ec6" Oct 08 18:37:36 crc kubenswrapper[4859]: I1008 18:37:36.311422 4859 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"56c96709e82513e92c3d9367fe8c571724d57b12903fa91ea473fc776eee2ec6"} err="failed to get container status \"56c96709e82513e92c3d9367fe8c571724d57b12903fa91ea473fc776eee2ec6\": rpc error: code = NotFound desc = could not find container \"56c96709e82513e92c3d9367fe8c571724d57b12903fa91ea473fc776eee2ec6\": container with ID starting with 56c96709e82513e92c3d9367fe8c571724d57b12903fa91ea473fc776eee2ec6 not found: ID does not exist" Oct 08 18:37:36 crc kubenswrapper[4859]: I1008 18:37:36.311523 4859 scope.go:117] "RemoveContainer" containerID="fecccb4aab6e4695efa36ed5ba960994a10727b9aac69914dd84867213d10810" Oct 08 18:37:36 crc kubenswrapper[4859]: E1008 18:37:36.312057 4859 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"fecccb4aab6e4695efa36ed5ba960994a10727b9aac69914dd84867213d10810\": container with ID starting with fecccb4aab6e4695efa36ed5ba960994a10727b9aac69914dd84867213d10810 not found: ID does not exist" containerID="fecccb4aab6e4695efa36ed5ba960994a10727b9aac69914dd84867213d10810" Oct 08 18:37:36 crc kubenswrapper[4859]: I1008 18:37:36.312172 4859 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"fecccb4aab6e4695efa36ed5ba960994a10727b9aac69914dd84867213d10810"} err="failed to get container status \"fecccb4aab6e4695efa36ed5ba960994a10727b9aac69914dd84867213d10810\": rpc error: code = NotFound desc = could not find container \"fecccb4aab6e4695efa36ed5ba960994a10727b9aac69914dd84867213d10810\": container with ID starting with fecccb4aab6e4695efa36ed5ba960994a10727b9aac69914dd84867213d10810 not found: ID does not exist" Oct 08 18:37:36 crc kubenswrapper[4859]: I1008 18:37:36.312254 4859 scope.go:117] "RemoveContainer" containerID="1ff56b52d2d31d1f67ef0e984e2373890024eeaf7a31c66677cb07168c88da23" Oct 08 18:37:36 crc kubenswrapper[4859]: I1008 18:37:36.312425 4859 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/rabbitmq-server-0"] Oct 08 18:37:36 crc kubenswrapper[4859]: E1008 18:37:36.312959 4859 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ba5a78be-e7b3-401a-bdb9-ba31071d7075" containerName="setup-container" Oct 08 18:37:36 crc kubenswrapper[4859]: I1008 18:37:36.313048 4859 state_mem.go:107] "Deleted CPUSet assignment" podUID="ba5a78be-e7b3-401a-bdb9-ba31071d7075" containerName="setup-container" Oct 08 18:37:36 crc kubenswrapper[4859]: E1008 18:37:36.313128 4859 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ba5a78be-e7b3-401a-bdb9-ba31071d7075" containerName="rabbitmq" Oct 08 18:37:36 crc kubenswrapper[4859]: I1008 18:37:36.313205 4859 state_mem.go:107] "Deleted CPUSet assignment" podUID="ba5a78be-e7b3-401a-bdb9-ba31071d7075" containerName="rabbitmq" Oct 08 18:37:36 crc kubenswrapper[4859]: E1008 18:37:36.313281 4859 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3075d5c2-3e74-4a4c-829f-248e87f45f3b" containerName="setup-container" Oct 08 18:37:36 crc kubenswrapper[4859]: I1008 18:37:36.313360 4859 state_mem.go:107] "Deleted CPUSet assignment" podUID="3075d5c2-3e74-4a4c-829f-248e87f45f3b" containerName="setup-container" Oct 08 18:37:36 crc kubenswrapper[4859]: E1008 18:37:36.313446 4859 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3075d5c2-3e74-4a4c-829f-248e87f45f3b" containerName="rabbitmq" Oct 08 18:37:36 crc kubenswrapper[4859]: I1008 18:37:36.313518 4859 state_mem.go:107] "Deleted CPUSet assignment" podUID="3075d5c2-3e74-4a4c-829f-248e87f45f3b" containerName="rabbitmq" Oct 08 18:37:36 crc kubenswrapper[4859]: I1008 18:37:36.313835 4859 memory_manager.go:354] "RemoveStaleState removing state" podUID="ba5a78be-e7b3-401a-bdb9-ba31071d7075" containerName="rabbitmq" Oct 08 18:37:36 crc kubenswrapper[4859]: I1008 18:37:36.313952 4859 memory_manager.go:354] "RemoveStaleState removing state" podUID="3075d5c2-3e74-4a4c-829f-248e87f45f3b" containerName="rabbitmq" Oct 08 18:37:36 crc kubenswrapper[4859]: I1008 18:37:36.315304 4859 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-server-0" Oct 08 18:37:36 crc kubenswrapper[4859]: I1008 18:37:36.319358 4859 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-rabbitmq-svc" Oct 08 18:37:36 crc kubenswrapper[4859]: I1008 18:37:36.319543 4859 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-server-dockercfg-rdvqq" Oct 08 18:37:36 crc kubenswrapper[4859]: I1008 18:37:36.319653 4859 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-plugins-conf" Oct 08 18:37:36 crc kubenswrapper[4859]: I1008 18:37:36.319823 4859 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-erlang-cookie" Oct 08 18:37:36 crc kubenswrapper[4859]: I1008 18:37:36.319925 4859 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-server-conf" Oct 08 18:37:36 crc kubenswrapper[4859]: I1008 18:37:36.320053 4859 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-default-user" Oct 08 18:37:36 crc kubenswrapper[4859]: I1008 18:37:36.320188 4859 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-config-data" Oct 08 18:37:36 crc kubenswrapper[4859]: I1008 18:37:36.320390 4859 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/rabbitmq-server-0"] Oct 08 18:37:36 crc kubenswrapper[4859]: I1008 18:37:36.342892 4859 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/rabbitmq-cell1-server-0"] Oct 08 18:37:36 crc kubenswrapper[4859]: I1008 18:37:36.344773 4859 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-cell1-server-0" Oct 08 18:37:36 crc kubenswrapper[4859]: I1008 18:37:36.348299 4859 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-cell1-erlang-cookie" Oct 08 18:37:36 crc kubenswrapper[4859]: I1008 18:37:36.348940 4859 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-cell1-server-dockercfg-rg8nn" Oct 08 18:37:36 crc kubenswrapper[4859]: I1008 18:37:36.349089 4859 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-cell1-config-data" Oct 08 18:37:36 crc kubenswrapper[4859]: I1008 18:37:36.349220 4859 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-cell1-plugins-conf" Oct 08 18:37:36 crc kubenswrapper[4859]: I1008 18:37:36.349301 4859 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-cell1-default-user" Oct 08 18:37:36 crc kubenswrapper[4859]: I1008 18:37:36.349563 4859 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-rabbitmq-cell1-svc" Oct 08 18:37:36 crc kubenswrapper[4859]: I1008 18:37:36.351024 4859 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-cell1-server-conf" Oct 08 18:37:36 crc kubenswrapper[4859]: I1008 18:37:36.351859 4859 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/rabbitmq-cell1-server-0"] Oct 08 18:37:36 crc kubenswrapper[4859]: I1008 18:37:36.380865 4859 scope.go:117] "RemoveContainer" containerID="6e2de3cebf597d7492590e385a7bc16bfc073ff1e91df7b11869d97d8960f7dd" Oct 08 18:37:36 crc kubenswrapper[4859]: I1008 18:37:36.476294 4859 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/2c6db854-3378-4bc4-b622-480bc8c42bf5-pod-info\") pod \"rabbitmq-cell1-server-0\" (UID: \"2c6db854-3378-4bc4-b622-480bc8c42bf5\") " pod="openstack/rabbitmq-cell1-server-0" Oct 08 18:37:36 crc kubenswrapper[4859]: I1008 18:37:36.476330 4859 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/2c6db854-3378-4bc4-b622-480bc8c42bf5-rabbitmq-confd\") pod \"rabbitmq-cell1-server-0\" (UID: \"2c6db854-3378-4bc4-b622-480bc8c42bf5\") " pod="openstack/rabbitmq-cell1-server-0" Oct 08 18:37:36 crc kubenswrapper[4859]: I1008 18:37:36.476356 4859 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage09-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage09-crc\") pod \"rabbitmq-cell1-server-0\" (UID: \"2c6db854-3378-4bc4-b622-480bc8c42bf5\") " pod="openstack/rabbitmq-cell1-server-0" Oct 08 18:37:36 crc kubenswrapper[4859]: I1008 18:37:36.476381 4859 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-2gmb7\" (UniqueName: \"kubernetes.io/projected/2c6db854-3378-4bc4-b622-480bc8c42bf5-kube-api-access-2gmb7\") pod \"rabbitmq-cell1-server-0\" (UID: \"2c6db854-3378-4bc4-b622-480bc8c42bf5\") " pod="openstack/rabbitmq-cell1-server-0" Oct 08 18:37:36 crc kubenswrapper[4859]: I1008 18:37:36.476432 4859 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/2c6db854-3378-4bc4-b622-480bc8c42bf5-erlang-cookie-secret\") pod \"rabbitmq-cell1-server-0\" (UID: \"2c6db854-3378-4bc4-b622-480bc8c42bf5\") " pod="openstack/rabbitmq-cell1-server-0" Oct 08 18:37:36 crc kubenswrapper[4859]: I1008 18:37:36.476454 4859 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/94450380-649a-4476-b646-018ed24aa703-plugins-conf\") pod \"rabbitmq-server-0\" (UID: \"94450380-649a-4476-b646-018ed24aa703\") " pod="openstack/rabbitmq-server-0" Oct 08 18:37:36 crc kubenswrapper[4859]: I1008 18:37:36.476478 4859 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-kxr59\" (UniqueName: \"kubernetes.io/projected/94450380-649a-4476-b646-018ed24aa703-kube-api-access-kxr59\") pod \"rabbitmq-server-0\" (UID: \"94450380-649a-4476-b646-018ed24aa703\") " pod="openstack/rabbitmq-server-0" Oct 08 18:37:36 crc kubenswrapper[4859]: I1008 18:37:36.476496 4859 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/94450380-649a-4476-b646-018ed24aa703-rabbitmq-plugins\") pod \"rabbitmq-server-0\" (UID: \"94450380-649a-4476-b646-018ed24aa703\") " pod="openstack/rabbitmq-server-0" Oct 08 18:37:36 crc kubenswrapper[4859]: I1008 18:37:36.476518 4859 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/2c6db854-3378-4bc4-b622-480bc8c42bf5-rabbitmq-plugins\") pod \"rabbitmq-cell1-server-0\" (UID: \"2c6db854-3378-4bc4-b622-480bc8c42bf5\") " pod="openstack/rabbitmq-cell1-server-0" Oct 08 18:37:36 crc kubenswrapper[4859]: I1008 18:37:36.476551 4859 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/94450380-649a-4476-b646-018ed24aa703-config-data\") pod \"rabbitmq-server-0\" (UID: \"94450380-649a-4476-b646-018ed24aa703\") " pod="openstack/rabbitmq-server-0" Oct 08 18:37:36 crc kubenswrapper[4859]: I1008 18:37:36.476572 4859 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/2c6db854-3378-4bc4-b622-480bc8c42bf5-rabbitmq-tls\") pod \"rabbitmq-cell1-server-0\" (UID: \"2c6db854-3378-4bc4-b622-480bc8c42bf5\") " pod="openstack/rabbitmq-cell1-server-0" Oct 08 18:37:36 crc kubenswrapper[4859]: I1008 18:37:36.476595 4859 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/94450380-649a-4476-b646-018ed24aa703-server-conf\") pod \"rabbitmq-server-0\" (UID: \"94450380-649a-4476-b646-018ed24aa703\") " pod="openstack/rabbitmq-server-0" Oct 08 18:37:36 crc kubenswrapper[4859]: I1008 18:37:36.477040 4859 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/2c6db854-3378-4bc4-b622-480bc8c42bf5-rabbitmq-erlang-cookie\") pod \"rabbitmq-cell1-server-0\" (UID: \"2c6db854-3378-4bc4-b622-480bc8c42bf5\") " pod="openstack/rabbitmq-cell1-server-0" Oct 08 18:37:36 crc kubenswrapper[4859]: I1008 18:37:36.477080 4859 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/94450380-649a-4476-b646-018ed24aa703-rabbitmq-confd\") pod \"rabbitmq-server-0\" (UID: \"94450380-649a-4476-b646-018ed24aa703\") " pod="openstack/rabbitmq-server-0" Oct 08 18:37:36 crc kubenswrapper[4859]: I1008 18:37:36.477106 4859 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/2c6db854-3378-4bc4-b622-480bc8c42bf5-server-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"2c6db854-3378-4bc4-b622-480bc8c42bf5\") " pod="openstack/rabbitmq-cell1-server-0" Oct 08 18:37:36 crc kubenswrapper[4859]: I1008 18:37:36.477126 4859 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage08-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage08-crc\") pod \"rabbitmq-server-0\" (UID: \"94450380-649a-4476-b646-018ed24aa703\") " pod="openstack/rabbitmq-server-0" Oct 08 18:37:36 crc kubenswrapper[4859]: I1008 18:37:36.477142 4859 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/2c6db854-3378-4bc4-b622-480bc8c42bf5-config-data\") pod \"rabbitmq-cell1-server-0\" (UID: \"2c6db854-3378-4bc4-b622-480bc8c42bf5\") " pod="openstack/rabbitmq-cell1-server-0" Oct 08 18:37:36 crc kubenswrapper[4859]: I1008 18:37:36.477162 4859 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/94450380-649a-4476-b646-018ed24aa703-erlang-cookie-secret\") pod \"rabbitmq-server-0\" (UID: \"94450380-649a-4476-b646-018ed24aa703\") " pod="openstack/rabbitmq-server-0" Oct 08 18:37:36 crc kubenswrapper[4859]: I1008 18:37:36.477191 4859 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/2c6db854-3378-4bc4-b622-480bc8c42bf5-plugins-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"2c6db854-3378-4bc4-b622-480bc8c42bf5\") " pod="openstack/rabbitmq-cell1-server-0" Oct 08 18:37:36 crc kubenswrapper[4859]: I1008 18:37:36.477209 4859 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/94450380-649a-4476-b646-018ed24aa703-rabbitmq-tls\") pod \"rabbitmq-server-0\" (UID: \"94450380-649a-4476-b646-018ed24aa703\") " pod="openstack/rabbitmq-server-0" Oct 08 18:37:36 crc kubenswrapper[4859]: I1008 18:37:36.477227 4859 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/94450380-649a-4476-b646-018ed24aa703-rabbitmq-erlang-cookie\") pod \"rabbitmq-server-0\" (UID: \"94450380-649a-4476-b646-018ed24aa703\") " pod="openstack/rabbitmq-server-0" Oct 08 18:37:36 crc kubenswrapper[4859]: I1008 18:37:36.477241 4859 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/94450380-649a-4476-b646-018ed24aa703-pod-info\") pod \"rabbitmq-server-0\" (UID: \"94450380-649a-4476-b646-018ed24aa703\") " pod="openstack/rabbitmq-server-0" Oct 08 18:37:36 crc kubenswrapper[4859]: I1008 18:37:36.482443 4859 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="3075d5c2-3e74-4a4c-829f-248e87f45f3b" path="/var/lib/kubelet/pods/3075d5c2-3e74-4a4c-829f-248e87f45f3b/volumes" Oct 08 18:37:36 crc kubenswrapper[4859]: I1008 18:37:36.484521 4859 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="ba5a78be-e7b3-401a-bdb9-ba31071d7075" path="/var/lib/kubelet/pods/ba5a78be-e7b3-401a-bdb9-ba31071d7075/volumes" Oct 08 18:37:36 crc kubenswrapper[4859]: I1008 18:37:36.581314 4859 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/94450380-649a-4476-b646-018ed24aa703-rabbitmq-confd\") pod \"rabbitmq-server-0\" (UID: \"94450380-649a-4476-b646-018ed24aa703\") " pod="openstack/rabbitmq-server-0" Oct 08 18:37:36 crc kubenswrapper[4859]: I1008 18:37:36.581361 4859 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/2c6db854-3378-4bc4-b622-480bc8c42bf5-server-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"2c6db854-3378-4bc4-b622-480bc8c42bf5\") " pod="openstack/rabbitmq-cell1-server-0" Oct 08 18:37:36 crc kubenswrapper[4859]: I1008 18:37:36.581389 4859 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage08-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage08-crc\") pod \"rabbitmq-server-0\" (UID: \"94450380-649a-4476-b646-018ed24aa703\") " pod="openstack/rabbitmq-server-0" Oct 08 18:37:36 crc kubenswrapper[4859]: I1008 18:37:36.581415 4859 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/2c6db854-3378-4bc4-b622-480bc8c42bf5-config-data\") pod \"rabbitmq-cell1-server-0\" (UID: \"2c6db854-3378-4bc4-b622-480bc8c42bf5\") " pod="openstack/rabbitmq-cell1-server-0" Oct 08 18:37:36 crc kubenswrapper[4859]: I1008 18:37:36.581452 4859 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/94450380-649a-4476-b646-018ed24aa703-erlang-cookie-secret\") pod \"rabbitmq-server-0\" (UID: \"94450380-649a-4476-b646-018ed24aa703\") " pod="openstack/rabbitmq-server-0" Oct 08 18:37:36 crc kubenswrapper[4859]: I1008 18:37:36.581511 4859 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/2c6db854-3378-4bc4-b622-480bc8c42bf5-plugins-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"2c6db854-3378-4bc4-b622-480bc8c42bf5\") " pod="openstack/rabbitmq-cell1-server-0" Oct 08 18:37:36 crc kubenswrapper[4859]: I1008 18:37:36.581540 4859 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/94450380-649a-4476-b646-018ed24aa703-rabbitmq-tls\") pod \"rabbitmq-server-0\" (UID: \"94450380-649a-4476-b646-018ed24aa703\") " pod="openstack/rabbitmq-server-0" Oct 08 18:37:36 crc kubenswrapper[4859]: I1008 18:37:36.581563 4859 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/94450380-649a-4476-b646-018ed24aa703-rabbitmq-erlang-cookie\") pod \"rabbitmq-server-0\" (UID: \"94450380-649a-4476-b646-018ed24aa703\") " pod="openstack/rabbitmq-server-0" Oct 08 18:37:36 crc kubenswrapper[4859]: I1008 18:37:36.581578 4859 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/94450380-649a-4476-b646-018ed24aa703-pod-info\") pod \"rabbitmq-server-0\" (UID: \"94450380-649a-4476-b646-018ed24aa703\") " pod="openstack/rabbitmq-server-0" Oct 08 18:37:36 crc kubenswrapper[4859]: I1008 18:37:36.581617 4859 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/2c6db854-3378-4bc4-b622-480bc8c42bf5-pod-info\") pod \"rabbitmq-cell1-server-0\" (UID: \"2c6db854-3378-4bc4-b622-480bc8c42bf5\") " pod="openstack/rabbitmq-cell1-server-0" Oct 08 18:37:36 crc kubenswrapper[4859]: I1008 18:37:36.581650 4859 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/2c6db854-3378-4bc4-b622-480bc8c42bf5-rabbitmq-confd\") pod \"rabbitmq-cell1-server-0\" (UID: \"2c6db854-3378-4bc4-b622-480bc8c42bf5\") " pod="openstack/rabbitmq-cell1-server-0" Oct 08 18:37:36 crc kubenswrapper[4859]: I1008 18:37:36.581701 4859 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage09-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage09-crc\") pod \"rabbitmq-cell1-server-0\" (UID: \"2c6db854-3378-4bc4-b622-480bc8c42bf5\") " pod="openstack/rabbitmq-cell1-server-0" Oct 08 18:37:36 crc kubenswrapper[4859]: I1008 18:37:36.581750 4859 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-2gmb7\" (UniqueName: \"kubernetes.io/projected/2c6db854-3378-4bc4-b622-480bc8c42bf5-kube-api-access-2gmb7\") pod \"rabbitmq-cell1-server-0\" (UID: \"2c6db854-3378-4bc4-b622-480bc8c42bf5\") " pod="openstack/rabbitmq-cell1-server-0" Oct 08 18:37:36 crc kubenswrapper[4859]: I1008 18:37:36.581780 4859 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/2c6db854-3378-4bc4-b622-480bc8c42bf5-erlang-cookie-secret\") pod \"rabbitmq-cell1-server-0\" (UID: \"2c6db854-3378-4bc4-b622-480bc8c42bf5\") " pod="openstack/rabbitmq-cell1-server-0" Oct 08 18:37:36 crc kubenswrapper[4859]: I1008 18:37:36.581836 4859 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/94450380-649a-4476-b646-018ed24aa703-plugins-conf\") pod \"rabbitmq-server-0\" (UID: \"94450380-649a-4476-b646-018ed24aa703\") " pod="openstack/rabbitmq-server-0" Oct 08 18:37:36 crc kubenswrapper[4859]: I1008 18:37:36.581886 4859 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-kxr59\" (UniqueName: \"kubernetes.io/projected/94450380-649a-4476-b646-018ed24aa703-kube-api-access-kxr59\") pod \"rabbitmq-server-0\" (UID: \"94450380-649a-4476-b646-018ed24aa703\") " pod="openstack/rabbitmq-server-0" Oct 08 18:37:36 crc kubenswrapper[4859]: I1008 18:37:36.581916 4859 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/94450380-649a-4476-b646-018ed24aa703-rabbitmq-plugins\") pod \"rabbitmq-server-0\" (UID: \"94450380-649a-4476-b646-018ed24aa703\") " pod="openstack/rabbitmq-server-0" Oct 08 18:37:36 crc kubenswrapper[4859]: I1008 18:37:36.581961 4859 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/2c6db854-3378-4bc4-b622-480bc8c42bf5-rabbitmq-plugins\") pod \"rabbitmq-cell1-server-0\" (UID: \"2c6db854-3378-4bc4-b622-480bc8c42bf5\") " pod="openstack/rabbitmq-cell1-server-0" Oct 08 18:37:36 crc kubenswrapper[4859]: I1008 18:37:36.582021 4859 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/94450380-649a-4476-b646-018ed24aa703-config-data\") pod \"rabbitmq-server-0\" (UID: \"94450380-649a-4476-b646-018ed24aa703\") " pod="openstack/rabbitmq-server-0" Oct 08 18:37:36 crc kubenswrapper[4859]: I1008 18:37:36.582045 4859 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/2c6db854-3378-4bc4-b622-480bc8c42bf5-rabbitmq-tls\") pod \"rabbitmq-cell1-server-0\" (UID: \"2c6db854-3378-4bc4-b622-480bc8c42bf5\") " pod="openstack/rabbitmq-cell1-server-0" Oct 08 18:37:36 crc kubenswrapper[4859]: I1008 18:37:36.582071 4859 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/94450380-649a-4476-b646-018ed24aa703-server-conf\") pod \"rabbitmq-server-0\" (UID: \"94450380-649a-4476-b646-018ed24aa703\") " pod="openstack/rabbitmq-server-0" Oct 08 18:37:36 crc kubenswrapper[4859]: I1008 18:37:36.582144 4859 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/2c6db854-3378-4bc4-b622-480bc8c42bf5-rabbitmq-erlang-cookie\") pod \"rabbitmq-cell1-server-0\" (UID: \"2c6db854-3378-4bc4-b622-480bc8c42bf5\") " pod="openstack/rabbitmq-cell1-server-0" Oct 08 18:37:36 crc kubenswrapper[4859]: I1008 18:37:36.582712 4859 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage08-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage08-crc\") pod \"rabbitmq-server-0\" (UID: \"94450380-649a-4476-b646-018ed24aa703\") device mount path \"/mnt/openstack/pv08\"" pod="openstack/rabbitmq-server-0" Oct 08 18:37:36 crc kubenswrapper[4859]: I1008 18:37:36.585160 4859 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/2c6db854-3378-4bc4-b622-480bc8c42bf5-config-data\") pod \"rabbitmq-cell1-server-0\" (UID: \"2c6db854-3378-4bc4-b622-480bc8c42bf5\") " pod="openstack/rabbitmq-cell1-server-0" Oct 08 18:37:36 crc kubenswrapper[4859]: I1008 18:37:36.586117 4859 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/94450380-649a-4476-b646-018ed24aa703-rabbitmq-plugins\") pod \"rabbitmq-server-0\" (UID: \"94450380-649a-4476-b646-018ed24aa703\") " pod="openstack/rabbitmq-server-0" Oct 08 18:37:36 crc kubenswrapper[4859]: I1008 18:37:36.586563 4859 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/2c6db854-3378-4bc4-b622-480bc8c42bf5-server-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"2c6db854-3378-4bc4-b622-480bc8c42bf5\") " pod="openstack/rabbitmq-cell1-server-0" Oct 08 18:37:36 crc kubenswrapper[4859]: I1008 18:37:36.586996 4859 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/94450380-649a-4476-b646-018ed24aa703-plugins-conf\") pod \"rabbitmq-server-0\" (UID: \"94450380-649a-4476-b646-018ed24aa703\") " pod="openstack/rabbitmq-server-0" Oct 08 18:37:36 crc kubenswrapper[4859]: I1008 18:37:36.587510 4859 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage09-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage09-crc\") pod \"rabbitmq-cell1-server-0\" (UID: \"2c6db854-3378-4bc4-b622-480bc8c42bf5\") device mount path \"/mnt/openstack/pv09\"" pod="openstack/rabbitmq-cell1-server-0" Oct 08 18:37:36 crc kubenswrapper[4859]: I1008 18:37:36.588124 4859 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/2c6db854-3378-4bc4-b622-480bc8c42bf5-rabbitmq-plugins\") pod \"rabbitmq-cell1-server-0\" (UID: \"2c6db854-3378-4bc4-b622-480bc8c42bf5\") " pod="openstack/rabbitmq-cell1-server-0" Oct 08 18:37:36 crc kubenswrapper[4859]: I1008 18:37:36.588434 4859 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/2c6db854-3378-4bc4-b622-480bc8c42bf5-rabbitmq-erlang-cookie\") pod \"rabbitmq-cell1-server-0\" (UID: \"2c6db854-3378-4bc4-b622-480bc8c42bf5\") " pod="openstack/rabbitmq-cell1-server-0" Oct 08 18:37:36 crc kubenswrapper[4859]: I1008 18:37:36.588525 4859 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/2c6db854-3378-4bc4-b622-480bc8c42bf5-plugins-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"2c6db854-3378-4bc4-b622-480bc8c42bf5\") " pod="openstack/rabbitmq-cell1-server-0" Oct 08 18:37:36 crc kubenswrapper[4859]: I1008 18:37:36.589046 4859 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/2c6db854-3378-4bc4-b622-480bc8c42bf5-pod-info\") pod \"rabbitmq-cell1-server-0\" (UID: \"2c6db854-3378-4bc4-b622-480bc8c42bf5\") " pod="openstack/rabbitmq-cell1-server-0" Oct 08 18:37:36 crc kubenswrapper[4859]: I1008 18:37:36.589741 4859 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/94450380-649a-4476-b646-018ed24aa703-config-data\") pod \"rabbitmq-server-0\" (UID: \"94450380-649a-4476-b646-018ed24aa703\") " pod="openstack/rabbitmq-server-0" Oct 08 18:37:36 crc kubenswrapper[4859]: I1008 18:37:36.589994 4859 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/94450380-649a-4476-b646-018ed24aa703-erlang-cookie-secret\") pod \"rabbitmq-server-0\" (UID: \"94450380-649a-4476-b646-018ed24aa703\") " pod="openstack/rabbitmq-server-0" Oct 08 18:37:36 crc kubenswrapper[4859]: I1008 18:37:36.590186 4859 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/94450380-649a-4476-b646-018ed24aa703-server-conf\") pod \"rabbitmq-server-0\" (UID: \"94450380-649a-4476-b646-018ed24aa703\") " pod="openstack/rabbitmq-server-0" Oct 08 18:37:36 crc kubenswrapper[4859]: I1008 18:37:36.590277 4859 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/94450380-649a-4476-b646-018ed24aa703-rabbitmq-erlang-cookie\") pod \"rabbitmq-server-0\" (UID: \"94450380-649a-4476-b646-018ed24aa703\") " pod="openstack/rabbitmq-server-0" Oct 08 18:37:36 crc kubenswrapper[4859]: I1008 18:37:36.592230 4859 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/94450380-649a-4476-b646-018ed24aa703-rabbitmq-tls\") pod \"rabbitmq-server-0\" (UID: \"94450380-649a-4476-b646-018ed24aa703\") " pod="openstack/rabbitmq-server-0" Oct 08 18:37:36 crc kubenswrapper[4859]: I1008 18:37:36.593607 4859 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/2c6db854-3378-4bc4-b622-480bc8c42bf5-rabbitmq-tls\") pod \"rabbitmq-cell1-server-0\" (UID: \"2c6db854-3378-4bc4-b622-480bc8c42bf5\") " pod="openstack/rabbitmq-cell1-server-0" Oct 08 18:37:36 crc kubenswrapper[4859]: I1008 18:37:36.601913 4859 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/2c6db854-3378-4bc4-b622-480bc8c42bf5-erlang-cookie-secret\") pod \"rabbitmq-cell1-server-0\" (UID: \"2c6db854-3378-4bc4-b622-480bc8c42bf5\") " pod="openstack/rabbitmq-cell1-server-0" Oct 08 18:37:36 crc kubenswrapper[4859]: I1008 18:37:36.602037 4859 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/94450380-649a-4476-b646-018ed24aa703-rabbitmq-confd\") pod \"rabbitmq-server-0\" (UID: \"94450380-649a-4476-b646-018ed24aa703\") " pod="openstack/rabbitmq-server-0" Oct 08 18:37:36 crc kubenswrapper[4859]: I1008 18:37:36.606434 4859 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-kxr59\" (UniqueName: \"kubernetes.io/projected/94450380-649a-4476-b646-018ed24aa703-kube-api-access-kxr59\") pod \"rabbitmq-server-0\" (UID: \"94450380-649a-4476-b646-018ed24aa703\") " pod="openstack/rabbitmq-server-0" Oct 08 18:37:36 crc kubenswrapper[4859]: I1008 18:37:36.607319 4859 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/94450380-649a-4476-b646-018ed24aa703-pod-info\") pod \"rabbitmq-server-0\" (UID: \"94450380-649a-4476-b646-018ed24aa703\") " pod="openstack/rabbitmq-server-0" Oct 08 18:37:36 crc kubenswrapper[4859]: I1008 18:37:36.608270 4859 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-2gmb7\" (UniqueName: \"kubernetes.io/projected/2c6db854-3378-4bc4-b622-480bc8c42bf5-kube-api-access-2gmb7\") pod \"rabbitmq-cell1-server-0\" (UID: \"2c6db854-3378-4bc4-b622-480bc8c42bf5\") " pod="openstack/rabbitmq-cell1-server-0" Oct 08 18:37:36 crc kubenswrapper[4859]: I1008 18:37:36.610729 4859 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/2c6db854-3378-4bc4-b622-480bc8c42bf5-rabbitmq-confd\") pod \"rabbitmq-cell1-server-0\" (UID: \"2c6db854-3378-4bc4-b622-480bc8c42bf5\") " pod="openstack/rabbitmq-cell1-server-0" Oct 08 18:37:36 crc kubenswrapper[4859]: I1008 18:37:36.629736 4859 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage09-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage09-crc\") pod \"rabbitmq-cell1-server-0\" (UID: \"2c6db854-3378-4bc4-b622-480bc8c42bf5\") " pod="openstack/rabbitmq-cell1-server-0" Oct 08 18:37:36 crc kubenswrapper[4859]: I1008 18:37:36.631074 4859 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage08-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage08-crc\") pod \"rabbitmq-server-0\" (UID: \"94450380-649a-4476-b646-018ed24aa703\") " pod="openstack/rabbitmq-server-0" Oct 08 18:37:36 crc kubenswrapper[4859]: I1008 18:37:36.654130 4859 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-server-0" Oct 08 18:37:36 crc kubenswrapper[4859]: I1008 18:37:36.679605 4859 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-cell1-server-0" Oct 08 18:37:37 crc kubenswrapper[4859]: I1008 18:37:37.129595 4859 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/rabbitmq-server-0"] Oct 08 18:37:37 crc kubenswrapper[4859]: I1008 18:37:37.241539 4859 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-server-0" event={"ID":"94450380-649a-4476-b646-018ed24aa703","Type":"ContainerStarted","Data":"18194fc529062dc06ab4227ff322958438d3142040b55368d78c3278a50d5b91"} Oct 08 18:37:37 crc kubenswrapper[4859]: W1008 18:37:37.253007 4859 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod2c6db854_3378_4bc4_b622_480bc8c42bf5.slice/crio-9edfb8c6b084281496a254c2c1be0f189106dc4e159c25c1d46153937b814e58 WatchSource:0}: Error finding container 9edfb8c6b084281496a254c2c1be0f189106dc4e159c25c1d46153937b814e58: Status 404 returned error can't find the container with id 9edfb8c6b084281496a254c2c1be0f189106dc4e159c25c1d46153937b814e58 Oct 08 18:37:37 crc kubenswrapper[4859]: I1008 18:37:37.253953 4859 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/rabbitmq-cell1-server-0"] Oct 08 18:37:37 crc kubenswrapper[4859]: I1008 18:37:37.271899 4859 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-84b4cc85c9-r4s54"] Oct 08 18:37:37 crc kubenswrapper[4859]: I1008 18:37:37.274045 4859 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-84b4cc85c9-r4s54" Oct 08 18:37:37 crc kubenswrapper[4859]: I1008 18:37:37.277995 4859 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-edpm-ipam" Oct 08 18:37:37 crc kubenswrapper[4859]: I1008 18:37:37.297267 4859 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-84b4cc85c9-r4s54"] Oct 08 18:37:37 crc kubenswrapper[4859]: I1008 18:37:37.398814 4859 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/f48d81b1-590d-43d8-9111-fdada4e99811-ovsdbserver-sb\") pod \"dnsmasq-dns-84b4cc85c9-r4s54\" (UID: \"f48d81b1-590d-43d8-9111-fdada4e99811\") " pod="openstack/dnsmasq-dns-84b4cc85c9-r4s54" Oct 08 18:37:37 crc kubenswrapper[4859]: I1008 18:37:37.398896 4859 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/configmap/f48d81b1-590d-43d8-9111-fdada4e99811-openstack-edpm-ipam\") pod \"dnsmasq-dns-84b4cc85c9-r4s54\" (UID: \"f48d81b1-590d-43d8-9111-fdada4e99811\") " pod="openstack/dnsmasq-dns-84b4cc85c9-r4s54" Oct 08 18:37:37 crc kubenswrapper[4859]: I1008 18:37:37.398923 4859 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-wfcrj\" (UniqueName: \"kubernetes.io/projected/f48d81b1-590d-43d8-9111-fdada4e99811-kube-api-access-wfcrj\") pod \"dnsmasq-dns-84b4cc85c9-r4s54\" (UID: \"f48d81b1-590d-43d8-9111-fdada4e99811\") " pod="openstack/dnsmasq-dns-84b4cc85c9-r4s54" Oct 08 18:37:37 crc kubenswrapper[4859]: I1008 18:37:37.398950 4859 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/f48d81b1-590d-43d8-9111-fdada4e99811-config\") pod \"dnsmasq-dns-84b4cc85c9-r4s54\" (UID: \"f48d81b1-590d-43d8-9111-fdada4e99811\") " pod="openstack/dnsmasq-dns-84b4cc85c9-r4s54" Oct 08 18:37:37 crc kubenswrapper[4859]: I1008 18:37:37.398978 4859 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/f48d81b1-590d-43d8-9111-fdada4e99811-dns-swift-storage-0\") pod \"dnsmasq-dns-84b4cc85c9-r4s54\" (UID: \"f48d81b1-590d-43d8-9111-fdada4e99811\") " pod="openstack/dnsmasq-dns-84b4cc85c9-r4s54" Oct 08 18:37:37 crc kubenswrapper[4859]: I1008 18:37:37.399050 4859 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/f48d81b1-590d-43d8-9111-fdada4e99811-ovsdbserver-nb\") pod \"dnsmasq-dns-84b4cc85c9-r4s54\" (UID: \"f48d81b1-590d-43d8-9111-fdada4e99811\") " pod="openstack/dnsmasq-dns-84b4cc85c9-r4s54" Oct 08 18:37:37 crc kubenswrapper[4859]: I1008 18:37:37.399070 4859 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/f48d81b1-590d-43d8-9111-fdada4e99811-dns-svc\") pod \"dnsmasq-dns-84b4cc85c9-r4s54\" (UID: \"f48d81b1-590d-43d8-9111-fdada4e99811\") " pod="openstack/dnsmasq-dns-84b4cc85c9-r4s54" Oct 08 18:37:37 crc kubenswrapper[4859]: I1008 18:37:37.500859 4859 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/f48d81b1-590d-43d8-9111-fdada4e99811-dns-swift-storage-0\") pod \"dnsmasq-dns-84b4cc85c9-r4s54\" (UID: \"f48d81b1-590d-43d8-9111-fdada4e99811\") " pod="openstack/dnsmasq-dns-84b4cc85c9-r4s54" Oct 08 18:37:37 crc kubenswrapper[4859]: I1008 18:37:37.500958 4859 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/f48d81b1-590d-43d8-9111-fdada4e99811-ovsdbserver-nb\") pod \"dnsmasq-dns-84b4cc85c9-r4s54\" (UID: \"f48d81b1-590d-43d8-9111-fdada4e99811\") " pod="openstack/dnsmasq-dns-84b4cc85c9-r4s54" Oct 08 18:37:37 crc kubenswrapper[4859]: I1008 18:37:37.500981 4859 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/f48d81b1-590d-43d8-9111-fdada4e99811-dns-svc\") pod \"dnsmasq-dns-84b4cc85c9-r4s54\" (UID: \"f48d81b1-590d-43d8-9111-fdada4e99811\") " pod="openstack/dnsmasq-dns-84b4cc85c9-r4s54" Oct 08 18:37:37 crc kubenswrapper[4859]: I1008 18:37:37.501067 4859 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/f48d81b1-590d-43d8-9111-fdada4e99811-ovsdbserver-sb\") pod \"dnsmasq-dns-84b4cc85c9-r4s54\" (UID: \"f48d81b1-590d-43d8-9111-fdada4e99811\") " pod="openstack/dnsmasq-dns-84b4cc85c9-r4s54" Oct 08 18:37:37 crc kubenswrapper[4859]: I1008 18:37:37.501101 4859 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/configmap/f48d81b1-590d-43d8-9111-fdada4e99811-openstack-edpm-ipam\") pod \"dnsmasq-dns-84b4cc85c9-r4s54\" (UID: \"f48d81b1-590d-43d8-9111-fdada4e99811\") " pod="openstack/dnsmasq-dns-84b4cc85c9-r4s54" Oct 08 18:37:37 crc kubenswrapper[4859]: I1008 18:37:37.501125 4859 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-wfcrj\" (UniqueName: \"kubernetes.io/projected/f48d81b1-590d-43d8-9111-fdada4e99811-kube-api-access-wfcrj\") pod \"dnsmasq-dns-84b4cc85c9-r4s54\" (UID: \"f48d81b1-590d-43d8-9111-fdada4e99811\") " pod="openstack/dnsmasq-dns-84b4cc85c9-r4s54" Oct 08 18:37:37 crc kubenswrapper[4859]: I1008 18:37:37.501146 4859 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/f48d81b1-590d-43d8-9111-fdada4e99811-config\") pod \"dnsmasq-dns-84b4cc85c9-r4s54\" (UID: \"f48d81b1-590d-43d8-9111-fdada4e99811\") " pod="openstack/dnsmasq-dns-84b4cc85c9-r4s54" Oct 08 18:37:37 crc kubenswrapper[4859]: I1008 18:37:37.501925 4859 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/f48d81b1-590d-43d8-9111-fdada4e99811-dns-swift-storage-0\") pod \"dnsmasq-dns-84b4cc85c9-r4s54\" (UID: \"f48d81b1-590d-43d8-9111-fdada4e99811\") " pod="openstack/dnsmasq-dns-84b4cc85c9-r4s54" Oct 08 18:37:37 crc kubenswrapper[4859]: I1008 18:37:37.501974 4859 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/f48d81b1-590d-43d8-9111-fdada4e99811-dns-svc\") pod \"dnsmasq-dns-84b4cc85c9-r4s54\" (UID: \"f48d81b1-590d-43d8-9111-fdada4e99811\") " pod="openstack/dnsmasq-dns-84b4cc85c9-r4s54" Oct 08 18:37:37 crc kubenswrapper[4859]: I1008 18:37:37.502053 4859 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/f48d81b1-590d-43d8-9111-fdada4e99811-config\") pod \"dnsmasq-dns-84b4cc85c9-r4s54\" (UID: \"f48d81b1-590d-43d8-9111-fdada4e99811\") " pod="openstack/dnsmasq-dns-84b4cc85c9-r4s54" Oct 08 18:37:37 crc kubenswrapper[4859]: I1008 18:37:37.502232 4859 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/configmap/f48d81b1-590d-43d8-9111-fdada4e99811-openstack-edpm-ipam\") pod \"dnsmasq-dns-84b4cc85c9-r4s54\" (UID: \"f48d81b1-590d-43d8-9111-fdada4e99811\") " pod="openstack/dnsmasq-dns-84b4cc85c9-r4s54" Oct 08 18:37:37 crc kubenswrapper[4859]: I1008 18:37:37.502262 4859 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/f48d81b1-590d-43d8-9111-fdada4e99811-ovsdbserver-sb\") pod \"dnsmasq-dns-84b4cc85c9-r4s54\" (UID: \"f48d81b1-590d-43d8-9111-fdada4e99811\") " pod="openstack/dnsmasq-dns-84b4cc85c9-r4s54" Oct 08 18:37:37 crc kubenswrapper[4859]: I1008 18:37:37.502709 4859 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/f48d81b1-590d-43d8-9111-fdada4e99811-ovsdbserver-nb\") pod \"dnsmasq-dns-84b4cc85c9-r4s54\" (UID: \"f48d81b1-590d-43d8-9111-fdada4e99811\") " pod="openstack/dnsmasq-dns-84b4cc85c9-r4s54" Oct 08 18:37:37 crc kubenswrapper[4859]: I1008 18:37:37.517630 4859 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-wfcrj\" (UniqueName: \"kubernetes.io/projected/f48d81b1-590d-43d8-9111-fdada4e99811-kube-api-access-wfcrj\") pod \"dnsmasq-dns-84b4cc85c9-r4s54\" (UID: \"f48d81b1-590d-43d8-9111-fdada4e99811\") " pod="openstack/dnsmasq-dns-84b4cc85c9-r4s54" Oct 08 18:37:37 crc kubenswrapper[4859]: I1008 18:37:37.749876 4859 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-84b4cc85c9-r4s54" Oct 08 18:37:38 crc kubenswrapper[4859]: I1008 18:37:38.234722 4859 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-84b4cc85c9-r4s54"] Oct 08 18:37:38 crc kubenswrapper[4859]: W1008 18:37:38.236298 4859 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podf48d81b1_590d_43d8_9111_fdada4e99811.slice/crio-0656f118bc2cbafa98b2f8318ba2a5da3a9da3b865f7f46f4355b06fba526311 WatchSource:0}: Error finding container 0656f118bc2cbafa98b2f8318ba2a5da3a9da3b865f7f46f4355b06fba526311: Status 404 returned error can't find the container with id 0656f118bc2cbafa98b2f8318ba2a5da3a9da3b865f7f46f4355b06fba526311 Oct 08 18:37:38 crc kubenswrapper[4859]: I1008 18:37:38.259786 4859 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-cell1-server-0" event={"ID":"2c6db854-3378-4bc4-b622-480bc8c42bf5","Type":"ContainerStarted","Data":"9edfb8c6b084281496a254c2c1be0f189106dc4e159c25c1d46153937b814e58"} Oct 08 18:37:38 crc kubenswrapper[4859]: I1008 18:37:38.263104 4859 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-84b4cc85c9-r4s54" event={"ID":"f48d81b1-590d-43d8-9111-fdada4e99811","Type":"ContainerStarted","Data":"0656f118bc2cbafa98b2f8318ba2a5da3a9da3b865f7f46f4355b06fba526311"} Oct 08 18:37:39 crc kubenswrapper[4859]: I1008 18:37:39.283892 4859 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-84b4cc85c9-r4s54" event={"ID":"f48d81b1-590d-43d8-9111-fdada4e99811","Type":"ContainerStarted","Data":"81a7b121948dc249a4250274264a13df49e74fba4bf3ace59f009db37adbc959"} Oct 08 18:37:39 crc kubenswrapper[4859]: I1008 18:37:39.289508 4859 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-cell1-server-0" event={"ID":"2c6db854-3378-4bc4-b622-480bc8c42bf5","Type":"ContainerStarted","Data":"c28efd71f628d332421e9bffaf65c2bc6cd1d1b494914b65831ddf2c8b82c9d2"} Oct 08 18:37:40 crc kubenswrapper[4859]: I1008 18:37:40.301944 4859 generic.go:334] "Generic (PLEG): container finished" podID="f48d81b1-590d-43d8-9111-fdada4e99811" containerID="81a7b121948dc249a4250274264a13df49e74fba4bf3ace59f009db37adbc959" exitCode=0 Oct 08 18:37:40 crc kubenswrapper[4859]: I1008 18:37:40.302016 4859 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-84b4cc85c9-r4s54" event={"ID":"f48d81b1-590d-43d8-9111-fdada4e99811","Type":"ContainerDied","Data":"81a7b121948dc249a4250274264a13df49e74fba4bf3ace59f009db37adbc959"} Oct 08 18:37:40 crc kubenswrapper[4859]: I1008 18:37:40.302430 4859 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-84b4cc85c9-r4s54" Oct 08 18:37:40 crc kubenswrapper[4859]: I1008 18:37:40.302450 4859 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-84b4cc85c9-r4s54" event={"ID":"f48d81b1-590d-43d8-9111-fdada4e99811","Type":"ContainerStarted","Data":"616f99b4770733c03bfa9d191274039b0e88b99543d2d268debeb4c299fba705"} Oct 08 18:37:40 crc kubenswrapper[4859]: I1008 18:37:40.304368 4859 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-server-0" event={"ID":"94450380-649a-4476-b646-018ed24aa703","Type":"ContainerStarted","Data":"0d06c4a1285c87136ae8add16f4fceb929ec3f7c13118a87352fc60e0db74c28"} Oct 08 18:37:40 crc kubenswrapper[4859]: I1008 18:37:40.340079 4859 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-84b4cc85c9-r4s54" podStartSLOduration=3.340059665 podStartE2EDuration="3.340059665s" podCreationTimestamp="2025-10-08 18:37:37 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-08 18:37:40.3293945 +0000 UTC m=+1230.576233889" watchObservedRunningTime="2025-10-08 18:37:40.340059665 +0000 UTC m=+1230.586899044" Oct 08 18:37:47 crc kubenswrapper[4859]: I1008 18:37:47.752921 4859 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-84b4cc85c9-r4s54" Oct 08 18:37:47 crc kubenswrapper[4859]: I1008 18:37:47.814161 4859 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-64986d45b9-sv5xb"] Oct 08 18:37:47 crc kubenswrapper[4859]: I1008 18:37:47.814413 4859 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-64986d45b9-sv5xb" podUID="93af9fa0-cdaa-4665-9120-870f91139ae6" containerName="dnsmasq-dns" containerID="cri-o://2f3570ff9aac41b4211ea2b76be831b98d152e30595b7dface9313ee3be68aed" gracePeriod=10 Oct 08 18:37:47 crc kubenswrapper[4859]: I1008 18:37:47.925133 4859 patch_prober.go:28] interesting pod/machine-config-daemon-82s52 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 08 18:37:47 crc kubenswrapper[4859]: I1008 18:37:47.925394 4859 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-82s52" podUID="b23a6a6c-9d92-4e7b-840e-55cfda873a2d" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 08 18:37:47 crc kubenswrapper[4859]: I1008 18:37:47.927518 4859 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-6f586dc697-ncnrs"] Oct 08 18:37:47 crc kubenswrapper[4859]: I1008 18:37:47.929143 4859 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-6f586dc697-ncnrs" Oct 08 18:37:47 crc kubenswrapper[4859]: I1008 18:37:47.947940 4859 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-6f586dc697-ncnrs"] Oct 08 18:37:48 crc kubenswrapper[4859]: I1008 18:37:48.018815 4859 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-l4wh8\" (UniqueName: \"kubernetes.io/projected/7f4d68a5-ccfe-48f2-a84f-e800d6718498-kube-api-access-l4wh8\") pod \"dnsmasq-dns-6f586dc697-ncnrs\" (UID: \"7f4d68a5-ccfe-48f2-a84f-e800d6718498\") " pod="openstack/dnsmasq-dns-6f586dc697-ncnrs" Oct 08 18:37:48 crc kubenswrapper[4859]: I1008 18:37:48.018896 4859 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7f4d68a5-ccfe-48f2-a84f-e800d6718498-config\") pod \"dnsmasq-dns-6f586dc697-ncnrs\" (UID: \"7f4d68a5-ccfe-48f2-a84f-e800d6718498\") " pod="openstack/dnsmasq-dns-6f586dc697-ncnrs" Oct 08 18:37:48 crc kubenswrapper[4859]: I1008 18:37:48.018970 4859 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/7f4d68a5-ccfe-48f2-a84f-e800d6718498-dns-swift-storage-0\") pod \"dnsmasq-dns-6f586dc697-ncnrs\" (UID: \"7f4d68a5-ccfe-48f2-a84f-e800d6718498\") " pod="openstack/dnsmasq-dns-6f586dc697-ncnrs" Oct 08 18:37:48 crc kubenswrapper[4859]: I1008 18:37:48.018998 4859 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/7f4d68a5-ccfe-48f2-a84f-e800d6718498-ovsdbserver-sb\") pod \"dnsmasq-dns-6f586dc697-ncnrs\" (UID: \"7f4d68a5-ccfe-48f2-a84f-e800d6718498\") " pod="openstack/dnsmasq-dns-6f586dc697-ncnrs" Oct 08 18:37:48 crc kubenswrapper[4859]: I1008 18:37:48.019019 4859 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/7f4d68a5-ccfe-48f2-a84f-e800d6718498-ovsdbserver-nb\") pod \"dnsmasq-dns-6f586dc697-ncnrs\" (UID: \"7f4d68a5-ccfe-48f2-a84f-e800d6718498\") " pod="openstack/dnsmasq-dns-6f586dc697-ncnrs" Oct 08 18:37:48 crc kubenswrapper[4859]: I1008 18:37:48.019038 4859 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/7f4d68a5-ccfe-48f2-a84f-e800d6718498-dns-svc\") pod \"dnsmasq-dns-6f586dc697-ncnrs\" (UID: \"7f4d68a5-ccfe-48f2-a84f-e800d6718498\") " pod="openstack/dnsmasq-dns-6f586dc697-ncnrs" Oct 08 18:37:48 crc kubenswrapper[4859]: I1008 18:37:48.019073 4859 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/configmap/7f4d68a5-ccfe-48f2-a84f-e800d6718498-openstack-edpm-ipam\") pod \"dnsmasq-dns-6f586dc697-ncnrs\" (UID: \"7f4d68a5-ccfe-48f2-a84f-e800d6718498\") " pod="openstack/dnsmasq-dns-6f586dc697-ncnrs" Oct 08 18:37:48 crc kubenswrapper[4859]: I1008 18:37:48.120802 4859 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/configmap/7f4d68a5-ccfe-48f2-a84f-e800d6718498-openstack-edpm-ipam\") pod \"dnsmasq-dns-6f586dc697-ncnrs\" (UID: \"7f4d68a5-ccfe-48f2-a84f-e800d6718498\") " pod="openstack/dnsmasq-dns-6f586dc697-ncnrs" Oct 08 18:37:48 crc kubenswrapper[4859]: I1008 18:37:48.121782 4859 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-l4wh8\" (UniqueName: \"kubernetes.io/projected/7f4d68a5-ccfe-48f2-a84f-e800d6718498-kube-api-access-l4wh8\") pod \"dnsmasq-dns-6f586dc697-ncnrs\" (UID: \"7f4d68a5-ccfe-48f2-a84f-e800d6718498\") " pod="openstack/dnsmasq-dns-6f586dc697-ncnrs" Oct 08 18:37:48 crc kubenswrapper[4859]: I1008 18:37:48.121794 4859 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/configmap/7f4d68a5-ccfe-48f2-a84f-e800d6718498-openstack-edpm-ipam\") pod \"dnsmasq-dns-6f586dc697-ncnrs\" (UID: \"7f4d68a5-ccfe-48f2-a84f-e800d6718498\") " pod="openstack/dnsmasq-dns-6f586dc697-ncnrs" Oct 08 18:37:48 crc kubenswrapper[4859]: I1008 18:37:48.121879 4859 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7f4d68a5-ccfe-48f2-a84f-e800d6718498-config\") pod \"dnsmasq-dns-6f586dc697-ncnrs\" (UID: \"7f4d68a5-ccfe-48f2-a84f-e800d6718498\") " pod="openstack/dnsmasq-dns-6f586dc697-ncnrs" Oct 08 18:37:48 crc kubenswrapper[4859]: I1008 18:37:48.121992 4859 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/7f4d68a5-ccfe-48f2-a84f-e800d6718498-dns-swift-storage-0\") pod \"dnsmasq-dns-6f586dc697-ncnrs\" (UID: \"7f4d68a5-ccfe-48f2-a84f-e800d6718498\") " pod="openstack/dnsmasq-dns-6f586dc697-ncnrs" Oct 08 18:37:48 crc kubenswrapper[4859]: I1008 18:37:48.122035 4859 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/7f4d68a5-ccfe-48f2-a84f-e800d6718498-ovsdbserver-sb\") pod \"dnsmasq-dns-6f586dc697-ncnrs\" (UID: \"7f4d68a5-ccfe-48f2-a84f-e800d6718498\") " pod="openstack/dnsmasq-dns-6f586dc697-ncnrs" Oct 08 18:37:48 crc kubenswrapper[4859]: I1008 18:37:48.122076 4859 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/7f4d68a5-ccfe-48f2-a84f-e800d6718498-ovsdbserver-nb\") pod \"dnsmasq-dns-6f586dc697-ncnrs\" (UID: \"7f4d68a5-ccfe-48f2-a84f-e800d6718498\") " pod="openstack/dnsmasq-dns-6f586dc697-ncnrs" Oct 08 18:37:48 crc kubenswrapper[4859]: I1008 18:37:48.122101 4859 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/7f4d68a5-ccfe-48f2-a84f-e800d6718498-dns-svc\") pod \"dnsmasq-dns-6f586dc697-ncnrs\" (UID: \"7f4d68a5-ccfe-48f2-a84f-e800d6718498\") " pod="openstack/dnsmasq-dns-6f586dc697-ncnrs" Oct 08 18:37:48 crc kubenswrapper[4859]: I1008 18:37:48.122731 4859 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7f4d68a5-ccfe-48f2-a84f-e800d6718498-config\") pod \"dnsmasq-dns-6f586dc697-ncnrs\" (UID: \"7f4d68a5-ccfe-48f2-a84f-e800d6718498\") " pod="openstack/dnsmasq-dns-6f586dc697-ncnrs" Oct 08 18:37:48 crc kubenswrapper[4859]: I1008 18:37:48.123350 4859 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/7f4d68a5-ccfe-48f2-a84f-e800d6718498-ovsdbserver-sb\") pod \"dnsmasq-dns-6f586dc697-ncnrs\" (UID: \"7f4d68a5-ccfe-48f2-a84f-e800d6718498\") " pod="openstack/dnsmasq-dns-6f586dc697-ncnrs" Oct 08 18:37:48 crc kubenswrapper[4859]: I1008 18:37:48.123575 4859 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/7f4d68a5-ccfe-48f2-a84f-e800d6718498-dns-svc\") pod \"dnsmasq-dns-6f586dc697-ncnrs\" (UID: \"7f4d68a5-ccfe-48f2-a84f-e800d6718498\") " pod="openstack/dnsmasq-dns-6f586dc697-ncnrs" Oct 08 18:37:48 crc kubenswrapper[4859]: I1008 18:37:48.124045 4859 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/7f4d68a5-ccfe-48f2-a84f-e800d6718498-dns-swift-storage-0\") pod \"dnsmasq-dns-6f586dc697-ncnrs\" (UID: \"7f4d68a5-ccfe-48f2-a84f-e800d6718498\") " pod="openstack/dnsmasq-dns-6f586dc697-ncnrs" Oct 08 18:37:48 crc kubenswrapper[4859]: I1008 18:37:48.124276 4859 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/7f4d68a5-ccfe-48f2-a84f-e800d6718498-ovsdbserver-nb\") pod \"dnsmasq-dns-6f586dc697-ncnrs\" (UID: \"7f4d68a5-ccfe-48f2-a84f-e800d6718498\") " pod="openstack/dnsmasq-dns-6f586dc697-ncnrs" Oct 08 18:37:48 crc kubenswrapper[4859]: I1008 18:37:48.150698 4859 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-l4wh8\" (UniqueName: \"kubernetes.io/projected/7f4d68a5-ccfe-48f2-a84f-e800d6718498-kube-api-access-l4wh8\") pod \"dnsmasq-dns-6f586dc697-ncnrs\" (UID: \"7f4d68a5-ccfe-48f2-a84f-e800d6718498\") " pod="openstack/dnsmasq-dns-6f586dc697-ncnrs" Oct 08 18:37:48 crc kubenswrapper[4859]: I1008 18:37:48.297220 4859 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-6f586dc697-ncnrs" Oct 08 18:37:48 crc kubenswrapper[4859]: I1008 18:37:48.381807 4859 generic.go:334] "Generic (PLEG): container finished" podID="93af9fa0-cdaa-4665-9120-870f91139ae6" containerID="2f3570ff9aac41b4211ea2b76be831b98d152e30595b7dface9313ee3be68aed" exitCode=0 Oct 08 18:37:48 crc kubenswrapper[4859]: I1008 18:37:48.381846 4859 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-64986d45b9-sv5xb" event={"ID":"93af9fa0-cdaa-4665-9120-870f91139ae6","Type":"ContainerDied","Data":"2f3570ff9aac41b4211ea2b76be831b98d152e30595b7dface9313ee3be68aed"} Oct 08 18:37:48 crc kubenswrapper[4859]: I1008 18:37:48.381874 4859 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-64986d45b9-sv5xb" event={"ID":"93af9fa0-cdaa-4665-9120-870f91139ae6","Type":"ContainerDied","Data":"5012775aaa2f789a4fee328b2d9dbbed704068459b990b56c1bba3e39026678b"} Oct 08 18:37:48 crc kubenswrapper[4859]: I1008 18:37:48.381884 4859 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="5012775aaa2f789a4fee328b2d9dbbed704068459b990b56c1bba3e39026678b" Oct 08 18:37:48 crc kubenswrapper[4859]: I1008 18:37:48.404796 4859 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-64986d45b9-sv5xb" Oct 08 18:37:48 crc kubenswrapper[4859]: I1008 18:37:48.544354 4859 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/93af9fa0-cdaa-4665-9120-870f91139ae6-dns-swift-storage-0\") pod \"93af9fa0-cdaa-4665-9120-870f91139ae6\" (UID: \"93af9fa0-cdaa-4665-9120-870f91139ae6\") " Oct 08 18:37:48 crc kubenswrapper[4859]: I1008 18:37:48.544471 4859 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-x4hs7\" (UniqueName: \"kubernetes.io/projected/93af9fa0-cdaa-4665-9120-870f91139ae6-kube-api-access-x4hs7\") pod \"93af9fa0-cdaa-4665-9120-870f91139ae6\" (UID: \"93af9fa0-cdaa-4665-9120-870f91139ae6\") " Oct 08 18:37:48 crc kubenswrapper[4859]: I1008 18:37:48.544629 4859 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/93af9fa0-cdaa-4665-9120-870f91139ae6-ovsdbserver-sb\") pod \"93af9fa0-cdaa-4665-9120-870f91139ae6\" (UID: \"93af9fa0-cdaa-4665-9120-870f91139ae6\") " Oct 08 18:37:48 crc kubenswrapper[4859]: I1008 18:37:48.544743 4859 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/93af9fa0-cdaa-4665-9120-870f91139ae6-config\") pod \"93af9fa0-cdaa-4665-9120-870f91139ae6\" (UID: \"93af9fa0-cdaa-4665-9120-870f91139ae6\") " Oct 08 18:37:48 crc kubenswrapper[4859]: I1008 18:37:48.544774 4859 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/93af9fa0-cdaa-4665-9120-870f91139ae6-ovsdbserver-nb\") pod \"93af9fa0-cdaa-4665-9120-870f91139ae6\" (UID: \"93af9fa0-cdaa-4665-9120-870f91139ae6\") " Oct 08 18:37:48 crc kubenswrapper[4859]: I1008 18:37:48.544838 4859 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/93af9fa0-cdaa-4665-9120-870f91139ae6-dns-svc\") pod \"93af9fa0-cdaa-4665-9120-870f91139ae6\" (UID: \"93af9fa0-cdaa-4665-9120-870f91139ae6\") " Oct 08 18:37:48 crc kubenswrapper[4859]: I1008 18:37:48.580924 4859 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/93af9fa0-cdaa-4665-9120-870f91139ae6-kube-api-access-x4hs7" (OuterVolumeSpecName: "kube-api-access-x4hs7") pod "93af9fa0-cdaa-4665-9120-870f91139ae6" (UID: "93af9fa0-cdaa-4665-9120-870f91139ae6"). InnerVolumeSpecName "kube-api-access-x4hs7". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 08 18:37:48 crc kubenswrapper[4859]: I1008 18:37:48.630500 4859 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/93af9fa0-cdaa-4665-9120-870f91139ae6-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "93af9fa0-cdaa-4665-9120-870f91139ae6" (UID: "93af9fa0-cdaa-4665-9120-870f91139ae6"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 08 18:37:48 crc kubenswrapper[4859]: I1008 18:37:48.647701 4859 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-x4hs7\" (UniqueName: \"kubernetes.io/projected/93af9fa0-cdaa-4665-9120-870f91139ae6-kube-api-access-x4hs7\") on node \"crc\" DevicePath \"\"" Oct 08 18:37:48 crc kubenswrapper[4859]: I1008 18:37:48.647736 4859 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/93af9fa0-cdaa-4665-9120-870f91139ae6-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Oct 08 18:37:48 crc kubenswrapper[4859]: I1008 18:37:48.687371 4859 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/93af9fa0-cdaa-4665-9120-870f91139ae6-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "93af9fa0-cdaa-4665-9120-870f91139ae6" (UID: "93af9fa0-cdaa-4665-9120-870f91139ae6"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 08 18:37:48 crc kubenswrapper[4859]: I1008 18:37:48.696352 4859 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-6f586dc697-ncnrs"] Oct 08 18:37:48 crc kubenswrapper[4859]: I1008 18:37:48.698963 4859 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/93af9fa0-cdaa-4665-9120-870f91139ae6-dns-swift-storage-0" (OuterVolumeSpecName: "dns-swift-storage-0") pod "93af9fa0-cdaa-4665-9120-870f91139ae6" (UID: "93af9fa0-cdaa-4665-9120-870f91139ae6"). InnerVolumeSpecName "dns-swift-storage-0". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 08 18:37:48 crc kubenswrapper[4859]: I1008 18:37:48.699284 4859 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/93af9fa0-cdaa-4665-9120-870f91139ae6-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "93af9fa0-cdaa-4665-9120-870f91139ae6" (UID: "93af9fa0-cdaa-4665-9120-870f91139ae6"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 08 18:37:48 crc kubenswrapper[4859]: W1008 18:37:48.709027 4859 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod7f4d68a5_ccfe_48f2_a84f_e800d6718498.slice/crio-5eb6c8408b55b5ccd729000eb21e4a46797f2d701a93240687edff1ca1b0e649 WatchSource:0}: Error finding container 5eb6c8408b55b5ccd729000eb21e4a46797f2d701a93240687edff1ca1b0e649: Status 404 returned error can't find the container with id 5eb6c8408b55b5ccd729000eb21e4a46797f2d701a93240687edff1ca1b0e649 Oct 08 18:37:48 crc kubenswrapper[4859]: I1008 18:37:48.709917 4859 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/93af9fa0-cdaa-4665-9120-870f91139ae6-config" (OuterVolumeSpecName: "config") pod "93af9fa0-cdaa-4665-9120-870f91139ae6" (UID: "93af9fa0-cdaa-4665-9120-870f91139ae6"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 08 18:37:48 crc kubenswrapper[4859]: I1008 18:37:48.749707 4859 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/93af9fa0-cdaa-4665-9120-870f91139ae6-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Oct 08 18:37:48 crc kubenswrapper[4859]: I1008 18:37:48.749742 4859 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/93af9fa0-cdaa-4665-9120-870f91139ae6-config\") on node \"crc\" DevicePath \"\"" Oct 08 18:37:48 crc kubenswrapper[4859]: I1008 18:37:48.749755 4859 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/93af9fa0-cdaa-4665-9120-870f91139ae6-dns-svc\") on node \"crc\" DevicePath \"\"" Oct 08 18:37:48 crc kubenswrapper[4859]: I1008 18:37:48.749767 4859 reconciler_common.go:293] "Volume detached for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/93af9fa0-cdaa-4665-9120-870f91139ae6-dns-swift-storage-0\") on node \"crc\" DevicePath \"\"" Oct 08 18:37:49 crc kubenswrapper[4859]: I1008 18:37:49.394518 4859 generic.go:334] "Generic (PLEG): container finished" podID="7f4d68a5-ccfe-48f2-a84f-e800d6718498" containerID="309f16953fdd30588911a8954e204be09f634d7e3abbc6b5df62ab261d1e482c" exitCode=0 Oct 08 18:37:49 crc kubenswrapper[4859]: I1008 18:37:49.394670 4859 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-6f586dc697-ncnrs" event={"ID":"7f4d68a5-ccfe-48f2-a84f-e800d6718498","Type":"ContainerDied","Data":"309f16953fdd30588911a8954e204be09f634d7e3abbc6b5df62ab261d1e482c"} Oct 08 18:37:49 crc kubenswrapper[4859]: I1008 18:37:49.394757 4859 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-6f586dc697-ncnrs" event={"ID":"7f4d68a5-ccfe-48f2-a84f-e800d6718498","Type":"ContainerStarted","Data":"5eb6c8408b55b5ccd729000eb21e4a46797f2d701a93240687edff1ca1b0e649"} Oct 08 18:37:49 crc kubenswrapper[4859]: I1008 18:37:49.394800 4859 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-64986d45b9-sv5xb" Oct 08 18:37:49 crc kubenswrapper[4859]: I1008 18:37:49.608940 4859 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-64986d45b9-sv5xb"] Oct 08 18:37:49 crc kubenswrapper[4859]: I1008 18:37:49.619561 4859 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-64986d45b9-sv5xb"] Oct 08 18:37:50 crc kubenswrapper[4859]: I1008 18:37:50.406285 4859 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-6f586dc697-ncnrs" event={"ID":"7f4d68a5-ccfe-48f2-a84f-e800d6718498","Type":"ContainerStarted","Data":"79104318d4769dfc436226a82be451f92b72e7d510f1051bbecf8cb0d7d0b7cb"} Oct 08 18:37:50 crc kubenswrapper[4859]: I1008 18:37:50.406594 4859 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-6f586dc697-ncnrs" Oct 08 18:37:50 crc kubenswrapper[4859]: I1008 18:37:50.431710 4859 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-6f586dc697-ncnrs" podStartSLOduration=3.43167885 podStartE2EDuration="3.43167885s" podCreationTimestamp="2025-10-08 18:37:47 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-08 18:37:50.426584504 +0000 UTC m=+1240.673423893" watchObservedRunningTime="2025-10-08 18:37:50.43167885 +0000 UTC m=+1240.678518229" Oct 08 18:37:50 crc kubenswrapper[4859]: I1008 18:37:50.482408 4859 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="93af9fa0-cdaa-4665-9120-870f91139ae6" path="/var/lib/kubelet/pods/93af9fa0-cdaa-4665-9120-870f91139ae6/volumes" Oct 08 18:37:58 crc kubenswrapper[4859]: I1008 18:37:58.299581 4859 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-6f586dc697-ncnrs" Oct 08 18:37:58 crc kubenswrapper[4859]: I1008 18:37:58.371323 4859 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-84b4cc85c9-r4s54"] Oct 08 18:37:58 crc kubenswrapper[4859]: I1008 18:37:58.371970 4859 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-84b4cc85c9-r4s54" podUID="f48d81b1-590d-43d8-9111-fdada4e99811" containerName="dnsmasq-dns" containerID="cri-o://616f99b4770733c03bfa9d191274039b0e88b99543d2d268debeb4c299fba705" gracePeriod=10 Oct 08 18:37:58 crc kubenswrapper[4859]: I1008 18:37:58.840563 4859 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-84b4cc85c9-r4s54" Oct 08 18:37:58 crc kubenswrapper[4859]: I1008 18:37:58.970367 4859 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/f48d81b1-590d-43d8-9111-fdada4e99811-dns-swift-storage-0\") pod \"f48d81b1-590d-43d8-9111-fdada4e99811\" (UID: \"f48d81b1-590d-43d8-9111-fdada4e99811\") " Oct 08 18:37:58 crc kubenswrapper[4859]: I1008 18:37:58.970447 4859 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/f48d81b1-590d-43d8-9111-fdada4e99811-ovsdbserver-sb\") pod \"f48d81b1-590d-43d8-9111-fdada4e99811\" (UID: \"f48d81b1-590d-43d8-9111-fdada4e99811\") " Oct 08 18:37:58 crc kubenswrapper[4859]: I1008 18:37:58.970476 4859 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-wfcrj\" (UniqueName: \"kubernetes.io/projected/f48d81b1-590d-43d8-9111-fdada4e99811-kube-api-access-wfcrj\") pod \"f48d81b1-590d-43d8-9111-fdada4e99811\" (UID: \"f48d81b1-590d-43d8-9111-fdada4e99811\") " Oct 08 18:37:58 crc kubenswrapper[4859]: I1008 18:37:58.970658 4859 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/f48d81b1-590d-43d8-9111-fdada4e99811-ovsdbserver-nb\") pod \"f48d81b1-590d-43d8-9111-fdada4e99811\" (UID: \"f48d81b1-590d-43d8-9111-fdada4e99811\") " Oct 08 18:37:58 crc kubenswrapper[4859]: I1008 18:37:58.971214 4859 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/f48d81b1-590d-43d8-9111-fdada4e99811-dns-svc\") pod \"f48d81b1-590d-43d8-9111-fdada4e99811\" (UID: \"f48d81b1-590d-43d8-9111-fdada4e99811\") " Oct 08 18:37:58 crc kubenswrapper[4859]: I1008 18:37:58.971591 4859 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/f48d81b1-590d-43d8-9111-fdada4e99811-config\") pod \"f48d81b1-590d-43d8-9111-fdada4e99811\" (UID: \"f48d81b1-590d-43d8-9111-fdada4e99811\") " Oct 08 18:37:58 crc kubenswrapper[4859]: I1008 18:37:58.971660 4859 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/configmap/f48d81b1-590d-43d8-9111-fdada4e99811-openstack-edpm-ipam\") pod \"f48d81b1-590d-43d8-9111-fdada4e99811\" (UID: \"f48d81b1-590d-43d8-9111-fdada4e99811\") " Oct 08 18:37:58 crc kubenswrapper[4859]: I1008 18:37:58.982454 4859 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/f48d81b1-590d-43d8-9111-fdada4e99811-kube-api-access-wfcrj" (OuterVolumeSpecName: "kube-api-access-wfcrj") pod "f48d81b1-590d-43d8-9111-fdada4e99811" (UID: "f48d81b1-590d-43d8-9111-fdada4e99811"). InnerVolumeSpecName "kube-api-access-wfcrj". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 08 18:37:59 crc kubenswrapper[4859]: I1008 18:37:59.022875 4859 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/f48d81b1-590d-43d8-9111-fdada4e99811-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "f48d81b1-590d-43d8-9111-fdada4e99811" (UID: "f48d81b1-590d-43d8-9111-fdada4e99811"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 08 18:37:59 crc kubenswrapper[4859]: I1008 18:37:59.023380 4859 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/f48d81b1-590d-43d8-9111-fdada4e99811-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "f48d81b1-590d-43d8-9111-fdada4e99811" (UID: "f48d81b1-590d-43d8-9111-fdada4e99811"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 08 18:37:59 crc kubenswrapper[4859]: I1008 18:37:59.027135 4859 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/f48d81b1-590d-43d8-9111-fdada4e99811-openstack-edpm-ipam" (OuterVolumeSpecName: "openstack-edpm-ipam") pod "f48d81b1-590d-43d8-9111-fdada4e99811" (UID: "f48d81b1-590d-43d8-9111-fdada4e99811"). InnerVolumeSpecName "openstack-edpm-ipam". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 08 18:37:59 crc kubenswrapper[4859]: I1008 18:37:59.029011 4859 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/f48d81b1-590d-43d8-9111-fdada4e99811-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "f48d81b1-590d-43d8-9111-fdada4e99811" (UID: "f48d81b1-590d-43d8-9111-fdada4e99811"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 08 18:37:59 crc kubenswrapper[4859]: I1008 18:37:59.029047 4859 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/f48d81b1-590d-43d8-9111-fdada4e99811-config" (OuterVolumeSpecName: "config") pod "f48d81b1-590d-43d8-9111-fdada4e99811" (UID: "f48d81b1-590d-43d8-9111-fdada4e99811"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 08 18:37:59 crc kubenswrapper[4859]: I1008 18:37:59.032229 4859 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/f48d81b1-590d-43d8-9111-fdada4e99811-dns-swift-storage-0" (OuterVolumeSpecName: "dns-swift-storage-0") pod "f48d81b1-590d-43d8-9111-fdada4e99811" (UID: "f48d81b1-590d-43d8-9111-fdada4e99811"). InnerVolumeSpecName "dns-swift-storage-0". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 08 18:37:59 crc kubenswrapper[4859]: I1008 18:37:59.074261 4859 reconciler_common.go:293] "Volume detached for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/f48d81b1-590d-43d8-9111-fdada4e99811-dns-swift-storage-0\") on node \"crc\" DevicePath \"\"" Oct 08 18:37:59 crc kubenswrapper[4859]: I1008 18:37:59.074301 4859 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/f48d81b1-590d-43d8-9111-fdada4e99811-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Oct 08 18:37:59 crc kubenswrapper[4859]: I1008 18:37:59.074310 4859 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-wfcrj\" (UniqueName: \"kubernetes.io/projected/f48d81b1-590d-43d8-9111-fdada4e99811-kube-api-access-wfcrj\") on node \"crc\" DevicePath \"\"" Oct 08 18:37:59 crc kubenswrapper[4859]: I1008 18:37:59.074320 4859 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/f48d81b1-590d-43d8-9111-fdada4e99811-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Oct 08 18:37:59 crc kubenswrapper[4859]: I1008 18:37:59.074329 4859 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/f48d81b1-590d-43d8-9111-fdada4e99811-dns-svc\") on node \"crc\" DevicePath \"\"" Oct 08 18:37:59 crc kubenswrapper[4859]: I1008 18:37:59.074338 4859 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/f48d81b1-590d-43d8-9111-fdada4e99811-config\") on node \"crc\" DevicePath \"\"" Oct 08 18:37:59 crc kubenswrapper[4859]: I1008 18:37:59.074346 4859 reconciler_common.go:293] "Volume detached for volume \"openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/configmap/f48d81b1-590d-43d8-9111-fdada4e99811-openstack-edpm-ipam\") on node \"crc\" DevicePath \"\"" Oct 08 18:37:59 crc kubenswrapper[4859]: I1008 18:37:59.506115 4859 generic.go:334] "Generic (PLEG): container finished" podID="f48d81b1-590d-43d8-9111-fdada4e99811" containerID="616f99b4770733c03bfa9d191274039b0e88b99543d2d268debeb4c299fba705" exitCode=0 Oct 08 18:37:59 crc kubenswrapper[4859]: I1008 18:37:59.506168 4859 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-84b4cc85c9-r4s54" event={"ID":"f48d81b1-590d-43d8-9111-fdada4e99811","Type":"ContainerDied","Data":"616f99b4770733c03bfa9d191274039b0e88b99543d2d268debeb4c299fba705"} Oct 08 18:37:59 crc kubenswrapper[4859]: I1008 18:37:59.506206 4859 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-84b4cc85c9-r4s54" event={"ID":"f48d81b1-590d-43d8-9111-fdada4e99811","Type":"ContainerDied","Data":"0656f118bc2cbafa98b2f8318ba2a5da3a9da3b865f7f46f4355b06fba526311"} Oct 08 18:37:59 crc kubenswrapper[4859]: I1008 18:37:59.506206 4859 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-84b4cc85c9-r4s54" Oct 08 18:37:59 crc kubenswrapper[4859]: I1008 18:37:59.506231 4859 scope.go:117] "RemoveContainer" containerID="616f99b4770733c03bfa9d191274039b0e88b99543d2d268debeb4c299fba705" Oct 08 18:37:59 crc kubenswrapper[4859]: I1008 18:37:59.544583 4859 scope.go:117] "RemoveContainer" containerID="81a7b121948dc249a4250274264a13df49e74fba4bf3ace59f009db37adbc959" Oct 08 18:37:59 crc kubenswrapper[4859]: I1008 18:37:59.559515 4859 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-84b4cc85c9-r4s54"] Oct 08 18:37:59 crc kubenswrapper[4859]: I1008 18:37:59.572615 4859 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-84b4cc85c9-r4s54"] Oct 08 18:37:59 crc kubenswrapper[4859]: I1008 18:37:59.594599 4859 scope.go:117] "RemoveContainer" containerID="616f99b4770733c03bfa9d191274039b0e88b99543d2d268debeb4c299fba705" Oct 08 18:37:59 crc kubenswrapper[4859]: E1008 18:37:59.594998 4859 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"616f99b4770733c03bfa9d191274039b0e88b99543d2d268debeb4c299fba705\": container with ID starting with 616f99b4770733c03bfa9d191274039b0e88b99543d2d268debeb4c299fba705 not found: ID does not exist" containerID="616f99b4770733c03bfa9d191274039b0e88b99543d2d268debeb4c299fba705" Oct 08 18:37:59 crc kubenswrapper[4859]: I1008 18:37:59.595048 4859 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"616f99b4770733c03bfa9d191274039b0e88b99543d2d268debeb4c299fba705"} err="failed to get container status \"616f99b4770733c03bfa9d191274039b0e88b99543d2d268debeb4c299fba705\": rpc error: code = NotFound desc = could not find container \"616f99b4770733c03bfa9d191274039b0e88b99543d2d268debeb4c299fba705\": container with ID starting with 616f99b4770733c03bfa9d191274039b0e88b99543d2d268debeb4c299fba705 not found: ID does not exist" Oct 08 18:37:59 crc kubenswrapper[4859]: I1008 18:37:59.595072 4859 scope.go:117] "RemoveContainer" containerID="81a7b121948dc249a4250274264a13df49e74fba4bf3ace59f009db37adbc959" Oct 08 18:37:59 crc kubenswrapper[4859]: E1008 18:37:59.595300 4859 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"81a7b121948dc249a4250274264a13df49e74fba4bf3ace59f009db37adbc959\": container with ID starting with 81a7b121948dc249a4250274264a13df49e74fba4bf3ace59f009db37adbc959 not found: ID does not exist" containerID="81a7b121948dc249a4250274264a13df49e74fba4bf3ace59f009db37adbc959" Oct 08 18:37:59 crc kubenswrapper[4859]: I1008 18:37:59.595322 4859 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"81a7b121948dc249a4250274264a13df49e74fba4bf3ace59f009db37adbc959"} err="failed to get container status \"81a7b121948dc249a4250274264a13df49e74fba4bf3ace59f009db37adbc959\": rpc error: code = NotFound desc = could not find container \"81a7b121948dc249a4250274264a13df49e74fba4bf3ace59f009db37adbc959\": container with ID starting with 81a7b121948dc249a4250274264a13df49e74fba4bf3ace59f009db37adbc959 not found: ID does not exist" Oct 08 18:38:00 crc kubenswrapper[4859]: I1008 18:38:00.479639 4859 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="f48d81b1-590d-43d8-9111-fdada4e99811" path="/var/lib/kubelet/pods/f48d81b1-590d-43d8-9111-fdada4e99811/volumes" Oct 08 18:38:11 crc kubenswrapper[4859]: I1008 18:38:11.328188 4859 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-n8rrq"] Oct 08 18:38:11 crc kubenswrapper[4859]: E1008 18:38:11.328846 4859 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="93af9fa0-cdaa-4665-9120-870f91139ae6" containerName="init" Oct 08 18:38:11 crc kubenswrapper[4859]: I1008 18:38:11.328859 4859 state_mem.go:107] "Deleted CPUSet assignment" podUID="93af9fa0-cdaa-4665-9120-870f91139ae6" containerName="init" Oct 08 18:38:11 crc kubenswrapper[4859]: E1008 18:38:11.328872 4859 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="93af9fa0-cdaa-4665-9120-870f91139ae6" containerName="dnsmasq-dns" Oct 08 18:38:11 crc kubenswrapper[4859]: I1008 18:38:11.328878 4859 state_mem.go:107] "Deleted CPUSet assignment" podUID="93af9fa0-cdaa-4665-9120-870f91139ae6" containerName="dnsmasq-dns" Oct 08 18:38:11 crc kubenswrapper[4859]: E1008 18:38:11.328900 4859 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f48d81b1-590d-43d8-9111-fdada4e99811" containerName="dnsmasq-dns" Oct 08 18:38:11 crc kubenswrapper[4859]: I1008 18:38:11.328906 4859 state_mem.go:107] "Deleted CPUSet assignment" podUID="f48d81b1-590d-43d8-9111-fdada4e99811" containerName="dnsmasq-dns" Oct 08 18:38:11 crc kubenswrapper[4859]: E1008 18:38:11.328918 4859 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f48d81b1-590d-43d8-9111-fdada4e99811" containerName="init" Oct 08 18:38:11 crc kubenswrapper[4859]: I1008 18:38:11.328924 4859 state_mem.go:107] "Deleted CPUSet assignment" podUID="f48d81b1-590d-43d8-9111-fdada4e99811" containerName="init" Oct 08 18:38:11 crc kubenswrapper[4859]: I1008 18:38:11.329106 4859 memory_manager.go:354] "RemoveStaleState removing state" podUID="f48d81b1-590d-43d8-9111-fdada4e99811" containerName="dnsmasq-dns" Oct 08 18:38:11 crc kubenswrapper[4859]: I1008 18:38:11.329117 4859 memory_manager.go:354] "RemoveStaleState removing state" podUID="93af9fa0-cdaa-4665-9120-870f91139ae6" containerName="dnsmasq-dns" Oct 08 18:38:11 crc kubenswrapper[4859]: I1008 18:38:11.329934 4859 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-n8rrq" Oct 08 18:38:11 crc kubenswrapper[4859]: I1008 18:38:11.332386 4859 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Oct 08 18:38:11 crc kubenswrapper[4859]: I1008 18:38:11.332783 4859 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Oct 08 18:38:11 crc kubenswrapper[4859]: I1008 18:38:11.333124 4859 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Oct 08 18:38:11 crc kubenswrapper[4859]: I1008 18:38:11.333789 4859 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-gxqst" Oct 08 18:38:11 crc kubenswrapper[4859]: I1008 18:38:11.358412 4859 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-n8rrq"] Oct 08 18:38:11 crc kubenswrapper[4859]: I1008 18:38:11.521842 4859 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/53d801a5-a432-44f3-970a-23519033d1e3-ssh-key\") pod \"repo-setup-edpm-deployment-openstack-edpm-ipam-n8rrq\" (UID: \"53d801a5-a432-44f3-970a-23519033d1e3\") " pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-n8rrq" Oct 08 18:38:11 crc kubenswrapper[4859]: I1008 18:38:11.521923 4859 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/53d801a5-a432-44f3-970a-23519033d1e3-inventory\") pod \"repo-setup-edpm-deployment-openstack-edpm-ipam-n8rrq\" (UID: \"53d801a5-a432-44f3-970a-23519033d1e3\") " pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-n8rrq" Oct 08 18:38:11 crc kubenswrapper[4859]: I1008 18:38:11.522105 4859 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"repo-setup-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/53d801a5-a432-44f3-970a-23519033d1e3-repo-setup-combined-ca-bundle\") pod \"repo-setup-edpm-deployment-openstack-edpm-ipam-n8rrq\" (UID: \"53d801a5-a432-44f3-970a-23519033d1e3\") " pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-n8rrq" Oct 08 18:38:11 crc kubenswrapper[4859]: I1008 18:38:11.522240 4859 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-hjmnp\" (UniqueName: \"kubernetes.io/projected/53d801a5-a432-44f3-970a-23519033d1e3-kube-api-access-hjmnp\") pod \"repo-setup-edpm-deployment-openstack-edpm-ipam-n8rrq\" (UID: \"53d801a5-a432-44f3-970a-23519033d1e3\") " pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-n8rrq" Oct 08 18:38:11 crc kubenswrapper[4859]: I1008 18:38:11.624518 4859 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/53d801a5-a432-44f3-970a-23519033d1e3-inventory\") pod \"repo-setup-edpm-deployment-openstack-edpm-ipam-n8rrq\" (UID: \"53d801a5-a432-44f3-970a-23519033d1e3\") " pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-n8rrq" Oct 08 18:38:11 crc kubenswrapper[4859]: I1008 18:38:11.624612 4859 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"repo-setup-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/53d801a5-a432-44f3-970a-23519033d1e3-repo-setup-combined-ca-bundle\") pod \"repo-setup-edpm-deployment-openstack-edpm-ipam-n8rrq\" (UID: \"53d801a5-a432-44f3-970a-23519033d1e3\") " pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-n8rrq" Oct 08 18:38:11 crc kubenswrapper[4859]: I1008 18:38:11.624640 4859 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-hjmnp\" (UniqueName: \"kubernetes.io/projected/53d801a5-a432-44f3-970a-23519033d1e3-kube-api-access-hjmnp\") pod \"repo-setup-edpm-deployment-openstack-edpm-ipam-n8rrq\" (UID: \"53d801a5-a432-44f3-970a-23519033d1e3\") " pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-n8rrq" Oct 08 18:38:11 crc kubenswrapper[4859]: I1008 18:38:11.624759 4859 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/53d801a5-a432-44f3-970a-23519033d1e3-ssh-key\") pod \"repo-setup-edpm-deployment-openstack-edpm-ipam-n8rrq\" (UID: \"53d801a5-a432-44f3-970a-23519033d1e3\") " pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-n8rrq" Oct 08 18:38:11 crc kubenswrapper[4859]: I1008 18:38:11.631821 4859 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"repo-setup-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/53d801a5-a432-44f3-970a-23519033d1e3-repo-setup-combined-ca-bundle\") pod \"repo-setup-edpm-deployment-openstack-edpm-ipam-n8rrq\" (UID: \"53d801a5-a432-44f3-970a-23519033d1e3\") " pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-n8rrq" Oct 08 18:38:11 crc kubenswrapper[4859]: I1008 18:38:11.633137 4859 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/53d801a5-a432-44f3-970a-23519033d1e3-inventory\") pod \"repo-setup-edpm-deployment-openstack-edpm-ipam-n8rrq\" (UID: \"53d801a5-a432-44f3-970a-23519033d1e3\") " pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-n8rrq" Oct 08 18:38:11 crc kubenswrapper[4859]: I1008 18:38:11.636058 4859 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/53d801a5-a432-44f3-970a-23519033d1e3-ssh-key\") pod \"repo-setup-edpm-deployment-openstack-edpm-ipam-n8rrq\" (UID: \"53d801a5-a432-44f3-970a-23519033d1e3\") " pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-n8rrq" Oct 08 18:38:11 crc kubenswrapper[4859]: I1008 18:38:11.643032 4859 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-hjmnp\" (UniqueName: \"kubernetes.io/projected/53d801a5-a432-44f3-970a-23519033d1e3-kube-api-access-hjmnp\") pod \"repo-setup-edpm-deployment-openstack-edpm-ipam-n8rrq\" (UID: \"53d801a5-a432-44f3-970a-23519033d1e3\") " pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-n8rrq" Oct 08 18:38:11 crc kubenswrapper[4859]: I1008 18:38:11.652965 4859 generic.go:334] "Generic (PLEG): container finished" podID="94450380-649a-4476-b646-018ed24aa703" containerID="0d06c4a1285c87136ae8add16f4fceb929ec3f7c13118a87352fc60e0db74c28" exitCode=0 Oct 08 18:38:11 crc kubenswrapper[4859]: I1008 18:38:11.653070 4859 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-server-0" event={"ID":"94450380-649a-4476-b646-018ed24aa703","Type":"ContainerDied","Data":"0d06c4a1285c87136ae8add16f4fceb929ec3f7c13118a87352fc60e0db74c28"} Oct 08 18:38:11 crc kubenswrapper[4859]: I1008 18:38:11.653131 4859 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-n8rrq" Oct 08 18:38:11 crc kubenswrapper[4859]: I1008 18:38:11.657925 4859 generic.go:334] "Generic (PLEG): container finished" podID="2c6db854-3378-4bc4-b622-480bc8c42bf5" containerID="c28efd71f628d332421e9bffaf65c2bc6cd1d1b494914b65831ddf2c8b82c9d2" exitCode=0 Oct 08 18:38:11 crc kubenswrapper[4859]: I1008 18:38:11.657987 4859 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-cell1-server-0" event={"ID":"2c6db854-3378-4bc4-b622-480bc8c42bf5","Type":"ContainerDied","Data":"c28efd71f628d332421e9bffaf65c2bc6cd1d1b494914b65831ddf2c8b82c9d2"} Oct 08 18:38:12 crc kubenswrapper[4859]: I1008 18:38:12.667075 4859 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-cell1-server-0" event={"ID":"2c6db854-3378-4bc4-b622-480bc8c42bf5","Type":"ContainerStarted","Data":"4f8bca3ac2ea4db68e3828ef8e6855a375e0cb1fc6af503f3954f386b1b4e1c0"} Oct 08 18:38:12 crc kubenswrapper[4859]: I1008 18:38:12.667648 4859 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/rabbitmq-cell1-server-0" Oct 08 18:38:12 crc kubenswrapper[4859]: I1008 18:38:12.668892 4859 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-server-0" event={"ID":"94450380-649a-4476-b646-018ed24aa703","Type":"ContainerStarted","Data":"5d06e2ed2d6f02b2fe3b158633d7b66d5227daaf47ff745bbc182e8143a5c82f"} Oct 08 18:38:12 crc kubenswrapper[4859]: I1008 18:38:12.669105 4859 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/rabbitmq-server-0" Oct 08 18:38:12 crc kubenswrapper[4859]: I1008 18:38:12.694106 4859 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/rabbitmq-cell1-server-0" podStartSLOduration=36.694072715 podStartE2EDuration="36.694072715s" podCreationTimestamp="2025-10-08 18:37:36 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-08 18:38:12.691942524 +0000 UTC m=+1262.938781903" watchObservedRunningTime="2025-10-08 18:38:12.694072715 +0000 UTC m=+1262.940912094" Oct 08 18:38:12 crc kubenswrapper[4859]: I1008 18:38:12.816822 4859 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/rabbitmq-server-0" podStartSLOduration=36.816804622 podStartE2EDuration="36.816804622s" podCreationTimestamp="2025-10-08 18:37:36 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-08 18:38:12.725340591 +0000 UTC m=+1262.972179970" watchObservedRunningTime="2025-10-08 18:38:12.816804622 +0000 UTC m=+1263.063644001" Oct 08 18:38:12 crc kubenswrapper[4859]: I1008 18:38:12.822258 4859 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-n8rrq"] Oct 08 18:38:12 crc kubenswrapper[4859]: W1008 18:38:12.822920 4859 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod53d801a5_a432_44f3_970a_23519033d1e3.slice/crio-c80daae8011ed3f7d955cefae4de7426e858f9be1b82c3a670df33505e42e5b5 WatchSource:0}: Error finding container c80daae8011ed3f7d955cefae4de7426e858f9be1b82c3a670df33505e42e5b5: Status 404 returned error can't find the container with id c80daae8011ed3f7d955cefae4de7426e858f9be1b82c3a670df33505e42e5b5 Oct 08 18:38:12 crc kubenswrapper[4859]: I1008 18:38:12.824870 4859 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Oct 08 18:38:13 crc kubenswrapper[4859]: I1008 18:38:13.684618 4859 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-n8rrq" event={"ID":"53d801a5-a432-44f3-970a-23519033d1e3","Type":"ContainerStarted","Data":"c80daae8011ed3f7d955cefae4de7426e858f9be1b82c3a670df33505e42e5b5"} Oct 08 18:38:17 crc kubenswrapper[4859]: I1008 18:38:17.924646 4859 patch_prober.go:28] interesting pod/machine-config-daemon-82s52 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 08 18:38:17 crc kubenswrapper[4859]: I1008 18:38:17.925282 4859 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-82s52" podUID="b23a6a6c-9d92-4e7b-840e-55cfda873a2d" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 08 18:38:17 crc kubenswrapper[4859]: I1008 18:38:17.925338 4859 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-82s52" Oct 08 18:38:17 crc kubenswrapper[4859]: I1008 18:38:17.926246 4859 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"7ed110461431af172203421fc6cf54521a9bc279642133f6f29a4bb290304252"} pod="openshift-machine-config-operator/machine-config-daemon-82s52" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Oct 08 18:38:17 crc kubenswrapper[4859]: I1008 18:38:17.926315 4859 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-82s52" podUID="b23a6a6c-9d92-4e7b-840e-55cfda873a2d" containerName="machine-config-daemon" containerID="cri-o://7ed110461431af172203421fc6cf54521a9bc279642133f6f29a4bb290304252" gracePeriod=600 Oct 08 18:38:18 crc kubenswrapper[4859]: I1008 18:38:18.766570 4859 generic.go:334] "Generic (PLEG): container finished" podID="b23a6a6c-9d92-4e7b-840e-55cfda873a2d" containerID="7ed110461431af172203421fc6cf54521a9bc279642133f6f29a4bb290304252" exitCode=0 Oct 08 18:38:18 crc kubenswrapper[4859]: I1008 18:38:18.766645 4859 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-82s52" event={"ID":"b23a6a6c-9d92-4e7b-840e-55cfda873a2d","Type":"ContainerDied","Data":"7ed110461431af172203421fc6cf54521a9bc279642133f6f29a4bb290304252"} Oct 08 18:38:18 crc kubenswrapper[4859]: I1008 18:38:18.766958 4859 scope.go:117] "RemoveContainer" containerID="091916d0315d4cd4d10d5ce64ba3d175b9444c9b90ffd7170b4b5fd2fb8bb62d" Oct 08 18:38:19 crc kubenswrapper[4859]: I1008 18:38:19.483640 4859 scope.go:117] "RemoveContainer" containerID="fb1976088f65191d06b2f0f31dffcf3d48a999de694d3b33c8208687385a83a9" Oct 08 18:38:21 crc kubenswrapper[4859]: I1008 18:38:21.558521 4859 scope.go:117] "RemoveContainer" containerID="9a36a11f754ff2f1288fc63ed49280399b579d6d8cc1a95aab5c7dbeed5c193b" Oct 08 18:38:22 crc kubenswrapper[4859]: I1008 18:38:22.810211 4859 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-82s52" event={"ID":"b23a6a6c-9d92-4e7b-840e-55cfda873a2d","Type":"ContainerStarted","Data":"8eec6d3743e0f820548e10238303377094560d6d193dfbfe39f2b47960c426e8"} Oct 08 18:38:22 crc kubenswrapper[4859]: I1008 18:38:22.812971 4859 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-n8rrq" event={"ID":"53d801a5-a432-44f3-970a-23519033d1e3","Type":"ContainerStarted","Data":"5455fc543390c7a789420f9bd0d9a21bab9737d004e0a3f8688862663450ec1f"} Oct 08 18:38:22 crc kubenswrapper[4859]: I1008 18:38:22.847397 4859 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-n8rrq" podStartSLOduration=2.987743865 podStartE2EDuration="11.847372605s" podCreationTimestamp="2025-10-08 18:38:11 +0000 UTC" firstStartedPulling="2025-10-08 18:38:12.824640256 +0000 UTC m=+1263.071479635" lastFinishedPulling="2025-10-08 18:38:21.684268996 +0000 UTC m=+1271.931108375" observedRunningTime="2025-10-08 18:38:22.840574951 +0000 UTC m=+1273.087414330" watchObservedRunningTime="2025-10-08 18:38:22.847372605 +0000 UTC m=+1273.094211984" Oct 08 18:38:26 crc kubenswrapper[4859]: I1008 18:38:26.658959 4859 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/rabbitmq-server-0" Oct 08 18:38:26 crc kubenswrapper[4859]: I1008 18:38:26.683799 4859 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/rabbitmq-cell1-server-0" Oct 08 18:38:34 crc kubenswrapper[4859]: I1008 18:38:34.934334 4859 generic.go:334] "Generic (PLEG): container finished" podID="53d801a5-a432-44f3-970a-23519033d1e3" containerID="5455fc543390c7a789420f9bd0d9a21bab9737d004e0a3f8688862663450ec1f" exitCode=0 Oct 08 18:38:34 crc kubenswrapper[4859]: I1008 18:38:34.934426 4859 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-n8rrq" event={"ID":"53d801a5-a432-44f3-970a-23519033d1e3","Type":"ContainerDied","Data":"5455fc543390c7a789420f9bd0d9a21bab9737d004e0a3f8688862663450ec1f"} Oct 08 18:38:36 crc kubenswrapper[4859]: I1008 18:38:36.354809 4859 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-n8rrq" Oct 08 18:38:36 crc kubenswrapper[4859]: I1008 18:38:36.489088 4859 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/53d801a5-a432-44f3-970a-23519033d1e3-ssh-key\") pod \"53d801a5-a432-44f3-970a-23519033d1e3\" (UID: \"53d801a5-a432-44f3-970a-23519033d1e3\") " Oct 08 18:38:36 crc kubenswrapper[4859]: I1008 18:38:36.489170 4859 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"repo-setup-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/53d801a5-a432-44f3-970a-23519033d1e3-repo-setup-combined-ca-bundle\") pod \"53d801a5-a432-44f3-970a-23519033d1e3\" (UID: \"53d801a5-a432-44f3-970a-23519033d1e3\") " Oct 08 18:38:36 crc kubenswrapper[4859]: I1008 18:38:36.489645 4859 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-hjmnp\" (UniqueName: \"kubernetes.io/projected/53d801a5-a432-44f3-970a-23519033d1e3-kube-api-access-hjmnp\") pod \"53d801a5-a432-44f3-970a-23519033d1e3\" (UID: \"53d801a5-a432-44f3-970a-23519033d1e3\") " Oct 08 18:38:36 crc kubenswrapper[4859]: I1008 18:38:36.489821 4859 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/53d801a5-a432-44f3-970a-23519033d1e3-inventory\") pod \"53d801a5-a432-44f3-970a-23519033d1e3\" (UID: \"53d801a5-a432-44f3-970a-23519033d1e3\") " Oct 08 18:38:36 crc kubenswrapper[4859]: I1008 18:38:36.494867 4859 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/53d801a5-a432-44f3-970a-23519033d1e3-repo-setup-combined-ca-bundle" (OuterVolumeSpecName: "repo-setup-combined-ca-bundle") pod "53d801a5-a432-44f3-970a-23519033d1e3" (UID: "53d801a5-a432-44f3-970a-23519033d1e3"). InnerVolumeSpecName "repo-setup-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 08 18:38:36 crc kubenswrapper[4859]: I1008 18:38:36.496204 4859 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/53d801a5-a432-44f3-970a-23519033d1e3-kube-api-access-hjmnp" (OuterVolumeSpecName: "kube-api-access-hjmnp") pod "53d801a5-a432-44f3-970a-23519033d1e3" (UID: "53d801a5-a432-44f3-970a-23519033d1e3"). InnerVolumeSpecName "kube-api-access-hjmnp". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 08 18:38:36 crc kubenswrapper[4859]: I1008 18:38:36.516912 4859 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/53d801a5-a432-44f3-970a-23519033d1e3-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "53d801a5-a432-44f3-970a-23519033d1e3" (UID: "53d801a5-a432-44f3-970a-23519033d1e3"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 08 18:38:36 crc kubenswrapper[4859]: I1008 18:38:36.521857 4859 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/53d801a5-a432-44f3-970a-23519033d1e3-inventory" (OuterVolumeSpecName: "inventory") pod "53d801a5-a432-44f3-970a-23519033d1e3" (UID: "53d801a5-a432-44f3-970a-23519033d1e3"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 08 18:38:36 crc kubenswrapper[4859]: I1008 18:38:36.592340 4859 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/53d801a5-a432-44f3-970a-23519033d1e3-ssh-key\") on node \"crc\" DevicePath \"\"" Oct 08 18:38:36 crc kubenswrapper[4859]: I1008 18:38:36.592396 4859 reconciler_common.go:293] "Volume detached for volume \"repo-setup-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/53d801a5-a432-44f3-970a-23519033d1e3-repo-setup-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 08 18:38:36 crc kubenswrapper[4859]: I1008 18:38:36.592421 4859 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-hjmnp\" (UniqueName: \"kubernetes.io/projected/53d801a5-a432-44f3-970a-23519033d1e3-kube-api-access-hjmnp\") on node \"crc\" DevicePath \"\"" Oct 08 18:38:36 crc kubenswrapper[4859]: I1008 18:38:36.592441 4859 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/53d801a5-a432-44f3-970a-23519033d1e3-inventory\") on node \"crc\" DevicePath \"\"" Oct 08 18:38:36 crc kubenswrapper[4859]: I1008 18:38:36.955766 4859 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-n8rrq" event={"ID":"53d801a5-a432-44f3-970a-23519033d1e3","Type":"ContainerDied","Data":"c80daae8011ed3f7d955cefae4de7426e858f9be1b82c3a670df33505e42e5b5"} Oct 08 18:38:36 crc kubenswrapper[4859]: I1008 18:38:36.955808 4859 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="c80daae8011ed3f7d955cefae4de7426e858f9be1b82c3a670df33505e42e5b5" Oct 08 18:38:36 crc kubenswrapper[4859]: I1008 18:38:36.955912 4859 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-n8rrq" Oct 08 18:38:37 crc kubenswrapper[4859]: I1008 18:38:37.047337 4859 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/redhat-edpm-deployment-openstack-edpm-ipam-fz6lm"] Oct 08 18:38:37 crc kubenswrapper[4859]: E1008 18:38:37.048073 4859 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="53d801a5-a432-44f3-970a-23519033d1e3" containerName="repo-setup-edpm-deployment-openstack-edpm-ipam" Oct 08 18:38:37 crc kubenswrapper[4859]: I1008 18:38:37.048171 4859 state_mem.go:107] "Deleted CPUSet assignment" podUID="53d801a5-a432-44f3-970a-23519033d1e3" containerName="repo-setup-edpm-deployment-openstack-edpm-ipam" Oct 08 18:38:37 crc kubenswrapper[4859]: I1008 18:38:37.048462 4859 memory_manager.go:354] "RemoveStaleState removing state" podUID="53d801a5-a432-44f3-970a-23519033d1e3" containerName="repo-setup-edpm-deployment-openstack-edpm-ipam" Oct 08 18:38:37 crc kubenswrapper[4859]: I1008 18:38:37.049268 4859 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/redhat-edpm-deployment-openstack-edpm-ipam-fz6lm" Oct 08 18:38:37 crc kubenswrapper[4859]: I1008 18:38:37.053194 4859 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Oct 08 18:38:37 crc kubenswrapper[4859]: I1008 18:38:37.053390 4859 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-gxqst" Oct 08 18:38:37 crc kubenswrapper[4859]: I1008 18:38:37.053662 4859 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Oct 08 18:38:37 crc kubenswrapper[4859]: I1008 18:38:37.053846 4859 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Oct 08 18:38:37 crc kubenswrapper[4859]: I1008 18:38:37.071664 4859 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/redhat-edpm-deployment-openstack-edpm-ipam-fz6lm"] Oct 08 18:38:37 crc kubenswrapper[4859]: I1008 18:38:37.206745 4859 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/02a2d3c9-f778-4465-b0aa-62c051299f2b-inventory\") pod \"redhat-edpm-deployment-openstack-edpm-ipam-fz6lm\" (UID: \"02a2d3c9-f778-4465-b0aa-62c051299f2b\") " pod="openstack/redhat-edpm-deployment-openstack-edpm-ipam-fz6lm" Oct 08 18:38:37 crc kubenswrapper[4859]: I1008 18:38:37.206851 4859 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-xxl8f\" (UniqueName: \"kubernetes.io/projected/02a2d3c9-f778-4465-b0aa-62c051299f2b-kube-api-access-xxl8f\") pod \"redhat-edpm-deployment-openstack-edpm-ipam-fz6lm\" (UID: \"02a2d3c9-f778-4465-b0aa-62c051299f2b\") " pod="openstack/redhat-edpm-deployment-openstack-edpm-ipam-fz6lm" Oct 08 18:38:37 crc kubenswrapper[4859]: I1008 18:38:37.207029 4859 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/02a2d3c9-f778-4465-b0aa-62c051299f2b-ssh-key\") pod \"redhat-edpm-deployment-openstack-edpm-ipam-fz6lm\" (UID: \"02a2d3c9-f778-4465-b0aa-62c051299f2b\") " pod="openstack/redhat-edpm-deployment-openstack-edpm-ipam-fz6lm" Oct 08 18:38:37 crc kubenswrapper[4859]: I1008 18:38:37.310651 4859 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/02a2d3c9-f778-4465-b0aa-62c051299f2b-ssh-key\") pod \"redhat-edpm-deployment-openstack-edpm-ipam-fz6lm\" (UID: \"02a2d3c9-f778-4465-b0aa-62c051299f2b\") " pod="openstack/redhat-edpm-deployment-openstack-edpm-ipam-fz6lm" Oct 08 18:38:37 crc kubenswrapper[4859]: I1008 18:38:37.310967 4859 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/02a2d3c9-f778-4465-b0aa-62c051299f2b-inventory\") pod \"redhat-edpm-deployment-openstack-edpm-ipam-fz6lm\" (UID: \"02a2d3c9-f778-4465-b0aa-62c051299f2b\") " pod="openstack/redhat-edpm-deployment-openstack-edpm-ipam-fz6lm" Oct 08 18:38:37 crc kubenswrapper[4859]: I1008 18:38:37.311131 4859 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-xxl8f\" (UniqueName: \"kubernetes.io/projected/02a2d3c9-f778-4465-b0aa-62c051299f2b-kube-api-access-xxl8f\") pod \"redhat-edpm-deployment-openstack-edpm-ipam-fz6lm\" (UID: \"02a2d3c9-f778-4465-b0aa-62c051299f2b\") " pod="openstack/redhat-edpm-deployment-openstack-edpm-ipam-fz6lm" Oct 08 18:38:37 crc kubenswrapper[4859]: I1008 18:38:37.315418 4859 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/02a2d3c9-f778-4465-b0aa-62c051299f2b-inventory\") pod \"redhat-edpm-deployment-openstack-edpm-ipam-fz6lm\" (UID: \"02a2d3c9-f778-4465-b0aa-62c051299f2b\") " pod="openstack/redhat-edpm-deployment-openstack-edpm-ipam-fz6lm" Oct 08 18:38:37 crc kubenswrapper[4859]: I1008 18:38:37.315478 4859 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/02a2d3c9-f778-4465-b0aa-62c051299f2b-ssh-key\") pod \"redhat-edpm-deployment-openstack-edpm-ipam-fz6lm\" (UID: \"02a2d3c9-f778-4465-b0aa-62c051299f2b\") " pod="openstack/redhat-edpm-deployment-openstack-edpm-ipam-fz6lm" Oct 08 18:38:37 crc kubenswrapper[4859]: I1008 18:38:37.329538 4859 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-xxl8f\" (UniqueName: \"kubernetes.io/projected/02a2d3c9-f778-4465-b0aa-62c051299f2b-kube-api-access-xxl8f\") pod \"redhat-edpm-deployment-openstack-edpm-ipam-fz6lm\" (UID: \"02a2d3c9-f778-4465-b0aa-62c051299f2b\") " pod="openstack/redhat-edpm-deployment-openstack-edpm-ipam-fz6lm" Oct 08 18:38:37 crc kubenswrapper[4859]: I1008 18:38:37.378016 4859 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/redhat-edpm-deployment-openstack-edpm-ipam-fz6lm" Oct 08 18:38:38 crc kubenswrapper[4859]: I1008 18:38:38.055282 4859 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/redhat-edpm-deployment-openstack-edpm-ipam-fz6lm"] Oct 08 18:38:38 crc kubenswrapper[4859]: I1008 18:38:38.982615 4859 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/redhat-edpm-deployment-openstack-edpm-ipam-fz6lm" event={"ID":"02a2d3c9-f778-4465-b0aa-62c051299f2b","Type":"ContainerStarted","Data":"23685d1dfe85cedf6ceb1388db5a1b192f021701b1743691bb4a6ad31bbe3968"} Oct 08 18:38:38 crc kubenswrapper[4859]: I1008 18:38:38.982903 4859 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/redhat-edpm-deployment-openstack-edpm-ipam-fz6lm" event={"ID":"02a2d3c9-f778-4465-b0aa-62c051299f2b","Type":"ContainerStarted","Data":"0659374cba7f6a571d648dc774a845cb76de0b10a231f0187dd5aeb24bf19aef"} Oct 08 18:38:38 crc kubenswrapper[4859]: I1008 18:38:38.998091 4859 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/redhat-edpm-deployment-openstack-edpm-ipam-fz6lm" podStartSLOduration=1.595003024 podStartE2EDuration="1.998072264s" podCreationTimestamp="2025-10-08 18:38:37 +0000 UTC" firstStartedPulling="2025-10-08 18:38:38.067508598 +0000 UTC m=+1288.314347977" lastFinishedPulling="2025-10-08 18:38:38.470577818 +0000 UTC m=+1288.717417217" observedRunningTime="2025-10-08 18:38:38.996016395 +0000 UTC m=+1289.242855774" watchObservedRunningTime="2025-10-08 18:38:38.998072264 +0000 UTC m=+1289.244911643" Oct 08 18:38:42 crc kubenswrapper[4859]: I1008 18:38:42.008642 4859 generic.go:334] "Generic (PLEG): container finished" podID="02a2d3c9-f778-4465-b0aa-62c051299f2b" containerID="23685d1dfe85cedf6ceb1388db5a1b192f021701b1743691bb4a6ad31bbe3968" exitCode=0 Oct 08 18:38:42 crc kubenswrapper[4859]: I1008 18:38:42.009154 4859 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/redhat-edpm-deployment-openstack-edpm-ipam-fz6lm" event={"ID":"02a2d3c9-f778-4465-b0aa-62c051299f2b","Type":"ContainerDied","Data":"23685d1dfe85cedf6ceb1388db5a1b192f021701b1743691bb4a6ad31bbe3968"} Oct 08 18:38:43 crc kubenswrapper[4859]: I1008 18:38:43.393982 4859 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/redhat-edpm-deployment-openstack-edpm-ipam-fz6lm" Oct 08 18:38:43 crc kubenswrapper[4859]: I1008 18:38:43.525250 4859 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-xxl8f\" (UniqueName: \"kubernetes.io/projected/02a2d3c9-f778-4465-b0aa-62c051299f2b-kube-api-access-xxl8f\") pod \"02a2d3c9-f778-4465-b0aa-62c051299f2b\" (UID: \"02a2d3c9-f778-4465-b0aa-62c051299f2b\") " Oct 08 18:38:43 crc kubenswrapper[4859]: I1008 18:38:43.525600 4859 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/02a2d3c9-f778-4465-b0aa-62c051299f2b-inventory\") pod \"02a2d3c9-f778-4465-b0aa-62c051299f2b\" (UID: \"02a2d3c9-f778-4465-b0aa-62c051299f2b\") " Oct 08 18:38:43 crc kubenswrapper[4859]: I1008 18:38:43.525717 4859 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/02a2d3c9-f778-4465-b0aa-62c051299f2b-ssh-key\") pod \"02a2d3c9-f778-4465-b0aa-62c051299f2b\" (UID: \"02a2d3c9-f778-4465-b0aa-62c051299f2b\") " Oct 08 18:38:43 crc kubenswrapper[4859]: I1008 18:38:43.530098 4859 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/02a2d3c9-f778-4465-b0aa-62c051299f2b-kube-api-access-xxl8f" (OuterVolumeSpecName: "kube-api-access-xxl8f") pod "02a2d3c9-f778-4465-b0aa-62c051299f2b" (UID: "02a2d3c9-f778-4465-b0aa-62c051299f2b"). InnerVolumeSpecName "kube-api-access-xxl8f". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 08 18:38:43 crc kubenswrapper[4859]: I1008 18:38:43.554152 4859 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/02a2d3c9-f778-4465-b0aa-62c051299f2b-inventory" (OuterVolumeSpecName: "inventory") pod "02a2d3c9-f778-4465-b0aa-62c051299f2b" (UID: "02a2d3c9-f778-4465-b0aa-62c051299f2b"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 08 18:38:43 crc kubenswrapper[4859]: I1008 18:38:43.558994 4859 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/02a2d3c9-f778-4465-b0aa-62c051299f2b-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "02a2d3c9-f778-4465-b0aa-62c051299f2b" (UID: "02a2d3c9-f778-4465-b0aa-62c051299f2b"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 08 18:38:43 crc kubenswrapper[4859]: I1008 18:38:43.627338 4859 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/02a2d3c9-f778-4465-b0aa-62c051299f2b-inventory\") on node \"crc\" DevicePath \"\"" Oct 08 18:38:43 crc kubenswrapper[4859]: I1008 18:38:43.627368 4859 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/02a2d3c9-f778-4465-b0aa-62c051299f2b-ssh-key\") on node \"crc\" DevicePath \"\"" Oct 08 18:38:43 crc kubenswrapper[4859]: I1008 18:38:43.627377 4859 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-xxl8f\" (UniqueName: \"kubernetes.io/projected/02a2d3c9-f778-4465-b0aa-62c051299f2b-kube-api-access-xxl8f\") on node \"crc\" DevicePath \"\"" Oct 08 18:38:44 crc kubenswrapper[4859]: I1008 18:38:44.027842 4859 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/redhat-edpm-deployment-openstack-edpm-ipam-fz6lm" event={"ID":"02a2d3c9-f778-4465-b0aa-62c051299f2b","Type":"ContainerDied","Data":"0659374cba7f6a571d648dc774a845cb76de0b10a231f0187dd5aeb24bf19aef"} Oct 08 18:38:44 crc kubenswrapper[4859]: I1008 18:38:44.027880 4859 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="0659374cba7f6a571d648dc774a845cb76de0b10a231f0187dd5aeb24bf19aef" Oct 08 18:38:44 crc kubenswrapper[4859]: I1008 18:38:44.027908 4859 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/redhat-edpm-deployment-openstack-edpm-ipam-fz6lm" Oct 08 18:38:44 crc kubenswrapper[4859]: I1008 18:38:44.104167 4859 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-nl8fp"] Oct 08 18:38:44 crc kubenswrapper[4859]: E1008 18:38:44.104632 4859 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="02a2d3c9-f778-4465-b0aa-62c051299f2b" containerName="redhat-edpm-deployment-openstack-edpm-ipam" Oct 08 18:38:44 crc kubenswrapper[4859]: I1008 18:38:44.104662 4859 state_mem.go:107] "Deleted CPUSet assignment" podUID="02a2d3c9-f778-4465-b0aa-62c051299f2b" containerName="redhat-edpm-deployment-openstack-edpm-ipam" Oct 08 18:38:44 crc kubenswrapper[4859]: I1008 18:38:44.104939 4859 memory_manager.go:354] "RemoveStaleState removing state" podUID="02a2d3c9-f778-4465-b0aa-62c051299f2b" containerName="redhat-edpm-deployment-openstack-edpm-ipam" Oct 08 18:38:44 crc kubenswrapper[4859]: I1008 18:38:44.108750 4859 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-nl8fp" Oct 08 18:38:44 crc kubenswrapper[4859]: I1008 18:38:44.113044 4859 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Oct 08 18:38:44 crc kubenswrapper[4859]: I1008 18:38:44.113262 4859 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Oct 08 18:38:44 crc kubenswrapper[4859]: I1008 18:38:44.113444 4859 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-gxqst" Oct 08 18:38:44 crc kubenswrapper[4859]: I1008 18:38:44.113766 4859 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Oct 08 18:38:44 crc kubenswrapper[4859]: I1008 18:38:44.120779 4859 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-nl8fp"] Oct 08 18:38:44 crc kubenswrapper[4859]: I1008 18:38:44.241688 4859 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/e0bf2482-ef25-48bb-9491-282a3eabfec0-inventory\") pod \"bootstrap-edpm-deployment-openstack-edpm-ipam-nl8fp\" (UID: \"e0bf2482-ef25-48bb-9491-282a3eabfec0\") " pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-nl8fp" Oct 08 18:38:44 crc kubenswrapper[4859]: I1008 18:38:44.242091 4859 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/e0bf2482-ef25-48bb-9491-282a3eabfec0-ssh-key\") pod \"bootstrap-edpm-deployment-openstack-edpm-ipam-nl8fp\" (UID: \"e0bf2482-ef25-48bb-9491-282a3eabfec0\") " pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-nl8fp" Oct 08 18:38:44 crc kubenswrapper[4859]: I1008 18:38:44.242878 4859 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bootstrap-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e0bf2482-ef25-48bb-9491-282a3eabfec0-bootstrap-combined-ca-bundle\") pod \"bootstrap-edpm-deployment-openstack-edpm-ipam-nl8fp\" (UID: \"e0bf2482-ef25-48bb-9491-282a3eabfec0\") " pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-nl8fp" Oct 08 18:38:44 crc kubenswrapper[4859]: I1008 18:38:44.243303 4859 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-2nktn\" (UniqueName: \"kubernetes.io/projected/e0bf2482-ef25-48bb-9491-282a3eabfec0-kube-api-access-2nktn\") pod \"bootstrap-edpm-deployment-openstack-edpm-ipam-nl8fp\" (UID: \"e0bf2482-ef25-48bb-9491-282a3eabfec0\") " pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-nl8fp" Oct 08 18:38:44 crc kubenswrapper[4859]: I1008 18:38:44.344775 4859 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bootstrap-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e0bf2482-ef25-48bb-9491-282a3eabfec0-bootstrap-combined-ca-bundle\") pod \"bootstrap-edpm-deployment-openstack-edpm-ipam-nl8fp\" (UID: \"e0bf2482-ef25-48bb-9491-282a3eabfec0\") " pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-nl8fp" Oct 08 18:38:44 crc kubenswrapper[4859]: I1008 18:38:44.344941 4859 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-2nktn\" (UniqueName: \"kubernetes.io/projected/e0bf2482-ef25-48bb-9491-282a3eabfec0-kube-api-access-2nktn\") pod \"bootstrap-edpm-deployment-openstack-edpm-ipam-nl8fp\" (UID: \"e0bf2482-ef25-48bb-9491-282a3eabfec0\") " pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-nl8fp" Oct 08 18:38:44 crc kubenswrapper[4859]: I1008 18:38:44.344991 4859 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/e0bf2482-ef25-48bb-9491-282a3eabfec0-inventory\") pod \"bootstrap-edpm-deployment-openstack-edpm-ipam-nl8fp\" (UID: \"e0bf2482-ef25-48bb-9491-282a3eabfec0\") " pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-nl8fp" Oct 08 18:38:44 crc kubenswrapper[4859]: I1008 18:38:44.345106 4859 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/e0bf2482-ef25-48bb-9491-282a3eabfec0-ssh-key\") pod \"bootstrap-edpm-deployment-openstack-edpm-ipam-nl8fp\" (UID: \"e0bf2482-ef25-48bb-9491-282a3eabfec0\") " pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-nl8fp" Oct 08 18:38:44 crc kubenswrapper[4859]: I1008 18:38:44.352173 4859 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bootstrap-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e0bf2482-ef25-48bb-9491-282a3eabfec0-bootstrap-combined-ca-bundle\") pod \"bootstrap-edpm-deployment-openstack-edpm-ipam-nl8fp\" (UID: \"e0bf2482-ef25-48bb-9491-282a3eabfec0\") " pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-nl8fp" Oct 08 18:38:44 crc kubenswrapper[4859]: I1008 18:38:44.353423 4859 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/e0bf2482-ef25-48bb-9491-282a3eabfec0-ssh-key\") pod \"bootstrap-edpm-deployment-openstack-edpm-ipam-nl8fp\" (UID: \"e0bf2482-ef25-48bb-9491-282a3eabfec0\") " pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-nl8fp" Oct 08 18:38:44 crc kubenswrapper[4859]: I1008 18:38:44.355031 4859 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/e0bf2482-ef25-48bb-9491-282a3eabfec0-inventory\") pod \"bootstrap-edpm-deployment-openstack-edpm-ipam-nl8fp\" (UID: \"e0bf2482-ef25-48bb-9491-282a3eabfec0\") " pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-nl8fp" Oct 08 18:38:44 crc kubenswrapper[4859]: I1008 18:38:44.375510 4859 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-2nktn\" (UniqueName: \"kubernetes.io/projected/e0bf2482-ef25-48bb-9491-282a3eabfec0-kube-api-access-2nktn\") pod \"bootstrap-edpm-deployment-openstack-edpm-ipam-nl8fp\" (UID: \"e0bf2482-ef25-48bb-9491-282a3eabfec0\") " pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-nl8fp" Oct 08 18:38:44 crc kubenswrapper[4859]: I1008 18:38:44.427340 4859 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-nl8fp" Oct 08 18:38:45 crc kubenswrapper[4859]: I1008 18:38:45.019465 4859 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-nl8fp"] Oct 08 18:38:45 crc kubenswrapper[4859]: W1008 18:38:45.022119 4859 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pode0bf2482_ef25_48bb_9491_282a3eabfec0.slice/crio-868c2bf1e0ba4f83a24a38305f6c6b6175ea0fedc6df3bb0c9468557bc959c19 WatchSource:0}: Error finding container 868c2bf1e0ba4f83a24a38305f6c6b6175ea0fedc6df3bb0c9468557bc959c19: Status 404 returned error can't find the container with id 868c2bf1e0ba4f83a24a38305f6c6b6175ea0fedc6df3bb0c9468557bc959c19 Oct 08 18:38:45 crc kubenswrapper[4859]: I1008 18:38:45.037871 4859 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-nl8fp" event={"ID":"e0bf2482-ef25-48bb-9491-282a3eabfec0","Type":"ContainerStarted","Data":"868c2bf1e0ba4f83a24a38305f6c6b6175ea0fedc6df3bb0c9468557bc959c19"} Oct 08 18:38:46 crc kubenswrapper[4859]: I1008 18:38:46.050868 4859 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-nl8fp" event={"ID":"e0bf2482-ef25-48bb-9491-282a3eabfec0","Type":"ContainerStarted","Data":"88550ff98a8d1bf38a7e727936e9209ff31952ae03e2be30c7269f8ca9ce7c22"} Oct 08 18:38:46 crc kubenswrapper[4859]: I1008 18:38:46.078121 4859 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-nl8fp" podStartSLOduration=1.479704419 podStartE2EDuration="2.078098737s" podCreationTimestamp="2025-10-08 18:38:44 +0000 UTC" firstStartedPulling="2025-10-08 18:38:45.025157484 +0000 UTC m=+1295.271996853" lastFinishedPulling="2025-10-08 18:38:45.623551792 +0000 UTC m=+1295.870391171" observedRunningTime="2025-10-08 18:38:46.069962154 +0000 UTC m=+1296.316801553" watchObservedRunningTime="2025-10-08 18:38:46.078098737 +0000 UTC m=+1296.324938126" Oct 08 18:39:21 crc kubenswrapper[4859]: I1008 18:39:21.755232 4859 scope.go:117] "RemoveContainer" containerID="f6db0e633c22c7dd66148f73868bba2294e2402d3003b5e69e1a17c426ad5c1c" Oct 08 18:39:53 crc kubenswrapper[4859]: I1008 18:39:53.654151 4859 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-cbkb9"] Oct 08 18:39:53 crc kubenswrapper[4859]: I1008 18:39:53.658357 4859 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-cbkb9" Oct 08 18:39:53 crc kubenswrapper[4859]: I1008 18:39:53.684364 4859 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-cbkb9"] Oct 08 18:39:53 crc kubenswrapper[4859]: I1008 18:39:53.797727 4859 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/048b339b-2d48-40b0-a569-fd44289d2047-utilities\") pod \"community-operators-cbkb9\" (UID: \"048b339b-2d48-40b0-a569-fd44289d2047\") " pod="openshift-marketplace/community-operators-cbkb9" Oct 08 18:39:53 crc kubenswrapper[4859]: I1008 18:39:53.797787 4859 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/048b339b-2d48-40b0-a569-fd44289d2047-catalog-content\") pod \"community-operators-cbkb9\" (UID: \"048b339b-2d48-40b0-a569-fd44289d2047\") " pod="openshift-marketplace/community-operators-cbkb9" Oct 08 18:39:53 crc kubenswrapper[4859]: I1008 18:39:53.798280 4859 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-4hvn5\" (UniqueName: \"kubernetes.io/projected/048b339b-2d48-40b0-a569-fd44289d2047-kube-api-access-4hvn5\") pod \"community-operators-cbkb9\" (UID: \"048b339b-2d48-40b0-a569-fd44289d2047\") " pod="openshift-marketplace/community-operators-cbkb9" Oct 08 18:39:53 crc kubenswrapper[4859]: I1008 18:39:53.900427 4859 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-4hvn5\" (UniqueName: \"kubernetes.io/projected/048b339b-2d48-40b0-a569-fd44289d2047-kube-api-access-4hvn5\") pod \"community-operators-cbkb9\" (UID: \"048b339b-2d48-40b0-a569-fd44289d2047\") " pod="openshift-marketplace/community-operators-cbkb9" Oct 08 18:39:53 crc kubenswrapper[4859]: I1008 18:39:53.900509 4859 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/048b339b-2d48-40b0-a569-fd44289d2047-utilities\") pod \"community-operators-cbkb9\" (UID: \"048b339b-2d48-40b0-a569-fd44289d2047\") " pod="openshift-marketplace/community-operators-cbkb9" Oct 08 18:39:53 crc kubenswrapper[4859]: I1008 18:39:53.900533 4859 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/048b339b-2d48-40b0-a569-fd44289d2047-catalog-content\") pod \"community-operators-cbkb9\" (UID: \"048b339b-2d48-40b0-a569-fd44289d2047\") " pod="openshift-marketplace/community-operators-cbkb9" Oct 08 18:39:53 crc kubenswrapper[4859]: I1008 18:39:53.901092 4859 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/048b339b-2d48-40b0-a569-fd44289d2047-utilities\") pod \"community-operators-cbkb9\" (UID: \"048b339b-2d48-40b0-a569-fd44289d2047\") " pod="openshift-marketplace/community-operators-cbkb9" Oct 08 18:39:53 crc kubenswrapper[4859]: I1008 18:39:53.901163 4859 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/048b339b-2d48-40b0-a569-fd44289d2047-catalog-content\") pod \"community-operators-cbkb9\" (UID: \"048b339b-2d48-40b0-a569-fd44289d2047\") " pod="openshift-marketplace/community-operators-cbkb9" Oct 08 18:39:53 crc kubenswrapper[4859]: I1008 18:39:53.937642 4859 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-4hvn5\" (UniqueName: \"kubernetes.io/projected/048b339b-2d48-40b0-a569-fd44289d2047-kube-api-access-4hvn5\") pod \"community-operators-cbkb9\" (UID: \"048b339b-2d48-40b0-a569-fd44289d2047\") " pod="openshift-marketplace/community-operators-cbkb9" Oct 08 18:39:53 crc kubenswrapper[4859]: I1008 18:39:53.987798 4859 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-cbkb9" Oct 08 18:39:54 crc kubenswrapper[4859]: I1008 18:39:54.545833 4859 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-cbkb9"] Oct 08 18:39:54 crc kubenswrapper[4859]: I1008 18:39:54.764904 4859 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-cbkb9" event={"ID":"048b339b-2d48-40b0-a569-fd44289d2047","Type":"ContainerStarted","Data":"5bd817513716dbb7f6f99c9b8a7c96f7a97dd24db18548f9f7bbea7ada4dd2e8"} Oct 08 18:39:54 crc kubenswrapper[4859]: I1008 18:39:54.764952 4859 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-cbkb9" event={"ID":"048b339b-2d48-40b0-a569-fd44289d2047","Type":"ContainerStarted","Data":"b857e41834be6ce4be64a4bddbe7c9b78987de5ae18df0ebf2d98b1e761cff6a"} Oct 08 18:39:55 crc kubenswrapper[4859]: I1008 18:39:55.777359 4859 generic.go:334] "Generic (PLEG): container finished" podID="048b339b-2d48-40b0-a569-fd44289d2047" containerID="5bd817513716dbb7f6f99c9b8a7c96f7a97dd24db18548f9f7bbea7ada4dd2e8" exitCode=0 Oct 08 18:39:55 crc kubenswrapper[4859]: I1008 18:39:55.777787 4859 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-cbkb9" event={"ID":"048b339b-2d48-40b0-a569-fd44289d2047","Type":"ContainerDied","Data":"5bd817513716dbb7f6f99c9b8a7c96f7a97dd24db18548f9f7bbea7ada4dd2e8"} Oct 08 18:39:56 crc kubenswrapper[4859]: I1008 18:39:56.787715 4859 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-cbkb9" event={"ID":"048b339b-2d48-40b0-a569-fd44289d2047","Type":"ContainerStarted","Data":"b1076ebc34e0ee0ed7776d751b5ccdcddc5c3b93901e9c877894f6d42afa674b"} Oct 08 18:39:57 crc kubenswrapper[4859]: I1008 18:39:57.797531 4859 generic.go:334] "Generic (PLEG): container finished" podID="048b339b-2d48-40b0-a569-fd44289d2047" containerID="b1076ebc34e0ee0ed7776d751b5ccdcddc5c3b93901e9c877894f6d42afa674b" exitCode=0 Oct 08 18:39:57 crc kubenswrapper[4859]: I1008 18:39:57.797667 4859 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-cbkb9" event={"ID":"048b339b-2d48-40b0-a569-fd44289d2047","Type":"ContainerDied","Data":"b1076ebc34e0ee0ed7776d751b5ccdcddc5c3b93901e9c877894f6d42afa674b"} Oct 08 18:39:58 crc kubenswrapper[4859]: I1008 18:39:58.809393 4859 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-cbkb9" event={"ID":"048b339b-2d48-40b0-a569-fd44289d2047","Type":"ContainerStarted","Data":"064f866c14e6d0e34945d126e116b0c59a1258f4b17484b3bc925a137fcb27b0"} Oct 08 18:39:58 crc kubenswrapper[4859]: I1008 18:39:58.828908 4859 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-cbkb9" podStartSLOduration=3.363182102 podStartE2EDuration="5.828888666s" podCreationTimestamp="2025-10-08 18:39:53 +0000 UTC" firstStartedPulling="2025-10-08 18:39:55.780495308 +0000 UTC m=+1366.027334697" lastFinishedPulling="2025-10-08 18:39:58.246201892 +0000 UTC m=+1368.493041261" observedRunningTime="2025-10-08 18:39:58.824824659 +0000 UTC m=+1369.071664048" watchObservedRunningTime="2025-10-08 18:39:58.828888666 +0000 UTC m=+1369.075728065" Oct 08 18:40:03 crc kubenswrapper[4859]: I1008 18:40:03.988864 4859 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-cbkb9" Oct 08 18:40:03 crc kubenswrapper[4859]: I1008 18:40:03.989375 4859 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-cbkb9" Oct 08 18:40:04 crc kubenswrapper[4859]: I1008 18:40:04.045823 4859 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-cbkb9" Oct 08 18:40:04 crc kubenswrapper[4859]: I1008 18:40:04.954967 4859 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-cbkb9" Oct 08 18:40:05 crc kubenswrapper[4859]: I1008 18:40:05.005070 4859 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-cbkb9"] Oct 08 18:40:06 crc kubenswrapper[4859]: I1008 18:40:06.887423 4859 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/community-operators-cbkb9" podUID="048b339b-2d48-40b0-a569-fd44289d2047" containerName="registry-server" containerID="cri-o://064f866c14e6d0e34945d126e116b0c59a1258f4b17484b3bc925a137fcb27b0" gracePeriod=2 Oct 08 18:40:07 crc kubenswrapper[4859]: I1008 18:40:07.389471 4859 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-cbkb9" Oct 08 18:40:07 crc kubenswrapper[4859]: I1008 18:40:07.451613 4859 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/048b339b-2d48-40b0-a569-fd44289d2047-utilities\") pod \"048b339b-2d48-40b0-a569-fd44289d2047\" (UID: \"048b339b-2d48-40b0-a569-fd44289d2047\") " Oct 08 18:40:07 crc kubenswrapper[4859]: I1008 18:40:07.451763 4859 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/048b339b-2d48-40b0-a569-fd44289d2047-catalog-content\") pod \"048b339b-2d48-40b0-a569-fd44289d2047\" (UID: \"048b339b-2d48-40b0-a569-fd44289d2047\") " Oct 08 18:40:07 crc kubenswrapper[4859]: I1008 18:40:07.451847 4859 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-4hvn5\" (UniqueName: \"kubernetes.io/projected/048b339b-2d48-40b0-a569-fd44289d2047-kube-api-access-4hvn5\") pod \"048b339b-2d48-40b0-a569-fd44289d2047\" (UID: \"048b339b-2d48-40b0-a569-fd44289d2047\") " Oct 08 18:40:07 crc kubenswrapper[4859]: I1008 18:40:07.452937 4859 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/048b339b-2d48-40b0-a569-fd44289d2047-utilities" (OuterVolumeSpecName: "utilities") pod "048b339b-2d48-40b0-a569-fd44289d2047" (UID: "048b339b-2d48-40b0-a569-fd44289d2047"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 08 18:40:07 crc kubenswrapper[4859]: I1008 18:40:07.457584 4859 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/048b339b-2d48-40b0-a569-fd44289d2047-kube-api-access-4hvn5" (OuterVolumeSpecName: "kube-api-access-4hvn5") pod "048b339b-2d48-40b0-a569-fd44289d2047" (UID: "048b339b-2d48-40b0-a569-fd44289d2047"). InnerVolumeSpecName "kube-api-access-4hvn5". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 08 18:40:07 crc kubenswrapper[4859]: I1008 18:40:07.497123 4859 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/048b339b-2d48-40b0-a569-fd44289d2047-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "048b339b-2d48-40b0-a569-fd44289d2047" (UID: "048b339b-2d48-40b0-a569-fd44289d2047"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 08 18:40:07 crc kubenswrapper[4859]: I1008 18:40:07.557411 4859 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/048b339b-2d48-40b0-a569-fd44289d2047-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 08 18:40:07 crc kubenswrapper[4859]: I1008 18:40:07.557819 4859 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-4hvn5\" (UniqueName: \"kubernetes.io/projected/048b339b-2d48-40b0-a569-fd44289d2047-kube-api-access-4hvn5\") on node \"crc\" DevicePath \"\"" Oct 08 18:40:07 crc kubenswrapper[4859]: I1008 18:40:07.557845 4859 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/048b339b-2d48-40b0-a569-fd44289d2047-utilities\") on node \"crc\" DevicePath \"\"" Oct 08 18:40:07 crc kubenswrapper[4859]: I1008 18:40:07.900511 4859 generic.go:334] "Generic (PLEG): container finished" podID="048b339b-2d48-40b0-a569-fd44289d2047" containerID="064f866c14e6d0e34945d126e116b0c59a1258f4b17484b3bc925a137fcb27b0" exitCode=0 Oct 08 18:40:07 crc kubenswrapper[4859]: I1008 18:40:07.900571 4859 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-cbkb9" event={"ID":"048b339b-2d48-40b0-a569-fd44289d2047","Type":"ContainerDied","Data":"064f866c14e6d0e34945d126e116b0c59a1258f4b17484b3bc925a137fcb27b0"} Oct 08 18:40:07 crc kubenswrapper[4859]: I1008 18:40:07.900614 4859 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-cbkb9" Oct 08 18:40:07 crc kubenswrapper[4859]: I1008 18:40:07.900640 4859 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-cbkb9" event={"ID":"048b339b-2d48-40b0-a569-fd44289d2047","Type":"ContainerDied","Data":"b857e41834be6ce4be64a4bddbe7c9b78987de5ae18df0ebf2d98b1e761cff6a"} Oct 08 18:40:07 crc kubenswrapper[4859]: I1008 18:40:07.900662 4859 scope.go:117] "RemoveContainer" containerID="064f866c14e6d0e34945d126e116b0c59a1258f4b17484b3bc925a137fcb27b0" Oct 08 18:40:07 crc kubenswrapper[4859]: I1008 18:40:07.926096 4859 scope.go:117] "RemoveContainer" containerID="b1076ebc34e0ee0ed7776d751b5ccdcddc5c3b93901e9c877894f6d42afa674b" Oct 08 18:40:07 crc kubenswrapper[4859]: I1008 18:40:07.949069 4859 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-cbkb9"] Oct 08 18:40:07 crc kubenswrapper[4859]: I1008 18:40:07.960448 4859 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/community-operators-cbkb9"] Oct 08 18:40:07 crc kubenswrapper[4859]: I1008 18:40:07.975182 4859 scope.go:117] "RemoveContainer" containerID="5bd817513716dbb7f6f99c9b8a7c96f7a97dd24db18548f9f7bbea7ada4dd2e8" Oct 08 18:40:08 crc kubenswrapper[4859]: I1008 18:40:08.029168 4859 scope.go:117] "RemoveContainer" containerID="064f866c14e6d0e34945d126e116b0c59a1258f4b17484b3bc925a137fcb27b0" Oct 08 18:40:08 crc kubenswrapper[4859]: E1008 18:40:08.029832 4859 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"064f866c14e6d0e34945d126e116b0c59a1258f4b17484b3bc925a137fcb27b0\": container with ID starting with 064f866c14e6d0e34945d126e116b0c59a1258f4b17484b3bc925a137fcb27b0 not found: ID does not exist" containerID="064f866c14e6d0e34945d126e116b0c59a1258f4b17484b3bc925a137fcb27b0" Oct 08 18:40:08 crc kubenswrapper[4859]: I1008 18:40:08.029890 4859 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"064f866c14e6d0e34945d126e116b0c59a1258f4b17484b3bc925a137fcb27b0"} err="failed to get container status \"064f866c14e6d0e34945d126e116b0c59a1258f4b17484b3bc925a137fcb27b0\": rpc error: code = NotFound desc = could not find container \"064f866c14e6d0e34945d126e116b0c59a1258f4b17484b3bc925a137fcb27b0\": container with ID starting with 064f866c14e6d0e34945d126e116b0c59a1258f4b17484b3bc925a137fcb27b0 not found: ID does not exist" Oct 08 18:40:08 crc kubenswrapper[4859]: I1008 18:40:08.029925 4859 scope.go:117] "RemoveContainer" containerID="b1076ebc34e0ee0ed7776d751b5ccdcddc5c3b93901e9c877894f6d42afa674b" Oct 08 18:40:08 crc kubenswrapper[4859]: E1008 18:40:08.030483 4859 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"b1076ebc34e0ee0ed7776d751b5ccdcddc5c3b93901e9c877894f6d42afa674b\": container with ID starting with b1076ebc34e0ee0ed7776d751b5ccdcddc5c3b93901e9c877894f6d42afa674b not found: ID does not exist" containerID="b1076ebc34e0ee0ed7776d751b5ccdcddc5c3b93901e9c877894f6d42afa674b" Oct 08 18:40:08 crc kubenswrapper[4859]: I1008 18:40:08.030516 4859 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"b1076ebc34e0ee0ed7776d751b5ccdcddc5c3b93901e9c877894f6d42afa674b"} err="failed to get container status \"b1076ebc34e0ee0ed7776d751b5ccdcddc5c3b93901e9c877894f6d42afa674b\": rpc error: code = NotFound desc = could not find container \"b1076ebc34e0ee0ed7776d751b5ccdcddc5c3b93901e9c877894f6d42afa674b\": container with ID starting with b1076ebc34e0ee0ed7776d751b5ccdcddc5c3b93901e9c877894f6d42afa674b not found: ID does not exist" Oct 08 18:40:08 crc kubenswrapper[4859]: I1008 18:40:08.030534 4859 scope.go:117] "RemoveContainer" containerID="5bd817513716dbb7f6f99c9b8a7c96f7a97dd24db18548f9f7bbea7ada4dd2e8" Oct 08 18:40:08 crc kubenswrapper[4859]: E1008 18:40:08.030906 4859 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"5bd817513716dbb7f6f99c9b8a7c96f7a97dd24db18548f9f7bbea7ada4dd2e8\": container with ID starting with 5bd817513716dbb7f6f99c9b8a7c96f7a97dd24db18548f9f7bbea7ada4dd2e8 not found: ID does not exist" containerID="5bd817513716dbb7f6f99c9b8a7c96f7a97dd24db18548f9f7bbea7ada4dd2e8" Oct 08 18:40:08 crc kubenswrapper[4859]: I1008 18:40:08.030937 4859 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"5bd817513716dbb7f6f99c9b8a7c96f7a97dd24db18548f9f7bbea7ada4dd2e8"} err="failed to get container status \"5bd817513716dbb7f6f99c9b8a7c96f7a97dd24db18548f9f7bbea7ada4dd2e8\": rpc error: code = NotFound desc = could not find container \"5bd817513716dbb7f6f99c9b8a7c96f7a97dd24db18548f9f7bbea7ada4dd2e8\": container with ID starting with 5bd817513716dbb7f6f99c9b8a7c96f7a97dd24db18548f9f7bbea7ada4dd2e8 not found: ID does not exist" Oct 08 18:40:08 crc kubenswrapper[4859]: I1008 18:40:08.526108 4859 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="048b339b-2d48-40b0-a569-fd44289d2047" path="/var/lib/kubelet/pods/048b339b-2d48-40b0-a569-fd44289d2047/volumes" Oct 08 18:40:21 crc kubenswrapper[4859]: I1008 18:40:21.832368 4859 scope.go:117] "RemoveContainer" containerID="e48d82fde98292fc3ac341cff6cc00d37119be8b99b5e5c33a5f112eb3f15c4a" Oct 08 18:40:21 crc kubenswrapper[4859]: I1008 18:40:21.867624 4859 scope.go:117] "RemoveContainer" containerID="d4f40d2b1c23ce93761ada628d8d73a724dc6702e886ec2cd59ef35c8b2b69d8" Oct 08 18:40:21 crc kubenswrapper[4859]: I1008 18:40:21.907773 4859 scope.go:117] "RemoveContainer" containerID="2a699d4ef1d913cf94815cb3b9d6ef92b6f7bcdf4ce58dbd6b5b6b78663afd86" Oct 08 18:40:21 crc kubenswrapper[4859]: I1008 18:40:21.926389 4859 scope.go:117] "RemoveContainer" containerID="c76586e5e90236f90858e341d25b5dde536c874c2bca5a632ffc704ce10dd611" Oct 08 18:40:47 crc kubenswrapper[4859]: I1008 18:40:47.925469 4859 patch_prober.go:28] interesting pod/machine-config-daemon-82s52 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 08 18:40:47 crc kubenswrapper[4859]: I1008 18:40:47.926152 4859 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-82s52" podUID="b23a6a6c-9d92-4e7b-840e-55cfda873a2d" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 08 18:40:56 crc kubenswrapper[4859]: I1008 18:40:56.714387 4859 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-hc7vd"] Oct 08 18:40:56 crc kubenswrapper[4859]: E1008 18:40:56.715794 4859 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="048b339b-2d48-40b0-a569-fd44289d2047" containerName="extract-content" Oct 08 18:40:56 crc kubenswrapper[4859]: I1008 18:40:56.715818 4859 state_mem.go:107] "Deleted CPUSet assignment" podUID="048b339b-2d48-40b0-a569-fd44289d2047" containerName="extract-content" Oct 08 18:40:56 crc kubenswrapper[4859]: E1008 18:40:56.715838 4859 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="048b339b-2d48-40b0-a569-fd44289d2047" containerName="extract-utilities" Oct 08 18:40:56 crc kubenswrapper[4859]: I1008 18:40:56.715852 4859 state_mem.go:107] "Deleted CPUSet assignment" podUID="048b339b-2d48-40b0-a569-fd44289d2047" containerName="extract-utilities" Oct 08 18:40:56 crc kubenswrapper[4859]: E1008 18:40:56.715873 4859 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="048b339b-2d48-40b0-a569-fd44289d2047" containerName="registry-server" Oct 08 18:40:56 crc kubenswrapper[4859]: I1008 18:40:56.715887 4859 state_mem.go:107] "Deleted CPUSet assignment" podUID="048b339b-2d48-40b0-a569-fd44289d2047" containerName="registry-server" Oct 08 18:40:56 crc kubenswrapper[4859]: I1008 18:40:56.716301 4859 memory_manager.go:354] "RemoveStaleState removing state" podUID="048b339b-2d48-40b0-a569-fd44289d2047" containerName="registry-server" Oct 08 18:40:56 crc kubenswrapper[4859]: I1008 18:40:56.718942 4859 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-hc7vd" Oct 08 18:40:56 crc kubenswrapper[4859]: I1008 18:40:56.725593 4859 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-hc7vd"] Oct 08 18:40:56 crc kubenswrapper[4859]: I1008 18:40:56.820861 4859 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-ncz2z\" (UniqueName: \"kubernetes.io/projected/ec0f51e7-a093-46eb-8cc5-e4b9e1a0fe44-kube-api-access-ncz2z\") pod \"certified-operators-hc7vd\" (UID: \"ec0f51e7-a093-46eb-8cc5-e4b9e1a0fe44\") " pod="openshift-marketplace/certified-operators-hc7vd" Oct 08 18:40:56 crc kubenswrapper[4859]: I1008 18:40:56.821016 4859 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/ec0f51e7-a093-46eb-8cc5-e4b9e1a0fe44-catalog-content\") pod \"certified-operators-hc7vd\" (UID: \"ec0f51e7-a093-46eb-8cc5-e4b9e1a0fe44\") " pod="openshift-marketplace/certified-operators-hc7vd" Oct 08 18:40:56 crc kubenswrapper[4859]: I1008 18:40:56.821045 4859 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/ec0f51e7-a093-46eb-8cc5-e4b9e1a0fe44-utilities\") pod \"certified-operators-hc7vd\" (UID: \"ec0f51e7-a093-46eb-8cc5-e4b9e1a0fe44\") " pod="openshift-marketplace/certified-operators-hc7vd" Oct 08 18:40:56 crc kubenswrapper[4859]: I1008 18:40:56.923144 4859 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/ec0f51e7-a093-46eb-8cc5-e4b9e1a0fe44-catalog-content\") pod \"certified-operators-hc7vd\" (UID: \"ec0f51e7-a093-46eb-8cc5-e4b9e1a0fe44\") " pod="openshift-marketplace/certified-operators-hc7vd" Oct 08 18:40:56 crc kubenswrapper[4859]: I1008 18:40:56.923199 4859 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/ec0f51e7-a093-46eb-8cc5-e4b9e1a0fe44-utilities\") pod \"certified-operators-hc7vd\" (UID: \"ec0f51e7-a093-46eb-8cc5-e4b9e1a0fe44\") " pod="openshift-marketplace/certified-operators-hc7vd" Oct 08 18:40:56 crc kubenswrapper[4859]: I1008 18:40:56.923278 4859 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-ncz2z\" (UniqueName: \"kubernetes.io/projected/ec0f51e7-a093-46eb-8cc5-e4b9e1a0fe44-kube-api-access-ncz2z\") pod \"certified-operators-hc7vd\" (UID: \"ec0f51e7-a093-46eb-8cc5-e4b9e1a0fe44\") " pod="openshift-marketplace/certified-operators-hc7vd" Oct 08 18:40:56 crc kubenswrapper[4859]: I1008 18:40:56.924069 4859 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/ec0f51e7-a093-46eb-8cc5-e4b9e1a0fe44-catalog-content\") pod \"certified-operators-hc7vd\" (UID: \"ec0f51e7-a093-46eb-8cc5-e4b9e1a0fe44\") " pod="openshift-marketplace/certified-operators-hc7vd" Oct 08 18:40:56 crc kubenswrapper[4859]: I1008 18:40:56.924267 4859 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/ec0f51e7-a093-46eb-8cc5-e4b9e1a0fe44-utilities\") pod \"certified-operators-hc7vd\" (UID: \"ec0f51e7-a093-46eb-8cc5-e4b9e1a0fe44\") " pod="openshift-marketplace/certified-operators-hc7vd" Oct 08 18:40:56 crc kubenswrapper[4859]: I1008 18:40:56.945156 4859 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-ncz2z\" (UniqueName: \"kubernetes.io/projected/ec0f51e7-a093-46eb-8cc5-e4b9e1a0fe44-kube-api-access-ncz2z\") pod \"certified-operators-hc7vd\" (UID: \"ec0f51e7-a093-46eb-8cc5-e4b9e1a0fe44\") " pod="openshift-marketplace/certified-operators-hc7vd" Oct 08 18:40:57 crc kubenswrapper[4859]: I1008 18:40:57.047617 4859 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-hc7vd" Oct 08 18:40:57 crc kubenswrapper[4859]: I1008 18:40:57.573338 4859 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-hc7vd"] Oct 08 18:40:58 crc kubenswrapper[4859]: I1008 18:40:58.511778 4859 generic.go:334] "Generic (PLEG): container finished" podID="ec0f51e7-a093-46eb-8cc5-e4b9e1a0fe44" containerID="61bca07c0abce38eff0ca53124ca48bbb597309552d5cd0c68b0f0c0c415aaea" exitCode=0 Oct 08 18:40:58 crc kubenswrapper[4859]: I1008 18:40:58.511815 4859 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-hc7vd" event={"ID":"ec0f51e7-a093-46eb-8cc5-e4b9e1a0fe44","Type":"ContainerDied","Data":"61bca07c0abce38eff0ca53124ca48bbb597309552d5cd0c68b0f0c0c415aaea"} Oct 08 18:40:58 crc kubenswrapper[4859]: I1008 18:40:58.512167 4859 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-hc7vd" event={"ID":"ec0f51e7-a093-46eb-8cc5-e4b9e1a0fe44","Type":"ContainerStarted","Data":"a060e9f3432339def437fcaa0e13850ac68536dd08e3efd382a08c3bf7ac7f51"} Oct 08 18:41:00 crc kubenswrapper[4859]: I1008 18:41:00.538106 4859 generic.go:334] "Generic (PLEG): container finished" podID="ec0f51e7-a093-46eb-8cc5-e4b9e1a0fe44" containerID="d11fcc2faca9bfc6b362592a97badbb5be6111273fbf417de785e9a0b0bff4c1" exitCode=0 Oct 08 18:41:00 crc kubenswrapper[4859]: I1008 18:41:00.538201 4859 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-hc7vd" event={"ID":"ec0f51e7-a093-46eb-8cc5-e4b9e1a0fe44","Type":"ContainerDied","Data":"d11fcc2faca9bfc6b362592a97badbb5be6111273fbf417de785e9a0b0bff4c1"} Oct 08 18:41:01 crc kubenswrapper[4859]: I1008 18:41:01.550200 4859 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-hc7vd" event={"ID":"ec0f51e7-a093-46eb-8cc5-e4b9e1a0fe44","Type":"ContainerStarted","Data":"b4a85e09f5b786988c02be2e412146a63e84815706bd04d92a46462c24cbc22e"} Oct 08 18:41:01 crc kubenswrapper[4859]: I1008 18:41:01.584047 4859 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-hc7vd" podStartSLOduration=3.190031585 podStartE2EDuration="5.58399946s" podCreationTimestamp="2025-10-08 18:40:56 +0000 UTC" firstStartedPulling="2025-10-08 18:40:58.513673086 +0000 UTC m=+1428.760512465" lastFinishedPulling="2025-10-08 18:41:00.907640951 +0000 UTC m=+1431.154480340" observedRunningTime="2025-10-08 18:41:01.56915805 +0000 UTC m=+1431.815997429" watchObservedRunningTime="2025-10-08 18:41:01.58399946 +0000 UTC m=+1431.830838839" Oct 08 18:41:07 crc kubenswrapper[4859]: I1008 18:41:07.048667 4859 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-hc7vd" Oct 08 18:41:07 crc kubenswrapper[4859]: I1008 18:41:07.049485 4859 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-hc7vd" Oct 08 18:41:07 crc kubenswrapper[4859]: I1008 18:41:07.099979 4859 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-hc7vd" Oct 08 18:41:07 crc kubenswrapper[4859]: I1008 18:41:07.669550 4859 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-hc7vd" Oct 08 18:41:07 crc kubenswrapper[4859]: I1008 18:41:07.718121 4859 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-hc7vd"] Oct 08 18:41:09 crc kubenswrapper[4859]: I1008 18:41:09.632070 4859 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/certified-operators-hc7vd" podUID="ec0f51e7-a093-46eb-8cc5-e4b9e1a0fe44" containerName="registry-server" containerID="cri-o://b4a85e09f5b786988c02be2e412146a63e84815706bd04d92a46462c24cbc22e" gracePeriod=2 Oct 08 18:41:10 crc kubenswrapper[4859]: I1008 18:41:10.155199 4859 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-hc7vd" Oct 08 18:41:10 crc kubenswrapper[4859]: I1008 18:41:10.307578 4859 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-ncz2z\" (UniqueName: \"kubernetes.io/projected/ec0f51e7-a093-46eb-8cc5-e4b9e1a0fe44-kube-api-access-ncz2z\") pod \"ec0f51e7-a093-46eb-8cc5-e4b9e1a0fe44\" (UID: \"ec0f51e7-a093-46eb-8cc5-e4b9e1a0fe44\") " Oct 08 18:41:10 crc kubenswrapper[4859]: I1008 18:41:10.308117 4859 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/ec0f51e7-a093-46eb-8cc5-e4b9e1a0fe44-utilities\") pod \"ec0f51e7-a093-46eb-8cc5-e4b9e1a0fe44\" (UID: \"ec0f51e7-a093-46eb-8cc5-e4b9e1a0fe44\") " Oct 08 18:41:10 crc kubenswrapper[4859]: I1008 18:41:10.308322 4859 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/ec0f51e7-a093-46eb-8cc5-e4b9e1a0fe44-catalog-content\") pod \"ec0f51e7-a093-46eb-8cc5-e4b9e1a0fe44\" (UID: \"ec0f51e7-a093-46eb-8cc5-e4b9e1a0fe44\") " Oct 08 18:41:10 crc kubenswrapper[4859]: I1008 18:41:10.309971 4859 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/ec0f51e7-a093-46eb-8cc5-e4b9e1a0fe44-utilities" (OuterVolumeSpecName: "utilities") pod "ec0f51e7-a093-46eb-8cc5-e4b9e1a0fe44" (UID: "ec0f51e7-a093-46eb-8cc5-e4b9e1a0fe44"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 08 18:41:10 crc kubenswrapper[4859]: I1008 18:41:10.314962 4859 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/ec0f51e7-a093-46eb-8cc5-e4b9e1a0fe44-kube-api-access-ncz2z" (OuterVolumeSpecName: "kube-api-access-ncz2z") pod "ec0f51e7-a093-46eb-8cc5-e4b9e1a0fe44" (UID: "ec0f51e7-a093-46eb-8cc5-e4b9e1a0fe44"). InnerVolumeSpecName "kube-api-access-ncz2z". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 08 18:41:10 crc kubenswrapper[4859]: I1008 18:41:10.371571 4859 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/ec0f51e7-a093-46eb-8cc5-e4b9e1a0fe44-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "ec0f51e7-a093-46eb-8cc5-e4b9e1a0fe44" (UID: "ec0f51e7-a093-46eb-8cc5-e4b9e1a0fe44"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 08 18:41:10 crc kubenswrapper[4859]: I1008 18:41:10.411106 4859 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/ec0f51e7-a093-46eb-8cc5-e4b9e1a0fe44-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 08 18:41:10 crc kubenswrapper[4859]: I1008 18:41:10.411135 4859 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-ncz2z\" (UniqueName: \"kubernetes.io/projected/ec0f51e7-a093-46eb-8cc5-e4b9e1a0fe44-kube-api-access-ncz2z\") on node \"crc\" DevicePath \"\"" Oct 08 18:41:10 crc kubenswrapper[4859]: I1008 18:41:10.411145 4859 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/ec0f51e7-a093-46eb-8cc5-e4b9e1a0fe44-utilities\") on node \"crc\" DevicePath \"\"" Oct 08 18:41:10 crc kubenswrapper[4859]: I1008 18:41:10.642092 4859 generic.go:334] "Generic (PLEG): container finished" podID="ec0f51e7-a093-46eb-8cc5-e4b9e1a0fe44" containerID="b4a85e09f5b786988c02be2e412146a63e84815706bd04d92a46462c24cbc22e" exitCode=0 Oct 08 18:41:10 crc kubenswrapper[4859]: I1008 18:41:10.642150 4859 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-hc7vd" event={"ID":"ec0f51e7-a093-46eb-8cc5-e4b9e1a0fe44","Type":"ContainerDied","Data":"b4a85e09f5b786988c02be2e412146a63e84815706bd04d92a46462c24cbc22e"} Oct 08 18:41:10 crc kubenswrapper[4859]: I1008 18:41:10.642190 4859 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-hc7vd" event={"ID":"ec0f51e7-a093-46eb-8cc5-e4b9e1a0fe44","Type":"ContainerDied","Data":"a060e9f3432339def437fcaa0e13850ac68536dd08e3efd382a08c3bf7ac7f51"} Oct 08 18:41:10 crc kubenswrapper[4859]: I1008 18:41:10.642205 4859 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-hc7vd" Oct 08 18:41:10 crc kubenswrapper[4859]: I1008 18:41:10.642212 4859 scope.go:117] "RemoveContainer" containerID="b4a85e09f5b786988c02be2e412146a63e84815706bd04d92a46462c24cbc22e" Oct 08 18:41:10 crc kubenswrapper[4859]: I1008 18:41:10.668504 4859 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-hc7vd"] Oct 08 18:41:10 crc kubenswrapper[4859]: I1008 18:41:10.669715 4859 scope.go:117] "RemoveContainer" containerID="d11fcc2faca9bfc6b362592a97badbb5be6111273fbf417de785e9a0b0bff4c1" Oct 08 18:41:10 crc kubenswrapper[4859]: I1008 18:41:10.676435 4859 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/certified-operators-hc7vd"] Oct 08 18:41:10 crc kubenswrapper[4859]: I1008 18:41:10.689406 4859 scope.go:117] "RemoveContainer" containerID="61bca07c0abce38eff0ca53124ca48bbb597309552d5cd0c68b0f0c0c415aaea" Oct 08 18:41:10 crc kubenswrapper[4859]: I1008 18:41:10.734172 4859 scope.go:117] "RemoveContainer" containerID="b4a85e09f5b786988c02be2e412146a63e84815706bd04d92a46462c24cbc22e" Oct 08 18:41:10 crc kubenswrapper[4859]: E1008 18:41:10.734599 4859 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"b4a85e09f5b786988c02be2e412146a63e84815706bd04d92a46462c24cbc22e\": container with ID starting with b4a85e09f5b786988c02be2e412146a63e84815706bd04d92a46462c24cbc22e not found: ID does not exist" containerID="b4a85e09f5b786988c02be2e412146a63e84815706bd04d92a46462c24cbc22e" Oct 08 18:41:10 crc kubenswrapper[4859]: I1008 18:41:10.734676 4859 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"b4a85e09f5b786988c02be2e412146a63e84815706bd04d92a46462c24cbc22e"} err="failed to get container status \"b4a85e09f5b786988c02be2e412146a63e84815706bd04d92a46462c24cbc22e\": rpc error: code = NotFound desc = could not find container \"b4a85e09f5b786988c02be2e412146a63e84815706bd04d92a46462c24cbc22e\": container with ID starting with b4a85e09f5b786988c02be2e412146a63e84815706bd04d92a46462c24cbc22e not found: ID does not exist" Oct 08 18:41:10 crc kubenswrapper[4859]: I1008 18:41:10.734737 4859 scope.go:117] "RemoveContainer" containerID="d11fcc2faca9bfc6b362592a97badbb5be6111273fbf417de785e9a0b0bff4c1" Oct 08 18:41:10 crc kubenswrapper[4859]: E1008 18:41:10.735119 4859 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"d11fcc2faca9bfc6b362592a97badbb5be6111273fbf417de785e9a0b0bff4c1\": container with ID starting with d11fcc2faca9bfc6b362592a97badbb5be6111273fbf417de785e9a0b0bff4c1 not found: ID does not exist" containerID="d11fcc2faca9bfc6b362592a97badbb5be6111273fbf417de785e9a0b0bff4c1" Oct 08 18:41:10 crc kubenswrapper[4859]: I1008 18:41:10.735150 4859 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"d11fcc2faca9bfc6b362592a97badbb5be6111273fbf417de785e9a0b0bff4c1"} err="failed to get container status \"d11fcc2faca9bfc6b362592a97badbb5be6111273fbf417de785e9a0b0bff4c1\": rpc error: code = NotFound desc = could not find container \"d11fcc2faca9bfc6b362592a97badbb5be6111273fbf417de785e9a0b0bff4c1\": container with ID starting with d11fcc2faca9bfc6b362592a97badbb5be6111273fbf417de785e9a0b0bff4c1 not found: ID does not exist" Oct 08 18:41:10 crc kubenswrapper[4859]: I1008 18:41:10.735172 4859 scope.go:117] "RemoveContainer" containerID="61bca07c0abce38eff0ca53124ca48bbb597309552d5cd0c68b0f0c0c415aaea" Oct 08 18:41:10 crc kubenswrapper[4859]: E1008 18:41:10.735428 4859 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"61bca07c0abce38eff0ca53124ca48bbb597309552d5cd0c68b0f0c0c415aaea\": container with ID starting with 61bca07c0abce38eff0ca53124ca48bbb597309552d5cd0c68b0f0c0c415aaea not found: ID does not exist" containerID="61bca07c0abce38eff0ca53124ca48bbb597309552d5cd0c68b0f0c0c415aaea" Oct 08 18:41:10 crc kubenswrapper[4859]: I1008 18:41:10.735472 4859 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"61bca07c0abce38eff0ca53124ca48bbb597309552d5cd0c68b0f0c0c415aaea"} err="failed to get container status \"61bca07c0abce38eff0ca53124ca48bbb597309552d5cd0c68b0f0c0c415aaea\": rpc error: code = NotFound desc = could not find container \"61bca07c0abce38eff0ca53124ca48bbb597309552d5cd0c68b0f0c0c415aaea\": container with ID starting with 61bca07c0abce38eff0ca53124ca48bbb597309552d5cd0c68b0f0c0c415aaea not found: ID does not exist" Oct 08 18:41:12 crc kubenswrapper[4859]: I1008 18:41:12.479482 4859 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="ec0f51e7-a093-46eb-8cc5-e4b9e1a0fe44" path="/var/lib/kubelet/pods/ec0f51e7-a093-46eb-8cc5-e4b9e1a0fe44/volumes" Oct 08 18:41:17 crc kubenswrapper[4859]: I1008 18:41:17.925196 4859 patch_prober.go:28] interesting pod/machine-config-daemon-82s52 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 08 18:41:17 crc kubenswrapper[4859]: I1008 18:41:17.925985 4859 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-82s52" podUID="b23a6a6c-9d92-4e7b-840e-55cfda873a2d" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 08 18:41:21 crc kubenswrapper[4859]: I1008 18:41:21.986327 4859 scope.go:117] "RemoveContainer" containerID="a912aba480685bc65f1310b409cf7397c5278c100778aa597be6b341c5f06d10" Oct 08 18:41:47 crc kubenswrapper[4859]: I1008 18:41:47.924937 4859 patch_prober.go:28] interesting pod/machine-config-daemon-82s52 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 08 18:41:47 crc kubenswrapper[4859]: I1008 18:41:47.925477 4859 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-82s52" podUID="b23a6a6c-9d92-4e7b-840e-55cfda873a2d" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 08 18:41:47 crc kubenswrapper[4859]: I1008 18:41:47.925528 4859 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-82s52" Oct 08 18:41:47 crc kubenswrapper[4859]: I1008 18:41:47.926234 4859 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"8eec6d3743e0f820548e10238303377094560d6d193dfbfe39f2b47960c426e8"} pod="openshift-machine-config-operator/machine-config-daemon-82s52" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Oct 08 18:41:47 crc kubenswrapper[4859]: I1008 18:41:47.926296 4859 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-82s52" podUID="b23a6a6c-9d92-4e7b-840e-55cfda873a2d" containerName="machine-config-daemon" containerID="cri-o://8eec6d3743e0f820548e10238303377094560d6d193dfbfe39f2b47960c426e8" gracePeriod=600 Oct 08 18:41:48 crc kubenswrapper[4859]: I1008 18:41:48.069715 4859 generic.go:334] "Generic (PLEG): container finished" podID="b23a6a6c-9d92-4e7b-840e-55cfda873a2d" containerID="8eec6d3743e0f820548e10238303377094560d6d193dfbfe39f2b47960c426e8" exitCode=0 Oct 08 18:41:48 crc kubenswrapper[4859]: I1008 18:41:48.069776 4859 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-82s52" event={"ID":"b23a6a6c-9d92-4e7b-840e-55cfda873a2d","Type":"ContainerDied","Data":"8eec6d3743e0f820548e10238303377094560d6d193dfbfe39f2b47960c426e8"} Oct 08 18:41:48 crc kubenswrapper[4859]: I1008 18:41:48.069809 4859 scope.go:117] "RemoveContainer" containerID="7ed110461431af172203421fc6cf54521a9bc279642133f6f29a4bb290304252" Oct 08 18:41:49 crc kubenswrapper[4859]: I1008 18:41:49.081738 4859 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-82s52" event={"ID":"b23a6a6c-9d92-4e7b-840e-55cfda873a2d","Type":"ContainerStarted","Data":"44aa7848c4f5acb003170608db1921ad40aa5b2f5464f1f85d423cd26d2e8155"} Oct 08 18:41:57 crc kubenswrapper[4859]: I1008 18:41:57.636841 4859 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-dr7lb"] Oct 08 18:41:57 crc kubenswrapper[4859]: E1008 18:41:57.637857 4859 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ec0f51e7-a093-46eb-8cc5-e4b9e1a0fe44" containerName="extract-utilities" Oct 08 18:41:57 crc kubenswrapper[4859]: I1008 18:41:57.637870 4859 state_mem.go:107] "Deleted CPUSet assignment" podUID="ec0f51e7-a093-46eb-8cc5-e4b9e1a0fe44" containerName="extract-utilities" Oct 08 18:41:57 crc kubenswrapper[4859]: E1008 18:41:57.637898 4859 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ec0f51e7-a093-46eb-8cc5-e4b9e1a0fe44" containerName="extract-content" Oct 08 18:41:57 crc kubenswrapper[4859]: I1008 18:41:57.637904 4859 state_mem.go:107] "Deleted CPUSet assignment" podUID="ec0f51e7-a093-46eb-8cc5-e4b9e1a0fe44" containerName="extract-content" Oct 08 18:41:57 crc kubenswrapper[4859]: E1008 18:41:57.637921 4859 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ec0f51e7-a093-46eb-8cc5-e4b9e1a0fe44" containerName="registry-server" Oct 08 18:41:57 crc kubenswrapper[4859]: I1008 18:41:57.637928 4859 state_mem.go:107] "Deleted CPUSet assignment" podUID="ec0f51e7-a093-46eb-8cc5-e4b9e1a0fe44" containerName="registry-server" Oct 08 18:41:57 crc kubenswrapper[4859]: I1008 18:41:57.638143 4859 memory_manager.go:354] "RemoveStaleState removing state" podUID="ec0f51e7-a093-46eb-8cc5-e4b9e1a0fe44" containerName="registry-server" Oct 08 18:41:57 crc kubenswrapper[4859]: I1008 18:41:57.639514 4859 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-dr7lb" Oct 08 18:41:57 crc kubenswrapper[4859]: I1008 18:41:57.652184 4859 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-dr7lb"] Oct 08 18:41:57 crc kubenswrapper[4859]: I1008 18:41:57.798104 4859 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-m29cj\" (UniqueName: \"kubernetes.io/projected/0b08244e-a6a0-4e1d-9951-699ab75ba8fa-kube-api-access-m29cj\") pod \"redhat-operators-dr7lb\" (UID: \"0b08244e-a6a0-4e1d-9951-699ab75ba8fa\") " pod="openshift-marketplace/redhat-operators-dr7lb" Oct 08 18:41:57 crc kubenswrapper[4859]: I1008 18:41:57.798363 4859 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/0b08244e-a6a0-4e1d-9951-699ab75ba8fa-catalog-content\") pod \"redhat-operators-dr7lb\" (UID: \"0b08244e-a6a0-4e1d-9951-699ab75ba8fa\") " pod="openshift-marketplace/redhat-operators-dr7lb" Oct 08 18:41:57 crc kubenswrapper[4859]: I1008 18:41:57.798579 4859 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/0b08244e-a6a0-4e1d-9951-699ab75ba8fa-utilities\") pod \"redhat-operators-dr7lb\" (UID: \"0b08244e-a6a0-4e1d-9951-699ab75ba8fa\") " pod="openshift-marketplace/redhat-operators-dr7lb" Oct 08 18:41:57 crc kubenswrapper[4859]: I1008 18:41:57.899990 4859 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-m29cj\" (UniqueName: \"kubernetes.io/projected/0b08244e-a6a0-4e1d-9951-699ab75ba8fa-kube-api-access-m29cj\") pod \"redhat-operators-dr7lb\" (UID: \"0b08244e-a6a0-4e1d-9951-699ab75ba8fa\") " pod="openshift-marketplace/redhat-operators-dr7lb" Oct 08 18:41:57 crc kubenswrapper[4859]: I1008 18:41:57.900091 4859 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/0b08244e-a6a0-4e1d-9951-699ab75ba8fa-catalog-content\") pod \"redhat-operators-dr7lb\" (UID: \"0b08244e-a6a0-4e1d-9951-699ab75ba8fa\") " pod="openshift-marketplace/redhat-operators-dr7lb" Oct 08 18:41:57 crc kubenswrapper[4859]: I1008 18:41:57.900144 4859 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/0b08244e-a6a0-4e1d-9951-699ab75ba8fa-utilities\") pod \"redhat-operators-dr7lb\" (UID: \"0b08244e-a6a0-4e1d-9951-699ab75ba8fa\") " pod="openshift-marketplace/redhat-operators-dr7lb" Oct 08 18:41:57 crc kubenswrapper[4859]: I1008 18:41:57.900659 4859 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/0b08244e-a6a0-4e1d-9951-699ab75ba8fa-catalog-content\") pod \"redhat-operators-dr7lb\" (UID: \"0b08244e-a6a0-4e1d-9951-699ab75ba8fa\") " pod="openshift-marketplace/redhat-operators-dr7lb" Oct 08 18:41:57 crc kubenswrapper[4859]: I1008 18:41:57.900723 4859 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/0b08244e-a6a0-4e1d-9951-699ab75ba8fa-utilities\") pod \"redhat-operators-dr7lb\" (UID: \"0b08244e-a6a0-4e1d-9951-699ab75ba8fa\") " pod="openshift-marketplace/redhat-operators-dr7lb" Oct 08 18:41:57 crc kubenswrapper[4859]: I1008 18:41:57.923631 4859 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-m29cj\" (UniqueName: \"kubernetes.io/projected/0b08244e-a6a0-4e1d-9951-699ab75ba8fa-kube-api-access-m29cj\") pod \"redhat-operators-dr7lb\" (UID: \"0b08244e-a6a0-4e1d-9951-699ab75ba8fa\") " pod="openshift-marketplace/redhat-operators-dr7lb" Oct 08 18:41:57 crc kubenswrapper[4859]: I1008 18:41:57.971772 4859 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-dr7lb" Oct 08 18:41:58 crc kubenswrapper[4859]: I1008 18:41:58.511846 4859 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-dr7lb"] Oct 08 18:41:59 crc kubenswrapper[4859]: I1008 18:41:59.170804 4859 generic.go:334] "Generic (PLEG): container finished" podID="0b08244e-a6a0-4e1d-9951-699ab75ba8fa" containerID="bfc990d78f4b69d12b3e60f4fa447f7b681c0c0e07129b0c97d99a2c1f98805c" exitCode=0 Oct 08 18:41:59 crc kubenswrapper[4859]: I1008 18:41:59.170847 4859 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-dr7lb" event={"ID":"0b08244e-a6a0-4e1d-9951-699ab75ba8fa","Type":"ContainerDied","Data":"bfc990d78f4b69d12b3e60f4fa447f7b681c0c0e07129b0c97d99a2c1f98805c"} Oct 08 18:41:59 crc kubenswrapper[4859]: I1008 18:41:59.171201 4859 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-dr7lb" event={"ID":"0b08244e-a6a0-4e1d-9951-699ab75ba8fa","Type":"ContainerStarted","Data":"ca010441e6b9309882a98aaf42b1fd17dd1958f7607387dbbfbebd3d6dfc2c35"} Oct 08 18:42:01 crc kubenswrapper[4859]: I1008 18:42:01.193442 4859 generic.go:334] "Generic (PLEG): container finished" podID="0b08244e-a6a0-4e1d-9951-699ab75ba8fa" containerID="c9a725b4adaf80ab36b9fa9280d211986718bcb1a371f0806e666b6770ae850f" exitCode=0 Oct 08 18:42:01 crc kubenswrapper[4859]: I1008 18:42:01.193707 4859 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-dr7lb" event={"ID":"0b08244e-a6a0-4e1d-9951-699ab75ba8fa","Type":"ContainerDied","Data":"c9a725b4adaf80ab36b9fa9280d211986718bcb1a371f0806e666b6770ae850f"} Oct 08 18:42:02 crc kubenswrapper[4859]: I1008 18:42:02.205293 4859 generic.go:334] "Generic (PLEG): container finished" podID="e0bf2482-ef25-48bb-9491-282a3eabfec0" containerID="88550ff98a8d1bf38a7e727936e9209ff31952ae03e2be30c7269f8ca9ce7c22" exitCode=0 Oct 08 18:42:02 crc kubenswrapper[4859]: I1008 18:42:02.205413 4859 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-nl8fp" event={"ID":"e0bf2482-ef25-48bb-9491-282a3eabfec0","Type":"ContainerDied","Data":"88550ff98a8d1bf38a7e727936e9209ff31952ae03e2be30c7269f8ca9ce7c22"} Oct 08 18:42:02 crc kubenswrapper[4859]: I1008 18:42:02.209176 4859 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-dr7lb" event={"ID":"0b08244e-a6a0-4e1d-9951-699ab75ba8fa","Type":"ContainerStarted","Data":"544f07a1e9a2b23ae301f3ec3eb340a1b875a1acecc3d12cb2c03e956a9ef768"} Oct 08 18:42:02 crc kubenswrapper[4859]: I1008 18:42:02.261026 4859 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-dr7lb" podStartSLOduration=2.838658744 podStartE2EDuration="5.261005274s" podCreationTimestamp="2025-10-08 18:41:57 +0000 UTC" firstStartedPulling="2025-10-08 18:41:59.172832272 +0000 UTC m=+1489.419671681" lastFinishedPulling="2025-10-08 18:42:01.595178832 +0000 UTC m=+1491.842018211" observedRunningTime="2025-10-08 18:42:02.253870355 +0000 UTC m=+1492.500709754" watchObservedRunningTime="2025-10-08 18:42:02.261005274 +0000 UTC m=+1492.507844673" Oct 08 18:42:03 crc kubenswrapper[4859]: I1008 18:42:03.604080 4859 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-nl8fp" Oct 08 18:42:03 crc kubenswrapper[4859]: I1008 18:42:03.716402 4859 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/e0bf2482-ef25-48bb-9491-282a3eabfec0-inventory\") pod \"e0bf2482-ef25-48bb-9491-282a3eabfec0\" (UID: \"e0bf2482-ef25-48bb-9491-282a3eabfec0\") " Oct 08 18:42:03 crc kubenswrapper[4859]: I1008 18:42:03.716462 4859 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bootstrap-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e0bf2482-ef25-48bb-9491-282a3eabfec0-bootstrap-combined-ca-bundle\") pod \"e0bf2482-ef25-48bb-9491-282a3eabfec0\" (UID: \"e0bf2482-ef25-48bb-9491-282a3eabfec0\") " Oct 08 18:42:03 crc kubenswrapper[4859]: I1008 18:42:03.716479 4859 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/e0bf2482-ef25-48bb-9491-282a3eabfec0-ssh-key\") pod \"e0bf2482-ef25-48bb-9491-282a3eabfec0\" (UID: \"e0bf2482-ef25-48bb-9491-282a3eabfec0\") " Oct 08 18:42:03 crc kubenswrapper[4859]: I1008 18:42:03.716638 4859 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-2nktn\" (UniqueName: \"kubernetes.io/projected/e0bf2482-ef25-48bb-9491-282a3eabfec0-kube-api-access-2nktn\") pod \"e0bf2482-ef25-48bb-9491-282a3eabfec0\" (UID: \"e0bf2482-ef25-48bb-9491-282a3eabfec0\") " Oct 08 18:42:03 crc kubenswrapper[4859]: I1008 18:42:03.721857 4859 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e0bf2482-ef25-48bb-9491-282a3eabfec0-bootstrap-combined-ca-bundle" (OuterVolumeSpecName: "bootstrap-combined-ca-bundle") pod "e0bf2482-ef25-48bb-9491-282a3eabfec0" (UID: "e0bf2482-ef25-48bb-9491-282a3eabfec0"). InnerVolumeSpecName "bootstrap-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 08 18:42:03 crc kubenswrapper[4859]: I1008 18:42:03.722262 4859 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/e0bf2482-ef25-48bb-9491-282a3eabfec0-kube-api-access-2nktn" (OuterVolumeSpecName: "kube-api-access-2nktn") pod "e0bf2482-ef25-48bb-9491-282a3eabfec0" (UID: "e0bf2482-ef25-48bb-9491-282a3eabfec0"). InnerVolumeSpecName "kube-api-access-2nktn". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 08 18:42:03 crc kubenswrapper[4859]: I1008 18:42:03.747997 4859 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e0bf2482-ef25-48bb-9491-282a3eabfec0-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "e0bf2482-ef25-48bb-9491-282a3eabfec0" (UID: "e0bf2482-ef25-48bb-9491-282a3eabfec0"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 08 18:42:03 crc kubenswrapper[4859]: I1008 18:42:03.749614 4859 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e0bf2482-ef25-48bb-9491-282a3eabfec0-inventory" (OuterVolumeSpecName: "inventory") pod "e0bf2482-ef25-48bb-9491-282a3eabfec0" (UID: "e0bf2482-ef25-48bb-9491-282a3eabfec0"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 08 18:42:03 crc kubenswrapper[4859]: I1008 18:42:03.818731 4859 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-2nktn\" (UniqueName: \"kubernetes.io/projected/e0bf2482-ef25-48bb-9491-282a3eabfec0-kube-api-access-2nktn\") on node \"crc\" DevicePath \"\"" Oct 08 18:42:03 crc kubenswrapper[4859]: I1008 18:42:03.818763 4859 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/e0bf2482-ef25-48bb-9491-282a3eabfec0-inventory\") on node \"crc\" DevicePath \"\"" Oct 08 18:42:03 crc kubenswrapper[4859]: I1008 18:42:03.818772 4859 reconciler_common.go:293] "Volume detached for volume \"bootstrap-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e0bf2482-ef25-48bb-9491-282a3eabfec0-bootstrap-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 08 18:42:03 crc kubenswrapper[4859]: I1008 18:42:03.818780 4859 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/e0bf2482-ef25-48bb-9491-282a3eabfec0-ssh-key\") on node \"crc\" DevicePath \"\"" Oct 08 18:42:04 crc kubenswrapper[4859]: I1008 18:42:04.231929 4859 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-nl8fp" event={"ID":"e0bf2482-ef25-48bb-9491-282a3eabfec0","Type":"ContainerDied","Data":"868c2bf1e0ba4f83a24a38305f6c6b6175ea0fedc6df3bb0c9468557bc959c19"} Oct 08 18:42:04 crc kubenswrapper[4859]: I1008 18:42:04.232503 4859 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="868c2bf1e0ba4f83a24a38305f6c6b6175ea0fedc6df3bb0c9468557bc959c19" Oct 08 18:42:04 crc kubenswrapper[4859]: I1008 18:42:04.232019 4859 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-nl8fp" Oct 08 18:42:04 crc kubenswrapper[4859]: I1008 18:42:04.335330 4859 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/download-cache-edpm-deployment-openstack-edpm-ipam-hb4tw"] Oct 08 18:42:04 crc kubenswrapper[4859]: E1008 18:42:04.335745 4859 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e0bf2482-ef25-48bb-9491-282a3eabfec0" containerName="bootstrap-edpm-deployment-openstack-edpm-ipam" Oct 08 18:42:04 crc kubenswrapper[4859]: I1008 18:42:04.335763 4859 state_mem.go:107] "Deleted CPUSet assignment" podUID="e0bf2482-ef25-48bb-9491-282a3eabfec0" containerName="bootstrap-edpm-deployment-openstack-edpm-ipam" Oct 08 18:42:04 crc kubenswrapper[4859]: I1008 18:42:04.335945 4859 memory_manager.go:354] "RemoveStaleState removing state" podUID="e0bf2482-ef25-48bb-9491-282a3eabfec0" containerName="bootstrap-edpm-deployment-openstack-edpm-ipam" Oct 08 18:42:04 crc kubenswrapper[4859]: I1008 18:42:04.336569 4859 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/download-cache-edpm-deployment-openstack-edpm-ipam-hb4tw" Oct 08 18:42:04 crc kubenswrapper[4859]: I1008 18:42:04.340451 4859 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Oct 08 18:42:04 crc kubenswrapper[4859]: I1008 18:42:04.340768 4859 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Oct 08 18:42:04 crc kubenswrapper[4859]: I1008 18:42:04.340903 4859 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Oct 08 18:42:04 crc kubenswrapper[4859]: I1008 18:42:04.341022 4859 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-gxqst" Oct 08 18:42:04 crc kubenswrapper[4859]: I1008 18:42:04.344781 4859 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/download-cache-edpm-deployment-openstack-edpm-ipam-hb4tw"] Oct 08 18:42:04 crc kubenswrapper[4859]: I1008 18:42:04.531087 4859 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/2db15fd4-e39e-474c-b274-53262e92bccb-inventory\") pod \"download-cache-edpm-deployment-openstack-edpm-ipam-hb4tw\" (UID: \"2db15fd4-e39e-474c-b274-53262e92bccb\") " pod="openstack/download-cache-edpm-deployment-openstack-edpm-ipam-hb4tw" Oct 08 18:42:04 crc kubenswrapper[4859]: I1008 18:42:04.531140 4859 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-llsx8\" (UniqueName: \"kubernetes.io/projected/2db15fd4-e39e-474c-b274-53262e92bccb-kube-api-access-llsx8\") pod \"download-cache-edpm-deployment-openstack-edpm-ipam-hb4tw\" (UID: \"2db15fd4-e39e-474c-b274-53262e92bccb\") " pod="openstack/download-cache-edpm-deployment-openstack-edpm-ipam-hb4tw" Oct 08 18:42:04 crc kubenswrapper[4859]: I1008 18:42:04.531221 4859 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/2db15fd4-e39e-474c-b274-53262e92bccb-ssh-key\") pod \"download-cache-edpm-deployment-openstack-edpm-ipam-hb4tw\" (UID: \"2db15fd4-e39e-474c-b274-53262e92bccb\") " pod="openstack/download-cache-edpm-deployment-openstack-edpm-ipam-hb4tw" Oct 08 18:42:04 crc kubenswrapper[4859]: I1008 18:42:04.633384 4859 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/2db15fd4-e39e-474c-b274-53262e92bccb-inventory\") pod \"download-cache-edpm-deployment-openstack-edpm-ipam-hb4tw\" (UID: \"2db15fd4-e39e-474c-b274-53262e92bccb\") " pod="openstack/download-cache-edpm-deployment-openstack-edpm-ipam-hb4tw" Oct 08 18:42:04 crc kubenswrapper[4859]: I1008 18:42:04.633454 4859 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-llsx8\" (UniqueName: \"kubernetes.io/projected/2db15fd4-e39e-474c-b274-53262e92bccb-kube-api-access-llsx8\") pod \"download-cache-edpm-deployment-openstack-edpm-ipam-hb4tw\" (UID: \"2db15fd4-e39e-474c-b274-53262e92bccb\") " pod="openstack/download-cache-edpm-deployment-openstack-edpm-ipam-hb4tw" Oct 08 18:42:04 crc kubenswrapper[4859]: I1008 18:42:04.633519 4859 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/2db15fd4-e39e-474c-b274-53262e92bccb-ssh-key\") pod \"download-cache-edpm-deployment-openstack-edpm-ipam-hb4tw\" (UID: \"2db15fd4-e39e-474c-b274-53262e92bccb\") " pod="openstack/download-cache-edpm-deployment-openstack-edpm-ipam-hb4tw" Oct 08 18:42:04 crc kubenswrapper[4859]: I1008 18:42:04.639834 4859 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/2db15fd4-e39e-474c-b274-53262e92bccb-inventory\") pod \"download-cache-edpm-deployment-openstack-edpm-ipam-hb4tw\" (UID: \"2db15fd4-e39e-474c-b274-53262e92bccb\") " pod="openstack/download-cache-edpm-deployment-openstack-edpm-ipam-hb4tw" Oct 08 18:42:04 crc kubenswrapper[4859]: I1008 18:42:04.639995 4859 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/2db15fd4-e39e-474c-b274-53262e92bccb-ssh-key\") pod \"download-cache-edpm-deployment-openstack-edpm-ipam-hb4tw\" (UID: \"2db15fd4-e39e-474c-b274-53262e92bccb\") " pod="openstack/download-cache-edpm-deployment-openstack-edpm-ipam-hb4tw" Oct 08 18:42:04 crc kubenswrapper[4859]: I1008 18:42:04.668646 4859 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-llsx8\" (UniqueName: \"kubernetes.io/projected/2db15fd4-e39e-474c-b274-53262e92bccb-kube-api-access-llsx8\") pod \"download-cache-edpm-deployment-openstack-edpm-ipam-hb4tw\" (UID: \"2db15fd4-e39e-474c-b274-53262e92bccb\") " pod="openstack/download-cache-edpm-deployment-openstack-edpm-ipam-hb4tw" Oct 08 18:42:04 crc kubenswrapper[4859]: I1008 18:42:04.699728 4859 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/download-cache-edpm-deployment-openstack-edpm-ipam-hb4tw" Oct 08 18:42:05 crc kubenswrapper[4859]: I1008 18:42:05.265019 4859 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/download-cache-edpm-deployment-openstack-edpm-ipam-hb4tw"] Oct 08 18:42:05 crc kubenswrapper[4859]: W1008 18:42:05.268926 4859 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod2db15fd4_e39e_474c_b274_53262e92bccb.slice/crio-47209fab83d608d93b7e9b762752e6cc265d877e8499d2ef665362c77c07cb48 WatchSource:0}: Error finding container 47209fab83d608d93b7e9b762752e6cc265d877e8499d2ef665362c77c07cb48: Status 404 returned error can't find the container with id 47209fab83d608d93b7e9b762752e6cc265d877e8499d2ef665362c77c07cb48 Oct 08 18:42:06 crc kubenswrapper[4859]: I1008 18:42:06.252293 4859 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/download-cache-edpm-deployment-openstack-edpm-ipam-hb4tw" event={"ID":"2db15fd4-e39e-474c-b274-53262e92bccb","Type":"ContainerStarted","Data":"3e9b49aaa1a459dc2787420eed44470f2a717ca606c2b15ba3dcfde033efe02c"} Oct 08 18:42:06 crc kubenswrapper[4859]: I1008 18:42:06.252891 4859 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/download-cache-edpm-deployment-openstack-edpm-ipam-hb4tw" event={"ID":"2db15fd4-e39e-474c-b274-53262e92bccb","Type":"ContainerStarted","Data":"47209fab83d608d93b7e9b762752e6cc265d877e8499d2ef665362c77c07cb48"} Oct 08 18:42:07 crc kubenswrapper[4859]: I1008 18:42:07.974274 4859 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-dr7lb" Oct 08 18:42:07 crc kubenswrapper[4859]: I1008 18:42:07.975200 4859 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-dr7lb" Oct 08 18:42:08 crc kubenswrapper[4859]: I1008 18:42:08.034486 4859 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-dr7lb" Oct 08 18:42:08 crc kubenswrapper[4859]: I1008 18:42:08.067159 4859 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/download-cache-edpm-deployment-openstack-edpm-ipam-hb4tw" podStartSLOduration=3.637949117 podStartE2EDuration="4.067137139s" podCreationTimestamp="2025-10-08 18:42:04 +0000 UTC" firstStartedPulling="2025-10-08 18:42:05.272031133 +0000 UTC m=+1495.518870512" lastFinishedPulling="2025-10-08 18:42:05.701219155 +0000 UTC m=+1495.948058534" observedRunningTime="2025-10-08 18:42:06.276011876 +0000 UTC m=+1496.522851255" watchObservedRunningTime="2025-10-08 18:42:08.067137139 +0000 UTC m=+1498.313976548" Oct 08 18:42:08 crc kubenswrapper[4859]: I1008 18:42:08.315558 4859 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-dr7lb" Oct 08 18:42:08 crc kubenswrapper[4859]: I1008 18:42:08.377155 4859 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-dr7lb"] Oct 08 18:42:10 crc kubenswrapper[4859]: I1008 18:42:10.290466 4859 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-operators-dr7lb" podUID="0b08244e-a6a0-4e1d-9951-699ab75ba8fa" containerName="registry-server" containerID="cri-o://544f07a1e9a2b23ae301f3ec3eb340a1b875a1acecc3d12cb2c03e956a9ef768" gracePeriod=2 Oct 08 18:42:10 crc kubenswrapper[4859]: I1008 18:42:10.739743 4859 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-dr7lb" Oct 08 18:42:10 crc kubenswrapper[4859]: I1008 18:42:10.869623 4859 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-m29cj\" (UniqueName: \"kubernetes.io/projected/0b08244e-a6a0-4e1d-9951-699ab75ba8fa-kube-api-access-m29cj\") pod \"0b08244e-a6a0-4e1d-9951-699ab75ba8fa\" (UID: \"0b08244e-a6a0-4e1d-9951-699ab75ba8fa\") " Oct 08 18:42:10 crc kubenswrapper[4859]: I1008 18:42:10.869772 4859 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/0b08244e-a6a0-4e1d-9951-699ab75ba8fa-utilities\") pod \"0b08244e-a6a0-4e1d-9951-699ab75ba8fa\" (UID: \"0b08244e-a6a0-4e1d-9951-699ab75ba8fa\") " Oct 08 18:42:10 crc kubenswrapper[4859]: I1008 18:42:10.869853 4859 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/0b08244e-a6a0-4e1d-9951-699ab75ba8fa-catalog-content\") pod \"0b08244e-a6a0-4e1d-9951-699ab75ba8fa\" (UID: \"0b08244e-a6a0-4e1d-9951-699ab75ba8fa\") " Oct 08 18:42:10 crc kubenswrapper[4859]: I1008 18:42:10.870997 4859 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/0b08244e-a6a0-4e1d-9951-699ab75ba8fa-utilities" (OuterVolumeSpecName: "utilities") pod "0b08244e-a6a0-4e1d-9951-699ab75ba8fa" (UID: "0b08244e-a6a0-4e1d-9951-699ab75ba8fa"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 08 18:42:10 crc kubenswrapper[4859]: I1008 18:42:10.883187 4859 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/0b08244e-a6a0-4e1d-9951-699ab75ba8fa-kube-api-access-m29cj" (OuterVolumeSpecName: "kube-api-access-m29cj") pod "0b08244e-a6a0-4e1d-9951-699ab75ba8fa" (UID: "0b08244e-a6a0-4e1d-9951-699ab75ba8fa"). InnerVolumeSpecName "kube-api-access-m29cj". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 08 18:42:10 crc kubenswrapper[4859]: I1008 18:42:10.949479 4859 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/0b08244e-a6a0-4e1d-9951-699ab75ba8fa-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "0b08244e-a6a0-4e1d-9951-699ab75ba8fa" (UID: "0b08244e-a6a0-4e1d-9951-699ab75ba8fa"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 08 18:42:10 crc kubenswrapper[4859]: I1008 18:42:10.972310 4859 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-m29cj\" (UniqueName: \"kubernetes.io/projected/0b08244e-a6a0-4e1d-9951-699ab75ba8fa-kube-api-access-m29cj\") on node \"crc\" DevicePath \"\"" Oct 08 18:42:10 crc kubenswrapper[4859]: I1008 18:42:10.972361 4859 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/0b08244e-a6a0-4e1d-9951-699ab75ba8fa-utilities\") on node \"crc\" DevicePath \"\"" Oct 08 18:42:10 crc kubenswrapper[4859]: I1008 18:42:10.972374 4859 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/0b08244e-a6a0-4e1d-9951-699ab75ba8fa-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 08 18:42:11 crc kubenswrapper[4859]: I1008 18:42:11.330833 4859 generic.go:334] "Generic (PLEG): container finished" podID="0b08244e-a6a0-4e1d-9951-699ab75ba8fa" containerID="544f07a1e9a2b23ae301f3ec3eb340a1b875a1acecc3d12cb2c03e956a9ef768" exitCode=0 Oct 08 18:42:11 crc kubenswrapper[4859]: I1008 18:42:11.330892 4859 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-dr7lb" Oct 08 18:42:11 crc kubenswrapper[4859]: I1008 18:42:11.330903 4859 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-dr7lb" event={"ID":"0b08244e-a6a0-4e1d-9951-699ab75ba8fa","Type":"ContainerDied","Data":"544f07a1e9a2b23ae301f3ec3eb340a1b875a1acecc3d12cb2c03e956a9ef768"} Oct 08 18:42:11 crc kubenswrapper[4859]: I1008 18:42:11.332427 4859 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-dr7lb" event={"ID":"0b08244e-a6a0-4e1d-9951-699ab75ba8fa","Type":"ContainerDied","Data":"ca010441e6b9309882a98aaf42b1fd17dd1958f7607387dbbfbebd3d6dfc2c35"} Oct 08 18:42:11 crc kubenswrapper[4859]: I1008 18:42:11.332480 4859 scope.go:117] "RemoveContainer" containerID="544f07a1e9a2b23ae301f3ec3eb340a1b875a1acecc3d12cb2c03e956a9ef768" Oct 08 18:42:11 crc kubenswrapper[4859]: I1008 18:42:11.380651 4859 scope.go:117] "RemoveContainer" containerID="c9a725b4adaf80ab36b9fa9280d211986718bcb1a371f0806e666b6770ae850f" Oct 08 18:42:11 crc kubenswrapper[4859]: I1008 18:42:11.389990 4859 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-dr7lb"] Oct 08 18:42:11 crc kubenswrapper[4859]: I1008 18:42:11.400189 4859 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-operators-dr7lb"] Oct 08 18:42:11 crc kubenswrapper[4859]: I1008 18:42:11.412344 4859 scope.go:117] "RemoveContainer" containerID="bfc990d78f4b69d12b3e60f4fa447f7b681c0c0e07129b0c97d99a2c1f98805c" Oct 08 18:42:11 crc kubenswrapper[4859]: I1008 18:42:11.457705 4859 scope.go:117] "RemoveContainer" containerID="544f07a1e9a2b23ae301f3ec3eb340a1b875a1acecc3d12cb2c03e956a9ef768" Oct 08 18:42:11 crc kubenswrapper[4859]: E1008 18:42:11.458141 4859 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"544f07a1e9a2b23ae301f3ec3eb340a1b875a1acecc3d12cb2c03e956a9ef768\": container with ID starting with 544f07a1e9a2b23ae301f3ec3eb340a1b875a1acecc3d12cb2c03e956a9ef768 not found: ID does not exist" containerID="544f07a1e9a2b23ae301f3ec3eb340a1b875a1acecc3d12cb2c03e956a9ef768" Oct 08 18:42:11 crc kubenswrapper[4859]: I1008 18:42:11.458176 4859 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"544f07a1e9a2b23ae301f3ec3eb340a1b875a1acecc3d12cb2c03e956a9ef768"} err="failed to get container status \"544f07a1e9a2b23ae301f3ec3eb340a1b875a1acecc3d12cb2c03e956a9ef768\": rpc error: code = NotFound desc = could not find container \"544f07a1e9a2b23ae301f3ec3eb340a1b875a1acecc3d12cb2c03e956a9ef768\": container with ID starting with 544f07a1e9a2b23ae301f3ec3eb340a1b875a1acecc3d12cb2c03e956a9ef768 not found: ID does not exist" Oct 08 18:42:11 crc kubenswrapper[4859]: I1008 18:42:11.458203 4859 scope.go:117] "RemoveContainer" containerID="c9a725b4adaf80ab36b9fa9280d211986718bcb1a371f0806e666b6770ae850f" Oct 08 18:42:11 crc kubenswrapper[4859]: E1008 18:42:11.458493 4859 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"c9a725b4adaf80ab36b9fa9280d211986718bcb1a371f0806e666b6770ae850f\": container with ID starting with c9a725b4adaf80ab36b9fa9280d211986718bcb1a371f0806e666b6770ae850f not found: ID does not exist" containerID="c9a725b4adaf80ab36b9fa9280d211986718bcb1a371f0806e666b6770ae850f" Oct 08 18:42:11 crc kubenswrapper[4859]: I1008 18:42:11.458518 4859 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"c9a725b4adaf80ab36b9fa9280d211986718bcb1a371f0806e666b6770ae850f"} err="failed to get container status \"c9a725b4adaf80ab36b9fa9280d211986718bcb1a371f0806e666b6770ae850f\": rpc error: code = NotFound desc = could not find container \"c9a725b4adaf80ab36b9fa9280d211986718bcb1a371f0806e666b6770ae850f\": container with ID starting with c9a725b4adaf80ab36b9fa9280d211986718bcb1a371f0806e666b6770ae850f not found: ID does not exist" Oct 08 18:42:11 crc kubenswrapper[4859]: I1008 18:42:11.458532 4859 scope.go:117] "RemoveContainer" containerID="bfc990d78f4b69d12b3e60f4fa447f7b681c0c0e07129b0c97d99a2c1f98805c" Oct 08 18:42:11 crc kubenswrapper[4859]: E1008 18:42:11.458810 4859 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"bfc990d78f4b69d12b3e60f4fa447f7b681c0c0e07129b0c97d99a2c1f98805c\": container with ID starting with bfc990d78f4b69d12b3e60f4fa447f7b681c0c0e07129b0c97d99a2c1f98805c not found: ID does not exist" containerID="bfc990d78f4b69d12b3e60f4fa447f7b681c0c0e07129b0c97d99a2c1f98805c" Oct 08 18:42:11 crc kubenswrapper[4859]: I1008 18:42:11.458830 4859 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"bfc990d78f4b69d12b3e60f4fa447f7b681c0c0e07129b0c97d99a2c1f98805c"} err="failed to get container status \"bfc990d78f4b69d12b3e60f4fa447f7b681c0c0e07129b0c97d99a2c1f98805c\": rpc error: code = NotFound desc = could not find container \"bfc990d78f4b69d12b3e60f4fa447f7b681c0c0e07129b0c97d99a2c1f98805c\": container with ID starting with bfc990d78f4b69d12b3e60f4fa447f7b681c0c0e07129b0c97d99a2c1f98805c not found: ID does not exist" Oct 08 18:42:12 crc kubenswrapper[4859]: I1008 18:42:12.481218 4859 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="0b08244e-a6a0-4e1d-9951-699ab75ba8fa" path="/var/lib/kubelet/pods/0b08244e-a6a0-4e1d-9951-699ab75ba8fa/volumes" Oct 08 18:42:42 crc kubenswrapper[4859]: I1008 18:42:42.041393 4859 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/placement-db-create-vrx2g"] Oct 08 18:42:42 crc kubenswrapper[4859]: I1008 18:42:42.053148 4859 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/placement-db-create-vrx2g"] Oct 08 18:42:42 crc kubenswrapper[4859]: I1008 18:42:42.490253 4859 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="5330b621-b843-410c-b41a-ffeeb10ea9e4" path="/var/lib/kubelet/pods/5330b621-b843-410c-b41a-ffeeb10ea9e4/volumes" Oct 08 18:42:44 crc kubenswrapper[4859]: I1008 18:42:44.025729 4859 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-db-create-cgx9g"] Oct 08 18:42:44 crc kubenswrapper[4859]: I1008 18:42:44.033953 4859 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/glance-db-create-cgx9g"] Oct 08 18:42:44 crc kubenswrapper[4859]: I1008 18:42:44.483651 4859 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="a380400a-453c-4cc9-b24c-540f762e1404" path="/var/lib/kubelet/pods/a380400a-453c-4cc9-b24c-540f762e1404/volumes" Oct 08 18:42:52 crc kubenswrapper[4859]: I1008 18:42:52.037119 4859 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/placement-b31f-account-create-r48bf"] Oct 08 18:42:52 crc kubenswrapper[4859]: I1008 18:42:52.048345 4859 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/keystone-db-create-9sdlj"] Oct 08 18:42:52 crc kubenswrapper[4859]: I1008 18:42:52.059074 4859 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/placement-b31f-account-create-r48bf"] Oct 08 18:42:52 crc kubenswrapper[4859]: I1008 18:42:52.069518 4859 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/keystone-db-create-9sdlj"] Oct 08 18:42:52 crc kubenswrapper[4859]: I1008 18:42:52.486162 4859 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="1e1f9094-586a-4f98-a7f6-32ca4816fa8b" path="/var/lib/kubelet/pods/1e1f9094-586a-4f98-a7f6-32ca4816fa8b/volumes" Oct 08 18:42:52 crc kubenswrapper[4859]: I1008 18:42:52.488464 4859 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="823f174d-4981-4db7-97cb-18d5ec2eead5" path="/var/lib/kubelet/pods/823f174d-4981-4db7-97cb-18d5ec2eead5/volumes" Oct 08 18:42:55 crc kubenswrapper[4859]: I1008 18:42:55.038424 4859 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-f62d-account-create-pcklr"] Oct 08 18:42:55 crc kubenswrapper[4859]: I1008 18:42:55.049945 4859 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/glance-f62d-account-create-pcklr"] Oct 08 18:42:56 crc kubenswrapper[4859]: I1008 18:42:56.480430 4859 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="c3879c1b-a548-4090-8709-777af70d9578" path="/var/lib/kubelet/pods/c3879c1b-a548-4090-8709-777af70d9578/volumes" Oct 08 18:43:02 crc kubenswrapper[4859]: I1008 18:43:02.041287 4859 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/keystone-f7b9-account-create-47kvw"] Oct 08 18:43:02 crc kubenswrapper[4859]: I1008 18:43:02.052734 4859 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/keystone-f7b9-account-create-47kvw"] Oct 08 18:43:02 crc kubenswrapper[4859]: I1008 18:43:02.488288 4859 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="b44043d2-d505-4e7f-8431-f42a3738ca8a" path="/var/lib/kubelet/pods/b44043d2-d505-4e7f-8431-f42a3738ca8a/volumes" Oct 08 18:43:19 crc kubenswrapper[4859]: I1008 18:43:19.032924 4859 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/cinder-db-create-wczl5"] Oct 08 18:43:19 crc kubenswrapper[4859]: I1008 18:43:19.048428 4859 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/neutron-db-create-p8mb2"] Oct 08 18:43:19 crc kubenswrapper[4859]: I1008 18:43:19.057681 4859 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/neutron-db-create-p8mb2"] Oct 08 18:43:19 crc kubenswrapper[4859]: I1008 18:43:19.064837 4859 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/barbican-db-create-dbftt"] Oct 08 18:43:19 crc kubenswrapper[4859]: I1008 18:43:19.071799 4859 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/cinder-db-create-wczl5"] Oct 08 18:43:19 crc kubenswrapper[4859]: I1008 18:43:19.080995 4859 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/barbican-db-create-dbftt"] Oct 08 18:43:20 crc kubenswrapper[4859]: I1008 18:43:20.482737 4859 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="84999f2e-5fbf-46d7-8334-436d1f9b12d7" path="/var/lib/kubelet/pods/84999f2e-5fbf-46d7-8334-436d1f9b12d7/volumes" Oct 08 18:43:20 crc kubenswrapper[4859]: I1008 18:43:20.483559 4859 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="b51059c4-e80f-4c66-be2d-c4ec6b9deaf9" path="/var/lib/kubelet/pods/b51059c4-e80f-4c66-be2d-c4ec6b9deaf9/volumes" Oct 08 18:43:20 crc kubenswrapper[4859]: I1008 18:43:20.484068 4859 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="c0da13c2-e693-4da8-83fa-ae3032d30ec5" path="/var/lib/kubelet/pods/c0da13c2-e693-4da8-83fa-ae3032d30ec5/volumes" Oct 08 18:43:22 crc kubenswrapper[4859]: I1008 18:43:22.094366 4859 scope.go:117] "RemoveContainer" containerID="fa34d069118c0da8f3d0adada988ca9e9482254ede94e90a33bd0b93333db806" Oct 08 18:43:22 crc kubenswrapper[4859]: I1008 18:43:22.126137 4859 scope.go:117] "RemoveContainer" containerID="c181b0f9e95d88022634d1baf22f6c42e474a38ddaee7084537b73bb541295e9" Oct 08 18:43:22 crc kubenswrapper[4859]: I1008 18:43:22.171598 4859 scope.go:117] "RemoveContainer" containerID="dca6842589b2acb69ce6753cc593071dc1f48239b1c6a978c540c3ac3a1cb9a3" Oct 08 18:43:22 crc kubenswrapper[4859]: I1008 18:43:22.224482 4859 scope.go:117] "RemoveContainer" containerID="166ef925709e4e3e1b6f80efc26c123ca3d703657de04aacb1d2659ec9dd5886" Oct 08 18:43:22 crc kubenswrapper[4859]: I1008 18:43:22.277924 4859 scope.go:117] "RemoveContainer" containerID="be9e6049d5bf7c698d69ee258fe38046579bf0fe9d4b9edec6903e4286455a14" Oct 08 18:43:22 crc kubenswrapper[4859]: I1008 18:43:22.315702 4859 scope.go:117] "RemoveContainer" containerID="cea203469b8abcb099d7c42979e0bdaf41e94673bba18642f9ca7349d2deaf3e" Oct 08 18:43:22 crc kubenswrapper[4859]: I1008 18:43:22.356583 4859 scope.go:117] "RemoveContainer" containerID="2f3570ff9aac41b4211ea2b76be831b98d152e30595b7dface9313ee3be68aed" Oct 08 18:43:22 crc kubenswrapper[4859]: I1008 18:43:22.384890 4859 scope.go:117] "RemoveContainer" containerID="2b8b1c1c7faf850e99249b3d02dba4845cbd57dde84a18c9f93e8f54883c8d9e" Oct 08 18:43:22 crc kubenswrapper[4859]: I1008 18:43:22.403974 4859 scope.go:117] "RemoveContainer" containerID="b402f3bb9b51f6d60caa8e68d4522dc322ffd9e059d829c1c745f190aaee7b39" Oct 08 18:43:22 crc kubenswrapper[4859]: I1008 18:43:22.438561 4859 scope.go:117] "RemoveContainer" containerID="74719fdf7321a3335282b9fa039ba753eb5cb2640408902b649f9ee2b83c0268" Oct 08 18:43:22 crc kubenswrapper[4859]: I1008 18:43:22.467756 4859 scope.go:117] "RemoveContainer" containerID="bbdcd7b0e44c3f12f0238a79fe0cc08a70a78cfa68fbe75279c10fe65fab927e" Oct 08 18:43:24 crc kubenswrapper[4859]: I1008 18:43:24.033352 4859 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-db-sync-xk8qg"] Oct 08 18:43:24 crc kubenswrapper[4859]: I1008 18:43:24.045484 4859 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/glance-db-sync-xk8qg"] Oct 08 18:43:24 crc kubenswrapper[4859]: I1008 18:43:24.488276 4859 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="9c2ff226-ee9a-4cb3-b66f-9df0c3497748" path="/var/lib/kubelet/pods/9c2ff226-ee9a-4cb3-b66f-9df0c3497748/volumes" Oct 08 18:43:32 crc kubenswrapper[4859]: I1008 18:43:32.053678 4859 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/barbican-c89b-account-create-bd78m"] Oct 08 18:43:32 crc kubenswrapper[4859]: I1008 18:43:32.063246 4859 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/keystone-db-sync-gsq5w"] Oct 08 18:43:32 crc kubenswrapper[4859]: I1008 18:43:32.075393 4859 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/keystone-db-sync-gsq5w"] Oct 08 18:43:32 crc kubenswrapper[4859]: I1008 18:43:32.082276 4859 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/cinder-158c-account-create-vvkvz"] Oct 08 18:43:32 crc kubenswrapper[4859]: I1008 18:43:32.091895 4859 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/neutron-d728-account-create-prt4m"] Oct 08 18:43:32 crc kubenswrapper[4859]: I1008 18:43:32.098886 4859 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/barbican-c89b-account-create-bd78m"] Oct 08 18:43:32 crc kubenswrapper[4859]: I1008 18:43:32.107170 4859 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/cinder-158c-account-create-vvkvz"] Oct 08 18:43:32 crc kubenswrapper[4859]: I1008 18:43:32.113654 4859 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/neutron-d728-account-create-prt4m"] Oct 08 18:43:32 crc kubenswrapper[4859]: I1008 18:43:32.481440 4859 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="498d1fd9-4fd1-4dce-a122-6f79d4274882" path="/var/lib/kubelet/pods/498d1fd9-4fd1-4dce-a122-6f79d4274882/volumes" Oct 08 18:43:32 crc kubenswrapper[4859]: I1008 18:43:32.482016 4859 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="4a64adc4-3689-4d56-b514-f86372be9938" path="/var/lib/kubelet/pods/4a64adc4-3689-4d56-b514-f86372be9938/volumes" Oct 08 18:43:32 crc kubenswrapper[4859]: I1008 18:43:32.482482 4859 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="8503bae7-a295-4935-bcc9-a19971c4f417" path="/var/lib/kubelet/pods/8503bae7-a295-4935-bcc9-a19971c4f417/volumes" Oct 08 18:43:32 crc kubenswrapper[4859]: I1008 18:43:32.483058 4859 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="ff748840-6d03-4e9a-8284-a4ca9d753278" path="/var/lib/kubelet/pods/ff748840-6d03-4e9a-8284-a4ca9d753278/volumes" Oct 08 18:43:44 crc kubenswrapper[4859]: I1008 18:43:44.268388 4859 generic.go:334] "Generic (PLEG): container finished" podID="2db15fd4-e39e-474c-b274-53262e92bccb" containerID="3e9b49aaa1a459dc2787420eed44470f2a717ca606c2b15ba3dcfde033efe02c" exitCode=0 Oct 08 18:43:44 crc kubenswrapper[4859]: I1008 18:43:44.268471 4859 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/download-cache-edpm-deployment-openstack-edpm-ipam-hb4tw" event={"ID":"2db15fd4-e39e-474c-b274-53262e92bccb","Type":"ContainerDied","Data":"3e9b49aaa1a459dc2787420eed44470f2a717ca606c2b15ba3dcfde033efe02c"} Oct 08 18:43:45 crc kubenswrapper[4859]: I1008 18:43:45.655527 4859 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/download-cache-edpm-deployment-openstack-edpm-ipam-hb4tw" Oct 08 18:43:45 crc kubenswrapper[4859]: I1008 18:43:45.718976 4859 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-llsx8\" (UniqueName: \"kubernetes.io/projected/2db15fd4-e39e-474c-b274-53262e92bccb-kube-api-access-llsx8\") pod \"2db15fd4-e39e-474c-b274-53262e92bccb\" (UID: \"2db15fd4-e39e-474c-b274-53262e92bccb\") " Oct 08 18:43:45 crc kubenswrapper[4859]: I1008 18:43:45.719094 4859 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/2db15fd4-e39e-474c-b274-53262e92bccb-ssh-key\") pod \"2db15fd4-e39e-474c-b274-53262e92bccb\" (UID: \"2db15fd4-e39e-474c-b274-53262e92bccb\") " Oct 08 18:43:45 crc kubenswrapper[4859]: I1008 18:43:45.719233 4859 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/2db15fd4-e39e-474c-b274-53262e92bccb-inventory\") pod \"2db15fd4-e39e-474c-b274-53262e92bccb\" (UID: \"2db15fd4-e39e-474c-b274-53262e92bccb\") " Oct 08 18:43:45 crc kubenswrapper[4859]: I1008 18:43:45.726904 4859 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/2db15fd4-e39e-474c-b274-53262e92bccb-kube-api-access-llsx8" (OuterVolumeSpecName: "kube-api-access-llsx8") pod "2db15fd4-e39e-474c-b274-53262e92bccb" (UID: "2db15fd4-e39e-474c-b274-53262e92bccb"). InnerVolumeSpecName "kube-api-access-llsx8". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 08 18:43:45 crc kubenswrapper[4859]: I1008 18:43:45.745436 4859 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/2db15fd4-e39e-474c-b274-53262e92bccb-inventory" (OuterVolumeSpecName: "inventory") pod "2db15fd4-e39e-474c-b274-53262e92bccb" (UID: "2db15fd4-e39e-474c-b274-53262e92bccb"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 08 18:43:45 crc kubenswrapper[4859]: I1008 18:43:45.755387 4859 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/2db15fd4-e39e-474c-b274-53262e92bccb-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "2db15fd4-e39e-474c-b274-53262e92bccb" (UID: "2db15fd4-e39e-474c-b274-53262e92bccb"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 08 18:43:45 crc kubenswrapper[4859]: I1008 18:43:45.820868 4859 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/2db15fd4-e39e-474c-b274-53262e92bccb-inventory\") on node \"crc\" DevicePath \"\"" Oct 08 18:43:45 crc kubenswrapper[4859]: I1008 18:43:45.821211 4859 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-llsx8\" (UniqueName: \"kubernetes.io/projected/2db15fd4-e39e-474c-b274-53262e92bccb-kube-api-access-llsx8\") on node \"crc\" DevicePath \"\"" Oct 08 18:43:45 crc kubenswrapper[4859]: I1008 18:43:45.821227 4859 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/2db15fd4-e39e-474c-b274-53262e92bccb-ssh-key\") on node \"crc\" DevicePath \"\"" Oct 08 18:43:46 crc kubenswrapper[4859]: I1008 18:43:46.292564 4859 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/download-cache-edpm-deployment-openstack-edpm-ipam-hb4tw" event={"ID":"2db15fd4-e39e-474c-b274-53262e92bccb","Type":"ContainerDied","Data":"47209fab83d608d93b7e9b762752e6cc265d877e8499d2ef665362c77c07cb48"} Oct 08 18:43:46 crc kubenswrapper[4859]: I1008 18:43:46.292623 4859 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="47209fab83d608d93b7e9b762752e6cc265d877e8499d2ef665362c77c07cb48" Oct 08 18:43:46 crc kubenswrapper[4859]: I1008 18:43:46.292713 4859 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/download-cache-edpm-deployment-openstack-edpm-ipam-hb4tw" Oct 08 18:43:46 crc kubenswrapper[4859]: I1008 18:43:46.382981 4859 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/configure-network-edpm-deployment-openstack-edpm-ipam-6n7d4"] Oct 08 18:43:46 crc kubenswrapper[4859]: E1008 18:43:46.383470 4859 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0b08244e-a6a0-4e1d-9951-699ab75ba8fa" containerName="registry-server" Oct 08 18:43:46 crc kubenswrapper[4859]: I1008 18:43:46.383503 4859 state_mem.go:107] "Deleted CPUSet assignment" podUID="0b08244e-a6a0-4e1d-9951-699ab75ba8fa" containerName="registry-server" Oct 08 18:43:46 crc kubenswrapper[4859]: E1008 18:43:46.383530 4859 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0b08244e-a6a0-4e1d-9951-699ab75ba8fa" containerName="extract-utilities" Oct 08 18:43:46 crc kubenswrapper[4859]: I1008 18:43:46.383537 4859 state_mem.go:107] "Deleted CPUSet assignment" podUID="0b08244e-a6a0-4e1d-9951-699ab75ba8fa" containerName="extract-utilities" Oct 08 18:43:46 crc kubenswrapper[4859]: E1008 18:43:46.383546 4859 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2db15fd4-e39e-474c-b274-53262e92bccb" containerName="download-cache-edpm-deployment-openstack-edpm-ipam" Oct 08 18:43:46 crc kubenswrapper[4859]: I1008 18:43:46.383555 4859 state_mem.go:107] "Deleted CPUSet assignment" podUID="2db15fd4-e39e-474c-b274-53262e92bccb" containerName="download-cache-edpm-deployment-openstack-edpm-ipam" Oct 08 18:43:46 crc kubenswrapper[4859]: E1008 18:43:46.383577 4859 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0b08244e-a6a0-4e1d-9951-699ab75ba8fa" containerName="extract-content" Oct 08 18:43:46 crc kubenswrapper[4859]: I1008 18:43:46.383583 4859 state_mem.go:107] "Deleted CPUSet assignment" podUID="0b08244e-a6a0-4e1d-9951-699ab75ba8fa" containerName="extract-content" Oct 08 18:43:46 crc kubenswrapper[4859]: I1008 18:43:46.383776 4859 memory_manager.go:354] "RemoveStaleState removing state" podUID="2db15fd4-e39e-474c-b274-53262e92bccb" containerName="download-cache-edpm-deployment-openstack-edpm-ipam" Oct 08 18:43:46 crc kubenswrapper[4859]: I1008 18:43:46.383796 4859 memory_manager.go:354] "RemoveStaleState removing state" podUID="0b08244e-a6a0-4e1d-9951-699ab75ba8fa" containerName="registry-server" Oct 08 18:43:46 crc kubenswrapper[4859]: I1008 18:43:46.384444 4859 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-6n7d4" Oct 08 18:43:46 crc kubenswrapper[4859]: I1008 18:43:46.387580 4859 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Oct 08 18:43:46 crc kubenswrapper[4859]: I1008 18:43:46.387591 4859 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Oct 08 18:43:46 crc kubenswrapper[4859]: I1008 18:43:46.387773 4859 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-gxqst" Oct 08 18:43:46 crc kubenswrapper[4859]: I1008 18:43:46.388295 4859 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Oct 08 18:43:46 crc kubenswrapper[4859]: I1008 18:43:46.393934 4859 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/configure-network-edpm-deployment-openstack-edpm-ipam-6n7d4"] Oct 08 18:43:46 crc kubenswrapper[4859]: I1008 18:43:46.433026 4859 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/70979534-d410-4768-95a9-5c9ff4ae1583-ssh-key\") pod \"configure-network-edpm-deployment-openstack-edpm-ipam-6n7d4\" (UID: \"70979534-d410-4768-95a9-5c9ff4ae1583\") " pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-6n7d4" Oct 08 18:43:46 crc kubenswrapper[4859]: I1008 18:43:46.433123 4859 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/70979534-d410-4768-95a9-5c9ff4ae1583-inventory\") pod \"configure-network-edpm-deployment-openstack-edpm-ipam-6n7d4\" (UID: \"70979534-d410-4768-95a9-5c9ff4ae1583\") " pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-6n7d4" Oct 08 18:43:46 crc kubenswrapper[4859]: I1008 18:43:46.433214 4859 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-8t2mm\" (UniqueName: \"kubernetes.io/projected/70979534-d410-4768-95a9-5c9ff4ae1583-kube-api-access-8t2mm\") pod \"configure-network-edpm-deployment-openstack-edpm-ipam-6n7d4\" (UID: \"70979534-d410-4768-95a9-5c9ff4ae1583\") " pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-6n7d4" Oct 08 18:43:46 crc kubenswrapper[4859]: I1008 18:43:46.535270 4859 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/70979534-d410-4768-95a9-5c9ff4ae1583-ssh-key\") pod \"configure-network-edpm-deployment-openstack-edpm-ipam-6n7d4\" (UID: \"70979534-d410-4768-95a9-5c9ff4ae1583\") " pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-6n7d4" Oct 08 18:43:46 crc kubenswrapper[4859]: I1008 18:43:46.535714 4859 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/70979534-d410-4768-95a9-5c9ff4ae1583-inventory\") pod \"configure-network-edpm-deployment-openstack-edpm-ipam-6n7d4\" (UID: \"70979534-d410-4768-95a9-5c9ff4ae1583\") " pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-6n7d4" Oct 08 18:43:46 crc kubenswrapper[4859]: I1008 18:43:46.536108 4859 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-8t2mm\" (UniqueName: \"kubernetes.io/projected/70979534-d410-4768-95a9-5c9ff4ae1583-kube-api-access-8t2mm\") pod \"configure-network-edpm-deployment-openstack-edpm-ipam-6n7d4\" (UID: \"70979534-d410-4768-95a9-5c9ff4ae1583\") " pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-6n7d4" Oct 08 18:43:46 crc kubenswrapper[4859]: I1008 18:43:46.540377 4859 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/70979534-d410-4768-95a9-5c9ff4ae1583-inventory\") pod \"configure-network-edpm-deployment-openstack-edpm-ipam-6n7d4\" (UID: \"70979534-d410-4768-95a9-5c9ff4ae1583\") " pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-6n7d4" Oct 08 18:43:46 crc kubenswrapper[4859]: I1008 18:43:46.540478 4859 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/70979534-d410-4768-95a9-5c9ff4ae1583-ssh-key\") pod \"configure-network-edpm-deployment-openstack-edpm-ipam-6n7d4\" (UID: \"70979534-d410-4768-95a9-5c9ff4ae1583\") " pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-6n7d4" Oct 08 18:43:46 crc kubenswrapper[4859]: I1008 18:43:46.557761 4859 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-8t2mm\" (UniqueName: \"kubernetes.io/projected/70979534-d410-4768-95a9-5c9ff4ae1583-kube-api-access-8t2mm\") pod \"configure-network-edpm-deployment-openstack-edpm-ipam-6n7d4\" (UID: \"70979534-d410-4768-95a9-5c9ff4ae1583\") " pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-6n7d4" Oct 08 18:43:46 crc kubenswrapper[4859]: I1008 18:43:46.701398 4859 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-6n7d4" Oct 08 18:43:47 crc kubenswrapper[4859]: I1008 18:43:47.201814 4859 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/configure-network-edpm-deployment-openstack-edpm-ipam-6n7d4"] Oct 08 18:43:47 crc kubenswrapper[4859]: I1008 18:43:47.215647 4859 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Oct 08 18:43:47 crc kubenswrapper[4859]: I1008 18:43:47.301006 4859 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-6n7d4" event={"ID":"70979534-d410-4768-95a9-5c9ff4ae1583","Type":"ContainerStarted","Data":"ffd20a88111beab025d35712f8a8dacd40613aef7ab57b9325eb2c2281415e66"} Oct 08 18:43:48 crc kubenswrapper[4859]: I1008 18:43:48.313489 4859 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-6n7d4" event={"ID":"70979534-d410-4768-95a9-5c9ff4ae1583","Type":"ContainerStarted","Data":"6c3e056d79813a7ac04abb5d6a0772bfe2686fcae6146cd5b3638c870b8fe122"} Oct 08 18:43:48 crc kubenswrapper[4859]: I1008 18:43:48.343355 4859 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-6n7d4" podStartSLOduration=1.8354425330000002 podStartE2EDuration="2.343332667s" podCreationTimestamp="2025-10-08 18:43:46 +0000 UTC" firstStartedPulling="2025-10-08 18:43:47.215459883 +0000 UTC m=+1597.462299262" lastFinishedPulling="2025-10-08 18:43:47.723350017 +0000 UTC m=+1597.970189396" observedRunningTime="2025-10-08 18:43:48.329341239 +0000 UTC m=+1598.576180618" watchObservedRunningTime="2025-10-08 18:43:48.343332667 +0000 UTC m=+1598.590172046" Oct 08 18:44:04 crc kubenswrapper[4859]: I1008 18:44:04.076304 4859 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/neutron-db-sync-cw6k2"] Oct 08 18:44:04 crc kubenswrapper[4859]: I1008 18:44:04.094455 4859 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/neutron-db-sync-cw6k2"] Oct 08 18:44:04 crc kubenswrapper[4859]: I1008 18:44:04.481643 4859 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="56c33135-38c1-4e29-9804-098312529f0f" path="/var/lib/kubelet/pods/56c33135-38c1-4e29-9804-098312529f0f/volumes" Oct 08 18:44:13 crc kubenswrapper[4859]: I1008 18:44:13.044486 4859 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/keystone-bootstrap-75br8"] Oct 08 18:44:13 crc kubenswrapper[4859]: I1008 18:44:13.056053 4859 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/keystone-bootstrap-75br8"] Oct 08 18:44:14 crc kubenswrapper[4859]: I1008 18:44:14.033594 4859 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/placement-db-sync-lsbrj"] Oct 08 18:44:14 crc kubenswrapper[4859]: I1008 18:44:14.040884 4859 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/placement-db-sync-lsbrj"] Oct 08 18:44:14 crc kubenswrapper[4859]: I1008 18:44:14.478308 4859 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="91a7b430-b4d6-49f6-8951-0d62d51d6685" path="/var/lib/kubelet/pods/91a7b430-b4d6-49f6-8951-0d62d51d6685/volumes" Oct 08 18:44:14 crc kubenswrapper[4859]: I1008 18:44:14.479320 4859 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="ed95b639-b2c8-4879-bedb-16681e0166cf" path="/var/lib/kubelet/pods/ed95b639-b2c8-4879-bedb-16681e0166cf/volumes" Oct 08 18:44:17 crc kubenswrapper[4859]: I1008 18:44:17.924782 4859 patch_prober.go:28] interesting pod/machine-config-daemon-82s52 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 08 18:44:17 crc kubenswrapper[4859]: I1008 18:44:17.925434 4859 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-82s52" podUID="b23a6a6c-9d92-4e7b-840e-55cfda873a2d" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 08 18:44:22 crc kubenswrapper[4859]: I1008 18:44:22.680644 4859 scope.go:117] "RemoveContainer" containerID="3817135e05ff0cedaf55bc7e268e2b05f26fca26eb2f615d768d6ce0b81c2cd2" Oct 08 18:44:22 crc kubenswrapper[4859]: I1008 18:44:22.712618 4859 scope.go:117] "RemoveContainer" containerID="5e20ee9937e1ddfebd11edf873c8c454923508a0da0ee8406c208dbf0b60306f" Oct 08 18:44:22 crc kubenswrapper[4859]: I1008 18:44:22.756844 4859 scope.go:117] "RemoveContainer" containerID="e90f0759e2ee9de59ae4c6f1fc51a484554c5850b56b12d4c621266811a5593d" Oct 08 18:44:22 crc kubenswrapper[4859]: I1008 18:44:22.832005 4859 scope.go:117] "RemoveContainer" containerID="912ebc52cf289ca586e23593221698f35f08e44164fc786655a89a1fa335f018" Oct 08 18:44:22 crc kubenswrapper[4859]: I1008 18:44:22.868956 4859 scope.go:117] "RemoveContainer" containerID="954a12574e8a4ca3fce7cfa866b30b0090d26d8e434f0ca540812c65bf12435d" Oct 08 18:44:22 crc kubenswrapper[4859]: I1008 18:44:22.920503 4859 scope.go:117] "RemoveContainer" containerID="ba60ab3185c0a8ffefed44fd91a99acbf6a00486ec248ab2c7753e30b2f4a4a9" Oct 08 18:44:22 crc kubenswrapper[4859]: I1008 18:44:22.952279 4859 scope.go:117] "RemoveContainer" containerID="9a099402c7e32c8080d9dd5a77c08917d15daa6b42628d4aa8b9826531ac0060" Oct 08 18:44:22 crc kubenswrapper[4859]: I1008 18:44:22.970461 4859 scope.go:117] "RemoveContainer" containerID="0188f80e29f014fdf763c155e5438e9a2bebbf2b89f5e7918a56107b4fe05af2" Oct 08 18:44:26 crc kubenswrapper[4859]: I1008 18:44:26.031781 4859 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/barbican-db-sync-x99jc"] Oct 08 18:44:26 crc kubenswrapper[4859]: I1008 18:44:26.038940 4859 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/barbican-db-sync-x99jc"] Oct 08 18:44:26 crc kubenswrapper[4859]: I1008 18:44:26.482298 4859 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="3b2953d1-3ebe-4d7d-b60c-6ad07d02af5d" path="/var/lib/kubelet/pods/3b2953d1-3ebe-4d7d-b60c-6ad07d02af5d/volumes" Oct 08 18:44:36 crc kubenswrapper[4859]: I1008 18:44:36.055227 4859 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/cinder-db-sync-jqwrh"] Oct 08 18:44:36 crc kubenswrapper[4859]: I1008 18:44:36.064283 4859 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/cinder-db-sync-jqwrh"] Oct 08 18:44:36 crc kubenswrapper[4859]: I1008 18:44:36.483556 4859 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="720db6c0-f18f-429b-9589-fa503d212139" path="/var/lib/kubelet/pods/720db6c0-f18f-429b-9589-fa503d212139/volumes" Oct 08 18:44:47 crc kubenswrapper[4859]: I1008 18:44:47.924445 4859 patch_prober.go:28] interesting pod/machine-config-daemon-82s52 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 08 18:44:47 crc kubenswrapper[4859]: I1008 18:44:47.924978 4859 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-82s52" podUID="b23a6a6c-9d92-4e7b-840e-55cfda873a2d" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 08 18:45:00 crc kubenswrapper[4859]: I1008 18:45:00.167099 4859 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29332485-lsgdn"] Oct 08 18:45:00 crc kubenswrapper[4859]: I1008 18:45:00.171304 4859 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29332485-lsgdn" Oct 08 18:45:00 crc kubenswrapper[4859]: I1008 18:45:00.174242 4859 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"collect-profiles-config" Oct 08 18:45:00 crc kubenswrapper[4859]: I1008 18:45:00.174416 4859 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"collect-profiles-dockercfg-kzf4t" Oct 08 18:45:00 crc kubenswrapper[4859]: I1008 18:45:00.182425 4859 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29332485-lsgdn"] Oct 08 18:45:00 crc kubenswrapper[4859]: I1008 18:45:00.339121 4859 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-wp7bt\" (UniqueName: \"kubernetes.io/projected/46b24a07-d369-4899-ad78-b6a8df8a77fe-kube-api-access-wp7bt\") pod \"collect-profiles-29332485-lsgdn\" (UID: \"46b24a07-d369-4899-ad78-b6a8df8a77fe\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29332485-lsgdn" Oct 08 18:45:00 crc kubenswrapper[4859]: I1008 18:45:00.339482 4859 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/46b24a07-d369-4899-ad78-b6a8df8a77fe-secret-volume\") pod \"collect-profiles-29332485-lsgdn\" (UID: \"46b24a07-d369-4899-ad78-b6a8df8a77fe\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29332485-lsgdn" Oct 08 18:45:00 crc kubenswrapper[4859]: I1008 18:45:00.339531 4859 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/46b24a07-d369-4899-ad78-b6a8df8a77fe-config-volume\") pod \"collect-profiles-29332485-lsgdn\" (UID: \"46b24a07-d369-4899-ad78-b6a8df8a77fe\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29332485-lsgdn" Oct 08 18:45:00 crc kubenswrapper[4859]: I1008 18:45:00.441434 4859 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-wp7bt\" (UniqueName: \"kubernetes.io/projected/46b24a07-d369-4899-ad78-b6a8df8a77fe-kube-api-access-wp7bt\") pod \"collect-profiles-29332485-lsgdn\" (UID: \"46b24a07-d369-4899-ad78-b6a8df8a77fe\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29332485-lsgdn" Oct 08 18:45:00 crc kubenswrapper[4859]: I1008 18:45:00.441473 4859 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/46b24a07-d369-4899-ad78-b6a8df8a77fe-secret-volume\") pod \"collect-profiles-29332485-lsgdn\" (UID: \"46b24a07-d369-4899-ad78-b6a8df8a77fe\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29332485-lsgdn" Oct 08 18:45:00 crc kubenswrapper[4859]: I1008 18:45:00.441520 4859 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/46b24a07-d369-4899-ad78-b6a8df8a77fe-config-volume\") pod \"collect-profiles-29332485-lsgdn\" (UID: \"46b24a07-d369-4899-ad78-b6a8df8a77fe\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29332485-lsgdn" Oct 08 18:45:00 crc kubenswrapper[4859]: I1008 18:45:00.442407 4859 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/46b24a07-d369-4899-ad78-b6a8df8a77fe-config-volume\") pod \"collect-profiles-29332485-lsgdn\" (UID: \"46b24a07-d369-4899-ad78-b6a8df8a77fe\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29332485-lsgdn" Oct 08 18:45:00 crc kubenswrapper[4859]: I1008 18:45:00.447799 4859 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/46b24a07-d369-4899-ad78-b6a8df8a77fe-secret-volume\") pod \"collect-profiles-29332485-lsgdn\" (UID: \"46b24a07-d369-4899-ad78-b6a8df8a77fe\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29332485-lsgdn" Oct 08 18:45:00 crc kubenswrapper[4859]: I1008 18:45:00.459565 4859 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-wp7bt\" (UniqueName: \"kubernetes.io/projected/46b24a07-d369-4899-ad78-b6a8df8a77fe-kube-api-access-wp7bt\") pod \"collect-profiles-29332485-lsgdn\" (UID: \"46b24a07-d369-4899-ad78-b6a8df8a77fe\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29332485-lsgdn" Oct 08 18:45:00 crc kubenswrapper[4859]: I1008 18:45:00.499827 4859 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29332485-lsgdn" Oct 08 18:45:00 crc kubenswrapper[4859]: I1008 18:45:00.979425 4859 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29332485-lsgdn"] Oct 08 18:45:01 crc kubenswrapper[4859]: I1008 18:45:01.096501 4859 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29332485-lsgdn" event={"ID":"46b24a07-d369-4899-ad78-b6a8df8a77fe","Type":"ContainerStarted","Data":"d365ebaa8315da01b702b4b01895d7a1d26c83fc31a60d30eae990986a5ece30"} Oct 08 18:45:02 crc kubenswrapper[4859]: I1008 18:45:02.106087 4859 generic.go:334] "Generic (PLEG): container finished" podID="70979534-d410-4768-95a9-5c9ff4ae1583" containerID="6c3e056d79813a7ac04abb5d6a0772bfe2686fcae6146cd5b3638c870b8fe122" exitCode=0 Oct 08 18:45:02 crc kubenswrapper[4859]: I1008 18:45:02.106169 4859 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-6n7d4" event={"ID":"70979534-d410-4768-95a9-5c9ff4ae1583","Type":"ContainerDied","Data":"6c3e056d79813a7ac04abb5d6a0772bfe2686fcae6146cd5b3638c870b8fe122"} Oct 08 18:45:02 crc kubenswrapper[4859]: I1008 18:45:02.108568 4859 generic.go:334] "Generic (PLEG): container finished" podID="46b24a07-d369-4899-ad78-b6a8df8a77fe" containerID="966ac395d3919248904554f9c71a57ce6ec546fc8e128120d1603cda8162d575" exitCode=0 Oct 08 18:45:02 crc kubenswrapper[4859]: I1008 18:45:02.108629 4859 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29332485-lsgdn" event={"ID":"46b24a07-d369-4899-ad78-b6a8df8a77fe","Type":"ContainerDied","Data":"966ac395d3919248904554f9c71a57ce6ec546fc8e128120d1603cda8162d575"} Oct 08 18:45:03 crc kubenswrapper[4859]: I1008 18:45:03.540197 4859 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-4p2dh"] Oct 08 18:45:03 crc kubenswrapper[4859]: I1008 18:45:03.551892 4859 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-4p2dh" Oct 08 18:45:03 crc kubenswrapper[4859]: I1008 18:45:03.555918 4859 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-4p2dh"] Oct 08 18:45:03 crc kubenswrapper[4859]: I1008 18:45:03.568707 4859 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29332485-lsgdn" Oct 08 18:45:03 crc kubenswrapper[4859]: I1008 18:45:03.577593 4859 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-6n7d4" Oct 08 18:45:03 crc kubenswrapper[4859]: I1008 18:45:03.712908 4859 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-wp7bt\" (UniqueName: \"kubernetes.io/projected/46b24a07-d369-4899-ad78-b6a8df8a77fe-kube-api-access-wp7bt\") pod \"46b24a07-d369-4899-ad78-b6a8df8a77fe\" (UID: \"46b24a07-d369-4899-ad78-b6a8df8a77fe\") " Oct 08 18:45:03 crc kubenswrapper[4859]: I1008 18:45:03.713052 4859 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-8t2mm\" (UniqueName: \"kubernetes.io/projected/70979534-d410-4768-95a9-5c9ff4ae1583-kube-api-access-8t2mm\") pod \"70979534-d410-4768-95a9-5c9ff4ae1583\" (UID: \"70979534-d410-4768-95a9-5c9ff4ae1583\") " Oct 08 18:45:03 crc kubenswrapper[4859]: I1008 18:45:03.713109 4859 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/46b24a07-d369-4899-ad78-b6a8df8a77fe-secret-volume\") pod \"46b24a07-d369-4899-ad78-b6a8df8a77fe\" (UID: \"46b24a07-d369-4899-ad78-b6a8df8a77fe\") " Oct 08 18:45:03 crc kubenswrapper[4859]: I1008 18:45:03.713132 4859 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/70979534-d410-4768-95a9-5c9ff4ae1583-ssh-key\") pod \"70979534-d410-4768-95a9-5c9ff4ae1583\" (UID: \"70979534-d410-4768-95a9-5c9ff4ae1583\") " Oct 08 18:45:03 crc kubenswrapper[4859]: I1008 18:45:03.713159 4859 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/46b24a07-d369-4899-ad78-b6a8df8a77fe-config-volume\") pod \"46b24a07-d369-4899-ad78-b6a8df8a77fe\" (UID: \"46b24a07-d369-4899-ad78-b6a8df8a77fe\") " Oct 08 18:45:03 crc kubenswrapper[4859]: I1008 18:45:03.713203 4859 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/70979534-d410-4768-95a9-5c9ff4ae1583-inventory\") pod \"70979534-d410-4768-95a9-5c9ff4ae1583\" (UID: \"70979534-d410-4768-95a9-5c9ff4ae1583\") " Oct 08 18:45:03 crc kubenswrapper[4859]: I1008 18:45:03.713408 4859 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/06f6f559-cc12-4cc9-99de-a5365aced4fc-catalog-content\") pod \"redhat-marketplace-4p2dh\" (UID: \"06f6f559-cc12-4cc9-99de-a5365aced4fc\") " pod="openshift-marketplace/redhat-marketplace-4p2dh" Oct 08 18:45:03 crc kubenswrapper[4859]: I1008 18:45:03.713512 4859 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/06f6f559-cc12-4cc9-99de-a5365aced4fc-utilities\") pod \"redhat-marketplace-4p2dh\" (UID: \"06f6f559-cc12-4cc9-99de-a5365aced4fc\") " pod="openshift-marketplace/redhat-marketplace-4p2dh" Oct 08 18:45:03 crc kubenswrapper[4859]: I1008 18:45:03.713587 4859 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-4kw56\" (UniqueName: \"kubernetes.io/projected/06f6f559-cc12-4cc9-99de-a5365aced4fc-kube-api-access-4kw56\") pod \"redhat-marketplace-4p2dh\" (UID: \"06f6f559-cc12-4cc9-99de-a5365aced4fc\") " pod="openshift-marketplace/redhat-marketplace-4p2dh" Oct 08 18:45:03 crc kubenswrapper[4859]: I1008 18:45:03.714630 4859 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/46b24a07-d369-4899-ad78-b6a8df8a77fe-config-volume" (OuterVolumeSpecName: "config-volume") pod "46b24a07-d369-4899-ad78-b6a8df8a77fe" (UID: "46b24a07-d369-4899-ad78-b6a8df8a77fe"). InnerVolumeSpecName "config-volume". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 08 18:45:03 crc kubenswrapper[4859]: I1008 18:45:03.719552 4859 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/46b24a07-d369-4899-ad78-b6a8df8a77fe-kube-api-access-wp7bt" (OuterVolumeSpecName: "kube-api-access-wp7bt") pod "46b24a07-d369-4899-ad78-b6a8df8a77fe" (UID: "46b24a07-d369-4899-ad78-b6a8df8a77fe"). InnerVolumeSpecName "kube-api-access-wp7bt". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 08 18:45:03 crc kubenswrapper[4859]: I1008 18:45:03.720043 4859 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/46b24a07-d369-4899-ad78-b6a8df8a77fe-secret-volume" (OuterVolumeSpecName: "secret-volume") pod "46b24a07-d369-4899-ad78-b6a8df8a77fe" (UID: "46b24a07-d369-4899-ad78-b6a8df8a77fe"). InnerVolumeSpecName "secret-volume". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 08 18:45:03 crc kubenswrapper[4859]: I1008 18:45:03.720069 4859 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/70979534-d410-4768-95a9-5c9ff4ae1583-kube-api-access-8t2mm" (OuterVolumeSpecName: "kube-api-access-8t2mm") pod "70979534-d410-4768-95a9-5c9ff4ae1583" (UID: "70979534-d410-4768-95a9-5c9ff4ae1583"). InnerVolumeSpecName "kube-api-access-8t2mm". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 08 18:45:03 crc kubenswrapper[4859]: I1008 18:45:03.748573 4859 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/70979534-d410-4768-95a9-5c9ff4ae1583-inventory" (OuterVolumeSpecName: "inventory") pod "70979534-d410-4768-95a9-5c9ff4ae1583" (UID: "70979534-d410-4768-95a9-5c9ff4ae1583"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 08 18:45:03 crc kubenswrapper[4859]: I1008 18:45:03.750182 4859 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/70979534-d410-4768-95a9-5c9ff4ae1583-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "70979534-d410-4768-95a9-5c9ff4ae1583" (UID: "70979534-d410-4768-95a9-5c9ff4ae1583"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 08 18:45:03 crc kubenswrapper[4859]: I1008 18:45:03.815491 4859 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/06f6f559-cc12-4cc9-99de-a5365aced4fc-utilities\") pod \"redhat-marketplace-4p2dh\" (UID: \"06f6f559-cc12-4cc9-99de-a5365aced4fc\") " pod="openshift-marketplace/redhat-marketplace-4p2dh" Oct 08 18:45:03 crc kubenswrapper[4859]: I1008 18:45:03.815605 4859 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-4kw56\" (UniqueName: \"kubernetes.io/projected/06f6f559-cc12-4cc9-99de-a5365aced4fc-kube-api-access-4kw56\") pod \"redhat-marketplace-4p2dh\" (UID: \"06f6f559-cc12-4cc9-99de-a5365aced4fc\") " pod="openshift-marketplace/redhat-marketplace-4p2dh" Oct 08 18:45:03 crc kubenswrapper[4859]: I1008 18:45:03.815645 4859 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/06f6f559-cc12-4cc9-99de-a5365aced4fc-catalog-content\") pod \"redhat-marketplace-4p2dh\" (UID: \"06f6f559-cc12-4cc9-99de-a5365aced4fc\") " pod="openshift-marketplace/redhat-marketplace-4p2dh" Oct 08 18:45:03 crc kubenswrapper[4859]: I1008 18:45:03.815717 4859 reconciler_common.go:293] "Volume detached for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/46b24a07-d369-4899-ad78-b6a8df8a77fe-secret-volume\") on node \"crc\" DevicePath \"\"" Oct 08 18:45:03 crc kubenswrapper[4859]: I1008 18:45:03.815729 4859 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/70979534-d410-4768-95a9-5c9ff4ae1583-ssh-key\") on node \"crc\" DevicePath \"\"" Oct 08 18:45:03 crc kubenswrapper[4859]: I1008 18:45:03.815737 4859 reconciler_common.go:293] "Volume detached for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/46b24a07-d369-4899-ad78-b6a8df8a77fe-config-volume\") on node \"crc\" DevicePath \"\"" Oct 08 18:45:03 crc kubenswrapper[4859]: I1008 18:45:03.815746 4859 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/70979534-d410-4768-95a9-5c9ff4ae1583-inventory\") on node \"crc\" DevicePath \"\"" Oct 08 18:45:03 crc kubenswrapper[4859]: I1008 18:45:03.815754 4859 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-wp7bt\" (UniqueName: \"kubernetes.io/projected/46b24a07-d369-4899-ad78-b6a8df8a77fe-kube-api-access-wp7bt\") on node \"crc\" DevicePath \"\"" Oct 08 18:45:03 crc kubenswrapper[4859]: I1008 18:45:03.815763 4859 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-8t2mm\" (UniqueName: \"kubernetes.io/projected/70979534-d410-4768-95a9-5c9ff4ae1583-kube-api-access-8t2mm\") on node \"crc\" DevicePath \"\"" Oct 08 18:45:03 crc kubenswrapper[4859]: I1008 18:45:03.816061 4859 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/06f6f559-cc12-4cc9-99de-a5365aced4fc-utilities\") pod \"redhat-marketplace-4p2dh\" (UID: \"06f6f559-cc12-4cc9-99de-a5365aced4fc\") " pod="openshift-marketplace/redhat-marketplace-4p2dh" Oct 08 18:45:03 crc kubenswrapper[4859]: I1008 18:45:03.816228 4859 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/06f6f559-cc12-4cc9-99de-a5365aced4fc-catalog-content\") pod \"redhat-marketplace-4p2dh\" (UID: \"06f6f559-cc12-4cc9-99de-a5365aced4fc\") " pod="openshift-marketplace/redhat-marketplace-4p2dh" Oct 08 18:45:03 crc kubenswrapper[4859]: I1008 18:45:03.835445 4859 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-4kw56\" (UniqueName: \"kubernetes.io/projected/06f6f559-cc12-4cc9-99de-a5365aced4fc-kube-api-access-4kw56\") pod \"redhat-marketplace-4p2dh\" (UID: \"06f6f559-cc12-4cc9-99de-a5365aced4fc\") " pod="openshift-marketplace/redhat-marketplace-4p2dh" Oct 08 18:45:03 crc kubenswrapper[4859]: I1008 18:45:03.900835 4859 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-4p2dh" Oct 08 18:45:04 crc kubenswrapper[4859]: I1008 18:45:04.135089 4859 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29332485-lsgdn" Oct 08 18:45:04 crc kubenswrapper[4859]: I1008 18:45:04.135095 4859 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29332485-lsgdn" event={"ID":"46b24a07-d369-4899-ad78-b6a8df8a77fe","Type":"ContainerDied","Data":"d365ebaa8315da01b702b4b01895d7a1d26c83fc31a60d30eae990986a5ece30"} Oct 08 18:45:04 crc kubenswrapper[4859]: I1008 18:45:04.135981 4859 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="d365ebaa8315da01b702b4b01895d7a1d26c83fc31a60d30eae990986a5ece30" Oct 08 18:45:04 crc kubenswrapper[4859]: I1008 18:45:04.138520 4859 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-6n7d4" event={"ID":"70979534-d410-4768-95a9-5c9ff4ae1583","Type":"ContainerDied","Data":"ffd20a88111beab025d35712f8a8dacd40613aef7ab57b9325eb2c2281415e66"} Oct 08 18:45:04 crc kubenswrapper[4859]: I1008 18:45:04.138610 4859 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="ffd20a88111beab025d35712f8a8dacd40613aef7ab57b9325eb2c2281415e66" Oct 08 18:45:04 crc kubenswrapper[4859]: I1008 18:45:04.138575 4859 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-6n7d4" Oct 08 18:45:04 crc kubenswrapper[4859]: I1008 18:45:04.141439 4859 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-4p2dh"] Oct 08 18:45:04 crc kubenswrapper[4859]: I1008 18:45:04.203586 4859 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/validate-network-edpm-deployment-openstack-edpm-ipam-2lvj9"] Oct 08 18:45:04 crc kubenswrapper[4859]: E1008 18:45:04.204021 4859 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="46b24a07-d369-4899-ad78-b6a8df8a77fe" containerName="collect-profiles" Oct 08 18:45:04 crc kubenswrapper[4859]: I1008 18:45:04.204045 4859 state_mem.go:107] "Deleted CPUSet assignment" podUID="46b24a07-d369-4899-ad78-b6a8df8a77fe" containerName="collect-profiles" Oct 08 18:45:04 crc kubenswrapper[4859]: E1008 18:45:04.204068 4859 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="70979534-d410-4768-95a9-5c9ff4ae1583" containerName="configure-network-edpm-deployment-openstack-edpm-ipam" Oct 08 18:45:04 crc kubenswrapper[4859]: I1008 18:45:04.204075 4859 state_mem.go:107] "Deleted CPUSet assignment" podUID="70979534-d410-4768-95a9-5c9ff4ae1583" containerName="configure-network-edpm-deployment-openstack-edpm-ipam" Oct 08 18:45:04 crc kubenswrapper[4859]: I1008 18:45:04.204246 4859 memory_manager.go:354] "RemoveStaleState removing state" podUID="46b24a07-d369-4899-ad78-b6a8df8a77fe" containerName="collect-profiles" Oct 08 18:45:04 crc kubenswrapper[4859]: I1008 18:45:04.204271 4859 memory_manager.go:354] "RemoveStaleState removing state" podUID="70979534-d410-4768-95a9-5c9ff4ae1583" containerName="configure-network-edpm-deployment-openstack-edpm-ipam" Oct 08 18:45:04 crc kubenswrapper[4859]: I1008 18:45:04.204836 4859 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-2lvj9" Oct 08 18:45:04 crc kubenswrapper[4859]: I1008 18:45:04.206920 4859 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Oct 08 18:45:04 crc kubenswrapper[4859]: I1008 18:45:04.207725 4859 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Oct 08 18:45:04 crc kubenswrapper[4859]: I1008 18:45:04.212013 4859 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-gxqst" Oct 08 18:45:04 crc kubenswrapper[4859]: I1008 18:45:04.212182 4859 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Oct 08 18:45:04 crc kubenswrapper[4859]: I1008 18:45:04.217358 4859 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/validate-network-edpm-deployment-openstack-edpm-ipam-2lvj9"] Oct 08 18:45:04 crc kubenswrapper[4859]: I1008 18:45:04.323379 4859 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/67e6fffe-7caa-42a6-812c-efa6a8f2162b-inventory\") pod \"validate-network-edpm-deployment-openstack-edpm-ipam-2lvj9\" (UID: \"67e6fffe-7caa-42a6-812c-efa6a8f2162b\") " pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-2lvj9" Oct 08 18:45:04 crc kubenswrapper[4859]: I1008 18:45:04.323798 4859 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/67e6fffe-7caa-42a6-812c-efa6a8f2162b-ssh-key\") pod \"validate-network-edpm-deployment-openstack-edpm-ipam-2lvj9\" (UID: \"67e6fffe-7caa-42a6-812c-efa6a8f2162b\") " pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-2lvj9" Oct 08 18:45:04 crc kubenswrapper[4859]: I1008 18:45:04.323940 4859 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-b7s6b\" (UniqueName: \"kubernetes.io/projected/67e6fffe-7caa-42a6-812c-efa6a8f2162b-kube-api-access-b7s6b\") pod \"validate-network-edpm-deployment-openstack-edpm-ipam-2lvj9\" (UID: \"67e6fffe-7caa-42a6-812c-efa6a8f2162b\") " pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-2lvj9" Oct 08 18:45:04 crc kubenswrapper[4859]: I1008 18:45:04.425553 4859 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-b7s6b\" (UniqueName: \"kubernetes.io/projected/67e6fffe-7caa-42a6-812c-efa6a8f2162b-kube-api-access-b7s6b\") pod \"validate-network-edpm-deployment-openstack-edpm-ipam-2lvj9\" (UID: \"67e6fffe-7caa-42a6-812c-efa6a8f2162b\") " pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-2lvj9" Oct 08 18:45:04 crc kubenswrapper[4859]: I1008 18:45:04.425735 4859 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/67e6fffe-7caa-42a6-812c-efa6a8f2162b-inventory\") pod \"validate-network-edpm-deployment-openstack-edpm-ipam-2lvj9\" (UID: \"67e6fffe-7caa-42a6-812c-efa6a8f2162b\") " pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-2lvj9" Oct 08 18:45:04 crc kubenswrapper[4859]: I1008 18:45:04.425773 4859 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/67e6fffe-7caa-42a6-812c-efa6a8f2162b-ssh-key\") pod \"validate-network-edpm-deployment-openstack-edpm-ipam-2lvj9\" (UID: \"67e6fffe-7caa-42a6-812c-efa6a8f2162b\") " pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-2lvj9" Oct 08 18:45:04 crc kubenswrapper[4859]: I1008 18:45:04.436237 4859 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/67e6fffe-7caa-42a6-812c-efa6a8f2162b-ssh-key\") pod \"validate-network-edpm-deployment-openstack-edpm-ipam-2lvj9\" (UID: \"67e6fffe-7caa-42a6-812c-efa6a8f2162b\") " pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-2lvj9" Oct 08 18:45:04 crc kubenswrapper[4859]: I1008 18:45:04.436345 4859 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/67e6fffe-7caa-42a6-812c-efa6a8f2162b-inventory\") pod \"validate-network-edpm-deployment-openstack-edpm-ipam-2lvj9\" (UID: \"67e6fffe-7caa-42a6-812c-efa6a8f2162b\") " pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-2lvj9" Oct 08 18:45:04 crc kubenswrapper[4859]: I1008 18:45:04.448804 4859 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-b7s6b\" (UniqueName: \"kubernetes.io/projected/67e6fffe-7caa-42a6-812c-efa6a8f2162b-kube-api-access-b7s6b\") pod \"validate-network-edpm-deployment-openstack-edpm-ipam-2lvj9\" (UID: \"67e6fffe-7caa-42a6-812c-efa6a8f2162b\") " pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-2lvj9" Oct 08 18:45:04 crc kubenswrapper[4859]: I1008 18:45:04.533422 4859 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-2lvj9" Oct 08 18:45:05 crc kubenswrapper[4859]: I1008 18:45:05.058165 4859 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/validate-network-edpm-deployment-openstack-edpm-ipam-2lvj9"] Oct 08 18:45:05 crc kubenswrapper[4859]: I1008 18:45:05.157195 4859 generic.go:334] "Generic (PLEG): container finished" podID="06f6f559-cc12-4cc9-99de-a5365aced4fc" containerID="ad7f8ac6e86b2712486521220bf0d274380236bd573c213a3c3eb48ea714a487" exitCode=0 Oct 08 18:45:05 crc kubenswrapper[4859]: I1008 18:45:05.157227 4859 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-4p2dh" event={"ID":"06f6f559-cc12-4cc9-99de-a5365aced4fc","Type":"ContainerDied","Data":"ad7f8ac6e86b2712486521220bf0d274380236bd573c213a3c3eb48ea714a487"} Oct 08 18:45:05 crc kubenswrapper[4859]: I1008 18:45:05.157291 4859 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-4p2dh" event={"ID":"06f6f559-cc12-4cc9-99de-a5365aced4fc","Type":"ContainerStarted","Data":"9b0e45332826e26e36e8149611245aeb428a5823c0f20115d1fb6711480439c8"} Oct 08 18:45:05 crc kubenswrapper[4859]: I1008 18:45:05.159147 4859 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-2lvj9" event={"ID":"67e6fffe-7caa-42a6-812c-efa6a8f2162b","Type":"ContainerStarted","Data":"d3bc0f34876176ed183e3bfe7df9dc08c67aef215c6bb5b8d60a251cd488c155"} Oct 08 18:45:06 crc kubenswrapper[4859]: I1008 18:45:06.170402 4859 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-4p2dh" event={"ID":"06f6f559-cc12-4cc9-99de-a5365aced4fc","Type":"ContainerStarted","Data":"5853ab82b7935a4a3961767c734e42c74fd1c3e6971ff1940636f55b4094bd97"} Oct 08 18:45:06 crc kubenswrapper[4859]: I1008 18:45:06.176305 4859 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-2lvj9" event={"ID":"67e6fffe-7caa-42a6-812c-efa6a8f2162b","Type":"ContainerStarted","Data":"de3d7f4339280d1cfb0d318ea70df5bf8f014c4a783cd06ce12d03bef081a221"} Oct 08 18:45:06 crc kubenswrapper[4859]: I1008 18:45:06.223928 4859 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-2lvj9" podStartSLOduration=1.788589548 podStartE2EDuration="2.223912389s" podCreationTimestamp="2025-10-08 18:45:04 +0000 UTC" firstStartedPulling="2025-10-08 18:45:05.063468926 +0000 UTC m=+1675.310308315" lastFinishedPulling="2025-10-08 18:45:05.498791737 +0000 UTC m=+1675.745631156" observedRunningTime="2025-10-08 18:45:06.218163241 +0000 UTC m=+1676.465002620" watchObservedRunningTime="2025-10-08 18:45:06.223912389 +0000 UTC m=+1676.470751768" Oct 08 18:45:07 crc kubenswrapper[4859]: I1008 18:45:07.187963 4859 generic.go:334] "Generic (PLEG): container finished" podID="06f6f559-cc12-4cc9-99de-a5365aced4fc" containerID="5853ab82b7935a4a3961767c734e42c74fd1c3e6971ff1940636f55b4094bd97" exitCode=0 Oct 08 18:45:07 crc kubenswrapper[4859]: I1008 18:45:07.188045 4859 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-4p2dh" event={"ID":"06f6f559-cc12-4cc9-99de-a5365aced4fc","Type":"ContainerDied","Data":"5853ab82b7935a4a3961767c734e42c74fd1c3e6971ff1940636f55b4094bd97"} Oct 08 18:45:08 crc kubenswrapper[4859]: I1008 18:45:08.040219 4859 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell0-db-create-drcrc"] Oct 08 18:45:08 crc kubenswrapper[4859]: I1008 18:45:08.050951 4859 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-api-db-create-2t4xd"] Oct 08 18:45:08 crc kubenswrapper[4859]: I1008 18:45:08.061022 4859 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-cell0-db-create-drcrc"] Oct 08 18:45:08 crc kubenswrapper[4859]: I1008 18:45:08.069072 4859 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell1-db-create-h7mqh"] Oct 08 18:45:08 crc kubenswrapper[4859]: I1008 18:45:08.077637 4859 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-api-db-create-2t4xd"] Oct 08 18:45:08 crc kubenswrapper[4859]: I1008 18:45:08.086994 4859 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-cell1-db-create-h7mqh"] Oct 08 18:45:08 crc kubenswrapper[4859]: I1008 18:45:08.198821 4859 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-4p2dh" event={"ID":"06f6f559-cc12-4cc9-99de-a5365aced4fc","Type":"ContainerStarted","Data":"af0f60497f48393cbd4e109dbbe049f8b905ac3f1b1ec4069728560ded5c1bd3"} Oct 08 18:45:08 crc kubenswrapper[4859]: I1008 18:45:08.219638 4859 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-4p2dh" podStartSLOduration=2.6271354430000002 podStartE2EDuration="5.219592722s" podCreationTimestamp="2025-10-08 18:45:03 +0000 UTC" firstStartedPulling="2025-10-08 18:45:05.159123922 +0000 UTC m=+1675.405963321" lastFinishedPulling="2025-10-08 18:45:07.751581221 +0000 UTC m=+1677.998420600" observedRunningTime="2025-10-08 18:45:08.216498432 +0000 UTC m=+1678.463337821" watchObservedRunningTime="2025-10-08 18:45:08.219592722 +0000 UTC m=+1678.466432091" Oct 08 18:45:08 crc kubenswrapper[4859]: I1008 18:45:08.496422 4859 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="3d62765f-7919-4338-b4c7-abc341837b2e" path="/var/lib/kubelet/pods/3d62765f-7919-4338-b4c7-abc341837b2e/volumes" Oct 08 18:45:08 crc kubenswrapper[4859]: I1008 18:45:08.496973 4859 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="d6886827-bf3f-4530-be76-a21a757a4c7f" path="/var/lib/kubelet/pods/d6886827-bf3f-4530-be76-a21a757a4c7f/volumes" Oct 08 18:45:08 crc kubenswrapper[4859]: I1008 18:45:08.497501 4859 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="fa8367eb-d900-45d2-90c0-4401b7d9cc89" path="/var/lib/kubelet/pods/fa8367eb-d900-45d2-90c0-4401b7d9cc89/volumes" Oct 08 18:45:11 crc kubenswrapper[4859]: I1008 18:45:11.235300 4859 generic.go:334] "Generic (PLEG): container finished" podID="67e6fffe-7caa-42a6-812c-efa6a8f2162b" containerID="de3d7f4339280d1cfb0d318ea70df5bf8f014c4a783cd06ce12d03bef081a221" exitCode=0 Oct 08 18:45:11 crc kubenswrapper[4859]: I1008 18:45:11.235382 4859 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-2lvj9" event={"ID":"67e6fffe-7caa-42a6-812c-efa6a8f2162b","Type":"ContainerDied","Data":"de3d7f4339280d1cfb0d318ea70df5bf8f014c4a783cd06ce12d03bef081a221"} Oct 08 18:45:12 crc kubenswrapper[4859]: I1008 18:45:12.029064 4859 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-api-dd49-account-create-cnv2d"] Oct 08 18:45:12 crc kubenswrapper[4859]: I1008 18:45:12.036514 4859 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-api-dd49-account-create-cnv2d"] Oct 08 18:45:12 crc kubenswrapper[4859]: I1008 18:45:12.484322 4859 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="4ee130b1-a5ee-428e-81f9-e4669ad95b11" path="/var/lib/kubelet/pods/4ee130b1-a5ee-428e-81f9-e4669ad95b11/volumes" Oct 08 18:45:12 crc kubenswrapper[4859]: I1008 18:45:12.663239 4859 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-2lvj9" Oct 08 18:45:12 crc kubenswrapper[4859]: I1008 18:45:12.799324 4859 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-b7s6b\" (UniqueName: \"kubernetes.io/projected/67e6fffe-7caa-42a6-812c-efa6a8f2162b-kube-api-access-b7s6b\") pod \"67e6fffe-7caa-42a6-812c-efa6a8f2162b\" (UID: \"67e6fffe-7caa-42a6-812c-efa6a8f2162b\") " Oct 08 18:45:12 crc kubenswrapper[4859]: I1008 18:45:12.799668 4859 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/67e6fffe-7caa-42a6-812c-efa6a8f2162b-ssh-key\") pod \"67e6fffe-7caa-42a6-812c-efa6a8f2162b\" (UID: \"67e6fffe-7caa-42a6-812c-efa6a8f2162b\") " Oct 08 18:45:12 crc kubenswrapper[4859]: I1008 18:45:12.799837 4859 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/67e6fffe-7caa-42a6-812c-efa6a8f2162b-inventory\") pod \"67e6fffe-7caa-42a6-812c-efa6a8f2162b\" (UID: \"67e6fffe-7caa-42a6-812c-efa6a8f2162b\") " Oct 08 18:45:12 crc kubenswrapper[4859]: I1008 18:45:12.805112 4859 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/67e6fffe-7caa-42a6-812c-efa6a8f2162b-kube-api-access-b7s6b" (OuterVolumeSpecName: "kube-api-access-b7s6b") pod "67e6fffe-7caa-42a6-812c-efa6a8f2162b" (UID: "67e6fffe-7caa-42a6-812c-efa6a8f2162b"). InnerVolumeSpecName "kube-api-access-b7s6b". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 08 18:45:12 crc kubenswrapper[4859]: I1008 18:45:12.826377 4859 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/67e6fffe-7caa-42a6-812c-efa6a8f2162b-inventory" (OuterVolumeSpecName: "inventory") pod "67e6fffe-7caa-42a6-812c-efa6a8f2162b" (UID: "67e6fffe-7caa-42a6-812c-efa6a8f2162b"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 08 18:45:12 crc kubenswrapper[4859]: I1008 18:45:12.826890 4859 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/67e6fffe-7caa-42a6-812c-efa6a8f2162b-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "67e6fffe-7caa-42a6-812c-efa6a8f2162b" (UID: "67e6fffe-7caa-42a6-812c-efa6a8f2162b"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 08 18:45:12 crc kubenswrapper[4859]: I1008 18:45:12.902274 4859 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-b7s6b\" (UniqueName: \"kubernetes.io/projected/67e6fffe-7caa-42a6-812c-efa6a8f2162b-kube-api-access-b7s6b\") on node \"crc\" DevicePath \"\"" Oct 08 18:45:12 crc kubenswrapper[4859]: I1008 18:45:12.902313 4859 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/67e6fffe-7caa-42a6-812c-efa6a8f2162b-ssh-key\") on node \"crc\" DevicePath \"\"" Oct 08 18:45:12 crc kubenswrapper[4859]: I1008 18:45:12.902323 4859 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/67e6fffe-7caa-42a6-812c-efa6a8f2162b-inventory\") on node \"crc\" DevicePath \"\"" Oct 08 18:45:13 crc kubenswrapper[4859]: I1008 18:45:13.259722 4859 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-2lvj9" event={"ID":"67e6fffe-7caa-42a6-812c-efa6a8f2162b","Type":"ContainerDied","Data":"d3bc0f34876176ed183e3bfe7df9dc08c67aef215c6bb5b8d60a251cd488c155"} Oct 08 18:45:13 crc kubenswrapper[4859]: I1008 18:45:13.260103 4859 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="d3bc0f34876176ed183e3bfe7df9dc08c67aef215c6bb5b8d60a251cd488c155" Oct 08 18:45:13 crc kubenswrapper[4859]: I1008 18:45:13.259764 4859 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-2lvj9" Oct 08 18:45:13 crc kubenswrapper[4859]: I1008 18:45:13.339738 4859 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/install-os-edpm-deployment-openstack-edpm-ipam-twjcj"] Oct 08 18:45:13 crc kubenswrapper[4859]: E1008 18:45:13.340100 4859 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="67e6fffe-7caa-42a6-812c-efa6a8f2162b" containerName="validate-network-edpm-deployment-openstack-edpm-ipam" Oct 08 18:45:13 crc kubenswrapper[4859]: I1008 18:45:13.340119 4859 state_mem.go:107] "Deleted CPUSet assignment" podUID="67e6fffe-7caa-42a6-812c-efa6a8f2162b" containerName="validate-network-edpm-deployment-openstack-edpm-ipam" Oct 08 18:45:13 crc kubenswrapper[4859]: I1008 18:45:13.340305 4859 memory_manager.go:354] "RemoveStaleState removing state" podUID="67e6fffe-7caa-42a6-812c-efa6a8f2162b" containerName="validate-network-edpm-deployment-openstack-edpm-ipam" Oct 08 18:45:13 crc kubenswrapper[4859]: I1008 18:45:13.340900 4859 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-twjcj" Oct 08 18:45:13 crc kubenswrapper[4859]: I1008 18:45:13.343631 4859 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Oct 08 18:45:13 crc kubenswrapper[4859]: I1008 18:45:13.343773 4859 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Oct 08 18:45:13 crc kubenswrapper[4859]: I1008 18:45:13.343907 4859 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-gxqst" Oct 08 18:45:13 crc kubenswrapper[4859]: I1008 18:45:13.344235 4859 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Oct 08 18:45:13 crc kubenswrapper[4859]: I1008 18:45:13.361760 4859 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/install-os-edpm-deployment-openstack-edpm-ipam-twjcj"] Oct 08 18:45:13 crc kubenswrapper[4859]: I1008 18:45:13.413059 4859 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/5d2f49c5-b079-4409-9ad7-e7c6b419e843-ssh-key\") pod \"install-os-edpm-deployment-openstack-edpm-ipam-twjcj\" (UID: \"5d2f49c5-b079-4409-9ad7-e7c6b419e843\") " pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-twjcj" Oct 08 18:45:13 crc kubenswrapper[4859]: I1008 18:45:13.413120 4859 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/5d2f49c5-b079-4409-9ad7-e7c6b419e843-inventory\") pod \"install-os-edpm-deployment-openstack-edpm-ipam-twjcj\" (UID: \"5d2f49c5-b079-4409-9ad7-e7c6b419e843\") " pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-twjcj" Oct 08 18:45:13 crc kubenswrapper[4859]: I1008 18:45:13.413199 4859 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-cw5ps\" (UniqueName: \"kubernetes.io/projected/5d2f49c5-b079-4409-9ad7-e7c6b419e843-kube-api-access-cw5ps\") pod \"install-os-edpm-deployment-openstack-edpm-ipam-twjcj\" (UID: \"5d2f49c5-b079-4409-9ad7-e7c6b419e843\") " pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-twjcj" Oct 08 18:45:13 crc kubenswrapper[4859]: I1008 18:45:13.515127 4859 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/5d2f49c5-b079-4409-9ad7-e7c6b419e843-ssh-key\") pod \"install-os-edpm-deployment-openstack-edpm-ipam-twjcj\" (UID: \"5d2f49c5-b079-4409-9ad7-e7c6b419e843\") " pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-twjcj" Oct 08 18:45:13 crc kubenswrapper[4859]: I1008 18:45:13.515222 4859 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/5d2f49c5-b079-4409-9ad7-e7c6b419e843-inventory\") pod \"install-os-edpm-deployment-openstack-edpm-ipam-twjcj\" (UID: \"5d2f49c5-b079-4409-9ad7-e7c6b419e843\") " pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-twjcj" Oct 08 18:45:13 crc kubenswrapper[4859]: I1008 18:45:13.515312 4859 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cw5ps\" (UniqueName: \"kubernetes.io/projected/5d2f49c5-b079-4409-9ad7-e7c6b419e843-kube-api-access-cw5ps\") pod \"install-os-edpm-deployment-openstack-edpm-ipam-twjcj\" (UID: \"5d2f49c5-b079-4409-9ad7-e7c6b419e843\") " pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-twjcj" Oct 08 18:45:13 crc kubenswrapper[4859]: I1008 18:45:13.520798 4859 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/5d2f49c5-b079-4409-9ad7-e7c6b419e843-ssh-key\") pod \"install-os-edpm-deployment-openstack-edpm-ipam-twjcj\" (UID: \"5d2f49c5-b079-4409-9ad7-e7c6b419e843\") " pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-twjcj" Oct 08 18:45:13 crc kubenswrapper[4859]: I1008 18:45:13.521965 4859 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/5d2f49c5-b079-4409-9ad7-e7c6b419e843-inventory\") pod \"install-os-edpm-deployment-openstack-edpm-ipam-twjcj\" (UID: \"5d2f49c5-b079-4409-9ad7-e7c6b419e843\") " pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-twjcj" Oct 08 18:45:13 crc kubenswrapper[4859]: I1008 18:45:13.544094 4859 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-cw5ps\" (UniqueName: \"kubernetes.io/projected/5d2f49c5-b079-4409-9ad7-e7c6b419e843-kube-api-access-cw5ps\") pod \"install-os-edpm-deployment-openstack-edpm-ipam-twjcj\" (UID: \"5d2f49c5-b079-4409-9ad7-e7c6b419e843\") " pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-twjcj" Oct 08 18:45:13 crc kubenswrapper[4859]: I1008 18:45:13.656753 4859 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-twjcj" Oct 08 18:45:13 crc kubenswrapper[4859]: I1008 18:45:13.902128 4859 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-4p2dh" Oct 08 18:45:13 crc kubenswrapper[4859]: I1008 18:45:13.902488 4859 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-4p2dh" Oct 08 18:45:13 crc kubenswrapper[4859]: I1008 18:45:13.962946 4859 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-4p2dh" Oct 08 18:45:14 crc kubenswrapper[4859]: I1008 18:45:14.189198 4859 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/install-os-edpm-deployment-openstack-edpm-ipam-twjcj"] Oct 08 18:45:14 crc kubenswrapper[4859]: I1008 18:45:14.269782 4859 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-twjcj" event={"ID":"5d2f49c5-b079-4409-9ad7-e7c6b419e843","Type":"ContainerStarted","Data":"23158bf901e3137aaa5347323f29637aebb4db0eda60638a8c15113889449c28"} Oct 08 18:45:14 crc kubenswrapper[4859]: I1008 18:45:14.321499 4859 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-4p2dh" Oct 08 18:45:14 crc kubenswrapper[4859]: I1008 18:45:14.375864 4859 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-4p2dh"] Oct 08 18:45:15 crc kubenswrapper[4859]: I1008 18:45:15.280371 4859 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-twjcj" event={"ID":"5d2f49c5-b079-4409-9ad7-e7c6b419e843","Type":"ContainerStarted","Data":"8a0f40a4552a80d9ee0980c95ec8cb52dd41b09e39d9dc3ef9c531805e3623a2"} Oct 08 18:45:16 crc kubenswrapper[4859]: I1008 18:45:16.294151 4859 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-marketplace-4p2dh" podUID="06f6f559-cc12-4cc9-99de-a5365aced4fc" containerName="registry-server" containerID="cri-o://af0f60497f48393cbd4e109dbbe049f8b905ac3f1b1ec4069728560ded5c1bd3" gracePeriod=2 Oct 08 18:45:16 crc kubenswrapper[4859]: I1008 18:45:16.814478 4859 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-4p2dh" Oct 08 18:45:16 crc kubenswrapper[4859]: I1008 18:45:16.836400 4859 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-twjcj" podStartSLOduration=3.440884053 podStartE2EDuration="3.836378439s" podCreationTimestamp="2025-10-08 18:45:13 +0000 UTC" firstStartedPulling="2025-10-08 18:45:14.203148538 +0000 UTC m=+1684.449987917" lastFinishedPulling="2025-10-08 18:45:14.598642924 +0000 UTC m=+1684.845482303" observedRunningTime="2025-10-08 18:45:15.300367135 +0000 UTC m=+1685.547206524" watchObservedRunningTime="2025-10-08 18:45:16.836378439 +0000 UTC m=+1687.083217818" Oct 08 18:45:16 crc kubenswrapper[4859]: I1008 18:45:16.893881 4859 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-4kw56\" (UniqueName: \"kubernetes.io/projected/06f6f559-cc12-4cc9-99de-a5365aced4fc-kube-api-access-4kw56\") pod \"06f6f559-cc12-4cc9-99de-a5365aced4fc\" (UID: \"06f6f559-cc12-4cc9-99de-a5365aced4fc\") " Oct 08 18:45:16 crc kubenswrapper[4859]: I1008 18:45:16.894064 4859 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/06f6f559-cc12-4cc9-99de-a5365aced4fc-catalog-content\") pod \"06f6f559-cc12-4cc9-99de-a5365aced4fc\" (UID: \"06f6f559-cc12-4cc9-99de-a5365aced4fc\") " Oct 08 18:45:16 crc kubenswrapper[4859]: I1008 18:45:16.894110 4859 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/06f6f559-cc12-4cc9-99de-a5365aced4fc-utilities\") pod \"06f6f559-cc12-4cc9-99de-a5365aced4fc\" (UID: \"06f6f559-cc12-4cc9-99de-a5365aced4fc\") " Oct 08 18:45:16 crc kubenswrapper[4859]: I1008 18:45:16.895266 4859 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/06f6f559-cc12-4cc9-99de-a5365aced4fc-utilities" (OuterVolumeSpecName: "utilities") pod "06f6f559-cc12-4cc9-99de-a5365aced4fc" (UID: "06f6f559-cc12-4cc9-99de-a5365aced4fc"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 08 18:45:16 crc kubenswrapper[4859]: I1008 18:45:16.898863 4859 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/06f6f559-cc12-4cc9-99de-a5365aced4fc-kube-api-access-4kw56" (OuterVolumeSpecName: "kube-api-access-4kw56") pod "06f6f559-cc12-4cc9-99de-a5365aced4fc" (UID: "06f6f559-cc12-4cc9-99de-a5365aced4fc"). InnerVolumeSpecName "kube-api-access-4kw56". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 08 18:45:16 crc kubenswrapper[4859]: I1008 18:45:16.906739 4859 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/06f6f559-cc12-4cc9-99de-a5365aced4fc-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "06f6f559-cc12-4cc9-99de-a5365aced4fc" (UID: "06f6f559-cc12-4cc9-99de-a5365aced4fc"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 08 18:45:16 crc kubenswrapper[4859]: I1008 18:45:16.996156 4859 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-4kw56\" (UniqueName: \"kubernetes.io/projected/06f6f559-cc12-4cc9-99de-a5365aced4fc-kube-api-access-4kw56\") on node \"crc\" DevicePath \"\"" Oct 08 18:45:16 crc kubenswrapper[4859]: I1008 18:45:16.996219 4859 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/06f6f559-cc12-4cc9-99de-a5365aced4fc-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 08 18:45:16 crc kubenswrapper[4859]: I1008 18:45:16.996234 4859 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/06f6f559-cc12-4cc9-99de-a5365aced4fc-utilities\") on node \"crc\" DevicePath \"\"" Oct 08 18:45:17 crc kubenswrapper[4859]: I1008 18:45:17.306720 4859 generic.go:334] "Generic (PLEG): container finished" podID="06f6f559-cc12-4cc9-99de-a5365aced4fc" containerID="af0f60497f48393cbd4e109dbbe049f8b905ac3f1b1ec4069728560ded5c1bd3" exitCode=0 Oct 08 18:45:17 crc kubenswrapper[4859]: I1008 18:45:17.306774 4859 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-4p2dh" event={"ID":"06f6f559-cc12-4cc9-99de-a5365aced4fc","Type":"ContainerDied","Data":"af0f60497f48393cbd4e109dbbe049f8b905ac3f1b1ec4069728560ded5c1bd3"} Oct 08 18:45:17 crc kubenswrapper[4859]: I1008 18:45:17.306801 4859 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-4p2dh" Oct 08 18:45:17 crc kubenswrapper[4859]: I1008 18:45:17.306830 4859 scope.go:117] "RemoveContainer" containerID="af0f60497f48393cbd4e109dbbe049f8b905ac3f1b1ec4069728560ded5c1bd3" Oct 08 18:45:17 crc kubenswrapper[4859]: I1008 18:45:17.306814 4859 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-4p2dh" event={"ID":"06f6f559-cc12-4cc9-99de-a5365aced4fc","Type":"ContainerDied","Data":"9b0e45332826e26e36e8149611245aeb428a5823c0f20115d1fb6711480439c8"} Oct 08 18:45:17 crc kubenswrapper[4859]: I1008 18:45:17.336653 4859 scope.go:117] "RemoveContainer" containerID="5853ab82b7935a4a3961767c734e42c74fd1c3e6971ff1940636f55b4094bd97" Oct 08 18:45:17 crc kubenswrapper[4859]: I1008 18:45:17.362547 4859 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-4p2dh"] Oct 08 18:45:17 crc kubenswrapper[4859]: I1008 18:45:17.377325 4859 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-marketplace-4p2dh"] Oct 08 18:45:17 crc kubenswrapper[4859]: I1008 18:45:17.380962 4859 scope.go:117] "RemoveContainer" containerID="ad7f8ac6e86b2712486521220bf0d274380236bd573c213a3c3eb48ea714a487" Oct 08 18:45:17 crc kubenswrapper[4859]: I1008 18:45:17.430867 4859 scope.go:117] "RemoveContainer" containerID="af0f60497f48393cbd4e109dbbe049f8b905ac3f1b1ec4069728560ded5c1bd3" Oct 08 18:45:17 crc kubenswrapper[4859]: E1008 18:45:17.431579 4859 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"af0f60497f48393cbd4e109dbbe049f8b905ac3f1b1ec4069728560ded5c1bd3\": container with ID starting with af0f60497f48393cbd4e109dbbe049f8b905ac3f1b1ec4069728560ded5c1bd3 not found: ID does not exist" containerID="af0f60497f48393cbd4e109dbbe049f8b905ac3f1b1ec4069728560ded5c1bd3" Oct 08 18:45:17 crc kubenswrapper[4859]: I1008 18:45:17.431662 4859 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"af0f60497f48393cbd4e109dbbe049f8b905ac3f1b1ec4069728560ded5c1bd3"} err="failed to get container status \"af0f60497f48393cbd4e109dbbe049f8b905ac3f1b1ec4069728560ded5c1bd3\": rpc error: code = NotFound desc = could not find container \"af0f60497f48393cbd4e109dbbe049f8b905ac3f1b1ec4069728560ded5c1bd3\": container with ID starting with af0f60497f48393cbd4e109dbbe049f8b905ac3f1b1ec4069728560ded5c1bd3 not found: ID does not exist" Oct 08 18:45:17 crc kubenswrapper[4859]: I1008 18:45:17.431864 4859 scope.go:117] "RemoveContainer" containerID="5853ab82b7935a4a3961767c734e42c74fd1c3e6971ff1940636f55b4094bd97" Oct 08 18:45:17 crc kubenswrapper[4859]: E1008 18:45:17.432421 4859 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"5853ab82b7935a4a3961767c734e42c74fd1c3e6971ff1940636f55b4094bd97\": container with ID starting with 5853ab82b7935a4a3961767c734e42c74fd1c3e6971ff1940636f55b4094bd97 not found: ID does not exist" containerID="5853ab82b7935a4a3961767c734e42c74fd1c3e6971ff1940636f55b4094bd97" Oct 08 18:45:17 crc kubenswrapper[4859]: I1008 18:45:17.432482 4859 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"5853ab82b7935a4a3961767c734e42c74fd1c3e6971ff1940636f55b4094bd97"} err="failed to get container status \"5853ab82b7935a4a3961767c734e42c74fd1c3e6971ff1940636f55b4094bd97\": rpc error: code = NotFound desc = could not find container \"5853ab82b7935a4a3961767c734e42c74fd1c3e6971ff1940636f55b4094bd97\": container with ID starting with 5853ab82b7935a4a3961767c734e42c74fd1c3e6971ff1940636f55b4094bd97 not found: ID does not exist" Oct 08 18:45:17 crc kubenswrapper[4859]: I1008 18:45:17.432521 4859 scope.go:117] "RemoveContainer" containerID="ad7f8ac6e86b2712486521220bf0d274380236bd573c213a3c3eb48ea714a487" Oct 08 18:45:17 crc kubenswrapper[4859]: E1008 18:45:17.432952 4859 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"ad7f8ac6e86b2712486521220bf0d274380236bd573c213a3c3eb48ea714a487\": container with ID starting with ad7f8ac6e86b2712486521220bf0d274380236bd573c213a3c3eb48ea714a487 not found: ID does not exist" containerID="ad7f8ac6e86b2712486521220bf0d274380236bd573c213a3c3eb48ea714a487" Oct 08 18:45:17 crc kubenswrapper[4859]: I1008 18:45:17.432989 4859 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"ad7f8ac6e86b2712486521220bf0d274380236bd573c213a3c3eb48ea714a487"} err="failed to get container status \"ad7f8ac6e86b2712486521220bf0d274380236bd573c213a3c3eb48ea714a487\": rpc error: code = NotFound desc = could not find container \"ad7f8ac6e86b2712486521220bf0d274380236bd573c213a3c3eb48ea714a487\": container with ID starting with ad7f8ac6e86b2712486521220bf0d274380236bd573c213a3c3eb48ea714a487 not found: ID does not exist" Oct 08 18:45:17 crc kubenswrapper[4859]: I1008 18:45:17.924977 4859 patch_prober.go:28] interesting pod/machine-config-daemon-82s52 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 08 18:45:17 crc kubenswrapper[4859]: I1008 18:45:17.925057 4859 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-82s52" podUID="b23a6a6c-9d92-4e7b-840e-55cfda873a2d" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 08 18:45:17 crc kubenswrapper[4859]: I1008 18:45:17.925121 4859 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-82s52" Oct 08 18:45:17 crc kubenswrapper[4859]: I1008 18:45:17.926113 4859 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"44aa7848c4f5acb003170608db1921ad40aa5b2f5464f1f85d423cd26d2e8155"} pod="openshift-machine-config-operator/machine-config-daemon-82s52" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Oct 08 18:45:17 crc kubenswrapper[4859]: I1008 18:45:17.926216 4859 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-82s52" podUID="b23a6a6c-9d92-4e7b-840e-55cfda873a2d" containerName="machine-config-daemon" containerID="cri-o://44aa7848c4f5acb003170608db1921ad40aa5b2f5464f1f85d423cd26d2e8155" gracePeriod=600 Oct 08 18:45:18 crc kubenswrapper[4859]: E1008 18:45:18.052923 4859 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-82s52_openshift-machine-config-operator(b23a6a6c-9d92-4e7b-840e-55cfda873a2d)\"" pod="openshift-machine-config-operator/machine-config-daemon-82s52" podUID="b23a6a6c-9d92-4e7b-840e-55cfda873a2d" Oct 08 18:45:18 crc kubenswrapper[4859]: I1008 18:45:18.320657 4859 generic.go:334] "Generic (PLEG): container finished" podID="b23a6a6c-9d92-4e7b-840e-55cfda873a2d" containerID="44aa7848c4f5acb003170608db1921ad40aa5b2f5464f1f85d423cd26d2e8155" exitCode=0 Oct 08 18:45:18 crc kubenswrapper[4859]: I1008 18:45:18.320724 4859 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-82s52" event={"ID":"b23a6a6c-9d92-4e7b-840e-55cfda873a2d","Type":"ContainerDied","Data":"44aa7848c4f5acb003170608db1921ad40aa5b2f5464f1f85d423cd26d2e8155"} Oct 08 18:45:18 crc kubenswrapper[4859]: I1008 18:45:18.320789 4859 scope.go:117] "RemoveContainer" containerID="8eec6d3743e0f820548e10238303377094560d6d193dfbfe39f2b47960c426e8" Oct 08 18:45:18 crc kubenswrapper[4859]: I1008 18:45:18.321131 4859 scope.go:117] "RemoveContainer" containerID="44aa7848c4f5acb003170608db1921ad40aa5b2f5464f1f85d423cd26d2e8155" Oct 08 18:45:18 crc kubenswrapper[4859]: E1008 18:45:18.321412 4859 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-82s52_openshift-machine-config-operator(b23a6a6c-9d92-4e7b-840e-55cfda873a2d)\"" pod="openshift-machine-config-operator/machine-config-daemon-82s52" podUID="b23a6a6c-9d92-4e7b-840e-55cfda873a2d" Oct 08 18:45:18 crc kubenswrapper[4859]: I1008 18:45:18.483603 4859 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="06f6f559-cc12-4cc9-99de-a5365aced4fc" path="/var/lib/kubelet/pods/06f6f559-cc12-4cc9-99de-a5365aced4fc/volumes" Oct 08 18:45:22 crc kubenswrapper[4859]: I1008 18:45:22.024714 4859 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell0-e838-account-create-6ndb4"] Oct 08 18:45:22 crc kubenswrapper[4859]: I1008 18:45:22.033536 4859 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-cell0-e838-account-create-6ndb4"] Oct 08 18:45:22 crc kubenswrapper[4859]: I1008 18:45:22.041791 4859 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell1-f8cf-account-create-ftsj5"] Oct 08 18:45:22 crc kubenswrapper[4859]: I1008 18:45:22.067351 4859 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-cell1-f8cf-account-create-ftsj5"] Oct 08 18:45:22 crc kubenswrapper[4859]: I1008 18:45:22.487504 4859 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="4f171e6e-3d04-4e18-9b70-b183e9fb222d" path="/var/lib/kubelet/pods/4f171e6e-3d04-4e18-9b70-b183e9fb222d/volumes" Oct 08 18:45:22 crc kubenswrapper[4859]: I1008 18:45:22.488559 4859 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="d68d0cff-c3b1-45bb-a00c-23febbb14aee" path="/var/lib/kubelet/pods/d68d0cff-c3b1-45bb-a00c-23febbb14aee/volumes" Oct 08 18:45:23 crc kubenswrapper[4859]: I1008 18:45:23.139032 4859 scope.go:117] "RemoveContainer" containerID="8647fd3df54b81ebc93e3e14fe2b59cbd0e77836f881418a614ac5f2c3c4c10f" Oct 08 18:45:23 crc kubenswrapper[4859]: I1008 18:45:23.182281 4859 scope.go:117] "RemoveContainer" containerID="a3312ac46e8b3275c032896bb6f261fcefd038ae943b940c85e22a90bff4e1d6" Oct 08 18:45:23 crc kubenswrapper[4859]: I1008 18:45:23.248579 4859 scope.go:117] "RemoveContainer" containerID="0986204b68ccd47e6d46c16f7ae883db00acc964ad708ef4f4f5b7c984d726f7" Oct 08 18:45:23 crc kubenswrapper[4859]: I1008 18:45:23.288575 4859 scope.go:117] "RemoveContainer" containerID="f705521531b392a51ca4b6b14b60c8d34d3ec8f7dbebf2794d813c0cd105ff2c" Oct 08 18:45:23 crc kubenswrapper[4859]: I1008 18:45:23.349338 4859 scope.go:117] "RemoveContainer" containerID="07779e470ce52d0f4ab06a584cd0ab7f58267d01ef6cd06afd05e3def2f0a59b" Oct 08 18:45:23 crc kubenswrapper[4859]: I1008 18:45:23.366957 4859 scope.go:117] "RemoveContainer" containerID="8d304c8b88fe273732cb1eb09623aa4d5469a346872436266c01a7b9eec8dd66" Oct 08 18:45:23 crc kubenswrapper[4859]: I1008 18:45:23.444271 4859 scope.go:117] "RemoveContainer" containerID="60e9bb3a4e7c180c91ff79487b9adc18ab0fc78b4c8d1eba75682ae1b02a173f" Oct 08 18:45:23 crc kubenswrapper[4859]: I1008 18:45:23.464205 4859 scope.go:117] "RemoveContainer" containerID="588f280ec9b7b2a0e1c06aeef03cd5b1088654dadd0f63f256a0d0a4de198212" Oct 08 18:45:33 crc kubenswrapper[4859]: I1008 18:45:33.469592 4859 scope.go:117] "RemoveContainer" containerID="44aa7848c4f5acb003170608db1921ad40aa5b2f5464f1f85d423cd26d2e8155" Oct 08 18:45:33 crc kubenswrapper[4859]: E1008 18:45:33.470234 4859 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-82s52_openshift-machine-config-operator(b23a6a6c-9d92-4e7b-840e-55cfda873a2d)\"" pod="openshift-machine-config-operator/machine-config-daemon-82s52" podUID="b23a6a6c-9d92-4e7b-840e-55cfda873a2d" Oct 08 18:45:48 crc kubenswrapper[4859]: I1008 18:45:48.056476 4859 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell0-conductor-db-sync-9hb79"] Oct 08 18:45:48 crc kubenswrapper[4859]: I1008 18:45:48.062538 4859 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-cell0-conductor-db-sync-9hb79"] Oct 08 18:45:48 crc kubenswrapper[4859]: I1008 18:45:48.471548 4859 scope.go:117] "RemoveContainer" containerID="44aa7848c4f5acb003170608db1921ad40aa5b2f5464f1f85d423cd26d2e8155" Oct 08 18:45:48 crc kubenswrapper[4859]: E1008 18:45:48.471825 4859 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-82s52_openshift-machine-config-operator(b23a6a6c-9d92-4e7b-840e-55cfda873a2d)\"" pod="openshift-machine-config-operator/machine-config-daemon-82s52" podUID="b23a6a6c-9d92-4e7b-840e-55cfda873a2d" Oct 08 18:45:48 crc kubenswrapper[4859]: I1008 18:45:48.483960 4859 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="b03ff9b0-98b5-4d0b-990a-253ada95b6d8" path="/var/lib/kubelet/pods/b03ff9b0-98b5-4d0b-990a-253ada95b6d8/volumes" Oct 08 18:45:54 crc kubenswrapper[4859]: I1008 18:45:54.752555 4859 generic.go:334] "Generic (PLEG): container finished" podID="5d2f49c5-b079-4409-9ad7-e7c6b419e843" containerID="8a0f40a4552a80d9ee0980c95ec8cb52dd41b09e39d9dc3ef9c531805e3623a2" exitCode=0 Oct 08 18:45:54 crc kubenswrapper[4859]: I1008 18:45:54.752672 4859 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-twjcj" event={"ID":"5d2f49c5-b079-4409-9ad7-e7c6b419e843","Type":"ContainerDied","Data":"8a0f40a4552a80d9ee0980c95ec8cb52dd41b09e39d9dc3ef9c531805e3623a2"} Oct 08 18:45:56 crc kubenswrapper[4859]: I1008 18:45:56.174760 4859 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-twjcj" Oct 08 18:45:56 crc kubenswrapper[4859]: I1008 18:45:56.330439 4859 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-cw5ps\" (UniqueName: \"kubernetes.io/projected/5d2f49c5-b079-4409-9ad7-e7c6b419e843-kube-api-access-cw5ps\") pod \"5d2f49c5-b079-4409-9ad7-e7c6b419e843\" (UID: \"5d2f49c5-b079-4409-9ad7-e7c6b419e843\") " Oct 08 18:45:56 crc kubenswrapper[4859]: I1008 18:45:56.330615 4859 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/5d2f49c5-b079-4409-9ad7-e7c6b419e843-ssh-key\") pod \"5d2f49c5-b079-4409-9ad7-e7c6b419e843\" (UID: \"5d2f49c5-b079-4409-9ad7-e7c6b419e843\") " Oct 08 18:45:56 crc kubenswrapper[4859]: I1008 18:45:56.330682 4859 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/5d2f49c5-b079-4409-9ad7-e7c6b419e843-inventory\") pod \"5d2f49c5-b079-4409-9ad7-e7c6b419e843\" (UID: \"5d2f49c5-b079-4409-9ad7-e7c6b419e843\") " Oct 08 18:45:56 crc kubenswrapper[4859]: I1008 18:45:56.340067 4859 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5d2f49c5-b079-4409-9ad7-e7c6b419e843-kube-api-access-cw5ps" (OuterVolumeSpecName: "kube-api-access-cw5ps") pod "5d2f49c5-b079-4409-9ad7-e7c6b419e843" (UID: "5d2f49c5-b079-4409-9ad7-e7c6b419e843"). InnerVolumeSpecName "kube-api-access-cw5ps". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 08 18:45:56 crc kubenswrapper[4859]: I1008 18:45:56.359144 4859 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5d2f49c5-b079-4409-9ad7-e7c6b419e843-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "5d2f49c5-b079-4409-9ad7-e7c6b419e843" (UID: "5d2f49c5-b079-4409-9ad7-e7c6b419e843"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 08 18:45:56 crc kubenswrapper[4859]: I1008 18:45:56.359552 4859 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5d2f49c5-b079-4409-9ad7-e7c6b419e843-inventory" (OuterVolumeSpecName: "inventory") pod "5d2f49c5-b079-4409-9ad7-e7c6b419e843" (UID: "5d2f49c5-b079-4409-9ad7-e7c6b419e843"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 08 18:45:56 crc kubenswrapper[4859]: I1008 18:45:56.433331 4859 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/5d2f49c5-b079-4409-9ad7-e7c6b419e843-ssh-key\") on node \"crc\" DevicePath \"\"" Oct 08 18:45:56 crc kubenswrapper[4859]: I1008 18:45:56.433358 4859 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/5d2f49c5-b079-4409-9ad7-e7c6b419e843-inventory\") on node \"crc\" DevicePath \"\"" Oct 08 18:45:56 crc kubenswrapper[4859]: I1008 18:45:56.433369 4859 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-cw5ps\" (UniqueName: \"kubernetes.io/projected/5d2f49c5-b079-4409-9ad7-e7c6b419e843-kube-api-access-cw5ps\") on node \"crc\" DevicePath \"\"" Oct 08 18:45:56 crc kubenswrapper[4859]: I1008 18:45:56.775867 4859 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-twjcj" event={"ID":"5d2f49c5-b079-4409-9ad7-e7c6b419e843","Type":"ContainerDied","Data":"23158bf901e3137aaa5347323f29637aebb4db0eda60638a8c15113889449c28"} Oct 08 18:45:56 crc kubenswrapper[4859]: I1008 18:45:56.776247 4859 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="23158bf901e3137aaa5347323f29637aebb4db0eda60638a8c15113889449c28" Oct 08 18:45:56 crc kubenswrapper[4859]: I1008 18:45:56.775956 4859 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-twjcj" Oct 08 18:45:56 crc kubenswrapper[4859]: I1008 18:45:56.898657 4859 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/configure-os-edpm-deployment-openstack-edpm-ipam-dnz8t"] Oct 08 18:45:56 crc kubenswrapper[4859]: E1008 18:45:56.899013 4859 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="06f6f559-cc12-4cc9-99de-a5365aced4fc" containerName="extract-utilities" Oct 08 18:45:56 crc kubenswrapper[4859]: I1008 18:45:56.899030 4859 state_mem.go:107] "Deleted CPUSet assignment" podUID="06f6f559-cc12-4cc9-99de-a5365aced4fc" containerName="extract-utilities" Oct 08 18:45:56 crc kubenswrapper[4859]: E1008 18:45:56.899054 4859 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="06f6f559-cc12-4cc9-99de-a5365aced4fc" containerName="extract-content" Oct 08 18:45:56 crc kubenswrapper[4859]: I1008 18:45:56.899061 4859 state_mem.go:107] "Deleted CPUSet assignment" podUID="06f6f559-cc12-4cc9-99de-a5365aced4fc" containerName="extract-content" Oct 08 18:45:56 crc kubenswrapper[4859]: E1008 18:45:56.899073 4859 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5d2f49c5-b079-4409-9ad7-e7c6b419e843" containerName="install-os-edpm-deployment-openstack-edpm-ipam" Oct 08 18:45:56 crc kubenswrapper[4859]: I1008 18:45:56.899080 4859 state_mem.go:107] "Deleted CPUSet assignment" podUID="5d2f49c5-b079-4409-9ad7-e7c6b419e843" containerName="install-os-edpm-deployment-openstack-edpm-ipam" Oct 08 18:45:56 crc kubenswrapper[4859]: E1008 18:45:56.899100 4859 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="06f6f559-cc12-4cc9-99de-a5365aced4fc" containerName="registry-server" Oct 08 18:45:56 crc kubenswrapper[4859]: I1008 18:45:56.899106 4859 state_mem.go:107] "Deleted CPUSet assignment" podUID="06f6f559-cc12-4cc9-99de-a5365aced4fc" containerName="registry-server" Oct 08 18:45:56 crc kubenswrapper[4859]: I1008 18:45:56.899267 4859 memory_manager.go:354] "RemoveStaleState removing state" podUID="06f6f559-cc12-4cc9-99de-a5365aced4fc" containerName="registry-server" Oct 08 18:45:56 crc kubenswrapper[4859]: I1008 18:45:56.899297 4859 memory_manager.go:354] "RemoveStaleState removing state" podUID="5d2f49c5-b079-4409-9ad7-e7c6b419e843" containerName="install-os-edpm-deployment-openstack-edpm-ipam" Oct 08 18:45:56 crc kubenswrapper[4859]: I1008 18:45:56.899867 4859 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-dnz8t" Oct 08 18:45:56 crc kubenswrapper[4859]: I1008 18:45:56.901305 4859 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Oct 08 18:45:56 crc kubenswrapper[4859]: I1008 18:45:56.904315 4859 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Oct 08 18:45:56 crc kubenswrapper[4859]: I1008 18:45:56.904779 4859 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Oct 08 18:45:56 crc kubenswrapper[4859]: I1008 18:45:56.905406 4859 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-gxqst" Oct 08 18:45:56 crc kubenswrapper[4859]: I1008 18:45:56.909621 4859 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/configure-os-edpm-deployment-openstack-edpm-ipam-dnz8t"] Oct 08 18:45:57 crc kubenswrapper[4859]: I1008 18:45:57.044027 4859 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/28feb067-adfe-4b48-aef3-321e18ac0e37-ssh-key\") pod \"configure-os-edpm-deployment-openstack-edpm-ipam-dnz8t\" (UID: \"28feb067-adfe-4b48-aef3-321e18ac0e37\") " pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-dnz8t" Oct 08 18:45:57 crc kubenswrapper[4859]: I1008 18:45:57.044410 4859 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-4nfmq\" (UniqueName: \"kubernetes.io/projected/28feb067-adfe-4b48-aef3-321e18ac0e37-kube-api-access-4nfmq\") pod \"configure-os-edpm-deployment-openstack-edpm-ipam-dnz8t\" (UID: \"28feb067-adfe-4b48-aef3-321e18ac0e37\") " pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-dnz8t" Oct 08 18:45:57 crc kubenswrapper[4859]: I1008 18:45:57.044574 4859 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/28feb067-adfe-4b48-aef3-321e18ac0e37-inventory\") pod \"configure-os-edpm-deployment-openstack-edpm-ipam-dnz8t\" (UID: \"28feb067-adfe-4b48-aef3-321e18ac0e37\") " pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-dnz8t" Oct 08 18:45:57 crc kubenswrapper[4859]: I1008 18:45:57.146387 4859 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-4nfmq\" (UniqueName: \"kubernetes.io/projected/28feb067-adfe-4b48-aef3-321e18ac0e37-kube-api-access-4nfmq\") pod \"configure-os-edpm-deployment-openstack-edpm-ipam-dnz8t\" (UID: \"28feb067-adfe-4b48-aef3-321e18ac0e37\") " pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-dnz8t" Oct 08 18:45:57 crc kubenswrapper[4859]: I1008 18:45:57.146465 4859 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/28feb067-adfe-4b48-aef3-321e18ac0e37-inventory\") pod \"configure-os-edpm-deployment-openstack-edpm-ipam-dnz8t\" (UID: \"28feb067-adfe-4b48-aef3-321e18ac0e37\") " pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-dnz8t" Oct 08 18:45:57 crc kubenswrapper[4859]: I1008 18:45:57.146538 4859 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/28feb067-adfe-4b48-aef3-321e18ac0e37-ssh-key\") pod \"configure-os-edpm-deployment-openstack-edpm-ipam-dnz8t\" (UID: \"28feb067-adfe-4b48-aef3-321e18ac0e37\") " pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-dnz8t" Oct 08 18:45:57 crc kubenswrapper[4859]: I1008 18:45:57.150947 4859 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/28feb067-adfe-4b48-aef3-321e18ac0e37-ssh-key\") pod \"configure-os-edpm-deployment-openstack-edpm-ipam-dnz8t\" (UID: \"28feb067-adfe-4b48-aef3-321e18ac0e37\") " pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-dnz8t" Oct 08 18:45:57 crc kubenswrapper[4859]: I1008 18:45:57.153845 4859 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/28feb067-adfe-4b48-aef3-321e18ac0e37-inventory\") pod \"configure-os-edpm-deployment-openstack-edpm-ipam-dnz8t\" (UID: \"28feb067-adfe-4b48-aef3-321e18ac0e37\") " pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-dnz8t" Oct 08 18:45:57 crc kubenswrapper[4859]: I1008 18:45:57.170393 4859 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-4nfmq\" (UniqueName: \"kubernetes.io/projected/28feb067-adfe-4b48-aef3-321e18ac0e37-kube-api-access-4nfmq\") pod \"configure-os-edpm-deployment-openstack-edpm-ipam-dnz8t\" (UID: \"28feb067-adfe-4b48-aef3-321e18ac0e37\") " pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-dnz8t" Oct 08 18:45:57 crc kubenswrapper[4859]: I1008 18:45:57.223988 4859 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-dnz8t" Oct 08 18:45:57 crc kubenswrapper[4859]: I1008 18:45:57.715857 4859 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/configure-os-edpm-deployment-openstack-edpm-ipam-dnz8t"] Oct 08 18:45:57 crc kubenswrapper[4859]: I1008 18:45:57.786167 4859 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-dnz8t" event={"ID":"28feb067-adfe-4b48-aef3-321e18ac0e37","Type":"ContainerStarted","Data":"d9b428c0d11bea8cbbe90fa09b16489cb58b78769acd2bef609a37097a8fff77"} Oct 08 18:45:58 crc kubenswrapper[4859]: I1008 18:45:58.798818 4859 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-dnz8t" event={"ID":"28feb067-adfe-4b48-aef3-321e18ac0e37","Type":"ContainerStarted","Data":"335cdbf4052b4a373a2464f020256f4bdfa0a5c1daf7fded6e3040f031b7a2a8"} Oct 08 18:45:58 crc kubenswrapper[4859]: I1008 18:45:58.829602 4859 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-dnz8t" podStartSLOduration=2.396627259 podStartE2EDuration="2.829586334s" podCreationTimestamp="2025-10-08 18:45:56 +0000 UTC" firstStartedPulling="2025-10-08 18:45:57.722673065 +0000 UTC m=+1727.969512444" lastFinishedPulling="2025-10-08 18:45:58.15563214 +0000 UTC m=+1728.402471519" observedRunningTime="2025-10-08 18:45:58.824186666 +0000 UTC m=+1729.071026065" watchObservedRunningTime="2025-10-08 18:45:58.829586334 +0000 UTC m=+1729.076425713" Oct 08 18:46:02 crc kubenswrapper[4859]: I1008 18:46:02.470540 4859 scope.go:117] "RemoveContainer" containerID="44aa7848c4f5acb003170608db1921ad40aa5b2f5464f1f85d423cd26d2e8155" Oct 08 18:46:02 crc kubenswrapper[4859]: E1008 18:46:02.471889 4859 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-82s52_openshift-machine-config-operator(b23a6a6c-9d92-4e7b-840e-55cfda873a2d)\"" pod="openshift-machine-config-operator/machine-config-daemon-82s52" podUID="b23a6a6c-9d92-4e7b-840e-55cfda873a2d" Oct 08 18:46:07 crc kubenswrapper[4859]: I1008 18:46:07.038442 4859 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell0-cell-mapping-r2mmw"] Oct 08 18:46:07 crc kubenswrapper[4859]: I1008 18:46:07.047220 4859 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-cell0-cell-mapping-r2mmw"] Oct 08 18:46:08 crc kubenswrapper[4859]: I1008 18:46:08.042379 4859 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell1-conductor-db-sync-28hwn"] Oct 08 18:46:08 crc kubenswrapper[4859]: I1008 18:46:08.052858 4859 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-cell1-conductor-db-sync-28hwn"] Oct 08 18:46:08 crc kubenswrapper[4859]: I1008 18:46:08.505508 4859 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="6b0220ac-1dc0-4ca9-a9ae-9335b57820fa" path="/var/lib/kubelet/pods/6b0220ac-1dc0-4ca9-a9ae-9335b57820fa/volumes" Oct 08 18:46:08 crc kubenswrapper[4859]: I1008 18:46:08.506685 4859 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="76588317-83a8-42b4-8a4f-502d6889c7b5" path="/var/lib/kubelet/pods/76588317-83a8-42b4-8a4f-502d6889c7b5/volumes" Oct 08 18:46:13 crc kubenswrapper[4859]: I1008 18:46:13.471880 4859 scope.go:117] "RemoveContainer" containerID="44aa7848c4f5acb003170608db1921ad40aa5b2f5464f1f85d423cd26d2e8155" Oct 08 18:46:13 crc kubenswrapper[4859]: E1008 18:46:13.473305 4859 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-82s52_openshift-machine-config-operator(b23a6a6c-9d92-4e7b-840e-55cfda873a2d)\"" pod="openshift-machine-config-operator/machine-config-daemon-82s52" podUID="b23a6a6c-9d92-4e7b-840e-55cfda873a2d" Oct 08 18:46:23 crc kubenswrapper[4859]: I1008 18:46:23.635242 4859 scope.go:117] "RemoveContainer" containerID="6fb9ae99d0df9addcb1423e79ad8723c8a3d60ad81028ad521e0f4c0280241b7" Oct 08 18:46:23 crc kubenswrapper[4859]: I1008 18:46:23.697801 4859 scope.go:117] "RemoveContainer" containerID="38db145a1acab79ead1902ab5f38916c3ba433b24d3074e51ed13550d67ec7b4" Oct 08 18:46:23 crc kubenswrapper[4859]: I1008 18:46:23.753126 4859 scope.go:117] "RemoveContainer" containerID="dbccd234b8f52eb65169180bc0bf5c99fd5abfe26faafb1542803fd76eede24d" Oct 08 18:46:24 crc kubenswrapper[4859]: I1008 18:46:24.483622 4859 scope.go:117] "RemoveContainer" containerID="44aa7848c4f5acb003170608db1921ad40aa5b2f5464f1f85d423cd26d2e8155" Oct 08 18:46:24 crc kubenswrapper[4859]: E1008 18:46:24.484158 4859 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-82s52_openshift-machine-config-operator(b23a6a6c-9d92-4e7b-840e-55cfda873a2d)\"" pod="openshift-machine-config-operator/machine-config-daemon-82s52" podUID="b23a6a6c-9d92-4e7b-840e-55cfda873a2d" Oct 08 18:46:35 crc kubenswrapper[4859]: I1008 18:46:35.470788 4859 scope.go:117] "RemoveContainer" containerID="44aa7848c4f5acb003170608db1921ad40aa5b2f5464f1f85d423cd26d2e8155" Oct 08 18:46:35 crc kubenswrapper[4859]: E1008 18:46:35.471539 4859 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-82s52_openshift-machine-config-operator(b23a6a6c-9d92-4e7b-840e-55cfda873a2d)\"" pod="openshift-machine-config-operator/machine-config-daemon-82s52" podUID="b23a6a6c-9d92-4e7b-840e-55cfda873a2d" Oct 08 18:46:50 crc kubenswrapper[4859]: I1008 18:46:50.038183 4859 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell1-cell-mapping-9bz5d"] Oct 08 18:46:50 crc kubenswrapper[4859]: I1008 18:46:50.045487 4859 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-cell1-cell-mapping-9bz5d"] Oct 08 18:46:50 crc kubenswrapper[4859]: I1008 18:46:50.483510 4859 scope.go:117] "RemoveContainer" containerID="44aa7848c4f5acb003170608db1921ad40aa5b2f5464f1f85d423cd26d2e8155" Oct 08 18:46:50 crc kubenswrapper[4859]: I1008 18:46:50.483824 4859 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="7a928d30-8b64-4daf-9731-abea3343acbc" path="/var/lib/kubelet/pods/7a928d30-8b64-4daf-9731-abea3343acbc/volumes" Oct 08 18:46:50 crc kubenswrapper[4859]: E1008 18:46:50.484126 4859 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-82s52_openshift-machine-config-operator(b23a6a6c-9d92-4e7b-840e-55cfda873a2d)\"" pod="openshift-machine-config-operator/machine-config-daemon-82s52" podUID="b23a6a6c-9d92-4e7b-840e-55cfda873a2d" Oct 08 18:46:58 crc kubenswrapper[4859]: I1008 18:46:58.388351 4859 generic.go:334] "Generic (PLEG): container finished" podID="28feb067-adfe-4b48-aef3-321e18ac0e37" containerID="335cdbf4052b4a373a2464f020256f4bdfa0a5c1daf7fded6e3040f031b7a2a8" exitCode=2 Oct 08 18:46:58 crc kubenswrapper[4859]: I1008 18:46:58.388656 4859 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-dnz8t" event={"ID":"28feb067-adfe-4b48-aef3-321e18ac0e37","Type":"ContainerDied","Data":"335cdbf4052b4a373a2464f020256f4bdfa0a5c1daf7fded6e3040f031b7a2a8"} Oct 08 18:46:59 crc kubenswrapper[4859]: I1008 18:46:59.876294 4859 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-dnz8t" Oct 08 18:47:00 crc kubenswrapper[4859]: I1008 18:47:00.039284 4859 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/28feb067-adfe-4b48-aef3-321e18ac0e37-ssh-key\") pod \"28feb067-adfe-4b48-aef3-321e18ac0e37\" (UID: \"28feb067-adfe-4b48-aef3-321e18ac0e37\") " Oct 08 18:47:00 crc kubenswrapper[4859]: I1008 18:47:00.039445 4859 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/28feb067-adfe-4b48-aef3-321e18ac0e37-inventory\") pod \"28feb067-adfe-4b48-aef3-321e18ac0e37\" (UID: \"28feb067-adfe-4b48-aef3-321e18ac0e37\") " Oct 08 18:47:00 crc kubenswrapper[4859]: I1008 18:47:00.039504 4859 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-4nfmq\" (UniqueName: \"kubernetes.io/projected/28feb067-adfe-4b48-aef3-321e18ac0e37-kube-api-access-4nfmq\") pod \"28feb067-adfe-4b48-aef3-321e18ac0e37\" (UID: \"28feb067-adfe-4b48-aef3-321e18ac0e37\") " Oct 08 18:47:00 crc kubenswrapper[4859]: I1008 18:47:00.052000 4859 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/28feb067-adfe-4b48-aef3-321e18ac0e37-kube-api-access-4nfmq" (OuterVolumeSpecName: "kube-api-access-4nfmq") pod "28feb067-adfe-4b48-aef3-321e18ac0e37" (UID: "28feb067-adfe-4b48-aef3-321e18ac0e37"). InnerVolumeSpecName "kube-api-access-4nfmq". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 08 18:47:00 crc kubenswrapper[4859]: I1008 18:47:00.067488 4859 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/28feb067-adfe-4b48-aef3-321e18ac0e37-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "28feb067-adfe-4b48-aef3-321e18ac0e37" (UID: "28feb067-adfe-4b48-aef3-321e18ac0e37"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 08 18:47:00 crc kubenswrapper[4859]: I1008 18:47:00.074246 4859 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/28feb067-adfe-4b48-aef3-321e18ac0e37-inventory" (OuterVolumeSpecName: "inventory") pod "28feb067-adfe-4b48-aef3-321e18ac0e37" (UID: "28feb067-adfe-4b48-aef3-321e18ac0e37"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 08 18:47:00 crc kubenswrapper[4859]: I1008 18:47:00.143004 4859 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/28feb067-adfe-4b48-aef3-321e18ac0e37-inventory\") on node \"crc\" DevicePath \"\"" Oct 08 18:47:00 crc kubenswrapper[4859]: I1008 18:47:00.143064 4859 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-4nfmq\" (UniqueName: \"kubernetes.io/projected/28feb067-adfe-4b48-aef3-321e18ac0e37-kube-api-access-4nfmq\") on node \"crc\" DevicePath \"\"" Oct 08 18:47:00 crc kubenswrapper[4859]: I1008 18:47:00.143085 4859 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/28feb067-adfe-4b48-aef3-321e18ac0e37-ssh-key\") on node \"crc\" DevicePath \"\"" Oct 08 18:47:00 crc kubenswrapper[4859]: I1008 18:47:00.412808 4859 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-dnz8t" event={"ID":"28feb067-adfe-4b48-aef3-321e18ac0e37","Type":"ContainerDied","Data":"d9b428c0d11bea8cbbe90fa09b16489cb58b78769acd2bef609a37097a8fff77"} Oct 08 18:47:00 crc kubenswrapper[4859]: I1008 18:47:00.412848 4859 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="d9b428c0d11bea8cbbe90fa09b16489cb58b78769acd2bef609a37097a8fff77" Oct 08 18:47:00 crc kubenswrapper[4859]: I1008 18:47:00.412907 4859 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-dnz8t" Oct 08 18:47:05 crc kubenswrapper[4859]: I1008 18:47:05.470354 4859 scope.go:117] "RemoveContainer" containerID="44aa7848c4f5acb003170608db1921ad40aa5b2f5464f1f85d423cd26d2e8155" Oct 08 18:47:05 crc kubenswrapper[4859]: E1008 18:47:05.472015 4859 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-82s52_openshift-machine-config-operator(b23a6a6c-9d92-4e7b-840e-55cfda873a2d)\"" pod="openshift-machine-config-operator/machine-config-daemon-82s52" podUID="b23a6a6c-9d92-4e7b-840e-55cfda873a2d" Oct 08 18:47:07 crc kubenswrapper[4859]: I1008 18:47:07.027803 4859 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/configure-os-edpm-deployment-openstack-edpm-ipam-5dz4q"] Oct 08 18:47:07 crc kubenswrapper[4859]: E1008 18:47:07.028608 4859 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="28feb067-adfe-4b48-aef3-321e18ac0e37" containerName="configure-os-edpm-deployment-openstack-edpm-ipam" Oct 08 18:47:07 crc kubenswrapper[4859]: I1008 18:47:07.028626 4859 state_mem.go:107] "Deleted CPUSet assignment" podUID="28feb067-adfe-4b48-aef3-321e18ac0e37" containerName="configure-os-edpm-deployment-openstack-edpm-ipam" Oct 08 18:47:07 crc kubenswrapper[4859]: I1008 18:47:07.028882 4859 memory_manager.go:354] "RemoveStaleState removing state" podUID="28feb067-adfe-4b48-aef3-321e18ac0e37" containerName="configure-os-edpm-deployment-openstack-edpm-ipam" Oct 08 18:47:07 crc kubenswrapper[4859]: I1008 18:47:07.029828 4859 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-5dz4q" Oct 08 18:47:07 crc kubenswrapper[4859]: I1008 18:47:07.032135 4859 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-gxqst" Oct 08 18:47:07 crc kubenswrapper[4859]: I1008 18:47:07.032718 4859 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Oct 08 18:47:07 crc kubenswrapper[4859]: I1008 18:47:07.034221 4859 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Oct 08 18:47:07 crc kubenswrapper[4859]: I1008 18:47:07.035227 4859 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Oct 08 18:47:07 crc kubenswrapper[4859]: I1008 18:47:07.063133 4859 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/configure-os-edpm-deployment-openstack-edpm-ipam-5dz4q"] Oct 08 18:47:07 crc kubenswrapper[4859]: I1008 18:47:07.184651 4859 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/ff5a590b-fac3-4cf3-b472-db2f43030032-inventory\") pod \"configure-os-edpm-deployment-openstack-edpm-ipam-5dz4q\" (UID: \"ff5a590b-fac3-4cf3-b472-db2f43030032\") " pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-5dz4q" Oct 08 18:47:07 crc kubenswrapper[4859]: I1008 18:47:07.184847 4859 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-68rpm\" (UniqueName: \"kubernetes.io/projected/ff5a590b-fac3-4cf3-b472-db2f43030032-kube-api-access-68rpm\") pod \"configure-os-edpm-deployment-openstack-edpm-ipam-5dz4q\" (UID: \"ff5a590b-fac3-4cf3-b472-db2f43030032\") " pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-5dz4q" Oct 08 18:47:07 crc kubenswrapper[4859]: I1008 18:47:07.184936 4859 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/ff5a590b-fac3-4cf3-b472-db2f43030032-ssh-key\") pod \"configure-os-edpm-deployment-openstack-edpm-ipam-5dz4q\" (UID: \"ff5a590b-fac3-4cf3-b472-db2f43030032\") " pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-5dz4q" Oct 08 18:47:07 crc kubenswrapper[4859]: I1008 18:47:07.287376 4859 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/ff5a590b-fac3-4cf3-b472-db2f43030032-inventory\") pod \"configure-os-edpm-deployment-openstack-edpm-ipam-5dz4q\" (UID: \"ff5a590b-fac3-4cf3-b472-db2f43030032\") " pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-5dz4q" Oct 08 18:47:07 crc kubenswrapper[4859]: I1008 18:47:07.287436 4859 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-68rpm\" (UniqueName: \"kubernetes.io/projected/ff5a590b-fac3-4cf3-b472-db2f43030032-kube-api-access-68rpm\") pod \"configure-os-edpm-deployment-openstack-edpm-ipam-5dz4q\" (UID: \"ff5a590b-fac3-4cf3-b472-db2f43030032\") " pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-5dz4q" Oct 08 18:47:07 crc kubenswrapper[4859]: I1008 18:47:07.287559 4859 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/ff5a590b-fac3-4cf3-b472-db2f43030032-ssh-key\") pod \"configure-os-edpm-deployment-openstack-edpm-ipam-5dz4q\" (UID: \"ff5a590b-fac3-4cf3-b472-db2f43030032\") " pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-5dz4q" Oct 08 18:47:07 crc kubenswrapper[4859]: I1008 18:47:07.293569 4859 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/ff5a590b-fac3-4cf3-b472-db2f43030032-ssh-key\") pod \"configure-os-edpm-deployment-openstack-edpm-ipam-5dz4q\" (UID: \"ff5a590b-fac3-4cf3-b472-db2f43030032\") " pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-5dz4q" Oct 08 18:47:07 crc kubenswrapper[4859]: I1008 18:47:07.293601 4859 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/ff5a590b-fac3-4cf3-b472-db2f43030032-inventory\") pod \"configure-os-edpm-deployment-openstack-edpm-ipam-5dz4q\" (UID: \"ff5a590b-fac3-4cf3-b472-db2f43030032\") " pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-5dz4q" Oct 08 18:47:07 crc kubenswrapper[4859]: I1008 18:47:07.305711 4859 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-68rpm\" (UniqueName: \"kubernetes.io/projected/ff5a590b-fac3-4cf3-b472-db2f43030032-kube-api-access-68rpm\") pod \"configure-os-edpm-deployment-openstack-edpm-ipam-5dz4q\" (UID: \"ff5a590b-fac3-4cf3-b472-db2f43030032\") " pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-5dz4q" Oct 08 18:47:07 crc kubenswrapper[4859]: I1008 18:47:07.374256 4859 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-5dz4q" Oct 08 18:47:07 crc kubenswrapper[4859]: I1008 18:47:07.889471 4859 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/configure-os-edpm-deployment-openstack-edpm-ipam-5dz4q"] Oct 08 18:47:07 crc kubenswrapper[4859]: W1008 18:47:07.897578 4859 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podff5a590b_fac3_4cf3_b472_db2f43030032.slice/crio-84890fa1fae95e09771b9f3a43b1725ec625c0b960155aee69f106135de530cc WatchSource:0}: Error finding container 84890fa1fae95e09771b9f3a43b1725ec625c0b960155aee69f106135de530cc: Status 404 returned error can't find the container with id 84890fa1fae95e09771b9f3a43b1725ec625c0b960155aee69f106135de530cc Oct 08 18:47:08 crc kubenswrapper[4859]: I1008 18:47:08.501314 4859 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-5dz4q" event={"ID":"ff5a590b-fac3-4cf3-b472-db2f43030032","Type":"ContainerStarted","Data":"84890fa1fae95e09771b9f3a43b1725ec625c0b960155aee69f106135de530cc"} Oct 08 18:47:09 crc kubenswrapper[4859]: I1008 18:47:09.510431 4859 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-5dz4q" event={"ID":"ff5a590b-fac3-4cf3-b472-db2f43030032","Type":"ContainerStarted","Data":"b91eca0a36c0d04e6f4ec2092f7a5bc1c1de8a6ac60119fbc049f0f9c11fd0e3"} Oct 08 18:47:09 crc kubenswrapper[4859]: I1008 18:47:09.529432 4859 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-5dz4q" podStartSLOduration=2.081576197 podStartE2EDuration="2.529409636s" podCreationTimestamp="2025-10-08 18:47:07 +0000 UTC" firstStartedPulling="2025-10-08 18:47:07.900464157 +0000 UTC m=+1798.147303536" lastFinishedPulling="2025-10-08 18:47:08.348297566 +0000 UTC m=+1798.595136975" observedRunningTime="2025-10-08 18:47:09.522635319 +0000 UTC m=+1799.769474698" watchObservedRunningTime="2025-10-08 18:47:09.529409636 +0000 UTC m=+1799.776249015" Oct 08 18:47:18 crc kubenswrapper[4859]: I1008 18:47:18.477508 4859 scope.go:117] "RemoveContainer" containerID="44aa7848c4f5acb003170608db1921ad40aa5b2f5464f1f85d423cd26d2e8155" Oct 08 18:47:18 crc kubenswrapper[4859]: E1008 18:47:18.478644 4859 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-82s52_openshift-machine-config-operator(b23a6a6c-9d92-4e7b-840e-55cfda873a2d)\"" pod="openshift-machine-config-operator/machine-config-daemon-82s52" podUID="b23a6a6c-9d92-4e7b-840e-55cfda873a2d" Oct 08 18:47:23 crc kubenswrapper[4859]: I1008 18:47:23.868974 4859 scope.go:117] "RemoveContainer" containerID="7d3cd58642ece2fbfeef885a1db9ce1905a67a3195f6b96dfdd563340c9cc28b" Oct 08 18:47:29 crc kubenswrapper[4859]: I1008 18:47:29.469565 4859 scope.go:117] "RemoveContainer" containerID="44aa7848c4f5acb003170608db1921ad40aa5b2f5464f1f85d423cd26d2e8155" Oct 08 18:47:29 crc kubenswrapper[4859]: E1008 18:47:29.470509 4859 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-82s52_openshift-machine-config-operator(b23a6a6c-9d92-4e7b-840e-55cfda873a2d)\"" pod="openshift-machine-config-operator/machine-config-daemon-82s52" podUID="b23a6a6c-9d92-4e7b-840e-55cfda873a2d" Oct 08 18:47:40 crc kubenswrapper[4859]: I1008 18:47:40.480653 4859 scope.go:117] "RemoveContainer" containerID="44aa7848c4f5acb003170608db1921ad40aa5b2f5464f1f85d423cd26d2e8155" Oct 08 18:47:40 crc kubenswrapper[4859]: E1008 18:47:40.481597 4859 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-82s52_openshift-machine-config-operator(b23a6a6c-9d92-4e7b-840e-55cfda873a2d)\"" pod="openshift-machine-config-operator/machine-config-daemon-82s52" podUID="b23a6a6c-9d92-4e7b-840e-55cfda873a2d" Oct 08 18:47:55 crc kubenswrapper[4859]: I1008 18:47:55.470143 4859 scope.go:117] "RemoveContainer" containerID="44aa7848c4f5acb003170608db1921ad40aa5b2f5464f1f85d423cd26d2e8155" Oct 08 18:47:55 crc kubenswrapper[4859]: E1008 18:47:55.470983 4859 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-82s52_openshift-machine-config-operator(b23a6a6c-9d92-4e7b-840e-55cfda873a2d)\"" pod="openshift-machine-config-operator/machine-config-daemon-82s52" podUID="b23a6a6c-9d92-4e7b-840e-55cfda873a2d" Oct 08 18:47:57 crc kubenswrapper[4859]: I1008 18:47:57.950494 4859 generic.go:334] "Generic (PLEG): container finished" podID="ff5a590b-fac3-4cf3-b472-db2f43030032" containerID="b91eca0a36c0d04e6f4ec2092f7a5bc1c1de8a6ac60119fbc049f0f9c11fd0e3" exitCode=0 Oct 08 18:47:57 crc kubenswrapper[4859]: I1008 18:47:57.950582 4859 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-5dz4q" event={"ID":"ff5a590b-fac3-4cf3-b472-db2f43030032","Type":"ContainerDied","Data":"b91eca0a36c0d04e6f4ec2092f7a5bc1c1de8a6ac60119fbc049f0f9c11fd0e3"} Oct 08 18:47:59 crc kubenswrapper[4859]: I1008 18:47:59.373560 4859 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-5dz4q" Oct 08 18:47:59 crc kubenswrapper[4859]: I1008 18:47:59.556580 4859 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/ff5a590b-fac3-4cf3-b472-db2f43030032-ssh-key\") pod \"ff5a590b-fac3-4cf3-b472-db2f43030032\" (UID: \"ff5a590b-fac3-4cf3-b472-db2f43030032\") " Oct 08 18:47:59 crc kubenswrapper[4859]: I1008 18:47:59.556665 4859 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/ff5a590b-fac3-4cf3-b472-db2f43030032-inventory\") pod \"ff5a590b-fac3-4cf3-b472-db2f43030032\" (UID: \"ff5a590b-fac3-4cf3-b472-db2f43030032\") " Oct 08 18:47:59 crc kubenswrapper[4859]: I1008 18:47:59.556750 4859 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-68rpm\" (UniqueName: \"kubernetes.io/projected/ff5a590b-fac3-4cf3-b472-db2f43030032-kube-api-access-68rpm\") pod \"ff5a590b-fac3-4cf3-b472-db2f43030032\" (UID: \"ff5a590b-fac3-4cf3-b472-db2f43030032\") " Oct 08 18:47:59 crc kubenswrapper[4859]: I1008 18:47:59.573000 4859 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/ff5a590b-fac3-4cf3-b472-db2f43030032-kube-api-access-68rpm" (OuterVolumeSpecName: "kube-api-access-68rpm") pod "ff5a590b-fac3-4cf3-b472-db2f43030032" (UID: "ff5a590b-fac3-4cf3-b472-db2f43030032"). InnerVolumeSpecName "kube-api-access-68rpm". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 08 18:47:59 crc kubenswrapper[4859]: I1008 18:47:59.593412 4859 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/ff5a590b-fac3-4cf3-b472-db2f43030032-inventory" (OuterVolumeSpecName: "inventory") pod "ff5a590b-fac3-4cf3-b472-db2f43030032" (UID: "ff5a590b-fac3-4cf3-b472-db2f43030032"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 08 18:47:59 crc kubenswrapper[4859]: I1008 18:47:59.596186 4859 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/ff5a590b-fac3-4cf3-b472-db2f43030032-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "ff5a590b-fac3-4cf3-b472-db2f43030032" (UID: "ff5a590b-fac3-4cf3-b472-db2f43030032"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 08 18:47:59 crc kubenswrapper[4859]: I1008 18:47:59.661006 4859 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/ff5a590b-fac3-4cf3-b472-db2f43030032-ssh-key\") on node \"crc\" DevicePath \"\"" Oct 08 18:47:59 crc kubenswrapper[4859]: I1008 18:47:59.661058 4859 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/ff5a590b-fac3-4cf3-b472-db2f43030032-inventory\") on node \"crc\" DevicePath \"\"" Oct 08 18:47:59 crc kubenswrapper[4859]: I1008 18:47:59.661070 4859 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-68rpm\" (UniqueName: \"kubernetes.io/projected/ff5a590b-fac3-4cf3-b472-db2f43030032-kube-api-access-68rpm\") on node \"crc\" DevicePath \"\"" Oct 08 18:47:59 crc kubenswrapper[4859]: I1008 18:47:59.972978 4859 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-5dz4q" event={"ID":"ff5a590b-fac3-4cf3-b472-db2f43030032","Type":"ContainerDied","Data":"84890fa1fae95e09771b9f3a43b1725ec625c0b960155aee69f106135de530cc"} Oct 08 18:47:59 crc kubenswrapper[4859]: I1008 18:47:59.973271 4859 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="84890fa1fae95e09771b9f3a43b1725ec625c0b960155aee69f106135de530cc" Oct 08 18:47:59 crc kubenswrapper[4859]: I1008 18:47:59.973206 4859 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-5dz4q" Oct 08 18:48:00 crc kubenswrapper[4859]: I1008 18:48:00.059560 4859 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ssh-known-hosts-edpm-deployment-f546c"] Oct 08 18:48:00 crc kubenswrapper[4859]: E1008 18:48:00.060373 4859 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ff5a590b-fac3-4cf3-b472-db2f43030032" containerName="configure-os-edpm-deployment-openstack-edpm-ipam" Oct 08 18:48:00 crc kubenswrapper[4859]: I1008 18:48:00.060486 4859 state_mem.go:107] "Deleted CPUSet assignment" podUID="ff5a590b-fac3-4cf3-b472-db2f43030032" containerName="configure-os-edpm-deployment-openstack-edpm-ipam" Oct 08 18:48:00 crc kubenswrapper[4859]: I1008 18:48:00.060843 4859 memory_manager.go:354] "RemoveStaleState removing state" podUID="ff5a590b-fac3-4cf3-b472-db2f43030032" containerName="configure-os-edpm-deployment-openstack-edpm-ipam" Oct 08 18:48:00 crc kubenswrapper[4859]: I1008 18:48:00.061958 4859 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ssh-known-hosts-edpm-deployment-f546c" Oct 08 18:48:00 crc kubenswrapper[4859]: I1008 18:48:00.064334 4859 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Oct 08 18:48:00 crc kubenswrapper[4859]: I1008 18:48:00.067945 4859 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Oct 08 18:48:00 crc kubenswrapper[4859]: I1008 18:48:00.068055 4859 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-gxqst" Oct 08 18:48:00 crc kubenswrapper[4859]: I1008 18:48:00.068065 4859 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Oct 08 18:48:00 crc kubenswrapper[4859]: I1008 18:48:00.073887 4859 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ssh-known-hosts-edpm-deployment-f546c"] Oct 08 18:48:00 crc kubenswrapper[4859]: I1008 18:48:00.170659 4859 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory-0\" (UniqueName: \"kubernetes.io/secret/252ee072-dfc8-4b4b-a8dc-c695f2be633a-inventory-0\") pod \"ssh-known-hosts-edpm-deployment-f546c\" (UID: \"252ee072-dfc8-4b4b-a8dc-c695f2be633a\") " pod="openstack/ssh-known-hosts-edpm-deployment-f546c" Oct 08 18:48:00 crc kubenswrapper[4859]: I1008 18:48:00.170894 4859 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/252ee072-dfc8-4b4b-a8dc-c695f2be633a-ssh-key-openstack-edpm-ipam\") pod \"ssh-known-hosts-edpm-deployment-f546c\" (UID: \"252ee072-dfc8-4b4b-a8dc-c695f2be633a\") " pod="openstack/ssh-known-hosts-edpm-deployment-f546c" Oct 08 18:48:00 crc kubenswrapper[4859]: I1008 18:48:00.170943 4859 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-d8xnm\" (UniqueName: \"kubernetes.io/projected/252ee072-dfc8-4b4b-a8dc-c695f2be633a-kube-api-access-d8xnm\") pod \"ssh-known-hosts-edpm-deployment-f546c\" (UID: \"252ee072-dfc8-4b4b-a8dc-c695f2be633a\") " pod="openstack/ssh-known-hosts-edpm-deployment-f546c" Oct 08 18:48:00 crc kubenswrapper[4859]: I1008 18:48:00.272275 4859 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/252ee072-dfc8-4b4b-a8dc-c695f2be633a-ssh-key-openstack-edpm-ipam\") pod \"ssh-known-hosts-edpm-deployment-f546c\" (UID: \"252ee072-dfc8-4b4b-a8dc-c695f2be633a\") " pod="openstack/ssh-known-hosts-edpm-deployment-f546c" Oct 08 18:48:00 crc kubenswrapper[4859]: I1008 18:48:00.272336 4859 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-d8xnm\" (UniqueName: \"kubernetes.io/projected/252ee072-dfc8-4b4b-a8dc-c695f2be633a-kube-api-access-d8xnm\") pod \"ssh-known-hosts-edpm-deployment-f546c\" (UID: \"252ee072-dfc8-4b4b-a8dc-c695f2be633a\") " pod="openstack/ssh-known-hosts-edpm-deployment-f546c" Oct 08 18:48:00 crc kubenswrapper[4859]: I1008 18:48:00.272442 4859 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory-0\" (UniqueName: \"kubernetes.io/secret/252ee072-dfc8-4b4b-a8dc-c695f2be633a-inventory-0\") pod \"ssh-known-hosts-edpm-deployment-f546c\" (UID: \"252ee072-dfc8-4b4b-a8dc-c695f2be633a\") " pod="openstack/ssh-known-hosts-edpm-deployment-f546c" Oct 08 18:48:00 crc kubenswrapper[4859]: I1008 18:48:00.277172 4859 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/252ee072-dfc8-4b4b-a8dc-c695f2be633a-ssh-key-openstack-edpm-ipam\") pod \"ssh-known-hosts-edpm-deployment-f546c\" (UID: \"252ee072-dfc8-4b4b-a8dc-c695f2be633a\") " pod="openstack/ssh-known-hosts-edpm-deployment-f546c" Oct 08 18:48:00 crc kubenswrapper[4859]: I1008 18:48:00.282314 4859 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory-0\" (UniqueName: \"kubernetes.io/secret/252ee072-dfc8-4b4b-a8dc-c695f2be633a-inventory-0\") pod \"ssh-known-hosts-edpm-deployment-f546c\" (UID: \"252ee072-dfc8-4b4b-a8dc-c695f2be633a\") " pod="openstack/ssh-known-hosts-edpm-deployment-f546c" Oct 08 18:48:00 crc kubenswrapper[4859]: I1008 18:48:00.290382 4859 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-d8xnm\" (UniqueName: \"kubernetes.io/projected/252ee072-dfc8-4b4b-a8dc-c695f2be633a-kube-api-access-d8xnm\") pod \"ssh-known-hosts-edpm-deployment-f546c\" (UID: \"252ee072-dfc8-4b4b-a8dc-c695f2be633a\") " pod="openstack/ssh-known-hosts-edpm-deployment-f546c" Oct 08 18:48:00 crc kubenswrapper[4859]: I1008 18:48:00.379165 4859 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ssh-known-hosts-edpm-deployment-f546c" Oct 08 18:48:00 crc kubenswrapper[4859]: I1008 18:48:00.903317 4859 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ssh-known-hosts-edpm-deployment-f546c"] Oct 08 18:48:00 crc kubenswrapper[4859]: I1008 18:48:00.983744 4859 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ssh-known-hosts-edpm-deployment-f546c" event={"ID":"252ee072-dfc8-4b4b-a8dc-c695f2be633a","Type":"ContainerStarted","Data":"4353b6b3697277c53a146fa0ee4977edbd0871e1d390f08246a2ad50e6c12cd5"} Oct 08 18:48:01 crc kubenswrapper[4859]: I1008 18:48:01.992763 4859 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ssh-known-hosts-edpm-deployment-f546c" event={"ID":"252ee072-dfc8-4b4b-a8dc-c695f2be633a","Type":"ContainerStarted","Data":"f23fff281179e8eefbd5ff4c297ada0c78ad16c8f3f3f2d9b92d5678f4d69168"} Oct 08 18:48:02 crc kubenswrapper[4859]: I1008 18:48:02.011507 4859 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ssh-known-hosts-edpm-deployment-f546c" podStartSLOduration=1.413836971 podStartE2EDuration="2.011488129s" podCreationTimestamp="2025-10-08 18:48:00 +0000 UTC" firstStartedPulling="2025-10-08 18:48:00.909979241 +0000 UTC m=+1851.156818620" lastFinishedPulling="2025-10-08 18:48:01.507630359 +0000 UTC m=+1851.754469778" observedRunningTime="2025-10-08 18:48:02.006364782 +0000 UTC m=+1852.253204171" watchObservedRunningTime="2025-10-08 18:48:02.011488129 +0000 UTC m=+1852.258327508" Oct 08 18:48:07 crc kubenswrapper[4859]: I1008 18:48:07.470615 4859 scope.go:117] "RemoveContainer" containerID="44aa7848c4f5acb003170608db1921ad40aa5b2f5464f1f85d423cd26d2e8155" Oct 08 18:48:07 crc kubenswrapper[4859]: E1008 18:48:07.471668 4859 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-82s52_openshift-machine-config-operator(b23a6a6c-9d92-4e7b-840e-55cfda873a2d)\"" pod="openshift-machine-config-operator/machine-config-daemon-82s52" podUID="b23a6a6c-9d92-4e7b-840e-55cfda873a2d" Oct 08 18:48:09 crc kubenswrapper[4859]: I1008 18:48:09.050810 4859 generic.go:334] "Generic (PLEG): container finished" podID="252ee072-dfc8-4b4b-a8dc-c695f2be633a" containerID="f23fff281179e8eefbd5ff4c297ada0c78ad16c8f3f3f2d9b92d5678f4d69168" exitCode=0 Oct 08 18:48:09 crc kubenswrapper[4859]: I1008 18:48:09.050958 4859 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ssh-known-hosts-edpm-deployment-f546c" event={"ID":"252ee072-dfc8-4b4b-a8dc-c695f2be633a","Type":"ContainerDied","Data":"f23fff281179e8eefbd5ff4c297ada0c78ad16c8f3f3f2d9b92d5678f4d69168"} Oct 08 18:48:10 crc kubenswrapper[4859]: I1008 18:48:10.416754 4859 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ssh-known-hosts-edpm-deployment-f546c" Oct 08 18:48:10 crc kubenswrapper[4859]: I1008 18:48:10.563013 4859 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-d8xnm\" (UniqueName: \"kubernetes.io/projected/252ee072-dfc8-4b4b-a8dc-c695f2be633a-kube-api-access-d8xnm\") pod \"252ee072-dfc8-4b4b-a8dc-c695f2be633a\" (UID: \"252ee072-dfc8-4b4b-a8dc-c695f2be633a\") " Oct 08 18:48:10 crc kubenswrapper[4859]: I1008 18:48:10.563308 4859 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/252ee072-dfc8-4b4b-a8dc-c695f2be633a-ssh-key-openstack-edpm-ipam\") pod \"252ee072-dfc8-4b4b-a8dc-c695f2be633a\" (UID: \"252ee072-dfc8-4b4b-a8dc-c695f2be633a\") " Oct 08 18:48:10 crc kubenswrapper[4859]: I1008 18:48:10.563341 4859 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory-0\" (UniqueName: \"kubernetes.io/secret/252ee072-dfc8-4b4b-a8dc-c695f2be633a-inventory-0\") pod \"252ee072-dfc8-4b4b-a8dc-c695f2be633a\" (UID: \"252ee072-dfc8-4b4b-a8dc-c695f2be633a\") " Oct 08 18:48:10 crc kubenswrapper[4859]: I1008 18:48:10.568832 4859 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/252ee072-dfc8-4b4b-a8dc-c695f2be633a-kube-api-access-d8xnm" (OuterVolumeSpecName: "kube-api-access-d8xnm") pod "252ee072-dfc8-4b4b-a8dc-c695f2be633a" (UID: "252ee072-dfc8-4b4b-a8dc-c695f2be633a"). InnerVolumeSpecName "kube-api-access-d8xnm". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 08 18:48:10 crc kubenswrapper[4859]: I1008 18:48:10.592531 4859 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/252ee072-dfc8-4b4b-a8dc-c695f2be633a-ssh-key-openstack-edpm-ipam" (OuterVolumeSpecName: "ssh-key-openstack-edpm-ipam") pod "252ee072-dfc8-4b4b-a8dc-c695f2be633a" (UID: "252ee072-dfc8-4b4b-a8dc-c695f2be633a"). InnerVolumeSpecName "ssh-key-openstack-edpm-ipam". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 08 18:48:10 crc kubenswrapper[4859]: I1008 18:48:10.597273 4859 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/252ee072-dfc8-4b4b-a8dc-c695f2be633a-inventory-0" (OuterVolumeSpecName: "inventory-0") pod "252ee072-dfc8-4b4b-a8dc-c695f2be633a" (UID: "252ee072-dfc8-4b4b-a8dc-c695f2be633a"). InnerVolumeSpecName "inventory-0". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 08 18:48:10 crc kubenswrapper[4859]: I1008 18:48:10.666349 4859 reconciler_common.go:293] "Volume detached for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/252ee072-dfc8-4b4b-a8dc-c695f2be633a-ssh-key-openstack-edpm-ipam\") on node \"crc\" DevicePath \"\"" Oct 08 18:48:10 crc kubenswrapper[4859]: I1008 18:48:10.666395 4859 reconciler_common.go:293] "Volume detached for volume \"inventory-0\" (UniqueName: \"kubernetes.io/secret/252ee072-dfc8-4b4b-a8dc-c695f2be633a-inventory-0\") on node \"crc\" DevicePath \"\"" Oct 08 18:48:10 crc kubenswrapper[4859]: I1008 18:48:10.666411 4859 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-d8xnm\" (UniqueName: \"kubernetes.io/projected/252ee072-dfc8-4b4b-a8dc-c695f2be633a-kube-api-access-d8xnm\") on node \"crc\" DevicePath \"\"" Oct 08 18:48:11 crc kubenswrapper[4859]: I1008 18:48:11.071127 4859 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ssh-known-hosts-edpm-deployment-f546c" event={"ID":"252ee072-dfc8-4b4b-a8dc-c695f2be633a","Type":"ContainerDied","Data":"4353b6b3697277c53a146fa0ee4977edbd0871e1d390f08246a2ad50e6c12cd5"} Oct 08 18:48:11 crc kubenswrapper[4859]: I1008 18:48:11.071405 4859 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="4353b6b3697277c53a146fa0ee4977edbd0871e1d390f08246a2ad50e6c12cd5" Oct 08 18:48:11 crc kubenswrapper[4859]: I1008 18:48:11.071312 4859 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ssh-known-hosts-edpm-deployment-f546c" Oct 08 18:48:11 crc kubenswrapper[4859]: I1008 18:48:11.144369 4859 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/run-os-edpm-deployment-openstack-edpm-ipam-746dn"] Oct 08 18:48:11 crc kubenswrapper[4859]: E1008 18:48:11.144917 4859 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="252ee072-dfc8-4b4b-a8dc-c695f2be633a" containerName="ssh-known-hosts-edpm-deployment" Oct 08 18:48:11 crc kubenswrapper[4859]: I1008 18:48:11.144941 4859 state_mem.go:107] "Deleted CPUSet assignment" podUID="252ee072-dfc8-4b4b-a8dc-c695f2be633a" containerName="ssh-known-hosts-edpm-deployment" Oct 08 18:48:11 crc kubenswrapper[4859]: I1008 18:48:11.145253 4859 memory_manager.go:354] "RemoveStaleState removing state" podUID="252ee072-dfc8-4b4b-a8dc-c695f2be633a" containerName="ssh-known-hosts-edpm-deployment" Oct 08 18:48:11 crc kubenswrapper[4859]: I1008 18:48:11.146135 4859 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-746dn" Oct 08 18:48:11 crc kubenswrapper[4859]: I1008 18:48:11.148313 4859 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-gxqst" Oct 08 18:48:11 crc kubenswrapper[4859]: I1008 18:48:11.148930 4859 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Oct 08 18:48:11 crc kubenswrapper[4859]: I1008 18:48:11.151480 4859 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Oct 08 18:48:11 crc kubenswrapper[4859]: I1008 18:48:11.156103 4859 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Oct 08 18:48:11 crc kubenswrapper[4859]: I1008 18:48:11.158137 4859 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/run-os-edpm-deployment-openstack-edpm-ipam-746dn"] Oct 08 18:48:11 crc kubenswrapper[4859]: I1008 18:48:11.281529 4859 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/e37b6e90-4b9a-41e8-9a83-db72281cc7f5-inventory\") pod \"run-os-edpm-deployment-openstack-edpm-ipam-746dn\" (UID: \"e37b6e90-4b9a-41e8-9a83-db72281cc7f5\") " pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-746dn" Oct 08 18:48:11 crc kubenswrapper[4859]: I1008 18:48:11.281624 4859 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/e37b6e90-4b9a-41e8-9a83-db72281cc7f5-ssh-key\") pod \"run-os-edpm-deployment-openstack-edpm-ipam-746dn\" (UID: \"e37b6e90-4b9a-41e8-9a83-db72281cc7f5\") " pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-746dn" Oct 08 18:48:11 crc kubenswrapper[4859]: I1008 18:48:11.281716 4859 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-rd24z\" (UniqueName: \"kubernetes.io/projected/e37b6e90-4b9a-41e8-9a83-db72281cc7f5-kube-api-access-rd24z\") pod \"run-os-edpm-deployment-openstack-edpm-ipam-746dn\" (UID: \"e37b6e90-4b9a-41e8-9a83-db72281cc7f5\") " pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-746dn" Oct 08 18:48:11 crc kubenswrapper[4859]: I1008 18:48:11.383730 4859 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rd24z\" (UniqueName: \"kubernetes.io/projected/e37b6e90-4b9a-41e8-9a83-db72281cc7f5-kube-api-access-rd24z\") pod \"run-os-edpm-deployment-openstack-edpm-ipam-746dn\" (UID: \"e37b6e90-4b9a-41e8-9a83-db72281cc7f5\") " pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-746dn" Oct 08 18:48:11 crc kubenswrapper[4859]: I1008 18:48:11.383890 4859 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/e37b6e90-4b9a-41e8-9a83-db72281cc7f5-inventory\") pod \"run-os-edpm-deployment-openstack-edpm-ipam-746dn\" (UID: \"e37b6e90-4b9a-41e8-9a83-db72281cc7f5\") " pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-746dn" Oct 08 18:48:11 crc kubenswrapper[4859]: I1008 18:48:11.384068 4859 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/e37b6e90-4b9a-41e8-9a83-db72281cc7f5-ssh-key\") pod \"run-os-edpm-deployment-openstack-edpm-ipam-746dn\" (UID: \"e37b6e90-4b9a-41e8-9a83-db72281cc7f5\") " pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-746dn" Oct 08 18:48:11 crc kubenswrapper[4859]: I1008 18:48:11.389037 4859 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/e37b6e90-4b9a-41e8-9a83-db72281cc7f5-inventory\") pod \"run-os-edpm-deployment-openstack-edpm-ipam-746dn\" (UID: \"e37b6e90-4b9a-41e8-9a83-db72281cc7f5\") " pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-746dn" Oct 08 18:48:11 crc kubenswrapper[4859]: I1008 18:48:11.389144 4859 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/e37b6e90-4b9a-41e8-9a83-db72281cc7f5-ssh-key\") pod \"run-os-edpm-deployment-openstack-edpm-ipam-746dn\" (UID: \"e37b6e90-4b9a-41e8-9a83-db72281cc7f5\") " pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-746dn" Oct 08 18:48:11 crc kubenswrapper[4859]: I1008 18:48:11.409781 4859 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rd24z\" (UniqueName: \"kubernetes.io/projected/e37b6e90-4b9a-41e8-9a83-db72281cc7f5-kube-api-access-rd24z\") pod \"run-os-edpm-deployment-openstack-edpm-ipam-746dn\" (UID: \"e37b6e90-4b9a-41e8-9a83-db72281cc7f5\") " pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-746dn" Oct 08 18:48:11 crc kubenswrapper[4859]: I1008 18:48:11.466631 4859 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-746dn" Oct 08 18:48:12 crc kubenswrapper[4859]: I1008 18:48:12.021154 4859 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/run-os-edpm-deployment-openstack-edpm-ipam-746dn"] Oct 08 18:48:12 crc kubenswrapper[4859]: I1008 18:48:12.097100 4859 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-746dn" event={"ID":"e37b6e90-4b9a-41e8-9a83-db72281cc7f5","Type":"ContainerStarted","Data":"268c81842aeef701519d904983cbba4ec4b006e0d32ed9f0304f7054374b2241"} Oct 08 18:48:13 crc kubenswrapper[4859]: I1008 18:48:13.106240 4859 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-746dn" event={"ID":"e37b6e90-4b9a-41e8-9a83-db72281cc7f5","Type":"ContainerStarted","Data":"18f808badf81367163fa817ad4c7026d190096840b5b6fbf684d6d9911cd6be0"} Oct 08 18:48:13 crc kubenswrapper[4859]: I1008 18:48:13.128215 4859 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-746dn" podStartSLOduration=1.673244169 podStartE2EDuration="2.128196057s" podCreationTimestamp="2025-10-08 18:48:11 +0000 UTC" firstStartedPulling="2025-10-08 18:48:12.026199975 +0000 UTC m=+1862.273039354" lastFinishedPulling="2025-10-08 18:48:12.481151863 +0000 UTC m=+1862.727991242" observedRunningTime="2025-10-08 18:48:13.124318015 +0000 UTC m=+1863.371157404" watchObservedRunningTime="2025-10-08 18:48:13.128196057 +0000 UTC m=+1863.375035436" Oct 08 18:48:21 crc kubenswrapper[4859]: I1008 18:48:21.469613 4859 scope.go:117] "RemoveContainer" containerID="44aa7848c4f5acb003170608db1921ad40aa5b2f5464f1f85d423cd26d2e8155" Oct 08 18:48:21 crc kubenswrapper[4859]: E1008 18:48:21.470269 4859 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-82s52_openshift-machine-config-operator(b23a6a6c-9d92-4e7b-840e-55cfda873a2d)\"" pod="openshift-machine-config-operator/machine-config-daemon-82s52" podUID="b23a6a6c-9d92-4e7b-840e-55cfda873a2d" Oct 08 18:48:22 crc kubenswrapper[4859]: I1008 18:48:22.191161 4859 generic.go:334] "Generic (PLEG): container finished" podID="e37b6e90-4b9a-41e8-9a83-db72281cc7f5" containerID="18f808badf81367163fa817ad4c7026d190096840b5b6fbf684d6d9911cd6be0" exitCode=0 Oct 08 18:48:22 crc kubenswrapper[4859]: I1008 18:48:22.191248 4859 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-746dn" event={"ID":"e37b6e90-4b9a-41e8-9a83-db72281cc7f5","Type":"ContainerDied","Data":"18f808badf81367163fa817ad4c7026d190096840b5b6fbf684d6d9911cd6be0"} Oct 08 18:48:23 crc kubenswrapper[4859]: I1008 18:48:23.585700 4859 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-746dn" Oct 08 18:48:23 crc kubenswrapper[4859]: I1008 18:48:23.765127 4859 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/e37b6e90-4b9a-41e8-9a83-db72281cc7f5-ssh-key\") pod \"e37b6e90-4b9a-41e8-9a83-db72281cc7f5\" (UID: \"e37b6e90-4b9a-41e8-9a83-db72281cc7f5\") " Oct 08 18:48:23 crc kubenswrapper[4859]: I1008 18:48:23.765413 4859 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-rd24z\" (UniqueName: \"kubernetes.io/projected/e37b6e90-4b9a-41e8-9a83-db72281cc7f5-kube-api-access-rd24z\") pod \"e37b6e90-4b9a-41e8-9a83-db72281cc7f5\" (UID: \"e37b6e90-4b9a-41e8-9a83-db72281cc7f5\") " Oct 08 18:48:23 crc kubenswrapper[4859]: I1008 18:48:23.765465 4859 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/e37b6e90-4b9a-41e8-9a83-db72281cc7f5-inventory\") pod \"e37b6e90-4b9a-41e8-9a83-db72281cc7f5\" (UID: \"e37b6e90-4b9a-41e8-9a83-db72281cc7f5\") " Oct 08 18:48:23 crc kubenswrapper[4859]: I1008 18:48:23.770998 4859 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/e37b6e90-4b9a-41e8-9a83-db72281cc7f5-kube-api-access-rd24z" (OuterVolumeSpecName: "kube-api-access-rd24z") pod "e37b6e90-4b9a-41e8-9a83-db72281cc7f5" (UID: "e37b6e90-4b9a-41e8-9a83-db72281cc7f5"). InnerVolumeSpecName "kube-api-access-rd24z". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 08 18:48:23 crc kubenswrapper[4859]: I1008 18:48:23.798032 4859 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e37b6e90-4b9a-41e8-9a83-db72281cc7f5-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "e37b6e90-4b9a-41e8-9a83-db72281cc7f5" (UID: "e37b6e90-4b9a-41e8-9a83-db72281cc7f5"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 08 18:48:23 crc kubenswrapper[4859]: I1008 18:48:23.798540 4859 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e37b6e90-4b9a-41e8-9a83-db72281cc7f5-inventory" (OuterVolumeSpecName: "inventory") pod "e37b6e90-4b9a-41e8-9a83-db72281cc7f5" (UID: "e37b6e90-4b9a-41e8-9a83-db72281cc7f5"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 08 18:48:23 crc kubenswrapper[4859]: I1008 18:48:23.867768 4859 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/e37b6e90-4b9a-41e8-9a83-db72281cc7f5-ssh-key\") on node \"crc\" DevicePath \"\"" Oct 08 18:48:23 crc kubenswrapper[4859]: I1008 18:48:23.867796 4859 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-rd24z\" (UniqueName: \"kubernetes.io/projected/e37b6e90-4b9a-41e8-9a83-db72281cc7f5-kube-api-access-rd24z\") on node \"crc\" DevicePath \"\"" Oct 08 18:48:23 crc kubenswrapper[4859]: I1008 18:48:23.867807 4859 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/e37b6e90-4b9a-41e8-9a83-db72281cc7f5-inventory\") on node \"crc\" DevicePath \"\"" Oct 08 18:48:24 crc kubenswrapper[4859]: I1008 18:48:24.209131 4859 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-746dn" event={"ID":"e37b6e90-4b9a-41e8-9a83-db72281cc7f5","Type":"ContainerDied","Data":"268c81842aeef701519d904983cbba4ec4b006e0d32ed9f0304f7054374b2241"} Oct 08 18:48:24 crc kubenswrapper[4859]: I1008 18:48:24.209176 4859 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="268c81842aeef701519d904983cbba4ec4b006e0d32ed9f0304f7054374b2241" Oct 08 18:48:24 crc kubenswrapper[4859]: I1008 18:48:24.209171 4859 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-746dn" Oct 08 18:48:24 crc kubenswrapper[4859]: I1008 18:48:24.285517 4859 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-cjdnl"] Oct 08 18:48:24 crc kubenswrapper[4859]: E1008 18:48:24.285997 4859 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e37b6e90-4b9a-41e8-9a83-db72281cc7f5" containerName="run-os-edpm-deployment-openstack-edpm-ipam" Oct 08 18:48:24 crc kubenswrapper[4859]: I1008 18:48:24.286023 4859 state_mem.go:107] "Deleted CPUSet assignment" podUID="e37b6e90-4b9a-41e8-9a83-db72281cc7f5" containerName="run-os-edpm-deployment-openstack-edpm-ipam" Oct 08 18:48:24 crc kubenswrapper[4859]: I1008 18:48:24.286293 4859 memory_manager.go:354] "RemoveStaleState removing state" podUID="e37b6e90-4b9a-41e8-9a83-db72281cc7f5" containerName="run-os-edpm-deployment-openstack-edpm-ipam" Oct 08 18:48:24 crc kubenswrapper[4859]: I1008 18:48:24.286986 4859 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-cjdnl" Oct 08 18:48:24 crc kubenswrapper[4859]: I1008 18:48:24.289318 4859 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-gxqst" Oct 08 18:48:24 crc kubenswrapper[4859]: I1008 18:48:24.289393 4859 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Oct 08 18:48:24 crc kubenswrapper[4859]: I1008 18:48:24.289615 4859 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Oct 08 18:48:24 crc kubenswrapper[4859]: I1008 18:48:24.289642 4859 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Oct 08 18:48:24 crc kubenswrapper[4859]: I1008 18:48:24.306187 4859 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-cjdnl"] Oct 08 18:48:24 crc kubenswrapper[4859]: I1008 18:48:24.481033 4859 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/5dc6ec8a-d555-43e6-b295-821ae62d0eb1-inventory\") pod \"reboot-os-edpm-deployment-openstack-edpm-ipam-cjdnl\" (UID: \"5dc6ec8a-d555-43e6-b295-821ae62d0eb1\") " pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-cjdnl" Oct 08 18:48:24 crc kubenswrapper[4859]: I1008 18:48:24.481394 4859 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-kr2f4\" (UniqueName: \"kubernetes.io/projected/5dc6ec8a-d555-43e6-b295-821ae62d0eb1-kube-api-access-kr2f4\") pod \"reboot-os-edpm-deployment-openstack-edpm-ipam-cjdnl\" (UID: \"5dc6ec8a-d555-43e6-b295-821ae62d0eb1\") " pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-cjdnl" Oct 08 18:48:24 crc kubenswrapper[4859]: I1008 18:48:24.481429 4859 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/5dc6ec8a-d555-43e6-b295-821ae62d0eb1-ssh-key\") pod \"reboot-os-edpm-deployment-openstack-edpm-ipam-cjdnl\" (UID: \"5dc6ec8a-d555-43e6-b295-821ae62d0eb1\") " pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-cjdnl" Oct 08 18:48:24 crc kubenswrapper[4859]: I1008 18:48:24.583188 4859 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/5dc6ec8a-d555-43e6-b295-821ae62d0eb1-inventory\") pod \"reboot-os-edpm-deployment-openstack-edpm-ipam-cjdnl\" (UID: \"5dc6ec8a-d555-43e6-b295-821ae62d0eb1\") " pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-cjdnl" Oct 08 18:48:24 crc kubenswrapper[4859]: I1008 18:48:24.583269 4859 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-kr2f4\" (UniqueName: \"kubernetes.io/projected/5dc6ec8a-d555-43e6-b295-821ae62d0eb1-kube-api-access-kr2f4\") pod \"reboot-os-edpm-deployment-openstack-edpm-ipam-cjdnl\" (UID: \"5dc6ec8a-d555-43e6-b295-821ae62d0eb1\") " pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-cjdnl" Oct 08 18:48:24 crc kubenswrapper[4859]: I1008 18:48:24.583310 4859 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/5dc6ec8a-d555-43e6-b295-821ae62d0eb1-ssh-key\") pod \"reboot-os-edpm-deployment-openstack-edpm-ipam-cjdnl\" (UID: \"5dc6ec8a-d555-43e6-b295-821ae62d0eb1\") " pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-cjdnl" Oct 08 18:48:24 crc kubenswrapper[4859]: I1008 18:48:24.591176 4859 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/5dc6ec8a-d555-43e6-b295-821ae62d0eb1-ssh-key\") pod \"reboot-os-edpm-deployment-openstack-edpm-ipam-cjdnl\" (UID: \"5dc6ec8a-d555-43e6-b295-821ae62d0eb1\") " pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-cjdnl" Oct 08 18:48:24 crc kubenswrapper[4859]: I1008 18:48:24.593291 4859 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/5dc6ec8a-d555-43e6-b295-821ae62d0eb1-inventory\") pod \"reboot-os-edpm-deployment-openstack-edpm-ipam-cjdnl\" (UID: \"5dc6ec8a-d555-43e6-b295-821ae62d0eb1\") " pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-cjdnl" Oct 08 18:48:24 crc kubenswrapper[4859]: I1008 18:48:24.603141 4859 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-kr2f4\" (UniqueName: \"kubernetes.io/projected/5dc6ec8a-d555-43e6-b295-821ae62d0eb1-kube-api-access-kr2f4\") pod \"reboot-os-edpm-deployment-openstack-edpm-ipam-cjdnl\" (UID: \"5dc6ec8a-d555-43e6-b295-821ae62d0eb1\") " pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-cjdnl" Oct 08 18:48:24 crc kubenswrapper[4859]: I1008 18:48:24.902730 4859 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-cjdnl" Oct 08 18:48:25 crc kubenswrapper[4859]: I1008 18:48:25.494978 4859 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-cjdnl"] Oct 08 18:48:26 crc kubenswrapper[4859]: I1008 18:48:26.226126 4859 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-cjdnl" event={"ID":"5dc6ec8a-d555-43e6-b295-821ae62d0eb1","Type":"ContainerStarted","Data":"19e19a56be5146c56364eca7acf65d46ee4f613866f131904211ecfd9314a004"} Oct 08 18:48:26 crc kubenswrapper[4859]: I1008 18:48:26.226450 4859 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-cjdnl" event={"ID":"5dc6ec8a-d555-43e6-b295-821ae62d0eb1","Type":"ContainerStarted","Data":"91c6ee39956c42f9b0a067d2b38fddc45c69a1d99f3855d51c4b2d73d60acf0f"} Oct 08 18:48:26 crc kubenswrapper[4859]: I1008 18:48:26.245877 4859 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-cjdnl" podStartSLOduration=1.820059005 podStartE2EDuration="2.245857818s" podCreationTimestamp="2025-10-08 18:48:24 +0000 UTC" firstStartedPulling="2025-10-08 18:48:25.497172851 +0000 UTC m=+1875.744012230" lastFinishedPulling="2025-10-08 18:48:25.922971624 +0000 UTC m=+1876.169811043" observedRunningTime="2025-10-08 18:48:26.238404964 +0000 UTC m=+1876.485244363" watchObservedRunningTime="2025-10-08 18:48:26.245857818 +0000 UTC m=+1876.492697197" Oct 08 18:48:36 crc kubenswrapper[4859]: I1008 18:48:36.470514 4859 scope.go:117] "RemoveContainer" containerID="44aa7848c4f5acb003170608db1921ad40aa5b2f5464f1f85d423cd26d2e8155" Oct 08 18:48:36 crc kubenswrapper[4859]: E1008 18:48:36.471729 4859 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-82s52_openshift-machine-config-operator(b23a6a6c-9d92-4e7b-840e-55cfda873a2d)\"" pod="openshift-machine-config-operator/machine-config-daemon-82s52" podUID="b23a6a6c-9d92-4e7b-840e-55cfda873a2d" Oct 08 18:48:49 crc kubenswrapper[4859]: I1008 18:48:49.470462 4859 scope.go:117] "RemoveContainer" containerID="44aa7848c4f5acb003170608db1921ad40aa5b2f5464f1f85d423cd26d2e8155" Oct 08 18:48:49 crc kubenswrapper[4859]: E1008 18:48:49.471518 4859 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-82s52_openshift-machine-config-operator(b23a6a6c-9d92-4e7b-840e-55cfda873a2d)\"" pod="openshift-machine-config-operator/machine-config-daemon-82s52" podUID="b23a6a6c-9d92-4e7b-840e-55cfda873a2d" Oct 08 18:49:02 crc kubenswrapper[4859]: I1008 18:49:02.470372 4859 scope.go:117] "RemoveContainer" containerID="44aa7848c4f5acb003170608db1921ad40aa5b2f5464f1f85d423cd26d2e8155" Oct 08 18:49:02 crc kubenswrapper[4859]: E1008 18:49:02.472452 4859 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-82s52_openshift-machine-config-operator(b23a6a6c-9d92-4e7b-840e-55cfda873a2d)\"" pod="openshift-machine-config-operator/machine-config-daemon-82s52" podUID="b23a6a6c-9d92-4e7b-840e-55cfda873a2d" Oct 08 18:49:16 crc kubenswrapper[4859]: I1008 18:49:16.471223 4859 scope.go:117] "RemoveContainer" containerID="44aa7848c4f5acb003170608db1921ad40aa5b2f5464f1f85d423cd26d2e8155" Oct 08 18:49:16 crc kubenswrapper[4859]: E1008 18:49:16.471986 4859 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-82s52_openshift-machine-config-operator(b23a6a6c-9d92-4e7b-840e-55cfda873a2d)\"" pod="openshift-machine-config-operator/machine-config-daemon-82s52" podUID="b23a6a6c-9d92-4e7b-840e-55cfda873a2d" Oct 08 18:49:28 crc kubenswrapper[4859]: I1008 18:49:28.470601 4859 scope.go:117] "RemoveContainer" containerID="44aa7848c4f5acb003170608db1921ad40aa5b2f5464f1f85d423cd26d2e8155" Oct 08 18:49:28 crc kubenswrapper[4859]: E1008 18:49:28.471425 4859 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-82s52_openshift-machine-config-operator(b23a6a6c-9d92-4e7b-840e-55cfda873a2d)\"" pod="openshift-machine-config-operator/machine-config-daemon-82s52" podUID="b23a6a6c-9d92-4e7b-840e-55cfda873a2d" Oct 08 18:49:41 crc kubenswrapper[4859]: I1008 18:49:41.938305 4859 generic.go:334] "Generic (PLEG): container finished" podID="5dc6ec8a-d555-43e6-b295-821ae62d0eb1" containerID="19e19a56be5146c56364eca7acf65d46ee4f613866f131904211ecfd9314a004" exitCode=0 Oct 08 18:49:41 crc kubenswrapper[4859]: I1008 18:49:41.938386 4859 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-cjdnl" event={"ID":"5dc6ec8a-d555-43e6-b295-821ae62d0eb1","Type":"ContainerDied","Data":"19e19a56be5146c56364eca7acf65d46ee4f613866f131904211ecfd9314a004"} Oct 08 18:49:42 crc kubenswrapper[4859]: I1008 18:49:42.470933 4859 scope.go:117] "RemoveContainer" containerID="44aa7848c4f5acb003170608db1921ad40aa5b2f5464f1f85d423cd26d2e8155" Oct 08 18:49:42 crc kubenswrapper[4859]: E1008 18:49:42.471850 4859 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-82s52_openshift-machine-config-operator(b23a6a6c-9d92-4e7b-840e-55cfda873a2d)\"" pod="openshift-machine-config-operator/machine-config-daemon-82s52" podUID="b23a6a6c-9d92-4e7b-840e-55cfda873a2d" Oct 08 18:49:43 crc kubenswrapper[4859]: I1008 18:49:43.375531 4859 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-cjdnl" Oct 08 18:49:43 crc kubenswrapper[4859]: I1008 18:49:43.540280 4859 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/5dc6ec8a-d555-43e6-b295-821ae62d0eb1-ssh-key\") pod \"5dc6ec8a-d555-43e6-b295-821ae62d0eb1\" (UID: \"5dc6ec8a-d555-43e6-b295-821ae62d0eb1\") " Oct 08 18:49:43 crc kubenswrapper[4859]: I1008 18:49:43.540807 4859 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-kr2f4\" (UniqueName: \"kubernetes.io/projected/5dc6ec8a-d555-43e6-b295-821ae62d0eb1-kube-api-access-kr2f4\") pod \"5dc6ec8a-d555-43e6-b295-821ae62d0eb1\" (UID: \"5dc6ec8a-d555-43e6-b295-821ae62d0eb1\") " Oct 08 18:49:43 crc kubenswrapper[4859]: I1008 18:49:43.540923 4859 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/5dc6ec8a-d555-43e6-b295-821ae62d0eb1-inventory\") pod \"5dc6ec8a-d555-43e6-b295-821ae62d0eb1\" (UID: \"5dc6ec8a-d555-43e6-b295-821ae62d0eb1\") " Oct 08 18:49:43 crc kubenswrapper[4859]: I1008 18:49:43.545911 4859 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5dc6ec8a-d555-43e6-b295-821ae62d0eb1-kube-api-access-kr2f4" (OuterVolumeSpecName: "kube-api-access-kr2f4") pod "5dc6ec8a-d555-43e6-b295-821ae62d0eb1" (UID: "5dc6ec8a-d555-43e6-b295-821ae62d0eb1"). InnerVolumeSpecName "kube-api-access-kr2f4". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 08 18:49:43 crc kubenswrapper[4859]: I1008 18:49:43.567897 4859 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5dc6ec8a-d555-43e6-b295-821ae62d0eb1-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "5dc6ec8a-d555-43e6-b295-821ae62d0eb1" (UID: "5dc6ec8a-d555-43e6-b295-821ae62d0eb1"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 08 18:49:43 crc kubenswrapper[4859]: I1008 18:49:43.571584 4859 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5dc6ec8a-d555-43e6-b295-821ae62d0eb1-inventory" (OuterVolumeSpecName: "inventory") pod "5dc6ec8a-d555-43e6-b295-821ae62d0eb1" (UID: "5dc6ec8a-d555-43e6-b295-821ae62d0eb1"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 08 18:49:43 crc kubenswrapper[4859]: I1008 18:49:43.643328 4859 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-kr2f4\" (UniqueName: \"kubernetes.io/projected/5dc6ec8a-d555-43e6-b295-821ae62d0eb1-kube-api-access-kr2f4\") on node \"crc\" DevicePath \"\"" Oct 08 18:49:43 crc kubenswrapper[4859]: I1008 18:49:43.643373 4859 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/5dc6ec8a-d555-43e6-b295-821ae62d0eb1-inventory\") on node \"crc\" DevicePath \"\"" Oct 08 18:49:43 crc kubenswrapper[4859]: I1008 18:49:43.643386 4859 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/5dc6ec8a-d555-43e6-b295-821ae62d0eb1-ssh-key\") on node \"crc\" DevicePath \"\"" Oct 08 18:49:43 crc kubenswrapper[4859]: I1008 18:49:43.956465 4859 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-cjdnl" event={"ID":"5dc6ec8a-d555-43e6-b295-821ae62d0eb1","Type":"ContainerDied","Data":"91c6ee39956c42f9b0a067d2b38fddc45c69a1d99f3855d51c4b2d73d60acf0f"} Oct 08 18:49:43 crc kubenswrapper[4859]: I1008 18:49:43.956520 4859 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="91c6ee39956c42f9b0a067d2b38fddc45c69a1d99f3855d51c4b2d73d60acf0f" Oct 08 18:49:43 crc kubenswrapper[4859]: I1008 18:49:43.956530 4859 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-cjdnl" Oct 08 18:49:44 crc kubenswrapper[4859]: I1008 18:49:44.062747 4859 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/install-certs-edpm-deployment-openstack-edpm-ipam-w95ff"] Oct 08 18:49:44 crc kubenswrapper[4859]: E1008 18:49:44.063270 4859 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5dc6ec8a-d555-43e6-b295-821ae62d0eb1" containerName="reboot-os-edpm-deployment-openstack-edpm-ipam" Oct 08 18:49:44 crc kubenswrapper[4859]: I1008 18:49:44.063301 4859 state_mem.go:107] "Deleted CPUSet assignment" podUID="5dc6ec8a-d555-43e6-b295-821ae62d0eb1" containerName="reboot-os-edpm-deployment-openstack-edpm-ipam" Oct 08 18:49:44 crc kubenswrapper[4859]: I1008 18:49:44.063585 4859 memory_manager.go:354] "RemoveStaleState removing state" podUID="5dc6ec8a-d555-43e6-b295-821ae62d0eb1" containerName="reboot-os-edpm-deployment-openstack-edpm-ipam" Oct 08 18:49:44 crc kubenswrapper[4859]: I1008 18:49:44.064431 4859 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-w95ff" Oct 08 18:49:44 crc kubenswrapper[4859]: I1008 18:49:44.068440 4859 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Oct 08 18:49:44 crc kubenswrapper[4859]: I1008 18:49:44.068615 4859 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-telemetry-default-certs-0" Oct 08 18:49:44 crc kubenswrapper[4859]: I1008 18:49:44.068912 4859 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Oct 08 18:49:44 crc kubenswrapper[4859]: I1008 18:49:44.069662 4859 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-libvirt-default-certs-0" Oct 08 18:49:44 crc kubenswrapper[4859]: I1008 18:49:44.069771 4859 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-ovn-default-certs-0" Oct 08 18:49:44 crc kubenswrapper[4859]: I1008 18:49:44.069943 4859 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Oct 08 18:49:44 crc kubenswrapper[4859]: I1008 18:49:44.075956 4859 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/install-certs-edpm-deployment-openstack-edpm-ipam-w95ff"] Oct 08 18:49:44 crc kubenswrapper[4859]: I1008 18:49:44.087130 4859 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-neutron-metadata-default-certs-0" Oct 08 18:49:44 crc kubenswrapper[4859]: I1008 18:49:44.087371 4859 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-gxqst" Oct 08 18:49:44 crc kubenswrapper[4859]: I1008 18:49:44.252623 4859 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"repo-setup-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3b4b0004-0407-4671-a67e-0c14ef73e6ff-repo-setup-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-w95ff\" (UID: \"3b4b0004-0407-4671-a67e-0c14ef73e6ff\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-w95ff" Oct 08 18:49:44 crc kubenswrapper[4859]: I1008 18:49:44.252684 4859 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/3b4b0004-0407-4671-a67e-0c14ef73e6ff-ssh-key\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-w95ff\" (UID: \"3b4b0004-0407-4671-a67e-0c14ef73e6ff\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-w95ff" Oct 08 18:49:44 crc kubenswrapper[4859]: I1008 18:49:44.252750 4859 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"openstack-edpm-ipam-neutron-metadata-default-certs-0\" (UniqueName: \"kubernetes.io/projected/3b4b0004-0407-4671-a67e-0c14ef73e6ff-openstack-edpm-ipam-neutron-metadata-default-certs-0\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-w95ff\" (UID: \"3b4b0004-0407-4671-a67e-0c14ef73e6ff\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-w95ff" Oct 08 18:49:44 crc kubenswrapper[4859]: I1008 18:49:44.252784 4859 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"telemetry-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3b4b0004-0407-4671-a67e-0c14ef73e6ff-telemetry-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-w95ff\" (UID: \"3b4b0004-0407-4671-a67e-0c14ef73e6ff\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-w95ff" Oct 08 18:49:44 crc kubenswrapper[4859]: I1008 18:49:44.253570 4859 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3b4b0004-0407-4671-a67e-0c14ef73e6ff-nova-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-w95ff\" (UID: \"3b4b0004-0407-4671-a67e-0c14ef73e6ff\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-w95ff" Oct 08 18:49:44 crc kubenswrapper[4859]: I1008 18:49:44.253668 4859 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"openstack-edpm-ipam-libvirt-default-certs-0\" (UniqueName: \"kubernetes.io/projected/3b4b0004-0407-4671-a67e-0c14ef73e6ff-openstack-edpm-ipam-libvirt-default-certs-0\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-w95ff\" (UID: \"3b4b0004-0407-4671-a67e-0c14ef73e6ff\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-w95ff" Oct 08 18:49:44 crc kubenswrapper[4859]: I1008 18:49:44.254023 4859 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"libvirt-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3b4b0004-0407-4671-a67e-0c14ef73e6ff-libvirt-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-w95ff\" (UID: \"3b4b0004-0407-4671-a67e-0c14ef73e6ff\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-w95ff" Oct 08 18:49:44 crc kubenswrapper[4859]: I1008 18:49:44.254086 4859 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"neutron-metadata-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3b4b0004-0407-4671-a67e-0c14ef73e6ff-neutron-metadata-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-w95ff\" (UID: \"3b4b0004-0407-4671-a67e-0c14ef73e6ff\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-w95ff" Oct 08 18:49:44 crc kubenswrapper[4859]: I1008 18:49:44.254117 4859 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/3b4b0004-0407-4671-a67e-0c14ef73e6ff-inventory\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-w95ff\" (UID: \"3b4b0004-0407-4671-a67e-0c14ef73e6ff\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-w95ff" Oct 08 18:49:44 crc kubenswrapper[4859]: I1008 18:49:44.254156 4859 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"openstack-edpm-ipam-telemetry-default-certs-0\" (UniqueName: \"kubernetes.io/projected/3b4b0004-0407-4671-a67e-0c14ef73e6ff-openstack-edpm-ipam-telemetry-default-certs-0\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-w95ff\" (UID: \"3b4b0004-0407-4671-a67e-0c14ef73e6ff\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-w95ff" Oct 08 18:49:44 crc kubenswrapper[4859]: I1008 18:49:44.254340 4859 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bootstrap-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3b4b0004-0407-4671-a67e-0c14ef73e6ff-bootstrap-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-w95ff\" (UID: \"3b4b0004-0407-4671-a67e-0c14ef73e6ff\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-w95ff" Oct 08 18:49:44 crc kubenswrapper[4859]: I1008 18:49:44.254392 4859 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-pz66v\" (UniqueName: \"kubernetes.io/projected/3b4b0004-0407-4671-a67e-0c14ef73e6ff-kube-api-access-pz66v\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-w95ff\" (UID: \"3b4b0004-0407-4671-a67e-0c14ef73e6ff\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-w95ff" Oct 08 18:49:44 crc kubenswrapper[4859]: I1008 18:49:44.254432 4859 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"openstack-edpm-ipam-ovn-default-certs-0\" (UniqueName: \"kubernetes.io/projected/3b4b0004-0407-4671-a67e-0c14ef73e6ff-openstack-edpm-ipam-ovn-default-certs-0\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-w95ff\" (UID: \"3b4b0004-0407-4671-a67e-0c14ef73e6ff\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-w95ff" Oct 08 18:49:44 crc kubenswrapper[4859]: I1008 18:49:44.254454 4859 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovn-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3b4b0004-0407-4671-a67e-0c14ef73e6ff-ovn-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-w95ff\" (UID: \"3b4b0004-0407-4671-a67e-0c14ef73e6ff\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-w95ff" Oct 08 18:49:44 crc kubenswrapper[4859]: I1008 18:49:44.356116 4859 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"libvirt-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3b4b0004-0407-4671-a67e-0c14ef73e6ff-libvirt-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-w95ff\" (UID: \"3b4b0004-0407-4671-a67e-0c14ef73e6ff\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-w95ff" Oct 08 18:49:44 crc kubenswrapper[4859]: I1008 18:49:44.356201 4859 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"neutron-metadata-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3b4b0004-0407-4671-a67e-0c14ef73e6ff-neutron-metadata-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-w95ff\" (UID: \"3b4b0004-0407-4671-a67e-0c14ef73e6ff\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-w95ff" Oct 08 18:49:44 crc kubenswrapper[4859]: I1008 18:49:44.356237 4859 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/3b4b0004-0407-4671-a67e-0c14ef73e6ff-inventory\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-w95ff\" (UID: \"3b4b0004-0407-4671-a67e-0c14ef73e6ff\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-w95ff" Oct 08 18:49:44 crc kubenswrapper[4859]: I1008 18:49:44.356278 4859 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"openstack-edpm-ipam-telemetry-default-certs-0\" (UniqueName: \"kubernetes.io/projected/3b4b0004-0407-4671-a67e-0c14ef73e6ff-openstack-edpm-ipam-telemetry-default-certs-0\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-w95ff\" (UID: \"3b4b0004-0407-4671-a67e-0c14ef73e6ff\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-w95ff" Oct 08 18:49:44 crc kubenswrapper[4859]: I1008 18:49:44.356376 4859 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bootstrap-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3b4b0004-0407-4671-a67e-0c14ef73e6ff-bootstrap-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-w95ff\" (UID: \"3b4b0004-0407-4671-a67e-0c14ef73e6ff\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-w95ff" Oct 08 18:49:44 crc kubenswrapper[4859]: I1008 18:49:44.356424 4859 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-pz66v\" (UniqueName: \"kubernetes.io/projected/3b4b0004-0407-4671-a67e-0c14ef73e6ff-kube-api-access-pz66v\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-w95ff\" (UID: \"3b4b0004-0407-4671-a67e-0c14ef73e6ff\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-w95ff" Oct 08 18:49:44 crc kubenswrapper[4859]: I1008 18:49:44.356463 4859 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"openstack-edpm-ipam-ovn-default-certs-0\" (UniqueName: \"kubernetes.io/projected/3b4b0004-0407-4671-a67e-0c14ef73e6ff-openstack-edpm-ipam-ovn-default-certs-0\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-w95ff\" (UID: \"3b4b0004-0407-4671-a67e-0c14ef73e6ff\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-w95ff" Oct 08 18:49:44 crc kubenswrapper[4859]: I1008 18:49:44.356492 4859 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovn-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3b4b0004-0407-4671-a67e-0c14ef73e6ff-ovn-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-w95ff\" (UID: \"3b4b0004-0407-4671-a67e-0c14ef73e6ff\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-w95ff" Oct 08 18:49:44 crc kubenswrapper[4859]: I1008 18:49:44.356539 4859 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"repo-setup-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3b4b0004-0407-4671-a67e-0c14ef73e6ff-repo-setup-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-w95ff\" (UID: \"3b4b0004-0407-4671-a67e-0c14ef73e6ff\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-w95ff" Oct 08 18:49:44 crc kubenswrapper[4859]: I1008 18:49:44.356575 4859 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/3b4b0004-0407-4671-a67e-0c14ef73e6ff-ssh-key\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-w95ff\" (UID: \"3b4b0004-0407-4671-a67e-0c14ef73e6ff\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-w95ff" Oct 08 18:49:44 crc kubenswrapper[4859]: I1008 18:49:44.356618 4859 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"openstack-edpm-ipam-neutron-metadata-default-certs-0\" (UniqueName: \"kubernetes.io/projected/3b4b0004-0407-4671-a67e-0c14ef73e6ff-openstack-edpm-ipam-neutron-metadata-default-certs-0\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-w95ff\" (UID: \"3b4b0004-0407-4671-a67e-0c14ef73e6ff\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-w95ff" Oct 08 18:49:44 crc kubenswrapper[4859]: I1008 18:49:44.356670 4859 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"telemetry-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3b4b0004-0407-4671-a67e-0c14ef73e6ff-telemetry-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-w95ff\" (UID: \"3b4b0004-0407-4671-a67e-0c14ef73e6ff\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-w95ff" Oct 08 18:49:44 crc kubenswrapper[4859]: I1008 18:49:44.356728 4859 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3b4b0004-0407-4671-a67e-0c14ef73e6ff-nova-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-w95ff\" (UID: \"3b4b0004-0407-4671-a67e-0c14ef73e6ff\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-w95ff" Oct 08 18:49:44 crc kubenswrapper[4859]: I1008 18:49:44.356763 4859 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"openstack-edpm-ipam-libvirt-default-certs-0\" (UniqueName: \"kubernetes.io/projected/3b4b0004-0407-4671-a67e-0c14ef73e6ff-openstack-edpm-ipam-libvirt-default-certs-0\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-w95ff\" (UID: \"3b4b0004-0407-4671-a67e-0c14ef73e6ff\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-w95ff" Oct 08 18:49:44 crc kubenswrapper[4859]: I1008 18:49:44.361222 4859 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/3b4b0004-0407-4671-a67e-0c14ef73e6ff-ssh-key\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-w95ff\" (UID: \"3b4b0004-0407-4671-a67e-0c14ef73e6ff\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-w95ff" Oct 08 18:49:44 crc kubenswrapper[4859]: I1008 18:49:44.361355 4859 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"telemetry-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3b4b0004-0407-4671-a67e-0c14ef73e6ff-telemetry-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-w95ff\" (UID: \"3b4b0004-0407-4671-a67e-0c14ef73e6ff\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-w95ff" Oct 08 18:49:44 crc kubenswrapper[4859]: I1008 18:49:44.361379 4859 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"openstack-edpm-ipam-ovn-default-certs-0\" (UniqueName: \"kubernetes.io/projected/3b4b0004-0407-4671-a67e-0c14ef73e6ff-openstack-edpm-ipam-ovn-default-certs-0\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-w95ff\" (UID: \"3b4b0004-0407-4671-a67e-0c14ef73e6ff\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-w95ff" Oct 08 18:49:44 crc kubenswrapper[4859]: I1008 18:49:44.362009 4859 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"openstack-edpm-ipam-neutron-metadata-default-certs-0\" (UniqueName: \"kubernetes.io/projected/3b4b0004-0407-4671-a67e-0c14ef73e6ff-openstack-edpm-ipam-neutron-metadata-default-certs-0\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-w95ff\" (UID: \"3b4b0004-0407-4671-a67e-0c14ef73e6ff\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-w95ff" Oct 08 18:49:44 crc kubenswrapper[4859]: I1008 18:49:44.364843 4859 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"openstack-edpm-ipam-telemetry-default-certs-0\" (UniqueName: \"kubernetes.io/projected/3b4b0004-0407-4671-a67e-0c14ef73e6ff-openstack-edpm-ipam-telemetry-default-certs-0\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-w95ff\" (UID: \"3b4b0004-0407-4671-a67e-0c14ef73e6ff\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-w95ff" Oct 08 18:49:44 crc kubenswrapper[4859]: I1008 18:49:44.365866 4859 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovn-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3b4b0004-0407-4671-a67e-0c14ef73e6ff-ovn-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-w95ff\" (UID: \"3b4b0004-0407-4671-a67e-0c14ef73e6ff\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-w95ff" Oct 08 18:49:44 crc kubenswrapper[4859]: I1008 18:49:44.366746 4859 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"libvirt-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3b4b0004-0407-4671-a67e-0c14ef73e6ff-libvirt-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-w95ff\" (UID: \"3b4b0004-0407-4671-a67e-0c14ef73e6ff\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-w95ff" Oct 08 18:49:44 crc kubenswrapper[4859]: I1008 18:49:44.377220 4859 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"openstack-edpm-ipam-libvirt-default-certs-0\" (UniqueName: \"kubernetes.io/projected/3b4b0004-0407-4671-a67e-0c14ef73e6ff-openstack-edpm-ipam-libvirt-default-certs-0\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-w95ff\" (UID: \"3b4b0004-0407-4671-a67e-0c14ef73e6ff\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-w95ff" Oct 08 18:49:44 crc kubenswrapper[4859]: I1008 18:49:44.377630 4859 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/3b4b0004-0407-4671-a67e-0c14ef73e6ff-inventory\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-w95ff\" (UID: \"3b4b0004-0407-4671-a67e-0c14ef73e6ff\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-w95ff" Oct 08 18:49:44 crc kubenswrapper[4859]: I1008 18:49:44.378486 4859 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3b4b0004-0407-4671-a67e-0c14ef73e6ff-nova-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-w95ff\" (UID: \"3b4b0004-0407-4671-a67e-0c14ef73e6ff\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-w95ff" Oct 08 18:49:44 crc kubenswrapper[4859]: I1008 18:49:44.380425 4859 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"neutron-metadata-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3b4b0004-0407-4671-a67e-0c14ef73e6ff-neutron-metadata-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-w95ff\" (UID: \"3b4b0004-0407-4671-a67e-0c14ef73e6ff\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-w95ff" Oct 08 18:49:44 crc kubenswrapper[4859]: I1008 18:49:44.384100 4859 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-pz66v\" (UniqueName: \"kubernetes.io/projected/3b4b0004-0407-4671-a67e-0c14ef73e6ff-kube-api-access-pz66v\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-w95ff\" (UID: \"3b4b0004-0407-4671-a67e-0c14ef73e6ff\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-w95ff" Oct 08 18:49:44 crc kubenswrapper[4859]: I1008 18:49:44.386568 4859 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"repo-setup-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3b4b0004-0407-4671-a67e-0c14ef73e6ff-repo-setup-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-w95ff\" (UID: \"3b4b0004-0407-4671-a67e-0c14ef73e6ff\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-w95ff" Oct 08 18:49:44 crc kubenswrapper[4859]: I1008 18:49:44.387338 4859 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bootstrap-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3b4b0004-0407-4671-a67e-0c14ef73e6ff-bootstrap-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-w95ff\" (UID: \"3b4b0004-0407-4671-a67e-0c14ef73e6ff\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-w95ff" Oct 08 18:49:44 crc kubenswrapper[4859]: I1008 18:49:44.394193 4859 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-w95ff" Oct 08 18:49:44 crc kubenswrapper[4859]: I1008 18:49:44.920586 4859 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/install-certs-edpm-deployment-openstack-edpm-ipam-w95ff"] Oct 08 18:49:44 crc kubenswrapper[4859]: I1008 18:49:44.937165 4859 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Oct 08 18:49:44 crc kubenswrapper[4859]: I1008 18:49:44.965669 4859 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-w95ff" event={"ID":"3b4b0004-0407-4671-a67e-0c14ef73e6ff","Type":"ContainerStarted","Data":"61845653ae03af086e1ba34e1cf18d45b284da88741ad489f8699c9fbbdd720d"} Oct 08 18:49:45 crc kubenswrapper[4859]: I1008 18:49:45.977967 4859 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-w95ff" event={"ID":"3b4b0004-0407-4671-a67e-0c14ef73e6ff","Type":"ContainerStarted","Data":"ec43a61eaa091debfe2d732de9b40ce9c2c24852c09f03e8ebed4593f17c22ee"} Oct 08 18:49:46 crc kubenswrapper[4859]: I1008 18:49:46.008068 4859 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-w95ff" podStartSLOduration=1.407966902 podStartE2EDuration="2.00805114s" podCreationTimestamp="2025-10-08 18:49:44 +0000 UTC" firstStartedPulling="2025-10-08 18:49:44.936872571 +0000 UTC m=+1955.183711950" lastFinishedPulling="2025-10-08 18:49:45.536956799 +0000 UTC m=+1955.783796188" observedRunningTime="2025-10-08 18:49:46.005478976 +0000 UTC m=+1956.252318355" watchObservedRunningTime="2025-10-08 18:49:46.00805114 +0000 UTC m=+1956.254890519" Oct 08 18:49:55 crc kubenswrapper[4859]: I1008 18:49:55.470954 4859 scope.go:117] "RemoveContainer" containerID="44aa7848c4f5acb003170608db1921ad40aa5b2f5464f1f85d423cd26d2e8155" Oct 08 18:49:55 crc kubenswrapper[4859]: E1008 18:49:55.472132 4859 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-82s52_openshift-machine-config-operator(b23a6a6c-9d92-4e7b-840e-55cfda873a2d)\"" pod="openshift-machine-config-operator/machine-config-daemon-82s52" podUID="b23a6a6c-9d92-4e7b-840e-55cfda873a2d" Oct 08 18:50:06 crc kubenswrapper[4859]: I1008 18:50:06.470477 4859 scope.go:117] "RemoveContainer" containerID="44aa7848c4f5acb003170608db1921ad40aa5b2f5464f1f85d423cd26d2e8155" Oct 08 18:50:06 crc kubenswrapper[4859]: E1008 18:50:06.471301 4859 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-82s52_openshift-machine-config-operator(b23a6a6c-9d92-4e7b-840e-55cfda873a2d)\"" pod="openshift-machine-config-operator/machine-config-daemon-82s52" podUID="b23a6a6c-9d92-4e7b-840e-55cfda873a2d" Oct 08 18:50:18 crc kubenswrapper[4859]: I1008 18:50:18.470318 4859 scope.go:117] "RemoveContainer" containerID="44aa7848c4f5acb003170608db1921ad40aa5b2f5464f1f85d423cd26d2e8155" Oct 08 18:50:19 crc kubenswrapper[4859]: I1008 18:50:19.317397 4859 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-82s52" event={"ID":"b23a6a6c-9d92-4e7b-840e-55cfda873a2d","Type":"ContainerStarted","Data":"3b2f671305bc6d52048373f0287719da3af01a967a70fafd0c557ca58ab22b42"} Oct 08 18:50:26 crc kubenswrapper[4859]: I1008 18:50:26.320743 4859 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-whpxg"] Oct 08 18:50:26 crc kubenswrapper[4859]: I1008 18:50:26.323269 4859 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-whpxg" Oct 08 18:50:26 crc kubenswrapper[4859]: I1008 18:50:26.350229 4859 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-whpxg"] Oct 08 18:50:26 crc kubenswrapper[4859]: I1008 18:50:26.435324 4859 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/09087880-5623-4793-90fc-a95e5cede3cf-catalog-content\") pod \"community-operators-whpxg\" (UID: \"09087880-5623-4793-90fc-a95e5cede3cf\") " pod="openshift-marketplace/community-operators-whpxg" Oct 08 18:50:26 crc kubenswrapper[4859]: I1008 18:50:26.435369 4859 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/09087880-5623-4793-90fc-a95e5cede3cf-utilities\") pod \"community-operators-whpxg\" (UID: \"09087880-5623-4793-90fc-a95e5cede3cf\") " pod="openshift-marketplace/community-operators-whpxg" Oct 08 18:50:26 crc kubenswrapper[4859]: I1008 18:50:26.435415 4859 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-z9glk\" (UniqueName: \"kubernetes.io/projected/09087880-5623-4793-90fc-a95e5cede3cf-kube-api-access-z9glk\") pod \"community-operators-whpxg\" (UID: \"09087880-5623-4793-90fc-a95e5cede3cf\") " pod="openshift-marketplace/community-operators-whpxg" Oct 08 18:50:26 crc kubenswrapper[4859]: I1008 18:50:26.537798 4859 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/09087880-5623-4793-90fc-a95e5cede3cf-catalog-content\") pod \"community-operators-whpxg\" (UID: \"09087880-5623-4793-90fc-a95e5cede3cf\") " pod="openshift-marketplace/community-operators-whpxg" Oct 08 18:50:26 crc kubenswrapper[4859]: I1008 18:50:26.537868 4859 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/09087880-5623-4793-90fc-a95e5cede3cf-utilities\") pod \"community-operators-whpxg\" (UID: \"09087880-5623-4793-90fc-a95e5cede3cf\") " pod="openshift-marketplace/community-operators-whpxg" Oct 08 18:50:26 crc kubenswrapper[4859]: I1008 18:50:26.537931 4859 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-z9glk\" (UniqueName: \"kubernetes.io/projected/09087880-5623-4793-90fc-a95e5cede3cf-kube-api-access-z9glk\") pod \"community-operators-whpxg\" (UID: \"09087880-5623-4793-90fc-a95e5cede3cf\") " pod="openshift-marketplace/community-operators-whpxg" Oct 08 18:50:26 crc kubenswrapper[4859]: I1008 18:50:26.538516 4859 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/09087880-5623-4793-90fc-a95e5cede3cf-catalog-content\") pod \"community-operators-whpxg\" (UID: \"09087880-5623-4793-90fc-a95e5cede3cf\") " pod="openshift-marketplace/community-operators-whpxg" Oct 08 18:50:26 crc kubenswrapper[4859]: I1008 18:50:26.538528 4859 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/09087880-5623-4793-90fc-a95e5cede3cf-utilities\") pod \"community-operators-whpxg\" (UID: \"09087880-5623-4793-90fc-a95e5cede3cf\") " pod="openshift-marketplace/community-operators-whpxg" Oct 08 18:50:26 crc kubenswrapper[4859]: I1008 18:50:26.561386 4859 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-z9glk\" (UniqueName: \"kubernetes.io/projected/09087880-5623-4793-90fc-a95e5cede3cf-kube-api-access-z9glk\") pod \"community-operators-whpxg\" (UID: \"09087880-5623-4793-90fc-a95e5cede3cf\") " pod="openshift-marketplace/community-operators-whpxg" Oct 08 18:50:26 crc kubenswrapper[4859]: I1008 18:50:26.642982 4859 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-whpxg" Oct 08 18:50:27 crc kubenswrapper[4859]: I1008 18:50:27.176675 4859 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-whpxg"] Oct 08 18:50:27 crc kubenswrapper[4859]: I1008 18:50:27.398406 4859 generic.go:334] "Generic (PLEG): container finished" podID="09087880-5623-4793-90fc-a95e5cede3cf" containerID="cdb370b05d0c5bb9073cec21971d34b2b84b2c94d45a44ead21d29827fefa020" exitCode=0 Oct 08 18:50:27 crc kubenswrapper[4859]: I1008 18:50:27.398473 4859 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-whpxg" event={"ID":"09087880-5623-4793-90fc-a95e5cede3cf","Type":"ContainerDied","Data":"cdb370b05d0c5bb9073cec21971d34b2b84b2c94d45a44ead21d29827fefa020"} Oct 08 18:50:27 crc kubenswrapper[4859]: I1008 18:50:27.398810 4859 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-whpxg" event={"ID":"09087880-5623-4793-90fc-a95e5cede3cf","Type":"ContainerStarted","Data":"2e6369cf39d0bcb71b00e5f2dc4b6dcfd9efbdae42fe9738fbc0aaf59c610fe0"} Oct 08 18:50:28 crc kubenswrapper[4859]: I1008 18:50:28.417116 4859 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-whpxg" event={"ID":"09087880-5623-4793-90fc-a95e5cede3cf","Type":"ContainerStarted","Data":"813e622a613b8b27197b73b07025aa30bddb2af0b07a91434a8680b37c98f08a"} Oct 08 18:50:28 crc kubenswrapper[4859]: I1008 18:50:28.420215 4859 generic.go:334] "Generic (PLEG): container finished" podID="3b4b0004-0407-4671-a67e-0c14ef73e6ff" containerID="ec43a61eaa091debfe2d732de9b40ce9c2c24852c09f03e8ebed4593f17c22ee" exitCode=0 Oct 08 18:50:28 crc kubenswrapper[4859]: I1008 18:50:28.420251 4859 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-w95ff" event={"ID":"3b4b0004-0407-4671-a67e-0c14ef73e6ff","Type":"ContainerDied","Data":"ec43a61eaa091debfe2d732de9b40ce9c2c24852c09f03e8ebed4593f17c22ee"} Oct 08 18:50:29 crc kubenswrapper[4859]: I1008 18:50:29.433184 4859 generic.go:334] "Generic (PLEG): container finished" podID="09087880-5623-4793-90fc-a95e5cede3cf" containerID="813e622a613b8b27197b73b07025aa30bddb2af0b07a91434a8680b37c98f08a" exitCode=0 Oct 08 18:50:29 crc kubenswrapper[4859]: I1008 18:50:29.433357 4859 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-whpxg" event={"ID":"09087880-5623-4793-90fc-a95e5cede3cf","Type":"ContainerDied","Data":"813e622a613b8b27197b73b07025aa30bddb2af0b07a91434a8680b37c98f08a"} Oct 08 18:50:29 crc kubenswrapper[4859]: I1008 18:50:29.821038 4859 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-w95ff" Oct 08 18:50:30 crc kubenswrapper[4859]: I1008 18:50:30.007481 4859 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"neutron-metadata-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3b4b0004-0407-4671-a67e-0c14ef73e6ff-neutron-metadata-combined-ca-bundle\") pod \"3b4b0004-0407-4671-a67e-0c14ef73e6ff\" (UID: \"3b4b0004-0407-4671-a67e-0c14ef73e6ff\") " Oct 08 18:50:30 crc kubenswrapper[4859]: I1008 18:50:30.007913 4859 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/3b4b0004-0407-4671-a67e-0c14ef73e6ff-ssh-key\") pod \"3b4b0004-0407-4671-a67e-0c14ef73e6ff\" (UID: \"3b4b0004-0407-4671-a67e-0c14ef73e6ff\") " Oct 08 18:50:30 crc kubenswrapper[4859]: I1008 18:50:30.007948 4859 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovn-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3b4b0004-0407-4671-a67e-0c14ef73e6ff-ovn-combined-ca-bundle\") pod \"3b4b0004-0407-4671-a67e-0c14ef73e6ff\" (UID: \"3b4b0004-0407-4671-a67e-0c14ef73e6ff\") " Oct 08 18:50:30 crc kubenswrapper[4859]: I1008 18:50:30.007988 4859 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/3b4b0004-0407-4671-a67e-0c14ef73e6ff-inventory\") pod \"3b4b0004-0407-4671-a67e-0c14ef73e6ff\" (UID: \"3b4b0004-0407-4671-a67e-0c14ef73e6ff\") " Oct 08 18:50:30 crc kubenswrapper[4859]: I1008 18:50:30.008049 4859 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"openstack-edpm-ipam-ovn-default-certs-0\" (UniqueName: \"kubernetes.io/projected/3b4b0004-0407-4671-a67e-0c14ef73e6ff-openstack-edpm-ipam-ovn-default-certs-0\") pod \"3b4b0004-0407-4671-a67e-0c14ef73e6ff\" (UID: \"3b4b0004-0407-4671-a67e-0c14ef73e6ff\") " Oct 08 18:50:30 crc kubenswrapper[4859]: I1008 18:50:30.008099 4859 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"libvirt-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3b4b0004-0407-4671-a67e-0c14ef73e6ff-libvirt-combined-ca-bundle\") pod \"3b4b0004-0407-4671-a67e-0c14ef73e6ff\" (UID: \"3b4b0004-0407-4671-a67e-0c14ef73e6ff\") " Oct 08 18:50:30 crc kubenswrapper[4859]: I1008 18:50:30.008161 4859 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bootstrap-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3b4b0004-0407-4671-a67e-0c14ef73e6ff-bootstrap-combined-ca-bundle\") pod \"3b4b0004-0407-4671-a67e-0c14ef73e6ff\" (UID: \"3b4b0004-0407-4671-a67e-0c14ef73e6ff\") " Oct 08 18:50:30 crc kubenswrapper[4859]: I1008 18:50:30.008187 4859 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"repo-setup-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3b4b0004-0407-4671-a67e-0c14ef73e6ff-repo-setup-combined-ca-bundle\") pod \"3b4b0004-0407-4671-a67e-0c14ef73e6ff\" (UID: \"3b4b0004-0407-4671-a67e-0c14ef73e6ff\") " Oct 08 18:50:30 crc kubenswrapper[4859]: I1008 18:50:30.008229 4859 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-pz66v\" (UniqueName: \"kubernetes.io/projected/3b4b0004-0407-4671-a67e-0c14ef73e6ff-kube-api-access-pz66v\") pod \"3b4b0004-0407-4671-a67e-0c14ef73e6ff\" (UID: \"3b4b0004-0407-4671-a67e-0c14ef73e6ff\") " Oct 08 18:50:30 crc kubenswrapper[4859]: I1008 18:50:30.008253 4859 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"nova-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3b4b0004-0407-4671-a67e-0c14ef73e6ff-nova-combined-ca-bundle\") pod \"3b4b0004-0407-4671-a67e-0c14ef73e6ff\" (UID: \"3b4b0004-0407-4671-a67e-0c14ef73e6ff\") " Oct 08 18:50:30 crc kubenswrapper[4859]: I1008 18:50:30.008281 4859 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"telemetry-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3b4b0004-0407-4671-a67e-0c14ef73e6ff-telemetry-combined-ca-bundle\") pod \"3b4b0004-0407-4671-a67e-0c14ef73e6ff\" (UID: \"3b4b0004-0407-4671-a67e-0c14ef73e6ff\") " Oct 08 18:50:30 crc kubenswrapper[4859]: I1008 18:50:30.008330 4859 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"openstack-edpm-ipam-telemetry-default-certs-0\" (UniqueName: \"kubernetes.io/projected/3b4b0004-0407-4671-a67e-0c14ef73e6ff-openstack-edpm-ipam-telemetry-default-certs-0\") pod \"3b4b0004-0407-4671-a67e-0c14ef73e6ff\" (UID: \"3b4b0004-0407-4671-a67e-0c14ef73e6ff\") " Oct 08 18:50:30 crc kubenswrapper[4859]: I1008 18:50:30.008388 4859 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"openstack-edpm-ipam-libvirt-default-certs-0\" (UniqueName: \"kubernetes.io/projected/3b4b0004-0407-4671-a67e-0c14ef73e6ff-openstack-edpm-ipam-libvirt-default-certs-0\") pod \"3b4b0004-0407-4671-a67e-0c14ef73e6ff\" (UID: \"3b4b0004-0407-4671-a67e-0c14ef73e6ff\") " Oct 08 18:50:30 crc kubenswrapper[4859]: I1008 18:50:30.008421 4859 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"openstack-edpm-ipam-neutron-metadata-default-certs-0\" (UniqueName: \"kubernetes.io/projected/3b4b0004-0407-4671-a67e-0c14ef73e6ff-openstack-edpm-ipam-neutron-metadata-default-certs-0\") pod \"3b4b0004-0407-4671-a67e-0c14ef73e6ff\" (UID: \"3b4b0004-0407-4671-a67e-0c14ef73e6ff\") " Oct 08 18:50:30 crc kubenswrapper[4859]: I1008 18:50:30.015102 4859 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/3b4b0004-0407-4671-a67e-0c14ef73e6ff-openstack-edpm-ipam-neutron-metadata-default-certs-0" (OuterVolumeSpecName: "openstack-edpm-ipam-neutron-metadata-default-certs-0") pod "3b4b0004-0407-4671-a67e-0c14ef73e6ff" (UID: "3b4b0004-0407-4671-a67e-0c14ef73e6ff"). InnerVolumeSpecName "openstack-edpm-ipam-neutron-metadata-default-certs-0". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 08 18:50:30 crc kubenswrapper[4859]: I1008 18:50:30.015677 4859 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/3b4b0004-0407-4671-a67e-0c14ef73e6ff-bootstrap-combined-ca-bundle" (OuterVolumeSpecName: "bootstrap-combined-ca-bundle") pod "3b4b0004-0407-4671-a67e-0c14ef73e6ff" (UID: "3b4b0004-0407-4671-a67e-0c14ef73e6ff"). InnerVolumeSpecName "bootstrap-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 08 18:50:30 crc kubenswrapper[4859]: I1008 18:50:30.015820 4859 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/3b4b0004-0407-4671-a67e-0c14ef73e6ff-repo-setup-combined-ca-bundle" (OuterVolumeSpecName: "repo-setup-combined-ca-bundle") pod "3b4b0004-0407-4671-a67e-0c14ef73e6ff" (UID: "3b4b0004-0407-4671-a67e-0c14ef73e6ff"). InnerVolumeSpecName "repo-setup-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 08 18:50:30 crc kubenswrapper[4859]: I1008 18:50:30.016224 4859 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/3b4b0004-0407-4671-a67e-0c14ef73e6ff-telemetry-combined-ca-bundle" (OuterVolumeSpecName: "telemetry-combined-ca-bundle") pod "3b4b0004-0407-4671-a67e-0c14ef73e6ff" (UID: "3b4b0004-0407-4671-a67e-0c14ef73e6ff"). InnerVolumeSpecName "telemetry-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 08 18:50:30 crc kubenswrapper[4859]: I1008 18:50:30.016247 4859 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/3b4b0004-0407-4671-a67e-0c14ef73e6ff-ovn-combined-ca-bundle" (OuterVolumeSpecName: "ovn-combined-ca-bundle") pod "3b4b0004-0407-4671-a67e-0c14ef73e6ff" (UID: "3b4b0004-0407-4671-a67e-0c14ef73e6ff"). InnerVolumeSpecName "ovn-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 08 18:50:30 crc kubenswrapper[4859]: I1008 18:50:30.016837 4859 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/3b4b0004-0407-4671-a67e-0c14ef73e6ff-openstack-edpm-ipam-libvirt-default-certs-0" (OuterVolumeSpecName: "openstack-edpm-ipam-libvirt-default-certs-0") pod "3b4b0004-0407-4671-a67e-0c14ef73e6ff" (UID: "3b4b0004-0407-4671-a67e-0c14ef73e6ff"). InnerVolumeSpecName "openstack-edpm-ipam-libvirt-default-certs-0". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 08 18:50:30 crc kubenswrapper[4859]: I1008 18:50:30.016864 4859 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/3b4b0004-0407-4671-a67e-0c14ef73e6ff-kube-api-access-pz66v" (OuterVolumeSpecName: "kube-api-access-pz66v") pod "3b4b0004-0407-4671-a67e-0c14ef73e6ff" (UID: "3b4b0004-0407-4671-a67e-0c14ef73e6ff"). InnerVolumeSpecName "kube-api-access-pz66v". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 08 18:50:30 crc kubenswrapper[4859]: I1008 18:50:30.018107 4859 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/3b4b0004-0407-4671-a67e-0c14ef73e6ff-libvirt-combined-ca-bundle" (OuterVolumeSpecName: "libvirt-combined-ca-bundle") pod "3b4b0004-0407-4671-a67e-0c14ef73e6ff" (UID: "3b4b0004-0407-4671-a67e-0c14ef73e6ff"). InnerVolumeSpecName "libvirt-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 08 18:50:30 crc kubenswrapper[4859]: I1008 18:50:30.018170 4859 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/3b4b0004-0407-4671-a67e-0c14ef73e6ff-nova-combined-ca-bundle" (OuterVolumeSpecName: "nova-combined-ca-bundle") pod "3b4b0004-0407-4671-a67e-0c14ef73e6ff" (UID: "3b4b0004-0407-4671-a67e-0c14ef73e6ff"). InnerVolumeSpecName "nova-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 08 18:50:30 crc kubenswrapper[4859]: I1008 18:50:30.019340 4859 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/3b4b0004-0407-4671-a67e-0c14ef73e6ff-openstack-edpm-ipam-ovn-default-certs-0" (OuterVolumeSpecName: "openstack-edpm-ipam-ovn-default-certs-0") pod "3b4b0004-0407-4671-a67e-0c14ef73e6ff" (UID: "3b4b0004-0407-4671-a67e-0c14ef73e6ff"). InnerVolumeSpecName "openstack-edpm-ipam-ovn-default-certs-0". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 08 18:50:30 crc kubenswrapper[4859]: I1008 18:50:30.023792 4859 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/3b4b0004-0407-4671-a67e-0c14ef73e6ff-neutron-metadata-combined-ca-bundle" (OuterVolumeSpecName: "neutron-metadata-combined-ca-bundle") pod "3b4b0004-0407-4671-a67e-0c14ef73e6ff" (UID: "3b4b0004-0407-4671-a67e-0c14ef73e6ff"). InnerVolumeSpecName "neutron-metadata-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 08 18:50:30 crc kubenswrapper[4859]: I1008 18:50:30.024832 4859 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/3b4b0004-0407-4671-a67e-0c14ef73e6ff-openstack-edpm-ipam-telemetry-default-certs-0" (OuterVolumeSpecName: "openstack-edpm-ipam-telemetry-default-certs-0") pod "3b4b0004-0407-4671-a67e-0c14ef73e6ff" (UID: "3b4b0004-0407-4671-a67e-0c14ef73e6ff"). InnerVolumeSpecName "openstack-edpm-ipam-telemetry-default-certs-0". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 08 18:50:30 crc kubenswrapper[4859]: I1008 18:50:30.039976 4859 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/3b4b0004-0407-4671-a67e-0c14ef73e6ff-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "3b4b0004-0407-4671-a67e-0c14ef73e6ff" (UID: "3b4b0004-0407-4671-a67e-0c14ef73e6ff"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 08 18:50:30 crc kubenswrapper[4859]: I1008 18:50:30.041392 4859 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/3b4b0004-0407-4671-a67e-0c14ef73e6ff-inventory" (OuterVolumeSpecName: "inventory") pod "3b4b0004-0407-4671-a67e-0c14ef73e6ff" (UID: "3b4b0004-0407-4671-a67e-0c14ef73e6ff"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 08 18:50:30 crc kubenswrapper[4859]: I1008 18:50:30.112155 4859 reconciler_common.go:293] "Volume detached for volume \"bootstrap-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3b4b0004-0407-4671-a67e-0c14ef73e6ff-bootstrap-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 08 18:50:30 crc kubenswrapper[4859]: I1008 18:50:30.112197 4859 reconciler_common.go:293] "Volume detached for volume \"repo-setup-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3b4b0004-0407-4671-a67e-0c14ef73e6ff-repo-setup-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 08 18:50:30 crc kubenswrapper[4859]: I1008 18:50:30.112211 4859 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-pz66v\" (UniqueName: \"kubernetes.io/projected/3b4b0004-0407-4671-a67e-0c14ef73e6ff-kube-api-access-pz66v\") on node \"crc\" DevicePath \"\"" Oct 08 18:50:30 crc kubenswrapper[4859]: I1008 18:50:30.112225 4859 reconciler_common.go:293] "Volume detached for volume \"nova-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3b4b0004-0407-4671-a67e-0c14ef73e6ff-nova-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 08 18:50:30 crc kubenswrapper[4859]: I1008 18:50:30.112235 4859 reconciler_common.go:293] "Volume detached for volume \"telemetry-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3b4b0004-0407-4671-a67e-0c14ef73e6ff-telemetry-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 08 18:50:30 crc kubenswrapper[4859]: I1008 18:50:30.112248 4859 reconciler_common.go:293] "Volume detached for volume \"openstack-edpm-ipam-telemetry-default-certs-0\" (UniqueName: \"kubernetes.io/projected/3b4b0004-0407-4671-a67e-0c14ef73e6ff-openstack-edpm-ipam-telemetry-default-certs-0\") on node \"crc\" DevicePath \"\"" Oct 08 18:50:30 crc kubenswrapper[4859]: I1008 18:50:30.112264 4859 reconciler_common.go:293] "Volume detached for volume \"openstack-edpm-ipam-libvirt-default-certs-0\" (UniqueName: \"kubernetes.io/projected/3b4b0004-0407-4671-a67e-0c14ef73e6ff-openstack-edpm-ipam-libvirt-default-certs-0\") on node \"crc\" DevicePath \"\"" Oct 08 18:50:30 crc kubenswrapper[4859]: I1008 18:50:30.112276 4859 reconciler_common.go:293] "Volume detached for volume \"openstack-edpm-ipam-neutron-metadata-default-certs-0\" (UniqueName: \"kubernetes.io/projected/3b4b0004-0407-4671-a67e-0c14ef73e6ff-openstack-edpm-ipam-neutron-metadata-default-certs-0\") on node \"crc\" DevicePath \"\"" Oct 08 18:50:30 crc kubenswrapper[4859]: I1008 18:50:30.112287 4859 reconciler_common.go:293] "Volume detached for volume \"neutron-metadata-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3b4b0004-0407-4671-a67e-0c14ef73e6ff-neutron-metadata-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 08 18:50:30 crc kubenswrapper[4859]: I1008 18:50:30.112298 4859 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/3b4b0004-0407-4671-a67e-0c14ef73e6ff-ssh-key\") on node \"crc\" DevicePath \"\"" Oct 08 18:50:30 crc kubenswrapper[4859]: I1008 18:50:30.112308 4859 reconciler_common.go:293] "Volume detached for volume \"ovn-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3b4b0004-0407-4671-a67e-0c14ef73e6ff-ovn-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 08 18:50:30 crc kubenswrapper[4859]: I1008 18:50:30.112318 4859 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/3b4b0004-0407-4671-a67e-0c14ef73e6ff-inventory\") on node \"crc\" DevicePath \"\"" Oct 08 18:50:30 crc kubenswrapper[4859]: I1008 18:50:30.112328 4859 reconciler_common.go:293] "Volume detached for volume \"openstack-edpm-ipam-ovn-default-certs-0\" (UniqueName: \"kubernetes.io/projected/3b4b0004-0407-4671-a67e-0c14ef73e6ff-openstack-edpm-ipam-ovn-default-certs-0\") on node \"crc\" DevicePath \"\"" Oct 08 18:50:30 crc kubenswrapper[4859]: I1008 18:50:30.112338 4859 reconciler_common.go:293] "Volume detached for volume \"libvirt-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3b4b0004-0407-4671-a67e-0c14ef73e6ff-libvirt-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 08 18:50:30 crc kubenswrapper[4859]: I1008 18:50:30.447335 4859 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-whpxg" event={"ID":"09087880-5623-4793-90fc-a95e5cede3cf","Type":"ContainerStarted","Data":"acaf4e392082bdc98e2a7a31c2c30e3d216a9f486e86ff5c08e007f31aaa1572"} Oct 08 18:50:30 crc kubenswrapper[4859]: I1008 18:50:30.453258 4859 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-w95ff" event={"ID":"3b4b0004-0407-4671-a67e-0c14ef73e6ff","Type":"ContainerDied","Data":"61845653ae03af086e1ba34e1cf18d45b284da88741ad489f8699c9fbbdd720d"} Oct 08 18:50:30 crc kubenswrapper[4859]: I1008 18:50:30.453379 4859 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="61845653ae03af086e1ba34e1cf18d45b284da88741ad489f8699c9fbbdd720d" Oct 08 18:50:30 crc kubenswrapper[4859]: I1008 18:50:30.453456 4859 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-w95ff" Oct 08 18:50:30 crc kubenswrapper[4859]: I1008 18:50:30.510150 4859 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-whpxg" podStartSLOduration=2.108799452 podStartE2EDuration="4.51010119s" podCreationTimestamp="2025-10-08 18:50:26 +0000 UTC" firstStartedPulling="2025-10-08 18:50:27.401453272 +0000 UTC m=+1997.648292691" lastFinishedPulling="2025-10-08 18:50:29.80275505 +0000 UTC m=+2000.049594429" observedRunningTime="2025-10-08 18:50:30.470254839 +0000 UTC m=+2000.717094238" watchObservedRunningTime="2025-10-08 18:50:30.51010119 +0000 UTC m=+2000.756940569" Oct 08 18:50:30 crc kubenswrapper[4859]: I1008 18:50:30.564796 4859 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ovn-edpm-deployment-openstack-edpm-ipam-gtbjg"] Oct 08 18:50:30 crc kubenswrapper[4859]: E1008 18:50:30.565598 4859 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3b4b0004-0407-4671-a67e-0c14ef73e6ff" containerName="install-certs-edpm-deployment-openstack-edpm-ipam" Oct 08 18:50:30 crc kubenswrapper[4859]: I1008 18:50:30.565622 4859 state_mem.go:107] "Deleted CPUSet assignment" podUID="3b4b0004-0407-4671-a67e-0c14ef73e6ff" containerName="install-certs-edpm-deployment-openstack-edpm-ipam" Oct 08 18:50:30 crc kubenswrapper[4859]: I1008 18:50:30.565958 4859 memory_manager.go:354] "RemoveStaleState removing state" podUID="3b4b0004-0407-4671-a67e-0c14ef73e6ff" containerName="install-certs-edpm-deployment-openstack-edpm-ipam" Oct 08 18:50:30 crc kubenswrapper[4859]: I1008 18:50:30.581564 4859 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-gtbjg" Oct 08 18:50:30 crc kubenswrapper[4859]: I1008 18:50:30.583788 4859 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovncontroller-config" Oct 08 18:50:30 crc kubenswrapper[4859]: I1008 18:50:30.587027 4859 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Oct 08 18:50:30 crc kubenswrapper[4859]: I1008 18:50:30.587036 4859 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-gxqst" Oct 08 18:50:30 crc kubenswrapper[4859]: I1008 18:50:30.587591 4859 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Oct 08 18:50:30 crc kubenswrapper[4859]: I1008 18:50:30.587777 4859 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Oct 08 18:50:30 crc kubenswrapper[4859]: I1008 18:50:30.588015 4859 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-edpm-deployment-openstack-edpm-ipam-gtbjg"] Oct 08 18:50:30 crc kubenswrapper[4859]: I1008 18:50:30.723089 4859 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-2j5qp\" (UniqueName: \"kubernetes.io/projected/4bbd12de-a0ea-4741-b6af-dc19464d2161-kube-api-access-2j5qp\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-gtbjg\" (UID: \"4bbd12de-a0ea-4741-b6af-dc19464d2161\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-gtbjg" Oct 08 18:50:30 crc kubenswrapper[4859]: I1008 18:50:30.723226 4859 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/4bbd12de-a0ea-4741-b6af-dc19464d2161-inventory\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-gtbjg\" (UID: \"4bbd12de-a0ea-4741-b6af-dc19464d2161\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-gtbjg" Oct 08 18:50:30 crc kubenswrapper[4859]: I1008 18:50:30.723328 4859 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovncontroller-config-0\" (UniqueName: \"kubernetes.io/configmap/4bbd12de-a0ea-4741-b6af-dc19464d2161-ovncontroller-config-0\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-gtbjg\" (UID: \"4bbd12de-a0ea-4741-b6af-dc19464d2161\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-gtbjg" Oct 08 18:50:30 crc kubenswrapper[4859]: I1008 18:50:30.723460 4859 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovn-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4bbd12de-a0ea-4741-b6af-dc19464d2161-ovn-combined-ca-bundle\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-gtbjg\" (UID: \"4bbd12de-a0ea-4741-b6af-dc19464d2161\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-gtbjg" Oct 08 18:50:30 crc kubenswrapper[4859]: I1008 18:50:30.723542 4859 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/4bbd12de-a0ea-4741-b6af-dc19464d2161-ssh-key\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-gtbjg\" (UID: \"4bbd12de-a0ea-4741-b6af-dc19464d2161\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-gtbjg" Oct 08 18:50:30 crc kubenswrapper[4859]: I1008 18:50:30.825619 4859 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovn-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4bbd12de-a0ea-4741-b6af-dc19464d2161-ovn-combined-ca-bundle\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-gtbjg\" (UID: \"4bbd12de-a0ea-4741-b6af-dc19464d2161\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-gtbjg" Oct 08 18:50:30 crc kubenswrapper[4859]: I1008 18:50:30.825699 4859 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/4bbd12de-a0ea-4741-b6af-dc19464d2161-ssh-key\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-gtbjg\" (UID: \"4bbd12de-a0ea-4741-b6af-dc19464d2161\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-gtbjg" Oct 08 18:50:30 crc kubenswrapper[4859]: I1008 18:50:30.825757 4859 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-2j5qp\" (UniqueName: \"kubernetes.io/projected/4bbd12de-a0ea-4741-b6af-dc19464d2161-kube-api-access-2j5qp\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-gtbjg\" (UID: \"4bbd12de-a0ea-4741-b6af-dc19464d2161\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-gtbjg" Oct 08 18:50:30 crc kubenswrapper[4859]: I1008 18:50:30.825789 4859 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/4bbd12de-a0ea-4741-b6af-dc19464d2161-inventory\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-gtbjg\" (UID: \"4bbd12de-a0ea-4741-b6af-dc19464d2161\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-gtbjg" Oct 08 18:50:30 crc kubenswrapper[4859]: I1008 18:50:30.825840 4859 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovncontroller-config-0\" (UniqueName: \"kubernetes.io/configmap/4bbd12de-a0ea-4741-b6af-dc19464d2161-ovncontroller-config-0\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-gtbjg\" (UID: \"4bbd12de-a0ea-4741-b6af-dc19464d2161\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-gtbjg" Oct 08 18:50:30 crc kubenswrapper[4859]: I1008 18:50:30.826532 4859 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovncontroller-config-0\" (UniqueName: \"kubernetes.io/configmap/4bbd12de-a0ea-4741-b6af-dc19464d2161-ovncontroller-config-0\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-gtbjg\" (UID: \"4bbd12de-a0ea-4741-b6af-dc19464d2161\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-gtbjg" Oct 08 18:50:30 crc kubenswrapper[4859]: I1008 18:50:30.830145 4859 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/4bbd12de-a0ea-4741-b6af-dc19464d2161-inventory\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-gtbjg\" (UID: \"4bbd12de-a0ea-4741-b6af-dc19464d2161\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-gtbjg" Oct 08 18:50:30 crc kubenswrapper[4859]: I1008 18:50:30.830773 4859 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovn-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4bbd12de-a0ea-4741-b6af-dc19464d2161-ovn-combined-ca-bundle\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-gtbjg\" (UID: \"4bbd12de-a0ea-4741-b6af-dc19464d2161\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-gtbjg" Oct 08 18:50:30 crc kubenswrapper[4859]: I1008 18:50:30.831000 4859 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/4bbd12de-a0ea-4741-b6af-dc19464d2161-ssh-key\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-gtbjg\" (UID: \"4bbd12de-a0ea-4741-b6af-dc19464d2161\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-gtbjg" Oct 08 18:50:30 crc kubenswrapper[4859]: I1008 18:50:30.849322 4859 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-2j5qp\" (UniqueName: \"kubernetes.io/projected/4bbd12de-a0ea-4741-b6af-dc19464d2161-kube-api-access-2j5qp\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-gtbjg\" (UID: \"4bbd12de-a0ea-4741-b6af-dc19464d2161\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-gtbjg" Oct 08 18:50:30 crc kubenswrapper[4859]: I1008 18:50:30.901508 4859 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-gtbjg" Oct 08 18:50:31 crc kubenswrapper[4859]: I1008 18:50:31.402645 4859 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-edpm-deployment-openstack-edpm-ipam-gtbjg"] Oct 08 18:50:31 crc kubenswrapper[4859]: W1008 18:50:31.411758 4859 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod4bbd12de_a0ea_4741_b6af_dc19464d2161.slice/crio-7dadc24bdb1b8a1ea38f668b1481cf6c799a80fea193c5ff0cea0d704490496d WatchSource:0}: Error finding container 7dadc24bdb1b8a1ea38f668b1481cf6c799a80fea193c5ff0cea0d704490496d: Status 404 returned error can't find the container with id 7dadc24bdb1b8a1ea38f668b1481cf6c799a80fea193c5ff0cea0d704490496d Oct 08 18:50:31 crc kubenswrapper[4859]: I1008 18:50:31.462381 4859 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-gtbjg" event={"ID":"4bbd12de-a0ea-4741-b6af-dc19464d2161","Type":"ContainerStarted","Data":"7dadc24bdb1b8a1ea38f668b1481cf6c799a80fea193c5ff0cea0d704490496d"} Oct 08 18:50:32 crc kubenswrapper[4859]: I1008 18:50:32.490407 4859 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-gtbjg" event={"ID":"4bbd12de-a0ea-4741-b6af-dc19464d2161","Type":"ContainerStarted","Data":"97969ff064abcdbbd5fe369fcfaf49f108561848104e7f55cc19b2a96346273c"} Oct 08 18:50:32 crc kubenswrapper[4859]: I1008 18:50:32.502481 4859 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-gtbjg" podStartSLOduration=1.8280937929999999 podStartE2EDuration="2.502463739s" podCreationTimestamp="2025-10-08 18:50:30 +0000 UTC" firstStartedPulling="2025-10-08 18:50:31.417787374 +0000 UTC m=+2001.664626793" lastFinishedPulling="2025-10-08 18:50:32.09215735 +0000 UTC m=+2002.338996739" observedRunningTime="2025-10-08 18:50:32.495998064 +0000 UTC m=+2002.742837443" watchObservedRunningTime="2025-10-08 18:50:32.502463739 +0000 UTC m=+2002.749303128" Oct 08 18:50:36 crc kubenswrapper[4859]: I1008 18:50:36.643835 4859 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-whpxg" Oct 08 18:50:36 crc kubenswrapper[4859]: I1008 18:50:36.644653 4859 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-whpxg" Oct 08 18:50:36 crc kubenswrapper[4859]: I1008 18:50:36.697469 4859 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-whpxg" Oct 08 18:50:37 crc kubenswrapper[4859]: I1008 18:50:37.577388 4859 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-whpxg" Oct 08 18:50:37 crc kubenswrapper[4859]: I1008 18:50:37.636919 4859 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-whpxg"] Oct 08 18:50:39 crc kubenswrapper[4859]: I1008 18:50:39.530833 4859 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/community-operators-whpxg" podUID="09087880-5623-4793-90fc-a95e5cede3cf" containerName="registry-server" containerID="cri-o://acaf4e392082bdc98e2a7a31c2c30e3d216a9f486e86ff5c08e007f31aaa1572" gracePeriod=2 Oct 08 18:50:40 crc kubenswrapper[4859]: I1008 18:50:40.530031 4859 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-whpxg" Oct 08 18:50:40 crc kubenswrapper[4859]: I1008 18:50:40.541889 4859 generic.go:334] "Generic (PLEG): container finished" podID="09087880-5623-4793-90fc-a95e5cede3cf" containerID="acaf4e392082bdc98e2a7a31c2c30e3d216a9f486e86ff5c08e007f31aaa1572" exitCode=0 Oct 08 18:50:40 crc kubenswrapper[4859]: I1008 18:50:40.541928 4859 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-whpxg" event={"ID":"09087880-5623-4793-90fc-a95e5cede3cf","Type":"ContainerDied","Data":"acaf4e392082bdc98e2a7a31c2c30e3d216a9f486e86ff5c08e007f31aaa1572"} Oct 08 18:50:40 crc kubenswrapper[4859]: I1008 18:50:40.541953 4859 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-whpxg" event={"ID":"09087880-5623-4793-90fc-a95e5cede3cf","Type":"ContainerDied","Data":"2e6369cf39d0bcb71b00e5f2dc4b6dcfd9efbdae42fe9738fbc0aaf59c610fe0"} Oct 08 18:50:40 crc kubenswrapper[4859]: I1008 18:50:40.541969 4859 scope.go:117] "RemoveContainer" containerID="acaf4e392082bdc98e2a7a31c2c30e3d216a9f486e86ff5c08e007f31aaa1572" Oct 08 18:50:40 crc kubenswrapper[4859]: I1008 18:50:40.541976 4859 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-whpxg" Oct 08 18:50:40 crc kubenswrapper[4859]: I1008 18:50:40.623875 4859 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/09087880-5623-4793-90fc-a95e5cede3cf-utilities\") pod \"09087880-5623-4793-90fc-a95e5cede3cf\" (UID: \"09087880-5623-4793-90fc-a95e5cede3cf\") " Oct 08 18:50:40 crc kubenswrapper[4859]: I1008 18:50:40.624056 4859 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/09087880-5623-4793-90fc-a95e5cede3cf-catalog-content\") pod \"09087880-5623-4793-90fc-a95e5cede3cf\" (UID: \"09087880-5623-4793-90fc-a95e5cede3cf\") " Oct 08 18:50:40 crc kubenswrapper[4859]: I1008 18:50:40.624196 4859 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-z9glk\" (UniqueName: \"kubernetes.io/projected/09087880-5623-4793-90fc-a95e5cede3cf-kube-api-access-z9glk\") pod \"09087880-5623-4793-90fc-a95e5cede3cf\" (UID: \"09087880-5623-4793-90fc-a95e5cede3cf\") " Oct 08 18:50:40 crc kubenswrapper[4859]: I1008 18:50:40.632222 4859 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/09087880-5623-4793-90fc-a95e5cede3cf-utilities" (OuterVolumeSpecName: "utilities") pod "09087880-5623-4793-90fc-a95e5cede3cf" (UID: "09087880-5623-4793-90fc-a95e5cede3cf"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 08 18:50:40 crc kubenswrapper[4859]: I1008 18:50:40.663935 4859 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/09087880-5623-4793-90fc-a95e5cede3cf-kube-api-access-z9glk" (OuterVolumeSpecName: "kube-api-access-z9glk") pod "09087880-5623-4793-90fc-a95e5cede3cf" (UID: "09087880-5623-4793-90fc-a95e5cede3cf"). InnerVolumeSpecName "kube-api-access-z9glk". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 08 18:50:40 crc kubenswrapper[4859]: I1008 18:50:40.695465 4859 scope.go:117] "RemoveContainer" containerID="813e622a613b8b27197b73b07025aa30bddb2af0b07a91434a8680b37c98f08a" Oct 08 18:50:40 crc kubenswrapper[4859]: I1008 18:50:40.725386 4859 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-z9glk\" (UniqueName: \"kubernetes.io/projected/09087880-5623-4793-90fc-a95e5cede3cf-kube-api-access-z9glk\") on node \"crc\" DevicePath \"\"" Oct 08 18:50:40 crc kubenswrapper[4859]: I1008 18:50:40.725412 4859 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/09087880-5623-4793-90fc-a95e5cede3cf-utilities\") on node \"crc\" DevicePath \"\"" Oct 08 18:50:40 crc kubenswrapper[4859]: I1008 18:50:40.731127 4859 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/09087880-5623-4793-90fc-a95e5cede3cf-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "09087880-5623-4793-90fc-a95e5cede3cf" (UID: "09087880-5623-4793-90fc-a95e5cede3cf"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 08 18:50:40 crc kubenswrapper[4859]: I1008 18:50:40.736737 4859 scope.go:117] "RemoveContainer" containerID="cdb370b05d0c5bb9073cec21971d34b2b84b2c94d45a44ead21d29827fefa020" Oct 08 18:50:40 crc kubenswrapper[4859]: I1008 18:50:40.776471 4859 scope.go:117] "RemoveContainer" containerID="acaf4e392082bdc98e2a7a31c2c30e3d216a9f486e86ff5c08e007f31aaa1572" Oct 08 18:50:40 crc kubenswrapper[4859]: E1008 18:50:40.777125 4859 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"acaf4e392082bdc98e2a7a31c2c30e3d216a9f486e86ff5c08e007f31aaa1572\": container with ID starting with acaf4e392082bdc98e2a7a31c2c30e3d216a9f486e86ff5c08e007f31aaa1572 not found: ID does not exist" containerID="acaf4e392082bdc98e2a7a31c2c30e3d216a9f486e86ff5c08e007f31aaa1572" Oct 08 18:50:40 crc kubenswrapper[4859]: I1008 18:50:40.777171 4859 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"acaf4e392082bdc98e2a7a31c2c30e3d216a9f486e86ff5c08e007f31aaa1572"} err="failed to get container status \"acaf4e392082bdc98e2a7a31c2c30e3d216a9f486e86ff5c08e007f31aaa1572\": rpc error: code = NotFound desc = could not find container \"acaf4e392082bdc98e2a7a31c2c30e3d216a9f486e86ff5c08e007f31aaa1572\": container with ID starting with acaf4e392082bdc98e2a7a31c2c30e3d216a9f486e86ff5c08e007f31aaa1572 not found: ID does not exist" Oct 08 18:50:40 crc kubenswrapper[4859]: I1008 18:50:40.777197 4859 scope.go:117] "RemoveContainer" containerID="813e622a613b8b27197b73b07025aa30bddb2af0b07a91434a8680b37c98f08a" Oct 08 18:50:40 crc kubenswrapper[4859]: E1008 18:50:40.777460 4859 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"813e622a613b8b27197b73b07025aa30bddb2af0b07a91434a8680b37c98f08a\": container with ID starting with 813e622a613b8b27197b73b07025aa30bddb2af0b07a91434a8680b37c98f08a not found: ID does not exist" containerID="813e622a613b8b27197b73b07025aa30bddb2af0b07a91434a8680b37c98f08a" Oct 08 18:50:40 crc kubenswrapper[4859]: I1008 18:50:40.777480 4859 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"813e622a613b8b27197b73b07025aa30bddb2af0b07a91434a8680b37c98f08a"} err="failed to get container status \"813e622a613b8b27197b73b07025aa30bddb2af0b07a91434a8680b37c98f08a\": rpc error: code = NotFound desc = could not find container \"813e622a613b8b27197b73b07025aa30bddb2af0b07a91434a8680b37c98f08a\": container with ID starting with 813e622a613b8b27197b73b07025aa30bddb2af0b07a91434a8680b37c98f08a not found: ID does not exist" Oct 08 18:50:40 crc kubenswrapper[4859]: I1008 18:50:40.777492 4859 scope.go:117] "RemoveContainer" containerID="cdb370b05d0c5bb9073cec21971d34b2b84b2c94d45a44ead21d29827fefa020" Oct 08 18:50:40 crc kubenswrapper[4859]: E1008 18:50:40.777804 4859 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"cdb370b05d0c5bb9073cec21971d34b2b84b2c94d45a44ead21d29827fefa020\": container with ID starting with cdb370b05d0c5bb9073cec21971d34b2b84b2c94d45a44ead21d29827fefa020 not found: ID does not exist" containerID="cdb370b05d0c5bb9073cec21971d34b2b84b2c94d45a44ead21d29827fefa020" Oct 08 18:50:40 crc kubenswrapper[4859]: I1008 18:50:40.777826 4859 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"cdb370b05d0c5bb9073cec21971d34b2b84b2c94d45a44ead21d29827fefa020"} err="failed to get container status \"cdb370b05d0c5bb9073cec21971d34b2b84b2c94d45a44ead21d29827fefa020\": rpc error: code = NotFound desc = could not find container \"cdb370b05d0c5bb9073cec21971d34b2b84b2c94d45a44ead21d29827fefa020\": container with ID starting with cdb370b05d0c5bb9073cec21971d34b2b84b2c94d45a44ead21d29827fefa020 not found: ID does not exist" Oct 08 18:50:40 crc kubenswrapper[4859]: I1008 18:50:40.827612 4859 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/09087880-5623-4793-90fc-a95e5cede3cf-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 08 18:50:40 crc kubenswrapper[4859]: I1008 18:50:40.880304 4859 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-whpxg"] Oct 08 18:50:40 crc kubenswrapper[4859]: I1008 18:50:40.894943 4859 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/community-operators-whpxg"] Oct 08 18:50:42 crc kubenswrapper[4859]: I1008 18:50:42.487705 4859 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="09087880-5623-4793-90fc-a95e5cede3cf" path="/var/lib/kubelet/pods/09087880-5623-4793-90fc-a95e5cede3cf/volumes" Oct 08 18:51:18 crc kubenswrapper[4859]: I1008 18:51:18.801809 4859 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-8rkwt"] Oct 08 18:51:18 crc kubenswrapper[4859]: E1008 18:51:18.803929 4859 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="09087880-5623-4793-90fc-a95e5cede3cf" containerName="extract-content" Oct 08 18:51:18 crc kubenswrapper[4859]: I1008 18:51:18.803964 4859 state_mem.go:107] "Deleted CPUSet assignment" podUID="09087880-5623-4793-90fc-a95e5cede3cf" containerName="extract-content" Oct 08 18:51:18 crc kubenswrapper[4859]: E1008 18:51:18.803985 4859 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="09087880-5623-4793-90fc-a95e5cede3cf" containerName="extract-utilities" Oct 08 18:51:18 crc kubenswrapper[4859]: I1008 18:51:18.803998 4859 state_mem.go:107] "Deleted CPUSet assignment" podUID="09087880-5623-4793-90fc-a95e5cede3cf" containerName="extract-utilities" Oct 08 18:51:18 crc kubenswrapper[4859]: E1008 18:51:18.804037 4859 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="09087880-5623-4793-90fc-a95e5cede3cf" containerName="registry-server" Oct 08 18:51:18 crc kubenswrapper[4859]: I1008 18:51:18.804050 4859 state_mem.go:107] "Deleted CPUSet assignment" podUID="09087880-5623-4793-90fc-a95e5cede3cf" containerName="registry-server" Oct 08 18:51:18 crc kubenswrapper[4859]: I1008 18:51:18.805021 4859 memory_manager.go:354] "RemoveStaleState removing state" podUID="09087880-5623-4793-90fc-a95e5cede3cf" containerName="registry-server" Oct 08 18:51:18 crc kubenswrapper[4859]: I1008 18:51:18.817889 4859 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-8rkwt" Oct 08 18:51:18 crc kubenswrapper[4859]: I1008 18:51:18.819182 4859 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-8rkwt"] Oct 08 18:51:18 crc kubenswrapper[4859]: I1008 18:51:18.986016 4859 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/7e90626d-d633-4991-a768-747d73a77fbe-catalog-content\") pod \"certified-operators-8rkwt\" (UID: \"7e90626d-d633-4991-a768-747d73a77fbe\") " pod="openshift-marketplace/certified-operators-8rkwt" Oct 08 18:51:18 crc kubenswrapper[4859]: I1008 18:51:18.986173 4859 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/7e90626d-d633-4991-a768-747d73a77fbe-utilities\") pod \"certified-operators-8rkwt\" (UID: \"7e90626d-d633-4991-a768-747d73a77fbe\") " pod="openshift-marketplace/certified-operators-8rkwt" Oct 08 18:51:18 crc kubenswrapper[4859]: I1008 18:51:18.986205 4859 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-hfxmz\" (UniqueName: \"kubernetes.io/projected/7e90626d-d633-4991-a768-747d73a77fbe-kube-api-access-hfxmz\") pod \"certified-operators-8rkwt\" (UID: \"7e90626d-d633-4991-a768-747d73a77fbe\") " pod="openshift-marketplace/certified-operators-8rkwt" Oct 08 18:51:19 crc kubenswrapper[4859]: I1008 18:51:19.088557 4859 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/7e90626d-d633-4991-a768-747d73a77fbe-utilities\") pod \"certified-operators-8rkwt\" (UID: \"7e90626d-d633-4991-a768-747d73a77fbe\") " pod="openshift-marketplace/certified-operators-8rkwt" Oct 08 18:51:19 crc kubenswrapper[4859]: I1008 18:51:19.088615 4859 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-hfxmz\" (UniqueName: \"kubernetes.io/projected/7e90626d-d633-4991-a768-747d73a77fbe-kube-api-access-hfxmz\") pod \"certified-operators-8rkwt\" (UID: \"7e90626d-d633-4991-a768-747d73a77fbe\") " pod="openshift-marketplace/certified-operators-8rkwt" Oct 08 18:51:19 crc kubenswrapper[4859]: I1008 18:51:19.088733 4859 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/7e90626d-d633-4991-a768-747d73a77fbe-catalog-content\") pod \"certified-operators-8rkwt\" (UID: \"7e90626d-d633-4991-a768-747d73a77fbe\") " pod="openshift-marketplace/certified-operators-8rkwt" Oct 08 18:51:19 crc kubenswrapper[4859]: I1008 18:51:19.089226 4859 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/7e90626d-d633-4991-a768-747d73a77fbe-utilities\") pod \"certified-operators-8rkwt\" (UID: \"7e90626d-d633-4991-a768-747d73a77fbe\") " pod="openshift-marketplace/certified-operators-8rkwt" Oct 08 18:51:19 crc kubenswrapper[4859]: I1008 18:51:19.089271 4859 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/7e90626d-d633-4991-a768-747d73a77fbe-catalog-content\") pod \"certified-operators-8rkwt\" (UID: \"7e90626d-d633-4991-a768-747d73a77fbe\") " pod="openshift-marketplace/certified-operators-8rkwt" Oct 08 18:51:19 crc kubenswrapper[4859]: I1008 18:51:19.107614 4859 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-hfxmz\" (UniqueName: \"kubernetes.io/projected/7e90626d-d633-4991-a768-747d73a77fbe-kube-api-access-hfxmz\") pod \"certified-operators-8rkwt\" (UID: \"7e90626d-d633-4991-a768-747d73a77fbe\") " pod="openshift-marketplace/certified-operators-8rkwt" Oct 08 18:51:19 crc kubenswrapper[4859]: I1008 18:51:19.143821 4859 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-8rkwt" Oct 08 18:51:19 crc kubenswrapper[4859]: I1008 18:51:19.673945 4859 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-8rkwt"] Oct 08 18:51:19 crc kubenswrapper[4859]: I1008 18:51:19.919572 4859 generic.go:334] "Generic (PLEG): container finished" podID="7e90626d-d633-4991-a768-747d73a77fbe" containerID="c3ab088bdb73ce250f3cd8941958752651b51463fe32e9c45b121661b7eeebdc" exitCode=0 Oct 08 18:51:19 crc kubenswrapper[4859]: I1008 18:51:19.919614 4859 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-8rkwt" event={"ID":"7e90626d-d633-4991-a768-747d73a77fbe","Type":"ContainerDied","Data":"c3ab088bdb73ce250f3cd8941958752651b51463fe32e9c45b121661b7eeebdc"} Oct 08 18:51:19 crc kubenswrapper[4859]: I1008 18:51:19.919637 4859 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-8rkwt" event={"ID":"7e90626d-d633-4991-a768-747d73a77fbe","Type":"ContainerStarted","Data":"101b58832eb066b6fcb19935e8305a473cd97cbe74998cd76b654ca35dbb945f"} Oct 08 18:51:20 crc kubenswrapper[4859]: I1008 18:51:20.929433 4859 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-8rkwt" event={"ID":"7e90626d-d633-4991-a768-747d73a77fbe","Type":"ContainerStarted","Data":"83d6a16b3664db0104bb0d388bc6ca2cb611debd34b888d17ba813de26484e36"} Oct 08 18:51:21 crc kubenswrapper[4859]: I1008 18:51:21.939638 4859 generic.go:334] "Generic (PLEG): container finished" podID="7e90626d-d633-4991-a768-747d73a77fbe" containerID="83d6a16b3664db0104bb0d388bc6ca2cb611debd34b888d17ba813de26484e36" exitCode=0 Oct 08 18:51:21 crc kubenswrapper[4859]: I1008 18:51:21.939746 4859 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-8rkwt" event={"ID":"7e90626d-d633-4991-a768-747d73a77fbe","Type":"ContainerDied","Data":"83d6a16b3664db0104bb0d388bc6ca2cb611debd34b888d17ba813de26484e36"} Oct 08 18:51:22 crc kubenswrapper[4859]: I1008 18:51:22.952772 4859 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-8rkwt" event={"ID":"7e90626d-d633-4991-a768-747d73a77fbe","Type":"ContainerStarted","Data":"7216fdb3f9b474ae2d7acfabc491d89b636cafc168bc0656e09d95b98a8c2cc2"} Oct 08 18:51:22 crc kubenswrapper[4859]: I1008 18:51:22.981661 4859 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-8rkwt" podStartSLOduration=2.4944485690000002 podStartE2EDuration="4.981642447s" podCreationTimestamp="2025-10-08 18:51:18 +0000 UTC" firstStartedPulling="2025-10-08 18:51:19.921661033 +0000 UTC m=+2050.168500412" lastFinishedPulling="2025-10-08 18:51:22.408854901 +0000 UTC m=+2052.655694290" observedRunningTime="2025-10-08 18:51:22.974378789 +0000 UTC m=+2053.221218168" watchObservedRunningTime="2025-10-08 18:51:22.981642447 +0000 UTC m=+2053.228481826" Oct 08 18:51:29 crc kubenswrapper[4859]: I1008 18:51:29.144799 4859 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-8rkwt" Oct 08 18:51:29 crc kubenswrapper[4859]: I1008 18:51:29.145468 4859 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-8rkwt" Oct 08 18:51:29 crc kubenswrapper[4859]: I1008 18:51:29.226176 4859 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-8rkwt" Oct 08 18:51:30 crc kubenswrapper[4859]: I1008 18:51:30.096808 4859 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-8rkwt" Oct 08 18:51:30 crc kubenswrapper[4859]: I1008 18:51:30.169612 4859 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-8rkwt"] Oct 08 18:51:32 crc kubenswrapper[4859]: I1008 18:51:32.044100 4859 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/certified-operators-8rkwt" podUID="7e90626d-d633-4991-a768-747d73a77fbe" containerName="registry-server" containerID="cri-o://7216fdb3f9b474ae2d7acfabc491d89b636cafc168bc0656e09d95b98a8c2cc2" gracePeriod=2 Oct 08 18:51:32 crc kubenswrapper[4859]: I1008 18:51:32.517027 4859 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-8rkwt" Oct 08 18:51:32 crc kubenswrapper[4859]: I1008 18:51:32.663318 4859 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/7e90626d-d633-4991-a768-747d73a77fbe-utilities\") pod \"7e90626d-d633-4991-a768-747d73a77fbe\" (UID: \"7e90626d-d633-4991-a768-747d73a77fbe\") " Oct 08 18:51:32 crc kubenswrapper[4859]: I1008 18:51:32.663823 4859 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-hfxmz\" (UniqueName: \"kubernetes.io/projected/7e90626d-d633-4991-a768-747d73a77fbe-kube-api-access-hfxmz\") pod \"7e90626d-d633-4991-a768-747d73a77fbe\" (UID: \"7e90626d-d633-4991-a768-747d73a77fbe\") " Oct 08 18:51:32 crc kubenswrapper[4859]: I1008 18:51:32.664100 4859 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/7e90626d-d633-4991-a768-747d73a77fbe-catalog-content\") pod \"7e90626d-d633-4991-a768-747d73a77fbe\" (UID: \"7e90626d-d633-4991-a768-747d73a77fbe\") " Oct 08 18:51:32 crc kubenswrapper[4859]: I1008 18:51:32.664447 4859 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/7e90626d-d633-4991-a768-747d73a77fbe-utilities" (OuterVolumeSpecName: "utilities") pod "7e90626d-d633-4991-a768-747d73a77fbe" (UID: "7e90626d-d633-4991-a768-747d73a77fbe"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 08 18:51:32 crc kubenswrapper[4859]: I1008 18:51:32.664616 4859 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/7e90626d-d633-4991-a768-747d73a77fbe-utilities\") on node \"crc\" DevicePath \"\"" Oct 08 18:51:32 crc kubenswrapper[4859]: I1008 18:51:32.669733 4859 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/7e90626d-d633-4991-a768-747d73a77fbe-kube-api-access-hfxmz" (OuterVolumeSpecName: "kube-api-access-hfxmz") pod "7e90626d-d633-4991-a768-747d73a77fbe" (UID: "7e90626d-d633-4991-a768-747d73a77fbe"). InnerVolumeSpecName "kube-api-access-hfxmz". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 08 18:51:32 crc kubenswrapper[4859]: I1008 18:51:32.706180 4859 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/7e90626d-d633-4991-a768-747d73a77fbe-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "7e90626d-d633-4991-a768-747d73a77fbe" (UID: "7e90626d-d633-4991-a768-747d73a77fbe"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 08 18:51:32 crc kubenswrapper[4859]: I1008 18:51:32.766902 4859 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-hfxmz\" (UniqueName: \"kubernetes.io/projected/7e90626d-d633-4991-a768-747d73a77fbe-kube-api-access-hfxmz\") on node \"crc\" DevicePath \"\"" Oct 08 18:51:32 crc kubenswrapper[4859]: I1008 18:51:32.766928 4859 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/7e90626d-d633-4991-a768-747d73a77fbe-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 08 18:51:33 crc kubenswrapper[4859]: I1008 18:51:33.057598 4859 generic.go:334] "Generic (PLEG): container finished" podID="7e90626d-d633-4991-a768-747d73a77fbe" containerID="7216fdb3f9b474ae2d7acfabc491d89b636cafc168bc0656e09d95b98a8c2cc2" exitCode=0 Oct 08 18:51:33 crc kubenswrapper[4859]: I1008 18:51:33.057641 4859 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-8rkwt" event={"ID":"7e90626d-d633-4991-a768-747d73a77fbe","Type":"ContainerDied","Data":"7216fdb3f9b474ae2d7acfabc491d89b636cafc168bc0656e09d95b98a8c2cc2"} Oct 08 18:51:33 crc kubenswrapper[4859]: I1008 18:51:33.057655 4859 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-8rkwt" Oct 08 18:51:33 crc kubenswrapper[4859]: I1008 18:51:33.057668 4859 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-8rkwt" event={"ID":"7e90626d-d633-4991-a768-747d73a77fbe","Type":"ContainerDied","Data":"101b58832eb066b6fcb19935e8305a473cd97cbe74998cd76b654ca35dbb945f"} Oct 08 18:51:33 crc kubenswrapper[4859]: I1008 18:51:33.057700 4859 scope.go:117] "RemoveContainer" containerID="7216fdb3f9b474ae2d7acfabc491d89b636cafc168bc0656e09d95b98a8c2cc2" Oct 08 18:51:33 crc kubenswrapper[4859]: I1008 18:51:33.088350 4859 scope.go:117] "RemoveContainer" containerID="83d6a16b3664db0104bb0d388bc6ca2cb611debd34b888d17ba813de26484e36" Oct 08 18:51:33 crc kubenswrapper[4859]: I1008 18:51:33.120307 4859 scope.go:117] "RemoveContainer" containerID="c3ab088bdb73ce250f3cd8941958752651b51463fe32e9c45b121661b7eeebdc" Oct 08 18:51:33 crc kubenswrapper[4859]: I1008 18:51:33.120576 4859 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-8rkwt"] Oct 08 18:51:33 crc kubenswrapper[4859]: I1008 18:51:33.128949 4859 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/certified-operators-8rkwt"] Oct 08 18:51:33 crc kubenswrapper[4859]: I1008 18:51:33.167498 4859 scope.go:117] "RemoveContainer" containerID="7216fdb3f9b474ae2d7acfabc491d89b636cafc168bc0656e09d95b98a8c2cc2" Oct 08 18:51:33 crc kubenswrapper[4859]: E1008 18:51:33.167917 4859 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"7216fdb3f9b474ae2d7acfabc491d89b636cafc168bc0656e09d95b98a8c2cc2\": container with ID starting with 7216fdb3f9b474ae2d7acfabc491d89b636cafc168bc0656e09d95b98a8c2cc2 not found: ID does not exist" containerID="7216fdb3f9b474ae2d7acfabc491d89b636cafc168bc0656e09d95b98a8c2cc2" Oct 08 18:51:33 crc kubenswrapper[4859]: I1008 18:51:33.167962 4859 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"7216fdb3f9b474ae2d7acfabc491d89b636cafc168bc0656e09d95b98a8c2cc2"} err="failed to get container status \"7216fdb3f9b474ae2d7acfabc491d89b636cafc168bc0656e09d95b98a8c2cc2\": rpc error: code = NotFound desc = could not find container \"7216fdb3f9b474ae2d7acfabc491d89b636cafc168bc0656e09d95b98a8c2cc2\": container with ID starting with 7216fdb3f9b474ae2d7acfabc491d89b636cafc168bc0656e09d95b98a8c2cc2 not found: ID does not exist" Oct 08 18:51:33 crc kubenswrapper[4859]: I1008 18:51:33.167988 4859 scope.go:117] "RemoveContainer" containerID="83d6a16b3664db0104bb0d388bc6ca2cb611debd34b888d17ba813de26484e36" Oct 08 18:51:33 crc kubenswrapper[4859]: E1008 18:51:33.168378 4859 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"83d6a16b3664db0104bb0d388bc6ca2cb611debd34b888d17ba813de26484e36\": container with ID starting with 83d6a16b3664db0104bb0d388bc6ca2cb611debd34b888d17ba813de26484e36 not found: ID does not exist" containerID="83d6a16b3664db0104bb0d388bc6ca2cb611debd34b888d17ba813de26484e36" Oct 08 18:51:33 crc kubenswrapper[4859]: I1008 18:51:33.168408 4859 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"83d6a16b3664db0104bb0d388bc6ca2cb611debd34b888d17ba813de26484e36"} err="failed to get container status \"83d6a16b3664db0104bb0d388bc6ca2cb611debd34b888d17ba813de26484e36\": rpc error: code = NotFound desc = could not find container \"83d6a16b3664db0104bb0d388bc6ca2cb611debd34b888d17ba813de26484e36\": container with ID starting with 83d6a16b3664db0104bb0d388bc6ca2cb611debd34b888d17ba813de26484e36 not found: ID does not exist" Oct 08 18:51:33 crc kubenswrapper[4859]: I1008 18:51:33.168425 4859 scope.go:117] "RemoveContainer" containerID="c3ab088bdb73ce250f3cd8941958752651b51463fe32e9c45b121661b7eeebdc" Oct 08 18:51:33 crc kubenswrapper[4859]: E1008 18:51:33.168887 4859 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"c3ab088bdb73ce250f3cd8941958752651b51463fe32e9c45b121661b7eeebdc\": container with ID starting with c3ab088bdb73ce250f3cd8941958752651b51463fe32e9c45b121661b7eeebdc not found: ID does not exist" containerID="c3ab088bdb73ce250f3cd8941958752651b51463fe32e9c45b121661b7eeebdc" Oct 08 18:51:33 crc kubenswrapper[4859]: I1008 18:51:33.168918 4859 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"c3ab088bdb73ce250f3cd8941958752651b51463fe32e9c45b121661b7eeebdc"} err="failed to get container status \"c3ab088bdb73ce250f3cd8941958752651b51463fe32e9c45b121661b7eeebdc\": rpc error: code = NotFound desc = could not find container \"c3ab088bdb73ce250f3cd8941958752651b51463fe32e9c45b121661b7eeebdc\": container with ID starting with c3ab088bdb73ce250f3cd8941958752651b51463fe32e9c45b121661b7eeebdc not found: ID does not exist" Oct 08 18:51:33 crc kubenswrapper[4859]: E1008 18:51:33.223663 4859 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod7e90626d_d633_4991_a768_747d73a77fbe.slice\": RecentStats: unable to find data in memory cache]" Oct 08 18:51:34 crc kubenswrapper[4859]: I1008 18:51:34.490367 4859 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="7e90626d-d633-4991-a768-747d73a77fbe" path="/var/lib/kubelet/pods/7e90626d-d633-4991-a768-747d73a77fbe/volumes" Oct 08 18:51:44 crc kubenswrapper[4859]: I1008 18:51:44.168732 4859 generic.go:334] "Generic (PLEG): container finished" podID="4bbd12de-a0ea-4741-b6af-dc19464d2161" containerID="97969ff064abcdbbd5fe369fcfaf49f108561848104e7f55cc19b2a96346273c" exitCode=0 Oct 08 18:51:44 crc kubenswrapper[4859]: I1008 18:51:44.168851 4859 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-gtbjg" event={"ID":"4bbd12de-a0ea-4741-b6af-dc19464d2161","Type":"ContainerDied","Data":"97969ff064abcdbbd5fe369fcfaf49f108561848104e7f55cc19b2a96346273c"} Oct 08 18:51:45 crc kubenswrapper[4859]: I1008 18:51:45.550853 4859 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-gtbjg" Oct 08 18:51:45 crc kubenswrapper[4859]: I1008 18:51:45.629605 4859 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovn-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4bbd12de-a0ea-4741-b6af-dc19464d2161-ovn-combined-ca-bundle\") pod \"4bbd12de-a0ea-4741-b6af-dc19464d2161\" (UID: \"4bbd12de-a0ea-4741-b6af-dc19464d2161\") " Oct 08 18:51:45 crc kubenswrapper[4859]: I1008 18:51:45.629782 4859 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/4bbd12de-a0ea-4741-b6af-dc19464d2161-inventory\") pod \"4bbd12de-a0ea-4741-b6af-dc19464d2161\" (UID: \"4bbd12de-a0ea-4741-b6af-dc19464d2161\") " Oct 08 18:51:45 crc kubenswrapper[4859]: I1008 18:51:45.629833 4859 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-2j5qp\" (UniqueName: \"kubernetes.io/projected/4bbd12de-a0ea-4741-b6af-dc19464d2161-kube-api-access-2j5qp\") pod \"4bbd12de-a0ea-4741-b6af-dc19464d2161\" (UID: \"4bbd12de-a0ea-4741-b6af-dc19464d2161\") " Oct 08 18:51:45 crc kubenswrapper[4859]: I1008 18:51:45.629932 4859 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/4bbd12de-a0ea-4741-b6af-dc19464d2161-ssh-key\") pod \"4bbd12de-a0ea-4741-b6af-dc19464d2161\" (UID: \"4bbd12de-a0ea-4741-b6af-dc19464d2161\") " Oct 08 18:51:45 crc kubenswrapper[4859]: I1008 18:51:45.630015 4859 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovncontroller-config-0\" (UniqueName: \"kubernetes.io/configmap/4bbd12de-a0ea-4741-b6af-dc19464d2161-ovncontroller-config-0\") pod \"4bbd12de-a0ea-4741-b6af-dc19464d2161\" (UID: \"4bbd12de-a0ea-4741-b6af-dc19464d2161\") " Oct 08 18:51:45 crc kubenswrapper[4859]: I1008 18:51:45.636841 4859 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/4bbd12de-a0ea-4741-b6af-dc19464d2161-kube-api-access-2j5qp" (OuterVolumeSpecName: "kube-api-access-2j5qp") pod "4bbd12de-a0ea-4741-b6af-dc19464d2161" (UID: "4bbd12de-a0ea-4741-b6af-dc19464d2161"). InnerVolumeSpecName "kube-api-access-2j5qp". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 08 18:51:45 crc kubenswrapper[4859]: I1008 18:51:45.650094 4859 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/4bbd12de-a0ea-4741-b6af-dc19464d2161-ovn-combined-ca-bundle" (OuterVolumeSpecName: "ovn-combined-ca-bundle") pod "4bbd12de-a0ea-4741-b6af-dc19464d2161" (UID: "4bbd12de-a0ea-4741-b6af-dc19464d2161"). InnerVolumeSpecName "ovn-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 08 18:51:45 crc kubenswrapper[4859]: I1008 18:51:45.660457 4859 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/4bbd12de-a0ea-4741-b6af-dc19464d2161-inventory" (OuterVolumeSpecName: "inventory") pod "4bbd12de-a0ea-4741-b6af-dc19464d2161" (UID: "4bbd12de-a0ea-4741-b6af-dc19464d2161"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 08 18:51:45 crc kubenswrapper[4859]: I1008 18:51:45.671273 4859 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/4bbd12de-a0ea-4741-b6af-dc19464d2161-ovncontroller-config-0" (OuterVolumeSpecName: "ovncontroller-config-0") pod "4bbd12de-a0ea-4741-b6af-dc19464d2161" (UID: "4bbd12de-a0ea-4741-b6af-dc19464d2161"). InnerVolumeSpecName "ovncontroller-config-0". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 08 18:51:45 crc kubenswrapper[4859]: I1008 18:51:45.672318 4859 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/4bbd12de-a0ea-4741-b6af-dc19464d2161-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "4bbd12de-a0ea-4741-b6af-dc19464d2161" (UID: "4bbd12de-a0ea-4741-b6af-dc19464d2161"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 08 18:51:45 crc kubenswrapper[4859]: I1008 18:51:45.732878 4859 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/4bbd12de-a0ea-4741-b6af-dc19464d2161-ssh-key\") on node \"crc\" DevicePath \"\"" Oct 08 18:51:45 crc kubenswrapper[4859]: I1008 18:51:45.732967 4859 reconciler_common.go:293] "Volume detached for volume \"ovncontroller-config-0\" (UniqueName: \"kubernetes.io/configmap/4bbd12de-a0ea-4741-b6af-dc19464d2161-ovncontroller-config-0\") on node \"crc\" DevicePath \"\"" Oct 08 18:51:45 crc kubenswrapper[4859]: I1008 18:51:45.732992 4859 reconciler_common.go:293] "Volume detached for volume \"ovn-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4bbd12de-a0ea-4741-b6af-dc19464d2161-ovn-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 08 18:51:45 crc kubenswrapper[4859]: I1008 18:51:45.733013 4859 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/4bbd12de-a0ea-4741-b6af-dc19464d2161-inventory\") on node \"crc\" DevicePath \"\"" Oct 08 18:51:45 crc kubenswrapper[4859]: I1008 18:51:45.733032 4859 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-2j5qp\" (UniqueName: \"kubernetes.io/projected/4bbd12de-a0ea-4741-b6af-dc19464d2161-kube-api-access-2j5qp\") on node \"crc\" DevicePath \"\"" Oct 08 18:51:46 crc kubenswrapper[4859]: I1008 18:51:46.191596 4859 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-gtbjg" event={"ID":"4bbd12de-a0ea-4741-b6af-dc19464d2161","Type":"ContainerDied","Data":"7dadc24bdb1b8a1ea38f668b1481cf6c799a80fea193c5ff0cea0d704490496d"} Oct 08 18:51:46 crc kubenswrapper[4859]: I1008 18:51:46.191664 4859 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="7dadc24bdb1b8a1ea38f668b1481cf6c799a80fea193c5ff0cea0d704490496d" Oct 08 18:51:46 crc kubenswrapper[4859]: I1008 18:51:46.191682 4859 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-gtbjg" Oct 08 18:51:46 crc kubenswrapper[4859]: I1008 18:51:46.297023 4859 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-fnxfs"] Oct 08 18:51:46 crc kubenswrapper[4859]: E1008 18:51:46.299410 4859 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7e90626d-d633-4991-a768-747d73a77fbe" containerName="extract-content" Oct 08 18:51:46 crc kubenswrapper[4859]: I1008 18:51:46.299442 4859 state_mem.go:107] "Deleted CPUSet assignment" podUID="7e90626d-d633-4991-a768-747d73a77fbe" containerName="extract-content" Oct 08 18:51:46 crc kubenswrapper[4859]: E1008 18:51:46.299456 4859 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="4bbd12de-a0ea-4741-b6af-dc19464d2161" containerName="ovn-edpm-deployment-openstack-edpm-ipam" Oct 08 18:51:46 crc kubenswrapper[4859]: I1008 18:51:46.299464 4859 state_mem.go:107] "Deleted CPUSet assignment" podUID="4bbd12de-a0ea-4741-b6af-dc19464d2161" containerName="ovn-edpm-deployment-openstack-edpm-ipam" Oct 08 18:51:46 crc kubenswrapper[4859]: E1008 18:51:46.299480 4859 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7e90626d-d633-4991-a768-747d73a77fbe" containerName="registry-server" Oct 08 18:51:46 crc kubenswrapper[4859]: I1008 18:51:46.299488 4859 state_mem.go:107] "Deleted CPUSet assignment" podUID="7e90626d-d633-4991-a768-747d73a77fbe" containerName="registry-server" Oct 08 18:51:46 crc kubenswrapper[4859]: E1008 18:51:46.299533 4859 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7e90626d-d633-4991-a768-747d73a77fbe" containerName="extract-utilities" Oct 08 18:51:46 crc kubenswrapper[4859]: I1008 18:51:46.299542 4859 state_mem.go:107] "Deleted CPUSet assignment" podUID="7e90626d-d633-4991-a768-747d73a77fbe" containerName="extract-utilities" Oct 08 18:51:46 crc kubenswrapper[4859]: I1008 18:51:46.299822 4859 memory_manager.go:354] "RemoveStaleState removing state" podUID="4bbd12de-a0ea-4741-b6af-dc19464d2161" containerName="ovn-edpm-deployment-openstack-edpm-ipam" Oct 08 18:51:46 crc kubenswrapper[4859]: I1008 18:51:46.299846 4859 memory_manager.go:354] "RemoveStaleState removing state" podUID="7e90626d-d633-4991-a768-747d73a77fbe" containerName="registry-server" Oct 08 18:51:46 crc kubenswrapper[4859]: I1008 18:51:46.300672 4859 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-fnxfs" Oct 08 18:51:46 crc kubenswrapper[4859]: I1008 18:51:46.306906 4859 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-metadata-neutron-config" Oct 08 18:51:46 crc kubenswrapper[4859]: I1008 18:51:46.307000 4859 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Oct 08 18:51:46 crc kubenswrapper[4859]: I1008 18:51:46.307197 4859 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"neutron-ovn-metadata-agent-neutron-config" Oct 08 18:51:46 crc kubenswrapper[4859]: I1008 18:51:46.307309 4859 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-gxqst" Oct 08 18:51:46 crc kubenswrapper[4859]: I1008 18:51:46.307411 4859 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Oct 08 18:51:46 crc kubenswrapper[4859]: I1008 18:51:46.308497 4859 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-fnxfs"] Oct 08 18:51:46 crc kubenswrapper[4859]: I1008 18:51:46.309504 4859 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Oct 08 18:51:46 crc kubenswrapper[4859]: I1008 18:51:46.446219 4859 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/3f1e49a1-2a8d-4d9e-9fde-529f8832d0dc-ssh-key\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-fnxfs\" (UID: \"3f1e49a1-2a8d-4d9e-9fde-529f8832d0dc\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-fnxfs" Oct 08 18:51:46 crc kubenswrapper[4859]: I1008 18:51:46.446294 4859 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-5zzdx\" (UniqueName: \"kubernetes.io/projected/3f1e49a1-2a8d-4d9e-9fde-529f8832d0dc-kube-api-access-5zzdx\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-fnxfs\" (UID: \"3f1e49a1-2a8d-4d9e-9fde-529f8832d0dc\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-fnxfs" Oct 08 18:51:46 crc kubenswrapper[4859]: I1008 18:51:46.446349 4859 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"neutron-ovn-metadata-agent-neutron-config-0\" (UniqueName: \"kubernetes.io/secret/3f1e49a1-2a8d-4d9e-9fde-529f8832d0dc-neutron-ovn-metadata-agent-neutron-config-0\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-fnxfs\" (UID: \"3f1e49a1-2a8d-4d9e-9fde-529f8832d0dc\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-fnxfs" Oct 08 18:51:46 crc kubenswrapper[4859]: I1008 18:51:46.446416 4859 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/3f1e49a1-2a8d-4d9e-9fde-529f8832d0dc-inventory\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-fnxfs\" (UID: \"3f1e49a1-2a8d-4d9e-9fde-529f8832d0dc\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-fnxfs" Oct 08 18:51:46 crc kubenswrapper[4859]: I1008 18:51:46.446586 4859 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-metadata-neutron-config-0\" (UniqueName: \"kubernetes.io/secret/3f1e49a1-2a8d-4d9e-9fde-529f8832d0dc-nova-metadata-neutron-config-0\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-fnxfs\" (UID: \"3f1e49a1-2a8d-4d9e-9fde-529f8832d0dc\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-fnxfs" Oct 08 18:51:46 crc kubenswrapper[4859]: I1008 18:51:46.446820 4859 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"neutron-metadata-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3f1e49a1-2a8d-4d9e-9fde-529f8832d0dc-neutron-metadata-combined-ca-bundle\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-fnxfs\" (UID: \"3f1e49a1-2a8d-4d9e-9fde-529f8832d0dc\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-fnxfs" Oct 08 18:51:46 crc kubenswrapper[4859]: I1008 18:51:46.549180 4859 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"neutron-metadata-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3f1e49a1-2a8d-4d9e-9fde-529f8832d0dc-neutron-metadata-combined-ca-bundle\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-fnxfs\" (UID: \"3f1e49a1-2a8d-4d9e-9fde-529f8832d0dc\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-fnxfs" Oct 08 18:51:46 crc kubenswrapper[4859]: I1008 18:51:46.549328 4859 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/3f1e49a1-2a8d-4d9e-9fde-529f8832d0dc-ssh-key\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-fnxfs\" (UID: \"3f1e49a1-2a8d-4d9e-9fde-529f8832d0dc\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-fnxfs" Oct 08 18:51:46 crc kubenswrapper[4859]: I1008 18:51:46.549374 4859 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-5zzdx\" (UniqueName: \"kubernetes.io/projected/3f1e49a1-2a8d-4d9e-9fde-529f8832d0dc-kube-api-access-5zzdx\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-fnxfs\" (UID: \"3f1e49a1-2a8d-4d9e-9fde-529f8832d0dc\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-fnxfs" Oct 08 18:51:46 crc kubenswrapper[4859]: I1008 18:51:46.549422 4859 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"neutron-ovn-metadata-agent-neutron-config-0\" (UniqueName: \"kubernetes.io/secret/3f1e49a1-2a8d-4d9e-9fde-529f8832d0dc-neutron-ovn-metadata-agent-neutron-config-0\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-fnxfs\" (UID: \"3f1e49a1-2a8d-4d9e-9fde-529f8832d0dc\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-fnxfs" Oct 08 18:51:46 crc kubenswrapper[4859]: I1008 18:51:46.549552 4859 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/3f1e49a1-2a8d-4d9e-9fde-529f8832d0dc-inventory\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-fnxfs\" (UID: \"3f1e49a1-2a8d-4d9e-9fde-529f8832d0dc\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-fnxfs" Oct 08 18:51:46 crc kubenswrapper[4859]: I1008 18:51:46.549859 4859 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-metadata-neutron-config-0\" (UniqueName: \"kubernetes.io/secret/3f1e49a1-2a8d-4d9e-9fde-529f8832d0dc-nova-metadata-neutron-config-0\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-fnxfs\" (UID: \"3f1e49a1-2a8d-4d9e-9fde-529f8832d0dc\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-fnxfs" Oct 08 18:51:46 crc kubenswrapper[4859]: I1008 18:51:46.556909 4859 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"neutron-metadata-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3f1e49a1-2a8d-4d9e-9fde-529f8832d0dc-neutron-metadata-combined-ca-bundle\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-fnxfs\" (UID: \"3f1e49a1-2a8d-4d9e-9fde-529f8832d0dc\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-fnxfs" Oct 08 18:51:46 crc kubenswrapper[4859]: I1008 18:51:46.557892 4859 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"neutron-ovn-metadata-agent-neutron-config-0\" (UniqueName: \"kubernetes.io/secret/3f1e49a1-2a8d-4d9e-9fde-529f8832d0dc-neutron-ovn-metadata-agent-neutron-config-0\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-fnxfs\" (UID: \"3f1e49a1-2a8d-4d9e-9fde-529f8832d0dc\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-fnxfs" Oct 08 18:51:46 crc kubenswrapper[4859]: I1008 18:51:46.563420 4859 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/3f1e49a1-2a8d-4d9e-9fde-529f8832d0dc-ssh-key\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-fnxfs\" (UID: \"3f1e49a1-2a8d-4d9e-9fde-529f8832d0dc\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-fnxfs" Oct 08 18:51:46 crc kubenswrapper[4859]: I1008 18:51:46.563600 4859 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-metadata-neutron-config-0\" (UniqueName: \"kubernetes.io/secret/3f1e49a1-2a8d-4d9e-9fde-529f8832d0dc-nova-metadata-neutron-config-0\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-fnxfs\" (UID: \"3f1e49a1-2a8d-4d9e-9fde-529f8832d0dc\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-fnxfs" Oct 08 18:51:46 crc kubenswrapper[4859]: I1008 18:51:46.563869 4859 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/3f1e49a1-2a8d-4d9e-9fde-529f8832d0dc-inventory\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-fnxfs\" (UID: \"3f1e49a1-2a8d-4d9e-9fde-529f8832d0dc\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-fnxfs" Oct 08 18:51:46 crc kubenswrapper[4859]: I1008 18:51:46.582735 4859 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-5zzdx\" (UniqueName: \"kubernetes.io/projected/3f1e49a1-2a8d-4d9e-9fde-529f8832d0dc-kube-api-access-5zzdx\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-fnxfs\" (UID: \"3f1e49a1-2a8d-4d9e-9fde-529f8832d0dc\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-fnxfs" Oct 08 18:51:46 crc kubenswrapper[4859]: I1008 18:51:46.631437 4859 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-fnxfs" Oct 08 18:51:47 crc kubenswrapper[4859]: I1008 18:51:47.189852 4859 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-fnxfs"] Oct 08 18:51:47 crc kubenswrapper[4859]: I1008 18:51:47.201715 4859 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-fnxfs" event={"ID":"3f1e49a1-2a8d-4d9e-9fde-529f8832d0dc","Type":"ContainerStarted","Data":"4239f1a8c0f25064717db1bc1ff8f0fa6533b9ce1fb167039d2c437b800f4ca9"} Oct 08 18:51:49 crc kubenswrapper[4859]: I1008 18:51:49.219974 4859 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-fnxfs" event={"ID":"3f1e49a1-2a8d-4d9e-9fde-529f8832d0dc","Type":"ContainerStarted","Data":"13dd5292c8a57db6e44c20c5aebbeabea079cfc3f663e7c1e9291f4cff72346d"} Oct 08 18:51:49 crc kubenswrapper[4859]: I1008 18:51:49.251990 4859 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-fnxfs" podStartSLOduration=2.677480499 podStartE2EDuration="3.251961656s" podCreationTimestamp="2025-10-08 18:51:46 +0000 UTC" firstStartedPulling="2025-10-08 18:51:47.194182513 +0000 UTC m=+2077.441021902" lastFinishedPulling="2025-10-08 18:51:47.76866367 +0000 UTC m=+2078.015503059" observedRunningTime="2025-10-08 18:51:49.241169074 +0000 UTC m=+2079.488008483" watchObservedRunningTime="2025-10-08 18:51:49.251961656 +0000 UTC m=+2079.498801055" Oct 08 18:52:41 crc kubenswrapper[4859]: I1008 18:52:41.742617 4859 generic.go:334] "Generic (PLEG): container finished" podID="3f1e49a1-2a8d-4d9e-9fde-529f8832d0dc" containerID="13dd5292c8a57db6e44c20c5aebbeabea079cfc3f663e7c1e9291f4cff72346d" exitCode=0 Oct 08 18:52:41 crc kubenswrapper[4859]: I1008 18:52:41.742709 4859 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-fnxfs" event={"ID":"3f1e49a1-2a8d-4d9e-9fde-529f8832d0dc","Type":"ContainerDied","Data":"13dd5292c8a57db6e44c20c5aebbeabea079cfc3f663e7c1e9291f4cff72346d"} Oct 08 18:52:43 crc kubenswrapper[4859]: I1008 18:52:43.219869 4859 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-fnxfs" Oct 08 18:52:43 crc kubenswrapper[4859]: I1008 18:52:43.313899 4859 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"nova-metadata-neutron-config-0\" (UniqueName: \"kubernetes.io/secret/3f1e49a1-2a8d-4d9e-9fde-529f8832d0dc-nova-metadata-neutron-config-0\") pod \"3f1e49a1-2a8d-4d9e-9fde-529f8832d0dc\" (UID: \"3f1e49a1-2a8d-4d9e-9fde-529f8832d0dc\") " Oct 08 18:52:43 crc kubenswrapper[4859]: I1008 18:52:43.313998 4859 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/3f1e49a1-2a8d-4d9e-9fde-529f8832d0dc-inventory\") pod \"3f1e49a1-2a8d-4d9e-9fde-529f8832d0dc\" (UID: \"3f1e49a1-2a8d-4d9e-9fde-529f8832d0dc\") " Oct 08 18:52:43 crc kubenswrapper[4859]: I1008 18:52:43.314046 4859 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-5zzdx\" (UniqueName: \"kubernetes.io/projected/3f1e49a1-2a8d-4d9e-9fde-529f8832d0dc-kube-api-access-5zzdx\") pod \"3f1e49a1-2a8d-4d9e-9fde-529f8832d0dc\" (UID: \"3f1e49a1-2a8d-4d9e-9fde-529f8832d0dc\") " Oct 08 18:52:43 crc kubenswrapper[4859]: I1008 18:52:43.314071 4859 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/3f1e49a1-2a8d-4d9e-9fde-529f8832d0dc-ssh-key\") pod \"3f1e49a1-2a8d-4d9e-9fde-529f8832d0dc\" (UID: \"3f1e49a1-2a8d-4d9e-9fde-529f8832d0dc\") " Oct 08 18:52:43 crc kubenswrapper[4859]: I1008 18:52:43.314142 4859 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"neutron-ovn-metadata-agent-neutron-config-0\" (UniqueName: \"kubernetes.io/secret/3f1e49a1-2a8d-4d9e-9fde-529f8832d0dc-neutron-ovn-metadata-agent-neutron-config-0\") pod \"3f1e49a1-2a8d-4d9e-9fde-529f8832d0dc\" (UID: \"3f1e49a1-2a8d-4d9e-9fde-529f8832d0dc\") " Oct 08 18:52:43 crc kubenswrapper[4859]: I1008 18:52:43.314770 4859 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"neutron-metadata-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3f1e49a1-2a8d-4d9e-9fde-529f8832d0dc-neutron-metadata-combined-ca-bundle\") pod \"3f1e49a1-2a8d-4d9e-9fde-529f8832d0dc\" (UID: \"3f1e49a1-2a8d-4d9e-9fde-529f8832d0dc\") " Oct 08 18:52:43 crc kubenswrapper[4859]: I1008 18:52:43.330039 4859 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/3f1e49a1-2a8d-4d9e-9fde-529f8832d0dc-neutron-metadata-combined-ca-bundle" (OuterVolumeSpecName: "neutron-metadata-combined-ca-bundle") pod "3f1e49a1-2a8d-4d9e-9fde-529f8832d0dc" (UID: "3f1e49a1-2a8d-4d9e-9fde-529f8832d0dc"). InnerVolumeSpecName "neutron-metadata-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 08 18:52:43 crc kubenswrapper[4859]: I1008 18:52:43.330227 4859 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/3f1e49a1-2a8d-4d9e-9fde-529f8832d0dc-kube-api-access-5zzdx" (OuterVolumeSpecName: "kube-api-access-5zzdx") pod "3f1e49a1-2a8d-4d9e-9fde-529f8832d0dc" (UID: "3f1e49a1-2a8d-4d9e-9fde-529f8832d0dc"). InnerVolumeSpecName "kube-api-access-5zzdx". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 08 18:52:43 crc kubenswrapper[4859]: I1008 18:52:43.343102 4859 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/3f1e49a1-2a8d-4d9e-9fde-529f8832d0dc-inventory" (OuterVolumeSpecName: "inventory") pod "3f1e49a1-2a8d-4d9e-9fde-529f8832d0dc" (UID: "3f1e49a1-2a8d-4d9e-9fde-529f8832d0dc"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 08 18:52:43 crc kubenswrapper[4859]: I1008 18:52:43.343266 4859 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/3f1e49a1-2a8d-4d9e-9fde-529f8832d0dc-neutron-ovn-metadata-agent-neutron-config-0" (OuterVolumeSpecName: "neutron-ovn-metadata-agent-neutron-config-0") pod "3f1e49a1-2a8d-4d9e-9fde-529f8832d0dc" (UID: "3f1e49a1-2a8d-4d9e-9fde-529f8832d0dc"). InnerVolumeSpecName "neutron-ovn-metadata-agent-neutron-config-0". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 08 18:52:43 crc kubenswrapper[4859]: I1008 18:52:43.348621 4859 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/3f1e49a1-2a8d-4d9e-9fde-529f8832d0dc-nova-metadata-neutron-config-0" (OuterVolumeSpecName: "nova-metadata-neutron-config-0") pod "3f1e49a1-2a8d-4d9e-9fde-529f8832d0dc" (UID: "3f1e49a1-2a8d-4d9e-9fde-529f8832d0dc"). InnerVolumeSpecName "nova-metadata-neutron-config-0". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 08 18:52:43 crc kubenswrapper[4859]: I1008 18:52:43.350093 4859 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/3f1e49a1-2a8d-4d9e-9fde-529f8832d0dc-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "3f1e49a1-2a8d-4d9e-9fde-529f8832d0dc" (UID: "3f1e49a1-2a8d-4d9e-9fde-529f8832d0dc"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 08 18:52:43 crc kubenswrapper[4859]: I1008 18:52:43.417760 4859 reconciler_common.go:293] "Volume detached for volume \"nova-metadata-neutron-config-0\" (UniqueName: \"kubernetes.io/secret/3f1e49a1-2a8d-4d9e-9fde-529f8832d0dc-nova-metadata-neutron-config-0\") on node \"crc\" DevicePath \"\"" Oct 08 18:52:43 crc kubenswrapper[4859]: I1008 18:52:43.417797 4859 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/3f1e49a1-2a8d-4d9e-9fde-529f8832d0dc-inventory\") on node \"crc\" DevicePath \"\"" Oct 08 18:52:43 crc kubenswrapper[4859]: I1008 18:52:43.417811 4859 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-5zzdx\" (UniqueName: \"kubernetes.io/projected/3f1e49a1-2a8d-4d9e-9fde-529f8832d0dc-kube-api-access-5zzdx\") on node \"crc\" DevicePath \"\"" Oct 08 18:52:43 crc kubenswrapper[4859]: I1008 18:52:43.417824 4859 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/3f1e49a1-2a8d-4d9e-9fde-529f8832d0dc-ssh-key\") on node \"crc\" DevicePath \"\"" Oct 08 18:52:43 crc kubenswrapper[4859]: I1008 18:52:43.417868 4859 reconciler_common.go:293] "Volume detached for volume \"neutron-ovn-metadata-agent-neutron-config-0\" (UniqueName: \"kubernetes.io/secret/3f1e49a1-2a8d-4d9e-9fde-529f8832d0dc-neutron-ovn-metadata-agent-neutron-config-0\") on node \"crc\" DevicePath \"\"" Oct 08 18:52:43 crc kubenswrapper[4859]: I1008 18:52:43.417881 4859 reconciler_common.go:293] "Volume detached for volume \"neutron-metadata-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3f1e49a1-2a8d-4d9e-9fde-529f8832d0dc-neutron-metadata-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 08 18:52:43 crc kubenswrapper[4859]: I1008 18:52:43.761549 4859 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-fnxfs" event={"ID":"3f1e49a1-2a8d-4d9e-9fde-529f8832d0dc","Type":"ContainerDied","Data":"4239f1a8c0f25064717db1bc1ff8f0fa6533b9ce1fb167039d2c437b800f4ca9"} Oct 08 18:52:43 crc kubenswrapper[4859]: I1008 18:52:43.761957 4859 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="4239f1a8c0f25064717db1bc1ff8f0fa6533b9ce1fb167039d2c437b800f4ca9" Oct 08 18:52:43 crc kubenswrapper[4859]: I1008 18:52:43.761588 4859 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-fnxfs" Oct 08 18:52:43 crc kubenswrapper[4859]: I1008 18:52:43.924955 4859 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/libvirt-edpm-deployment-openstack-edpm-ipam-rrrnw"] Oct 08 18:52:43 crc kubenswrapper[4859]: E1008 18:52:43.925720 4859 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3f1e49a1-2a8d-4d9e-9fde-529f8832d0dc" containerName="neutron-metadata-edpm-deployment-openstack-edpm-ipam" Oct 08 18:52:43 crc kubenswrapper[4859]: I1008 18:52:43.925892 4859 state_mem.go:107] "Deleted CPUSet assignment" podUID="3f1e49a1-2a8d-4d9e-9fde-529f8832d0dc" containerName="neutron-metadata-edpm-deployment-openstack-edpm-ipam" Oct 08 18:52:43 crc kubenswrapper[4859]: I1008 18:52:43.926334 4859 memory_manager.go:354] "RemoveStaleState removing state" podUID="3f1e49a1-2a8d-4d9e-9fde-529f8832d0dc" containerName="neutron-metadata-edpm-deployment-openstack-edpm-ipam" Oct 08 18:52:43 crc kubenswrapper[4859]: I1008 18:52:43.927406 4859 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-rrrnw" Oct 08 18:52:43 crc kubenswrapper[4859]: I1008 18:52:43.933269 4859 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Oct 08 18:52:43 crc kubenswrapper[4859]: I1008 18:52:43.933269 4859 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"libvirt-secret" Oct 08 18:52:43 crc kubenswrapper[4859]: I1008 18:52:43.933428 4859 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Oct 08 18:52:43 crc kubenswrapper[4859]: I1008 18:52:43.933770 4859 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Oct 08 18:52:43 crc kubenswrapper[4859]: I1008 18:52:43.933915 4859 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-gxqst" Oct 08 18:52:43 crc kubenswrapper[4859]: I1008 18:52:43.942334 4859 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/libvirt-edpm-deployment-openstack-edpm-ipam-rrrnw"] Oct 08 18:52:44 crc kubenswrapper[4859]: I1008 18:52:44.030102 4859 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/a7cd7054-3a66-49ef-bcd5-405c815f8bfb-inventory\") pod \"libvirt-edpm-deployment-openstack-edpm-ipam-rrrnw\" (UID: \"a7cd7054-3a66-49ef-bcd5-405c815f8bfb\") " pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-rrrnw" Oct 08 18:52:44 crc kubenswrapper[4859]: I1008 18:52:44.030219 4859 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"libvirt-secret-0\" (UniqueName: \"kubernetes.io/secret/a7cd7054-3a66-49ef-bcd5-405c815f8bfb-libvirt-secret-0\") pod \"libvirt-edpm-deployment-openstack-edpm-ipam-rrrnw\" (UID: \"a7cd7054-3a66-49ef-bcd5-405c815f8bfb\") " pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-rrrnw" Oct 08 18:52:44 crc kubenswrapper[4859]: I1008 18:52:44.030354 4859 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"libvirt-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a7cd7054-3a66-49ef-bcd5-405c815f8bfb-libvirt-combined-ca-bundle\") pod \"libvirt-edpm-deployment-openstack-edpm-ipam-rrrnw\" (UID: \"a7cd7054-3a66-49ef-bcd5-405c815f8bfb\") " pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-rrrnw" Oct 08 18:52:44 crc kubenswrapper[4859]: I1008 18:52:44.030405 4859 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/a7cd7054-3a66-49ef-bcd5-405c815f8bfb-ssh-key\") pod \"libvirt-edpm-deployment-openstack-edpm-ipam-rrrnw\" (UID: \"a7cd7054-3a66-49ef-bcd5-405c815f8bfb\") " pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-rrrnw" Oct 08 18:52:44 crc kubenswrapper[4859]: I1008 18:52:44.030473 4859 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-td7zd\" (UniqueName: \"kubernetes.io/projected/a7cd7054-3a66-49ef-bcd5-405c815f8bfb-kube-api-access-td7zd\") pod \"libvirt-edpm-deployment-openstack-edpm-ipam-rrrnw\" (UID: \"a7cd7054-3a66-49ef-bcd5-405c815f8bfb\") " pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-rrrnw" Oct 08 18:52:44 crc kubenswrapper[4859]: I1008 18:52:44.132195 4859 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"libvirt-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a7cd7054-3a66-49ef-bcd5-405c815f8bfb-libvirt-combined-ca-bundle\") pod \"libvirt-edpm-deployment-openstack-edpm-ipam-rrrnw\" (UID: \"a7cd7054-3a66-49ef-bcd5-405c815f8bfb\") " pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-rrrnw" Oct 08 18:52:44 crc kubenswrapper[4859]: I1008 18:52:44.132260 4859 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/a7cd7054-3a66-49ef-bcd5-405c815f8bfb-ssh-key\") pod \"libvirt-edpm-deployment-openstack-edpm-ipam-rrrnw\" (UID: \"a7cd7054-3a66-49ef-bcd5-405c815f8bfb\") " pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-rrrnw" Oct 08 18:52:44 crc kubenswrapper[4859]: I1008 18:52:44.132329 4859 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-td7zd\" (UniqueName: \"kubernetes.io/projected/a7cd7054-3a66-49ef-bcd5-405c815f8bfb-kube-api-access-td7zd\") pod \"libvirt-edpm-deployment-openstack-edpm-ipam-rrrnw\" (UID: \"a7cd7054-3a66-49ef-bcd5-405c815f8bfb\") " pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-rrrnw" Oct 08 18:52:44 crc kubenswrapper[4859]: I1008 18:52:44.132378 4859 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/a7cd7054-3a66-49ef-bcd5-405c815f8bfb-inventory\") pod \"libvirt-edpm-deployment-openstack-edpm-ipam-rrrnw\" (UID: \"a7cd7054-3a66-49ef-bcd5-405c815f8bfb\") " pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-rrrnw" Oct 08 18:52:44 crc kubenswrapper[4859]: I1008 18:52:44.132440 4859 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"libvirt-secret-0\" (UniqueName: \"kubernetes.io/secret/a7cd7054-3a66-49ef-bcd5-405c815f8bfb-libvirt-secret-0\") pod \"libvirt-edpm-deployment-openstack-edpm-ipam-rrrnw\" (UID: \"a7cd7054-3a66-49ef-bcd5-405c815f8bfb\") " pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-rrrnw" Oct 08 18:52:44 crc kubenswrapper[4859]: I1008 18:52:44.137551 4859 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/a7cd7054-3a66-49ef-bcd5-405c815f8bfb-inventory\") pod \"libvirt-edpm-deployment-openstack-edpm-ipam-rrrnw\" (UID: \"a7cd7054-3a66-49ef-bcd5-405c815f8bfb\") " pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-rrrnw" Oct 08 18:52:44 crc kubenswrapper[4859]: I1008 18:52:44.137730 4859 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"libvirt-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a7cd7054-3a66-49ef-bcd5-405c815f8bfb-libvirt-combined-ca-bundle\") pod \"libvirt-edpm-deployment-openstack-edpm-ipam-rrrnw\" (UID: \"a7cd7054-3a66-49ef-bcd5-405c815f8bfb\") " pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-rrrnw" Oct 08 18:52:44 crc kubenswrapper[4859]: I1008 18:52:44.138060 4859 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"libvirt-secret-0\" (UniqueName: \"kubernetes.io/secret/a7cd7054-3a66-49ef-bcd5-405c815f8bfb-libvirt-secret-0\") pod \"libvirt-edpm-deployment-openstack-edpm-ipam-rrrnw\" (UID: \"a7cd7054-3a66-49ef-bcd5-405c815f8bfb\") " pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-rrrnw" Oct 08 18:52:44 crc kubenswrapper[4859]: I1008 18:52:44.139081 4859 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/a7cd7054-3a66-49ef-bcd5-405c815f8bfb-ssh-key\") pod \"libvirt-edpm-deployment-openstack-edpm-ipam-rrrnw\" (UID: \"a7cd7054-3a66-49ef-bcd5-405c815f8bfb\") " pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-rrrnw" Oct 08 18:52:44 crc kubenswrapper[4859]: I1008 18:52:44.157247 4859 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-td7zd\" (UniqueName: \"kubernetes.io/projected/a7cd7054-3a66-49ef-bcd5-405c815f8bfb-kube-api-access-td7zd\") pod \"libvirt-edpm-deployment-openstack-edpm-ipam-rrrnw\" (UID: \"a7cd7054-3a66-49ef-bcd5-405c815f8bfb\") " pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-rrrnw" Oct 08 18:52:44 crc kubenswrapper[4859]: I1008 18:52:44.251025 4859 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-rrrnw" Oct 08 18:52:44 crc kubenswrapper[4859]: I1008 18:52:44.808835 4859 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/libvirt-edpm-deployment-openstack-edpm-ipam-rrrnw"] Oct 08 18:52:45 crc kubenswrapper[4859]: I1008 18:52:45.780789 4859 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-rrrnw" event={"ID":"a7cd7054-3a66-49ef-bcd5-405c815f8bfb","Type":"ContainerStarted","Data":"31064c7eae0239f68af0386e905cf003a9f8b4c7407be99ba84c85d1a8c2edb2"} Oct 08 18:52:45 crc kubenswrapper[4859]: I1008 18:52:45.781794 4859 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-rrrnw" event={"ID":"a7cd7054-3a66-49ef-bcd5-405c815f8bfb","Type":"ContainerStarted","Data":"e924d2b3e4bdda0faa4884c01c379db5e2d78083a8d161c127b83d184e74f02e"} Oct 08 18:52:45 crc kubenswrapper[4859]: I1008 18:52:45.800052 4859 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-rrrnw" podStartSLOduration=2.233850589 podStartE2EDuration="2.800028336s" podCreationTimestamp="2025-10-08 18:52:43 +0000 UTC" firstStartedPulling="2025-10-08 18:52:44.827209784 +0000 UTC m=+2135.074049163" lastFinishedPulling="2025-10-08 18:52:45.393387531 +0000 UTC m=+2135.640226910" observedRunningTime="2025-10-08 18:52:45.798010627 +0000 UTC m=+2136.044850016" watchObservedRunningTime="2025-10-08 18:52:45.800028336 +0000 UTC m=+2136.046867715" Oct 08 18:52:47 crc kubenswrapper[4859]: I1008 18:52:47.925008 4859 patch_prober.go:28] interesting pod/machine-config-daemon-82s52 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 08 18:52:47 crc kubenswrapper[4859]: I1008 18:52:47.925327 4859 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-82s52" podUID="b23a6a6c-9d92-4e7b-840e-55cfda873a2d" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 08 18:52:52 crc kubenswrapper[4859]: I1008 18:52:52.213614 4859 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-4qpbl"] Oct 08 18:52:52 crc kubenswrapper[4859]: I1008 18:52:52.215847 4859 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-4qpbl" Oct 08 18:52:52 crc kubenswrapper[4859]: I1008 18:52:52.224123 4859 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-4qpbl"] Oct 08 18:52:52 crc kubenswrapper[4859]: I1008 18:52:52.313367 4859 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/c8ee999d-ee8b-4573-8a93-55a39ac7a855-catalog-content\") pod \"redhat-operators-4qpbl\" (UID: \"c8ee999d-ee8b-4573-8a93-55a39ac7a855\") " pod="openshift-marketplace/redhat-operators-4qpbl" Oct 08 18:52:52 crc kubenswrapper[4859]: I1008 18:52:52.313462 4859 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-9cr57\" (UniqueName: \"kubernetes.io/projected/c8ee999d-ee8b-4573-8a93-55a39ac7a855-kube-api-access-9cr57\") pod \"redhat-operators-4qpbl\" (UID: \"c8ee999d-ee8b-4573-8a93-55a39ac7a855\") " pod="openshift-marketplace/redhat-operators-4qpbl" Oct 08 18:52:52 crc kubenswrapper[4859]: I1008 18:52:52.313658 4859 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/c8ee999d-ee8b-4573-8a93-55a39ac7a855-utilities\") pod \"redhat-operators-4qpbl\" (UID: \"c8ee999d-ee8b-4573-8a93-55a39ac7a855\") " pod="openshift-marketplace/redhat-operators-4qpbl" Oct 08 18:52:52 crc kubenswrapper[4859]: I1008 18:52:52.414834 4859 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/c8ee999d-ee8b-4573-8a93-55a39ac7a855-catalog-content\") pod \"redhat-operators-4qpbl\" (UID: \"c8ee999d-ee8b-4573-8a93-55a39ac7a855\") " pod="openshift-marketplace/redhat-operators-4qpbl" Oct 08 18:52:52 crc kubenswrapper[4859]: I1008 18:52:52.414948 4859 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-9cr57\" (UniqueName: \"kubernetes.io/projected/c8ee999d-ee8b-4573-8a93-55a39ac7a855-kube-api-access-9cr57\") pod \"redhat-operators-4qpbl\" (UID: \"c8ee999d-ee8b-4573-8a93-55a39ac7a855\") " pod="openshift-marketplace/redhat-operators-4qpbl" Oct 08 18:52:52 crc kubenswrapper[4859]: I1008 18:52:52.415063 4859 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/c8ee999d-ee8b-4573-8a93-55a39ac7a855-utilities\") pod \"redhat-operators-4qpbl\" (UID: \"c8ee999d-ee8b-4573-8a93-55a39ac7a855\") " pod="openshift-marketplace/redhat-operators-4qpbl" Oct 08 18:52:52 crc kubenswrapper[4859]: I1008 18:52:52.415572 4859 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/c8ee999d-ee8b-4573-8a93-55a39ac7a855-utilities\") pod \"redhat-operators-4qpbl\" (UID: \"c8ee999d-ee8b-4573-8a93-55a39ac7a855\") " pod="openshift-marketplace/redhat-operators-4qpbl" Oct 08 18:52:52 crc kubenswrapper[4859]: I1008 18:52:52.415820 4859 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/c8ee999d-ee8b-4573-8a93-55a39ac7a855-catalog-content\") pod \"redhat-operators-4qpbl\" (UID: \"c8ee999d-ee8b-4573-8a93-55a39ac7a855\") " pod="openshift-marketplace/redhat-operators-4qpbl" Oct 08 18:52:52 crc kubenswrapper[4859]: I1008 18:52:52.435797 4859 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-9cr57\" (UniqueName: \"kubernetes.io/projected/c8ee999d-ee8b-4573-8a93-55a39ac7a855-kube-api-access-9cr57\") pod \"redhat-operators-4qpbl\" (UID: \"c8ee999d-ee8b-4573-8a93-55a39ac7a855\") " pod="openshift-marketplace/redhat-operators-4qpbl" Oct 08 18:52:52 crc kubenswrapper[4859]: I1008 18:52:52.565176 4859 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-4qpbl" Oct 08 18:52:53 crc kubenswrapper[4859]: I1008 18:52:53.042544 4859 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-4qpbl"] Oct 08 18:52:53 crc kubenswrapper[4859]: I1008 18:52:53.854093 4859 generic.go:334] "Generic (PLEG): container finished" podID="c8ee999d-ee8b-4573-8a93-55a39ac7a855" containerID="a01b6b47606575b6acfa2062ba696dba83daac7e2ba5ec5eb94b5ae6274ccf75" exitCode=0 Oct 08 18:52:53 crc kubenswrapper[4859]: I1008 18:52:53.854139 4859 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-4qpbl" event={"ID":"c8ee999d-ee8b-4573-8a93-55a39ac7a855","Type":"ContainerDied","Data":"a01b6b47606575b6acfa2062ba696dba83daac7e2ba5ec5eb94b5ae6274ccf75"} Oct 08 18:52:53 crc kubenswrapper[4859]: I1008 18:52:53.854444 4859 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-4qpbl" event={"ID":"c8ee999d-ee8b-4573-8a93-55a39ac7a855","Type":"ContainerStarted","Data":"664bd5b89071222336824dbf06e2c5f2646628c85feae32712618191107c12de"} Oct 08 18:52:54 crc kubenswrapper[4859]: I1008 18:52:54.865499 4859 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-4qpbl" event={"ID":"c8ee999d-ee8b-4573-8a93-55a39ac7a855","Type":"ContainerStarted","Data":"33efd467087a54ae412b090d698aea6bdf297a54c5d685663b5b9a887594e09d"} Oct 08 18:52:55 crc kubenswrapper[4859]: I1008 18:52:55.876100 4859 generic.go:334] "Generic (PLEG): container finished" podID="c8ee999d-ee8b-4573-8a93-55a39ac7a855" containerID="33efd467087a54ae412b090d698aea6bdf297a54c5d685663b5b9a887594e09d" exitCode=0 Oct 08 18:52:55 crc kubenswrapper[4859]: I1008 18:52:55.876200 4859 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-4qpbl" event={"ID":"c8ee999d-ee8b-4573-8a93-55a39ac7a855","Type":"ContainerDied","Data":"33efd467087a54ae412b090d698aea6bdf297a54c5d685663b5b9a887594e09d"} Oct 08 18:52:57 crc kubenswrapper[4859]: I1008 18:52:57.894417 4859 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-4qpbl" event={"ID":"c8ee999d-ee8b-4573-8a93-55a39ac7a855","Type":"ContainerStarted","Data":"6da7002b34086afc4eab2fe3fa162d3141eedd2424ef8cac95eccc8bc954b48f"} Oct 08 18:52:57 crc kubenswrapper[4859]: I1008 18:52:57.918532 4859 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-4qpbl" podStartSLOduration=2.787877714 podStartE2EDuration="5.918514491s" podCreationTimestamp="2025-10-08 18:52:52 +0000 UTC" firstStartedPulling="2025-10-08 18:52:53.858298583 +0000 UTC m=+2144.105137962" lastFinishedPulling="2025-10-08 18:52:56.98893536 +0000 UTC m=+2147.235774739" observedRunningTime="2025-10-08 18:52:57.91226981 +0000 UTC m=+2148.159109209" watchObservedRunningTime="2025-10-08 18:52:57.918514491 +0000 UTC m=+2148.165353870" Oct 08 18:53:02 crc kubenswrapper[4859]: I1008 18:53:02.565915 4859 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-4qpbl" Oct 08 18:53:02 crc kubenswrapper[4859]: I1008 18:53:02.566657 4859 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-4qpbl" Oct 08 18:53:03 crc kubenswrapper[4859]: I1008 18:53:03.616577 4859 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/redhat-operators-4qpbl" podUID="c8ee999d-ee8b-4573-8a93-55a39ac7a855" containerName="registry-server" probeResult="failure" output=< Oct 08 18:53:03 crc kubenswrapper[4859]: timeout: failed to connect service ":50051" within 1s Oct 08 18:53:03 crc kubenswrapper[4859]: > Oct 08 18:53:12 crc kubenswrapper[4859]: I1008 18:53:12.631755 4859 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-4qpbl" Oct 08 18:53:12 crc kubenswrapper[4859]: I1008 18:53:12.678336 4859 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-4qpbl" Oct 08 18:53:12 crc kubenswrapper[4859]: I1008 18:53:12.873521 4859 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-4qpbl"] Oct 08 18:53:14 crc kubenswrapper[4859]: I1008 18:53:14.058828 4859 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-operators-4qpbl" podUID="c8ee999d-ee8b-4573-8a93-55a39ac7a855" containerName="registry-server" containerID="cri-o://6da7002b34086afc4eab2fe3fa162d3141eedd2424ef8cac95eccc8bc954b48f" gracePeriod=2 Oct 08 18:53:14 crc kubenswrapper[4859]: I1008 18:53:14.510239 4859 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-4qpbl" Oct 08 18:53:14 crc kubenswrapper[4859]: I1008 18:53:14.551833 4859 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/c8ee999d-ee8b-4573-8a93-55a39ac7a855-catalog-content\") pod \"c8ee999d-ee8b-4573-8a93-55a39ac7a855\" (UID: \"c8ee999d-ee8b-4573-8a93-55a39ac7a855\") " Oct 08 18:53:14 crc kubenswrapper[4859]: I1008 18:53:14.551921 4859 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-9cr57\" (UniqueName: \"kubernetes.io/projected/c8ee999d-ee8b-4573-8a93-55a39ac7a855-kube-api-access-9cr57\") pod \"c8ee999d-ee8b-4573-8a93-55a39ac7a855\" (UID: \"c8ee999d-ee8b-4573-8a93-55a39ac7a855\") " Oct 08 18:53:14 crc kubenswrapper[4859]: I1008 18:53:14.551958 4859 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/c8ee999d-ee8b-4573-8a93-55a39ac7a855-utilities\") pod \"c8ee999d-ee8b-4573-8a93-55a39ac7a855\" (UID: \"c8ee999d-ee8b-4573-8a93-55a39ac7a855\") " Oct 08 18:53:14 crc kubenswrapper[4859]: I1008 18:53:14.553541 4859 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/c8ee999d-ee8b-4573-8a93-55a39ac7a855-utilities" (OuterVolumeSpecName: "utilities") pod "c8ee999d-ee8b-4573-8a93-55a39ac7a855" (UID: "c8ee999d-ee8b-4573-8a93-55a39ac7a855"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 08 18:53:14 crc kubenswrapper[4859]: I1008 18:53:14.557632 4859 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/c8ee999d-ee8b-4573-8a93-55a39ac7a855-kube-api-access-9cr57" (OuterVolumeSpecName: "kube-api-access-9cr57") pod "c8ee999d-ee8b-4573-8a93-55a39ac7a855" (UID: "c8ee999d-ee8b-4573-8a93-55a39ac7a855"). InnerVolumeSpecName "kube-api-access-9cr57". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 08 18:53:14 crc kubenswrapper[4859]: I1008 18:53:14.620550 4859 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/c8ee999d-ee8b-4573-8a93-55a39ac7a855-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "c8ee999d-ee8b-4573-8a93-55a39ac7a855" (UID: "c8ee999d-ee8b-4573-8a93-55a39ac7a855"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 08 18:53:14 crc kubenswrapper[4859]: I1008 18:53:14.654263 4859 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/c8ee999d-ee8b-4573-8a93-55a39ac7a855-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 08 18:53:14 crc kubenswrapper[4859]: I1008 18:53:14.654298 4859 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-9cr57\" (UniqueName: \"kubernetes.io/projected/c8ee999d-ee8b-4573-8a93-55a39ac7a855-kube-api-access-9cr57\") on node \"crc\" DevicePath \"\"" Oct 08 18:53:14 crc kubenswrapper[4859]: I1008 18:53:14.654313 4859 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/c8ee999d-ee8b-4573-8a93-55a39ac7a855-utilities\") on node \"crc\" DevicePath \"\"" Oct 08 18:53:15 crc kubenswrapper[4859]: I1008 18:53:15.071371 4859 generic.go:334] "Generic (PLEG): container finished" podID="c8ee999d-ee8b-4573-8a93-55a39ac7a855" containerID="6da7002b34086afc4eab2fe3fa162d3141eedd2424ef8cac95eccc8bc954b48f" exitCode=0 Oct 08 18:53:15 crc kubenswrapper[4859]: I1008 18:53:15.071427 4859 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-4qpbl" event={"ID":"c8ee999d-ee8b-4573-8a93-55a39ac7a855","Type":"ContainerDied","Data":"6da7002b34086afc4eab2fe3fa162d3141eedd2424ef8cac95eccc8bc954b48f"} Oct 08 18:53:15 crc kubenswrapper[4859]: I1008 18:53:15.071437 4859 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-4qpbl" Oct 08 18:53:15 crc kubenswrapper[4859]: I1008 18:53:15.071493 4859 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-4qpbl" event={"ID":"c8ee999d-ee8b-4573-8a93-55a39ac7a855","Type":"ContainerDied","Data":"664bd5b89071222336824dbf06e2c5f2646628c85feae32712618191107c12de"} Oct 08 18:53:15 crc kubenswrapper[4859]: I1008 18:53:15.071550 4859 scope.go:117] "RemoveContainer" containerID="6da7002b34086afc4eab2fe3fa162d3141eedd2424ef8cac95eccc8bc954b48f" Oct 08 18:53:15 crc kubenswrapper[4859]: I1008 18:53:15.111902 4859 scope.go:117] "RemoveContainer" containerID="33efd467087a54ae412b090d698aea6bdf297a54c5d685663b5b9a887594e09d" Oct 08 18:53:15 crc kubenswrapper[4859]: I1008 18:53:15.130143 4859 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-4qpbl"] Oct 08 18:53:15 crc kubenswrapper[4859]: I1008 18:53:15.139409 4859 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-operators-4qpbl"] Oct 08 18:53:15 crc kubenswrapper[4859]: I1008 18:53:15.143844 4859 scope.go:117] "RemoveContainer" containerID="a01b6b47606575b6acfa2062ba696dba83daac7e2ba5ec5eb94b5ae6274ccf75" Oct 08 18:53:15 crc kubenswrapper[4859]: I1008 18:53:15.191739 4859 scope.go:117] "RemoveContainer" containerID="6da7002b34086afc4eab2fe3fa162d3141eedd2424ef8cac95eccc8bc954b48f" Oct 08 18:53:15 crc kubenswrapper[4859]: E1008 18:53:15.192334 4859 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"6da7002b34086afc4eab2fe3fa162d3141eedd2424ef8cac95eccc8bc954b48f\": container with ID starting with 6da7002b34086afc4eab2fe3fa162d3141eedd2424ef8cac95eccc8bc954b48f not found: ID does not exist" containerID="6da7002b34086afc4eab2fe3fa162d3141eedd2424ef8cac95eccc8bc954b48f" Oct 08 18:53:15 crc kubenswrapper[4859]: I1008 18:53:15.192363 4859 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"6da7002b34086afc4eab2fe3fa162d3141eedd2424ef8cac95eccc8bc954b48f"} err="failed to get container status \"6da7002b34086afc4eab2fe3fa162d3141eedd2424ef8cac95eccc8bc954b48f\": rpc error: code = NotFound desc = could not find container \"6da7002b34086afc4eab2fe3fa162d3141eedd2424ef8cac95eccc8bc954b48f\": container with ID starting with 6da7002b34086afc4eab2fe3fa162d3141eedd2424ef8cac95eccc8bc954b48f not found: ID does not exist" Oct 08 18:53:15 crc kubenswrapper[4859]: I1008 18:53:15.192385 4859 scope.go:117] "RemoveContainer" containerID="33efd467087a54ae412b090d698aea6bdf297a54c5d685663b5b9a887594e09d" Oct 08 18:53:15 crc kubenswrapper[4859]: E1008 18:53:15.192599 4859 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"33efd467087a54ae412b090d698aea6bdf297a54c5d685663b5b9a887594e09d\": container with ID starting with 33efd467087a54ae412b090d698aea6bdf297a54c5d685663b5b9a887594e09d not found: ID does not exist" containerID="33efd467087a54ae412b090d698aea6bdf297a54c5d685663b5b9a887594e09d" Oct 08 18:53:15 crc kubenswrapper[4859]: I1008 18:53:15.192620 4859 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"33efd467087a54ae412b090d698aea6bdf297a54c5d685663b5b9a887594e09d"} err="failed to get container status \"33efd467087a54ae412b090d698aea6bdf297a54c5d685663b5b9a887594e09d\": rpc error: code = NotFound desc = could not find container \"33efd467087a54ae412b090d698aea6bdf297a54c5d685663b5b9a887594e09d\": container with ID starting with 33efd467087a54ae412b090d698aea6bdf297a54c5d685663b5b9a887594e09d not found: ID does not exist" Oct 08 18:53:15 crc kubenswrapper[4859]: I1008 18:53:15.192634 4859 scope.go:117] "RemoveContainer" containerID="a01b6b47606575b6acfa2062ba696dba83daac7e2ba5ec5eb94b5ae6274ccf75" Oct 08 18:53:15 crc kubenswrapper[4859]: E1008 18:53:15.192859 4859 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"a01b6b47606575b6acfa2062ba696dba83daac7e2ba5ec5eb94b5ae6274ccf75\": container with ID starting with a01b6b47606575b6acfa2062ba696dba83daac7e2ba5ec5eb94b5ae6274ccf75 not found: ID does not exist" containerID="a01b6b47606575b6acfa2062ba696dba83daac7e2ba5ec5eb94b5ae6274ccf75" Oct 08 18:53:15 crc kubenswrapper[4859]: I1008 18:53:15.192878 4859 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"a01b6b47606575b6acfa2062ba696dba83daac7e2ba5ec5eb94b5ae6274ccf75"} err="failed to get container status \"a01b6b47606575b6acfa2062ba696dba83daac7e2ba5ec5eb94b5ae6274ccf75\": rpc error: code = NotFound desc = could not find container \"a01b6b47606575b6acfa2062ba696dba83daac7e2ba5ec5eb94b5ae6274ccf75\": container with ID starting with a01b6b47606575b6acfa2062ba696dba83daac7e2ba5ec5eb94b5ae6274ccf75 not found: ID does not exist" Oct 08 18:53:16 crc kubenswrapper[4859]: I1008 18:53:16.493018 4859 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="c8ee999d-ee8b-4573-8a93-55a39ac7a855" path="/var/lib/kubelet/pods/c8ee999d-ee8b-4573-8a93-55a39ac7a855/volumes" Oct 08 18:53:17 crc kubenswrapper[4859]: I1008 18:53:17.925100 4859 patch_prober.go:28] interesting pod/machine-config-daemon-82s52 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 08 18:53:17 crc kubenswrapper[4859]: I1008 18:53:17.925541 4859 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-82s52" podUID="b23a6a6c-9d92-4e7b-840e-55cfda873a2d" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 08 18:53:47 crc kubenswrapper[4859]: I1008 18:53:47.924376 4859 patch_prober.go:28] interesting pod/machine-config-daemon-82s52 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 08 18:53:47 crc kubenswrapper[4859]: I1008 18:53:47.925269 4859 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-82s52" podUID="b23a6a6c-9d92-4e7b-840e-55cfda873a2d" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 08 18:53:47 crc kubenswrapper[4859]: I1008 18:53:47.925329 4859 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-82s52" Oct 08 18:53:47 crc kubenswrapper[4859]: I1008 18:53:47.926251 4859 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"3b2f671305bc6d52048373f0287719da3af01a967a70fafd0c557ca58ab22b42"} pod="openshift-machine-config-operator/machine-config-daemon-82s52" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Oct 08 18:53:47 crc kubenswrapper[4859]: I1008 18:53:47.926317 4859 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-82s52" podUID="b23a6a6c-9d92-4e7b-840e-55cfda873a2d" containerName="machine-config-daemon" containerID="cri-o://3b2f671305bc6d52048373f0287719da3af01a967a70fafd0c557ca58ab22b42" gracePeriod=600 Oct 08 18:53:48 crc kubenswrapper[4859]: I1008 18:53:48.391658 4859 generic.go:334] "Generic (PLEG): container finished" podID="b23a6a6c-9d92-4e7b-840e-55cfda873a2d" containerID="3b2f671305bc6d52048373f0287719da3af01a967a70fafd0c557ca58ab22b42" exitCode=0 Oct 08 18:53:48 crc kubenswrapper[4859]: I1008 18:53:48.391732 4859 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-82s52" event={"ID":"b23a6a6c-9d92-4e7b-840e-55cfda873a2d","Type":"ContainerDied","Data":"3b2f671305bc6d52048373f0287719da3af01a967a70fafd0c557ca58ab22b42"} Oct 08 18:53:48 crc kubenswrapper[4859]: I1008 18:53:48.391990 4859 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-82s52" event={"ID":"b23a6a6c-9d92-4e7b-840e-55cfda873a2d","Type":"ContainerStarted","Data":"265e61e43a0f4edee3adcaf87995c3c44fad0b135f61f14030c7ee297b24d748"} Oct 08 18:53:48 crc kubenswrapper[4859]: I1008 18:53:48.392021 4859 scope.go:117] "RemoveContainer" containerID="44aa7848c4f5acb003170608db1921ad40aa5b2f5464f1f85d423cd26d2e8155" Oct 08 18:55:05 crc kubenswrapper[4859]: I1008 18:55:05.136988 4859 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-kftmr"] Oct 08 18:55:05 crc kubenswrapper[4859]: E1008 18:55:05.138974 4859 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c8ee999d-ee8b-4573-8a93-55a39ac7a855" containerName="extract-content" Oct 08 18:55:05 crc kubenswrapper[4859]: I1008 18:55:05.139122 4859 state_mem.go:107] "Deleted CPUSet assignment" podUID="c8ee999d-ee8b-4573-8a93-55a39ac7a855" containerName="extract-content" Oct 08 18:55:05 crc kubenswrapper[4859]: E1008 18:55:05.139154 4859 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c8ee999d-ee8b-4573-8a93-55a39ac7a855" containerName="registry-server" Oct 08 18:55:05 crc kubenswrapper[4859]: I1008 18:55:05.139168 4859 state_mem.go:107] "Deleted CPUSet assignment" podUID="c8ee999d-ee8b-4573-8a93-55a39ac7a855" containerName="registry-server" Oct 08 18:55:05 crc kubenswrapper[4859]: E1008 18:55:05.139245 4859 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c8ee999d-ee8b-4573-8a93-55a39ac7a855" containerName="extract-utilities" Oct 08 18:55:05 crc kubenswrapper[4859]: I1008 18:55:05.139260 4859 state_mem.go:107] "Deleted CPUSet assignment" podUID="c8ee999d-ee8b-4573-8a93-55a39ac7a855" containerName="extract-utilities" Oct 08 18:55:05 crc kubenswrapper[4859]: I1008 18:55:05.140053 4859 memory_manager.go:354] "RemoveStaleState removing state" podUID="c8ee999d-ee8b-4573-8a93-55a39ac7a855" containerName="registry-server" Oct 08 18:55:05 crc kubenswrapper[4859]: I1008 18:55:05.145562 4859 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-kftmr" Oct 08 18:55:05 crc kubenswrapper[4859]: I1008 18:55:05.169167 4859 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-kftmr"] Oct 08 18:55:05 crc kubenswrapper[4859]: I1008 18:55:05.320166 4859 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-c459r\" (UniqueName: \"kubernetes.io/projected/72acafcc-032d-442c-b2fd-2eb00c8e4570-kube-api-access-c459r\") pod \"redhat-marketplace-kftmr\" (UID: \"72acafcc-032d-442c-b2fd-2eb00c8e4570\") " pod="openshift-marketplace/redhat-marketplace-kftmr" Oct 08 18:55:05 crc kubenswrapper[4859]: I1008 18:55:05.320495 4859 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/72acafcc-032d-442c-b2fd-2eb00c8e4570-utilities\") pod \"redhat-marketplace-kftmr\" (UID: \"72acafcc-032d-442c-b2fd-2eb00c8e4570\") " pod="openshift-marketplace/redhat-marketplace-kftmr" Oct 08 18:55:05 crc kubenswrapper[4859]: I1008 18:55:05.320588 4859 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/72acafcc-032d-442c-b2fd-2eb00c8e4570-catalog-content\") pod \"redhat-marketplace-kftmr\" (UID: \"72acafcc-032d-442c-b2fd-2eb00c8e4570\") " pod="openshift-marketplace/redhat-marketplace-kftmr" Oct 08 18:55:05 crc kubenswrapper[4859]: I1008 18:55:05.422281 4859 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/72acafcc-032d-442c-b2fd-2eb00c8e4570-utilities\") pod \"redhat-marketplace-kftmr\" (UID: \"72acafcc-032d-442c-b2fd-2eb00c8e4570\") " pod="openshift-marketplace/redhat-marketplace-kftmr" Oct 08 18:55:05 crc kubenswrapper[4859]: I1008 18:55:05.422336 4859 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/72acafcc-032d-442c-b2fd-2eb00c8e4570-catalog-content\") pod \"redhat-marketplace-kftmr\" (UID: \"72acafcc-032d-442c-b2fd-2eb00c8e4570\") " pod="openshift-marketplace/redhat-marketplace-kftmr" Oct 08 18:55:05 crc kubenswrapper[4859]: I1008 18:55:05.422506 4859 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-c459r\" (UniqueName: \"kubernetes.io/projected/72acafcc-032d-442c-b2fd-2eb00c8e4570-kube-api-access-c459r\") pod \"redhat-marketplace-kftmr\" (UID: \"72acafcc-032d-442c-b2fd-2eb00c8e4570\") " pod="openshift-marketplace/redhat-marketplace-kftmr" Oct 08 18:55:05 crc kubenswrapper[4859]: I1008 18:55:05.423171 4859 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/72acafcc-032d-442c-b2fd-2eb00c8e4570-utilities\") pod \"redhat-marketplace-kftmr\" (UID: \"72acafcc-032d-442c-b2fd-2eb00c8e4570\") " pod="openshift-marketplace/redhat-marketplace-kftmr" Oct 08 18:55:05 crc kubenswrapper[4859]: I1008 18:55:05.423258 4859 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/72acafcc-032d-442c-b2fd-2eb00c8e4570-catalog-content\") pod \"redhat-marketplace-kftmr\" (UID: \"72acafcc-032d-442c-b2fd-2eb00c8e4570\") " pod="openshift-marketplace/redhat-marketplace-kftmr" Oct 08 18:55:05 crc kubenswrapper[4859]: I1008 18:55:05.463671 4859 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-c459r\" (UniqueName: \"kubernetes.io/projected/72acafcc-032d-442c-b2fd-2eb00c8e4570-kube-api-access-c459r\") pod \"redhat-marketplace-kftmr\" (UID: \"72acafcc-032d-442c-b2fd-2eb00c8e4570\") " pod="openshift-marketplace/redhat-marketplace-kftmr" Oct 08 18:55:05 crc kubenswrapper[4859]: I1008 18:55:05.473725 4859 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-kftmr" Oct 08 18:55:05 crc kubenswrapper[4859]: I1008 18:55:05.911865 4859 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-kftmr"] Oct 08 18:55:05 crc kubenswrapper[4859]: W1008 18:55:05.920583 4859 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod72acafcc_032d_442c_b2fd_2eb00c8e4570.slice/crio-4de1913bc264ef4eab533f8aa146c424e8df027e4a3635684f1bdff31475aef8 WatchSource:0}: Error finding container 4de1913bc264ef4eab533f8aa146c424e8df027e4a3635684f1bdff31475aef8: Status 404 returned error can't find the container with id 4de1913bc264ef4eab533f8aa146c424e8df027e4a3635684f1bdff31475aef8 Oct 08 18:55:06 crc kubenswrapper[4859]: I1008 18:55:06.190564 4859 generic.go:334] "Generic (PLEG): container finished" podID="72acafcc-032d-442c-b2fd-2eb00c8e4570" containerID="fff8d8bcc4eb7ee1253a182fe0e0807deac9abc96a23fa32e547d363f25662eb" exitCode=0 Oct 08 18:55:06 crc kubenswrapper[4859]: I1008 18:55:06.190614 4859 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-kftmr" event={"ID":"72acafcc-032d-442c-b2fd-2eb00c8e4570","Type":"ContainerDied","Data":"fff8d8bcc4eb7ee1253a182fe0e0807deac9abc96a23fa32e547d363f25662eb"} Oct 08 18:55:06 crc kubenswrapper[4859]: I1008 18:55:06.190881 4859 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-kftmr" event={"ID":"72acafcc-032d-442c-b2fd-2eb00c8e4570","Type":"ContainerStarted","Data":"4de1913bc264ef4eab533f8aa146c424e8df027e4a3635684f1bdff31475aef8"} Oct 08 18:55:06 crc kubenswrapper[4859]: I1008 18:55:06.192674 4859 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Oct 08 18:55:07 crc kubenswrapper[4859]: I1008 18:55:07.201820 4859 generic.go:334] "Generic (PLEG): container finished" podID="72acafcc-032d-442c-b2fd-2eb00c8e4570" containerID="0ca11af64a14bf65c3b9eba69a5ca122ca73e8efc7c410b4df7e0eff9d3fd1ac" exitCode=0 Oct 08 18:55:07 crc kubenswrapper[4859]: I1008 18:55:07.201919 4859 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-kftmr" event={"ID":"72acafcc-032d-442c-b2fd-2eb00c8e4570","Type":"ContainerDied","Data":"0ca11af64a14bf65c3b9eba69a5ca122ca73e8efc7c410b4df7e0eff9d3fd1ac"} Oct 08 18:55:08 crc kubenswrapper[4859]: I1008 18:55:08.212456 4859 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-kftmr" event={"ID":"72acafcc-032d-442c-b2fd-2eb00c8e4570","Type":"ContainerStarted","Data":"5f42131f099eab3e79bed9598d14fee73f63a2e4f57e880b1219bc1a2351ab9a"} Oct 08 18:55:08 crc kubenswrapper[4859]: I1008 18:55:08.233650 4859 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-kftmr" podStartSLOduration=1.765904605 podStartE2EDuration="3.233632882s" podCreationTimestamp="2025-10-08 18:55:05 +0000 UTC" firstStartedPulling="2025-10-08 18:55:06.192412693 +0000 UTC m=+2276.439252072" lastFinishedPulling="2025-10-08 18:55:07.66014097 +0000 UTC m=+2277.906980349" observedRunningTime="2025-10-08 18:55:08.230152981 +0000 UTC m=+2278.476992360" watchObservedRunningTime="2025-10-08 18:55:08.233632882 +0000 UTC m=+2278.480472261" Oct 08 18:55:15 crc kubenswrapper[4859]: I1008 18:55:15.474445 4859 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-kftmr" Oct 08 18:55:15 crc kubenswrapper[4859]: I1008 18:55:15.475207 4859 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-kftmr" Oct 08 18:55:15 crc kubenswrapper[4859]: I1008 18:55:15.548569 4859 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-kftmr" Oct 08 18:55:16 crc kubenswrapper[4859]: I1008 18:55:16.377299 4859 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-kftmr" Oct 08 18:55:16 crc kubenswrapper[4859]: I1008 18:55:16.429089 4859 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-kftmr"] Oct 08 18:55:18 crc kubenswrapper[4859]: I1008 18:55:18.334234 4859 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-marketplace-kftmr" podUID="72acafcc-032d-442c-b2fd-2eb00c8e4570" containerName="registry-server" containerID="cri-o://5f42131f099eab3e79bed9598d14fee73f63a2e4f57e880b1219bc1a2351ab9a" gracePeriod=2 Oct 08 18:55:18 crc kubenswrapper[4859]: I1008 18:55:18.814398 4859 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-kftmr" Oct 08 18:55:19 crc kubenswrapper[4859]: I1008 18:55:19.006961 4859 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-c459r\" (UniqueName: \"kubernetes.io/projected/72acafcc-032d-442c-b2fd-2eb00c8e4570-kube-api-access-c459r\") pod \"72acafcc-032d-442c-b2fd-2eb00c8e4570\" (UID: \"72acafcc-032d-442c-b2fd-2eb00c8e4570\") " Oct 08 18:55:19 crc kubenswrapper[4859]: I1008 18:55:19.007417 4859 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/72acafcc-032d-442c-b2fd-2eb00c8e4570-catalog-content\") pod \"72acafcc-032d-442c-b2fd-2eb00c8e4570\" (UID: \"72acafcc-032d-442c-b2fd-2eb00c8e4570\") " Oct 08 18:55:19 crc kubenswrapper[4859]: I1008 18:55:19.007591 4859 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/72acafcc-032d-442c-b2fd-2eb00c8e4570-utilities\") pod \"72acafcc-032d-442c-b2fd-2eb00c8e4570\" (UID: \"72acafcc-032d-442c-b2fd-2eb00c8e4570\") " Oct 08 18:55:19 crc kubenswrapper[4859]: I1008 18:55:19.008277 4859 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/72acafcc-032d-442c-b2fd-2eb00c8e4570-utilities" (OuterVolumeSpecName: "utilities") pod "72acafcc-032d-442c-b2fd-2eb00c8e4570" (UID: "72acafcc-032d-442c-b2fd-2eb00c8e4570"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 08 18:55:19 crc kubenswrapper[4859]: I1008 18:55:19.008912 4859 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/72acafcc-032d-442c-b2fd-2eb00c8e4570-utilities\") on node \"crc\" DevicePath \"\"" Oct 08 18:55:19 crc kubenswrapper[4859]: I1008 18:55:19.012414 4859 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/72acafcc-032d-442c-b2fd-2eb00c8e4570-kube-api-access-c459r" (OuterVolumeSpecName: "kube-api-access-c459r") pod "72acafcc-032d-442c-b2fd-2eb00c8e4570" (UID: "72acafcc-032d-442c-b2fd-2eb00c8e4570"). InnerVolumeSpecName "kube-api-access-c459r". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 08 18:55:19 crc kubenswrapper[4859]: I1008 18:55:19.021493 4859 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/72acafcc-032d-442c-b2fd-2eb00c8e4570-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "72acafcc-032d-442c-b2fd-2eb00c8e4570" (UID: "72acafcc-032d-442c-b2fd-2eb00c8e4570"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 08 18:55:19 crc kubenswrapper[4859]: I1008 18:55:19.110025 4859 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-c459r\" (UniqueName: \"kubernetes.io/projected/72acafcc-032d-442c-b2fd-2eb00c8e4570-kube-api-access-c459r\") on node \"crc\" DevicePath \"\"" Oct 08 18:55:19 crc kubenswrapper[4859]: I1008 18:55:19.110072 4859 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/72acafcc-032d-442c-b2fd-2eb00c8e4570-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 08 18:55:19 crc kubenswrapper[4859]: I1008 18:55:19.347083 4859 generic.go:334] "Generic (PLEG): container finished" podID="72acafcc-032d-442c-b2fd-2eb00c8e4570" containerID="5f42131f099eab3e79bed9598d14fee73f63a2e4f57e880b1219bc1a2351ab9a" exitCode=0 Oct 08 18:55:19 crc kubenswrapper[4859]: I1008 18:55:19.347158 4859 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-kftmr" Oct 08 18:55:19 crc kubenswrapper[4859]: I1008 18:55:19.347188 4859 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-kftmr" event={"ID":"72acafcc-032d-442c-b2fd-2eb00c8e4570","Type":"ContainerDied","Data":"5f42131f099eab3e79bed9598d14fee73f63a2e4f57e880b1219bc1a2351ab9a"} Oct 08 18:55:19 crc kubenswrapper[4859]: I1008 18:55:19.347539 4859 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-kftmr" event={"ID":"72acafcc-032d-442c-b2fd-2eb00c8e4570","Type":"ContainerDied","Data":"4de1913bc264ef4eab533f8aa146c424e8df027e4a3635684f1bdff31475aef8"} Oct 08 18:55:19 crc kubenswrapper[4859]: I1008 18:55:19.347560 4859 scope.go:117] "RemoveContainer" containerID="5f42131f099eab3e79bed9598d14fee73f63a2e4f57e880b1219bc1a2351ab9a" Oct 08 18:55:19 crc kubenswrapper[4859]: I1008 18:55:19.385088 4859 scope.go:117] "RemoveContainer" containerID="0ca11af64a14bf65c3b9eba69a5ca122ca73e8efc7c410b4df7e0eff9d3fd1ac" Oct 08 18:55:19 crc kubenswrapper[4859]: I1008 18:55:19.387921 4859 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-kftmr"] Oct 08 18:55:19 crc kubenswrapper[4859]: I1008 18:55:19.395292 4859 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-marketplace-kftmr"] Oct 08 18:55:19 crc kubenswrapper[4859]: I1008 18:55:19.420015 4859 scope.go:117] "RemoveContainer" containerID="fff8d8bcc4eb7ee1253a182fe0e0807deac9abc96a23fa32e547d363f25662eb" Oct 08 18:55:19 crc kubenswrapper[4859]: I1008 18:55:19.475352 4859 scope.go:117] "RemoveContainer" containerID="5f42131f099eab3e79bed9598d14fee73f63a2e4f57e880b1219bc1a2351ab9a" Oct 08 18:55:19 crc kubenswrapper[4859]: E1008 18:55:19.475772 4859 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"5f42131f099eab3e79bed9598d14fee73f63a2e4f57e880b1219bc1a2351ab9a\": container with ID starting with 5f42131f099eab3e79bed9598d14fee73f63a2e4f57e880b1219bc1a2351ab9a not found: ID does not exist" containerID="5f42131f099eab3e79bed9598d14fee73f63a2e4f57e880b1219bc1a2351ab9a" Oct 08 18:55:19 crc kubenswrapper[4859]: I1008 18:55:19.475816 4859 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"5f42131f099eab3e79bed9598d14fee73f63a2e4f57e880b1219bc1a2351ab9a"} err="failed to get container status \"5f42131f099eab3e79bed9598d14fee73f63a2e4f57e880b1219bc1a2351ab9a\": rpc error: code = NotFound desc = could not find container \"5f42131f099eab3e79bed9598d14fee73f63a2e4f57e880b1219bc1a2351ab9a\": container with ID starting with 5f42131f099eab3e79bed9598d14fee73f63a2e4f57e880b1219bc1a2351ab9a not found: ID does not exist" Oct 08 18:55:19 crc kubenswrapper[4859]: I1008 18:55:19.475835 4859 scope.go:117] "RemoveContainer" containerID="0ca11af64a14bf65c3b9eba69a5ca122ca73e8efc7c410b4df7e0eff9d3fd1ac" Oct 08 18:55:19 crc kubenswrapper[4859]: E1008 18:55:19.476347 4859 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"0ca11af64a14bf65c3b9eba69a5ca122ca73e8efc7c410b4df7e0eff9d3fd1ac\": container with ID starting with 0ca11af64a14bf65c3b9eba69a5ca122ca73e8efc7c410b4df7e0eff9d3fd1ac not found: ID does not exist" containerID="0ca11af64a14bf65c3b9eba69a5ca122ca73e8efc7c410b4df7e0eff9d3fd1ac" Oct 08 18:55:19 crc kubenswrapper[4859]: I1008 18:55:19.476373 4859 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"0ca11af64a14bf65c3b9eba69a5ca122ca73e8efc7c410b4df7e0eff9d3fd1ac"} err="failed to get container status \"0ca11af64a14bf65c3b9eba69a5ca122ca73e8efc7c410b4df7e0eff9d3fd1ac\": rpc error: code = NotFound desc = could not find container \"0ca11af64a14bf65c3b9eba69a5ca122ca73e8efc7c410b4df7e0eff9d3fd1ac\": container with ID starting with 0ca11af64a14bf65c3b9eba69a5ca122ca73e8efc7c410b4df7e0eff9d3fd1ac not found: ID does not exist" Oct 08 18:55:19 crc kubenswrapper[4859]: I1008 18:55:19.476390 4859 scope.go:117] "RemoveContainer" containerID="fff8d8bcc4eb7ee1253a182fe0e0807deac9abc96a23fa32e547d363f25662eb" Oct 08 18:55:19 crc kubenswrapper[4859]: E1008 18:55:19.476699 4859 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"fff8d8bcc4eb7ee1253a182fe0e0807deac9abc96a23fa32e547d363f25662eb\": container with ID starting with fff8d8bcc4eb7ee1253a182fe0e0807deac9abc96a23fa32e547d363f25662eb not found: ID does not exist" containerID="fff8d8bcc4eb7ee1253a182fe0e0807deac9abc96a23fa32e547d363f25662eb" Oct 08 18:55:19 crc kubenswrapper[4859]: I1008 18:55:19.476731 4859 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"fff8d8bcc4eb7ee1253a182fe0e0807deac9abc96a23fa32e547d363f25662eb"} err="failed to get container status \"fff8d8bcc4eb7ee1253a182fe0e0807deac9abc96a23fa32e547d363f25662eb\": rpc error: code = NotFound desc = could not find container \"fff8d8bcc4eb7ee1253a182fe0e0807deac9abc96a23fa32e547d363f25662eb\": container with ID starting with fff8d8bcc4eb7ee1253a182fe0e0807deac9abc96a23fa32e547d363f25662eb not found: ID does not exist" Oct 08 18:55:20 crc kubenswrapper[4859]: I1008 18:55:20.494630 4859 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="72acafcc-032d-442c-b2fd-2eb00c8e4570" path="/var/lib/kubelet/pods/72acafcc-032d-442c-b2fd-2eb00c8e4570/volumes" Oct 08 18:56:17 crc kubenswrapper[4859]: I1008 18:56:17.926302 4859 patch_prober.go:28] interesting pod/machine-config-daemon-82s52 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 08 18:56:17 crc kubenswrapper[4859]: I1008 18:56:17.926888 4859 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-82s52" podUID="b23a6a6c-9d92-4e7b-840e-55cfda873a2d" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 08 18:56:47 crc kubenswrapper[4859]: I1008 18:56:47.924594 4859 patch_prober.go:28] interesting pod/machine-config-daemon-82s52 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 08 18:56:47 crc kubenswrapper[4859]: I1008 18:56:47.925249 4859 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-82s52" podUID="b23a6a6c-9d92-4e7b-840e-55cfda873a2d" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 08 18:57:17 crc kubenswrapper[4859]: I1008 18:57:17.924479 4859 patch_prober.go:28] interesting pod/machine-config-daemon-82s52 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 08 18:57:17 crc kubenswrapper[4859]: I1008 18:57:17.924990 4859 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-82s52" podUID="b23a6a6c-9d92-4e7b-840e-55cfda873a2d" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 08 18:57:17 crc kubenswrapper[4859]: I1008 18:57:17.925046 4859 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-82s52" Oct 08 18:57:17 crc kubenswrapper[4859]: I1008 18:57:17.925826 4859 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"265e61e43a0f4edee3adcaf87995c3c44fad0b135f61f14030c7ee297b24d748"} pod="openshift-machine-config-operator/machine-config-daemon-82s52" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Oct 08 18:57:17 crc kubenswrapper[4859]: I1008 18:57:17.925880 4859 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-82s52" podUID="b23a6a6c-9d92-4e7b-840e-55cfda873a2d" containerName="machine-config-daemon" containerID="cri-o://265e61e43a0f4edee3adcaf87995c3c44fad0b135f61f14030c7ee297b24d748" gracePeriod=600 Oct 08 18:57:18 crc kubenswrapper[4859]: E1008 18:57:18.057164 4859 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-82s52_openshift-machine-config-operator(b23a6a6c-9d92-4e7b-840e-55cfda873a2d)\"" pod="openshift-machine-config-operator/machine-config-daemon-82s52" podUID="b23a6a6c-9d92-4e7b-840e-55cfda873a2d" Oct 08 18:57:18 crc kubenswrapper[4859]: I1008 18:57:18.537473 4859 generic.go:334] "Generic (PLEG): container finished" podID="b23a6a6c-9d92-4e7b-840e-55cfda873a2d" containerID="265e61e43a0f4edee3adcaf87995c3c44fad0b135f61f14030c7ee297b24d748" exitCode=0 Oct 08 18:57:18 crc kubenswrapper[4859]: I1008 18:57:18.537526 4859 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-82s52" event={"ID":"b23a6a6c-9d92-4e7b-840e-55cfda873a2d","Type":"ContainerDied","Data":"265e61e43a0f4edee3adcaf87995c3c44fad0b135f61f14030c7ee297b24d748"} Oct 08 18:57:18 crc kubenswrapper[4859]: I1008 18:57:18.537844 4859 scope.go:117] "RemoveContainer" containerID="3b2f671305bc6d52048373f0287719da3af01a967a70fafd0c557ca58ab22b42" Oct 08 18:57:18 crc kubenswrapper[4859]: I1008 18:57:18.538418 4859 scope.go:117] "RemoveContainer" containerID="265e61e43a0f4edee3adcaf87995c3c44fad0b135f61f14030c7ee297b24d748" Oct 08 18:57:18 crc kubenswrapper[4859]: E1008 18:57:18.539507 4859 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-82s52_openshift-machine-config-operator(b23a6a6c-9d92-4e7b-840e-55cfda873a2d)\"" pod="openshift-machine-config-operator/machine-config-daemon-82s52" podUID="b23a6a6c-9d92-4e7b-840e-55cfda873a2d" Oct 08 18:57:19 crc kubenswrapper[4859]: I1008 18:57:19.549814 4859 generic.go:334] "Generic (PLEG): container finished" podID="a7cd7054-3a66-49ef-bcd5-405c815f8bfb" containerID="31064c7eae0239f68af0386e905cf003a9f8b4c7407be99ba84c85d1a8c2edb2" exitCode=0 Oct 08 18:57:19 crc kubenswrapper[4859]: I1008 18:57:19.549863 4859 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-rrrnw" event={"ID":"a7cd7054-3a66-49ef-bcd5-405c815f8bfb","Type":"ContainerDied","Data":"31064c7eae0239f68af0386e905cf003a9f8b4c7407be99ba84c85d1a8c2edb2"} Oct 08 18:57:20 crc kubenswrapper[4859]: I1008 18:57:20.967016 4859 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-rrrnw" Oct 08 18:57:21 crc kubenswrapper[4859]: I1008 18:57:21.131642 4859 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/a7cd7054-3a66-49ef-bcd5-405c815f8bfb-ssh-key\") pod \"a7cd7054-3a66-49ef-bcd5-405c815f8bfb\" (UID: \"a7cd7054-3a66-49ef-bcd5-405c815f8bfb\") " Oct 08 18:57:21 crc kubenswrapper[4859]: I1008 18:57:21.132010 4859 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"libvirt-secret-0\" (UniqueName: \"kubernetes.io/secret/a7cd7054-3a66-49ef-bcd5-405c815f8bfb-libvirt-secret-0\") pod \"a7cd7054-3a66-49ef-bcd5-405c815f8bfb\" (UID: \"a7cd7054-3a66-49ef-bcd5-405c815f8bfb\") " Oct 08 18:57:21 crc kubenswrapper[4859]: I1008 18:57:21.132045 4859 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/a7cd7054-3a66-49ef-bcd5-405c815f8bfb-inventory\") pod \"a7cd7054-3a66-49ef-bcd5-405c815f8bfb\" (UID: \"a7cd7054-3a66-49ef-bcd5-405c815f8bfb\") " Oct 08 18:57:21 crc kubenswrapper[4859]: I1008 18:57:21.132089 4859 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"libvirt-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a7cd7054-3a66-49ef-bcd5-405c815f8bfb-libvirt-combined-ca-bundle\") pod \"a7cd7054-3a66-49ef-bcd5-405c815f8bfb\" (UID: \"a7cd7054-3a66-49ef-bcd5-405c815f8bfb\") " Oct 08 18:57:21 crc kubenswrapper[4859]: I1008 18:57:21.132135 4859 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-td7zd\" (UniqueName: \"kubernetes.io/projected/a7cd7054-3a66-49ef-bcd5-405c815f8bfb-kube-api-access-td7zd\") pod \"a7cd7054-3a66-49ef-bcd5-405c815f8bfb\" (UID: \"a7cd7054-3a66-49ef-bcd5-405c815f8bfb\") " Oct 08 18:57:21 crc kubenswrapper[4859]: I1008 18:57:21.139127 4859 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a7cd7054-3a66-49ef-bcd5-405c815f8bfb-kube-api-access-td7zd" (OuterVolumeSpecName: "kube-api-access-td7zd") pod "a7cd7054-3a66-49ef-bcd5-405c815f8bfb" (UID: "a7cd7054-3a66-49ef-bcd5-405c815f8bfb"). InnerVolumeSpecName "kube-api-access-td7zd". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 08 18:57:21 crc kubenswrapper[4859]: I1008 18:57:21.140428 4859 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a7cd7054-3a66-49ef-bcd5-405c815f8bfb-libvirt-combined-ca-bundle" (OuterVolumeSpecName: "libvirt-combined-ca-bundle") pod "a7cd7054-3a66-49ef-bcd5-405c815f8bfb" (UID: "a7cd7054-3a66-49ef-bcd5-405c815f8bfb"). InnerVolumeSpecName "libvirt-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 08 18:57:21 crc kubenswrapper[4859]: I1008 18:57:21.171121 4859 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a7cd7054-3a66-49ef-bcd5-405c815f8bfb-inventory" (OuterVolumeSpecName: "inventory") pod "a7cd7054-3a66-49ef-bcd5-405c815f8bfb" (UID: "a7cd7054-3a66-49ef-bcd5-405c815f8bfb"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 08 18:57:21 crc kubenswrapper[4859]: I1008 18:57:21.176238 4859 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a7cd7054-3a66-49ef-bcd5-405c815f8bfb-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "a7cd7054-3a66-49ef-bcd5-405c815f8bfb" (UID: "a7cd7054-3a66-49ef-bcd5-405c815f8bfb"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 08 18:57:21 crc kubenswrapper[4859]: I1008 18:57:21.178822 4859 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a7cd7054-3a66-49ef-bcd5-405c815f8bfb-libvirt-secret-0" (OuterVolumeSpecName: "libvirt-secret-0") pod "a7cd7054-3a66-49ef-bcd5-405c815f8bfb" (UID: "a7cd7054-3a66-49ef-bcd5-405c815f8bfb"). InnerVolumeSpecName "libvirt-secret-0". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 08 18:57:21 crc kubenswrapper[4859]: I1008 18:57:21.235340 4859 reconciler_common.go:293] "Volume detached for volume \"libvirt-secret-0\" (UniqueName: \"kubernetes.io/secret/a7cd7054-3a66-49ef-bcd5-405c815f8bfb-libvirt-secret-0\") on node \"crc\" DevicePath \"\"" Oct 08 18:57:21 crc kubenswrapper[4859]: I1008 18:57:21.235379 4859 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/a7cd7054-3a66-49ef-bcd5-405c815f8bfb-inventory\") on node \"crc\" DevicePath \"\"" Oct 08 18:57:21 crc kubenswrapper[4859]: I1008 18:57:21.235391 4859 reconciler_common.go:293] "Volume detached for volume \"libvirt-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a7cd7054-3a66-49ef-bcd5-405c815f8bfb-libvirt-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 08 18:57:21 crc kubenswrapper[4859]: I1008 18:57:21.235402 4859 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-td7zd\" (UniqueName: \"kubernetes.io/projected/a7cd7054-3a66-49ef-bcd5-405c815f8bfb-kube-api-access-td7zd\") on node \"crc\" DevicePath \"\"" Oct 08 18:57:21 crc kubenswrapper[4859]: I1008 18:57:21.235410 4859 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/a7cd7054-3a66-49ef-bcd5-405c815f8bfb-ssh-key\") on node \"crc\" DevicePath \"\"" Oct 08 18:57:21 crc kubenswrapper[4859]: I1008 18:57:21.573642 4859 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-rrrnw" event={"ID":"a7cd7054-3a66-49ef-bcd5-405c815f8bfb","Type":"ContainerDied","Data":"e924d2b3e4bdda0faa4884c01c379db5e2d78083a8d161c127b83d184e74f02e"} Oct 08 18:57:21 crc kubenswrapper[4859]: I1008 18:57:21.574071 4859 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="e924d2b3e4bdda0faa4884c01c379db5e2d78083a8d161c127b83d184e74f02e" Oct 08 18:57:21 crc kubenswrapper[4859]: I1008 18:57:21.573792 4859 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-rrrnw" Oct 08 18:57:21 crc kubenswrapper[4859]: I1008 18:57:21.679611 4859 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-edpm-deployment-openstack-edpm-ipam-sg6sh"] Oct 08 18:57:21 crc kubenswrapper[4859]: E1008 18:57:21.680312 4859 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a7cd7054-3a66-49ef-bcd5-405c815f8bfb" containerName="libvirt-edpm-deployment-openstack-edpm-ipam" Oct 08 18:57:21 crc kubenswrapper[4859]: I1008 18:57:21.680417 4859 state_mem.go:107] "Deleted CPUSet assignment" podUID="a7cd7054-3a66-49ef-bcd5-405c815f8bfb" containerName="libvirt-edpm-deployment-openstack-edpm-ipam" Oct 08 18:57:21 crc kubenswrapper[4859]: E1008 18:57:21.680504 4859 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="72acafcc-032d-442c-b2fd-2eb00c8e4570" containerName="registry-server" Oct 08 18:57:21 crc kubenswrapper[4859]: I1008 18:57:21.680580 4859 state_mem.go:107] "Deleted CPUSet assignment" podUID="72acafcc-032d-442c-b2fd-2eb00c8e4570" containerName="registry-server" Oct 08 18:57:21 crc kubenswrapper[4859]: E1008 18:57:21.680667 4859 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="72acafcc-032d-442c-b2fd-2eb00c8e4570" containerName="extract-utilities" Oct 08 18:57:21 crc kubenswrapper[4859]: I1008 18:57:21.680757 4859 state_mem.go:107] "Deleted CPUSet assignment" podUID="72acafcc-032d-442c-b2fd-2eb00c8e4570" containerName="extract-utilities" Oct 08 18:57:21 crc kubenswrapper[4859]: E1008 18:57:21.680835 4859 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="72acafcc-032d-442c-b2fd-2eb00c8e4570" containerName="extract-content" Oct 08 18:57:21 crc kubenswrapper[4859]: I1008 18:57:21.680914 4859 state_mem.go:107] "Deleted CPUSet assignment" podUID="72acafcc-032d-442c-b2fd-2eb00c8e4570" containerName="extract-content" Oct 08 18:57:21 crc kubenswrapper[4859]: I1008 18:57:21.681205 4859 memory_manager.go:354] "RemoveStaleState removing state" podUID="a7cd7054-3a66-49ef-bcd5-405c815f8bfb" containerName="libvirt-edpm-deployment-openstack-edpm-ipam" Oct 08 18:57:21 crc kubenswrapper[4859]: I1008 18:57:21.681300 4859 memory_manager.go:354] "RemoveStaleState removing state" podUID="72acafcc-032d-442c-b2fd-2eb00c8e4570" containerName="registry-server" Oct 08 18:57:21 crc kubenswrapper[4859]: I1008 18:57:21.682194 4859 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-sg6sh" Oct 08 18:57:21 crc kubenswrapper[4859]: I1008 18:57:21.684766 4859 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Oct 08 18:57:21 crc kubenswrapper[4859]: I1008 18:57:21.685245 4859 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-gxqst" Oct 08 18:57:21 crc kubenswrapper[4859]: I1008 18:57:21.685274 4859 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-migration-ssh-key" Oct 08 18:57:21 crc kubenswrapper[4859]: I1008 18:57:21.685440 4859 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Oct 08 18:57:21 crc kubenswrapper[4859]: I1008 18:57:21.685648 4859 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"nova-extra-config" Oct 08 18:57:21 crc kubenswrapper[4859]: I1008 18:57:21.685716 4859 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Oct 08 18:57:21 crc kubenswrapper[4859]: I1008 18:57:21.685835 4859 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell1-compute-config" Oct 08 18:57:21 crc kubenswrapper[4859]: I1008 18:57:21.696270 4859 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-edpm-deployment-openstack-edpm-ipam-sg6sh"] Oct 08 18:57:21 crc kubenswrapper[4859]: I1008 18:57:21.745800 4859 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4035f4f1-5231-4112-89f6-d9d1c7c7eeb1-nova-combined-ca-bundle\") pod \"nova-edpm-deployment-openstack-edpm-ipam-sg6sh\" (UID: \"4035f4f1-5231-4112-89f6-d9d1c7c7eeb1\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-sg6sh" Oct 08 18:57:21 crc kubenswrapper[4859]: I1008 18:57:21.745864 4859 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-bgz5p\" (UniqueName: \"kubernetes.io/projected/4035f4f1-5231-4112-89f6-d9d1c7c7eeb1-kube-api-access-bgz5p\") pod \"nova-edpm-deployment-openstack-edpm-ipam-sg6sh\" (UID: \"4035f4f1-5231-4112-89f6-d9d1c7c7eeb1\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-sg6sh" Oct 08 18:57:21 crc kubenswrapper[4859]: I1008 18:57:21.745902 4859 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-cell1-compute-config-1\" (UniqueName: \"kubernetes.io/secret/4035f4f1-5231-4112-89f6-d9d1c7c7eeb1-nova-cell1-compute-config-1\") pod \"nova-edpm-deployment-openstack-edpm-ipam-sg6sh\" (UID: \"4035f4f1-5231-4112-89f6-d9d1c7c7eeb1\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-sg6sh" Oct 08 18:57:21 crc kubenswrapper[4859]: I1008 18:57:21.746000 4859 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/4035f4f1-5231-4112-89f6-d9d1c7c7eeb1-ssh-key\") pod \"nova-edpm-deployment-openstack-edpm-ipam-sg6sh\" (UID: \"4035f4f1-5231-4112-89f6-d9d1c7c7eeb1\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-sg6sh" Oct 08 18:57:21 crc kubenswrapper[4859]: I1008 18:57:21.746059 4859 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-migration-ssh-key-1\" (UniqueName: \"kubernetes.io/secret/4035f4f1-5231-4112-89f6-d9d1c7c7eeb1-nova-migration-ssh-key-1\") pod \"nova-edpm-deployment-openstack-edpm-ipam-sg6sh\" (UID: \"4035f4f1-5231-4112-89f6-d9d1c7c7eeb1\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-sg6sh" Oct 08 18:57:21 crc kubenswrapper[4859]: I1008 18:57:21.746292 4859 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/4035f4f1-5231-4112-89f6-d9d1c7c7eeb1-inventory\") pod \"nova-edpm-deployment-openstack-edpm-ipam-sg6sh\" (UID: \"4035f4f1-5231-4112-89f6-d9d1c7c7eeb1\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-sg6sh" Oct 08 18:57:21 crc kubenswrapper[4859]: I1008 18:57:21.746354 4859 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-migration-ssh-key-0\" (UniqueName: \"kubernetes.io/secret/4035f4f1-5231-4112-89f6-d9d1c7c7eeb1-nova-migration-ssh-key-0\") pod \"nova-edpm-deployment-openstack-edpm-ipam-sg6sh\" (UID: \"4035f4f1-5231-4112-89f6-d9d1c7c7eeb1\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-sg6sh" Oct 08 18:57:21 crc kubenswrapper[4859]: I1008 18:57:21.746383 4859 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-extra-config-0\" (UniqueName: \"kubernetes.io/configmap/4035f4f1-5231-4112-89f6-d9d1c7c7eeb1-nova-extra-config-0\") pod \"nova-edpm-deployment-openstack-edpm-ipam-sg6sh\" (UID: \"4035f4f1-5231-4112-89f6-d9d1c7c7eeb1\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-sg6sh" Oct 08 18:57:21 crc kubenswrapper[4859]: I1008 18:57:21.746549 4859 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-cell1-compute-config-0\" (UniqueName: \"kubernetes.io/secret/4035f4f1-5231-4112-89f6-d9d1c7c7eeb1-nova-cell1-compute-config-0\") pod \"nova-edpm-deployment-openstack-edpm-ipam-sg6sh\" (UID: \"4035f4f1-5231-4112-89f6-d9d1c7c7eeb1\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-sg6sh" Oct 08 18:57:21 crc kubenswrapper[4859]: I1008 18:57:21.847477 4859 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-cell1-compute-config-0\" (UniqueName: \"kubernetes.io/secret/4035f4f1-5231-4112-89f6-d9d1c7c7eeb1-nova-cell1-compute-config-0\") pod \"nova-edpm-deployment-openstack-edpm-ipam-sg6sh\" (UID: \"4035f4f1-5231-4112-89f6-d9d1c7c7eeb1\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-sg6sh" Oct 08 18:57:21 crc kubenswrapper[4859]: I1008 18:57:21.847571 4859 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4035f4f1-5231-4112-89f6-d9d1c7c7eeb1-nova-combined-ca-bundle\") pod \"nova-edpm-deployment-openstack-edpm-ipam-sg6sh\" (UID: \"4035f4f1-5231-4112-89f6-d9d1c7c7eeb1\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-sg6sh" Oct 08 18:57:21 crc kubenswrapper[4859]: I1008 18:57:21.847610 4859 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-bgz5p\" (UniqueName: \"kubernetes.io/projected/4035f4f1-5231-4112-89f6-d9d1c7c7eeb1-kube-api-access-bgz5p\") pod \"nova-edpm-deployment-openstack-edpm-ipam-sg6sh\" (UID: \"4035f4f1-5231-4112-89f6-d9d1c7c7eeb1\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-sg6sh" Oct 08 18:57:21 crc kubenswrapper[4859]: I1008 18:57:21.847645 4859 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-cell1-compute-config-1\" (UniqueName: \"kubernetes.io/secret/4035f4f1-5231-4112-89f6-d9d1c7c7eeb1-nova-cell1-compute-config-1\") pod \"nova-edpm-deployment-openstack-edpm-ipam-sg6sh\" (UID: \"4035f4f1-5231-4112-89f6-d9d1c7c7eeb1\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-sg6sh" Oct 08 18:57:21 crc kubenswrapper[4859]: I1008 18:57:21.847682 4859 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/4035f4f1-5231-4112-89f6-d9d1c7c7eeb1-ssh-key\") pod \"nova-edpm-deployment-openstack-edpm-ipam-sg6sh\" (UID: \"4035f4f1-5231-4112-89f6-d9d1c7c7eeb1\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-sg6sh" Oct 08 18:57:21 crc kubenswrapper[4859]: I1008 18:57:21.847733 4859 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-migration-ssh-key-1\" (UniqueName: \"kubernetes.io/secret/4035f4f1-5231-4112-89f6-d9d1c7c7eeb1-nova-migration-ssh-key-1\") pod \"nova-edpm-deployment-openstack-edpm-ipam-sg6sh\" (UID: \"4035f4f1-5231-4112-89f6-d9d1c7c7eeb1\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-sg6sh" Oct 08 18:57:21 crc kubenswrapper[4859]: I1008 18:57:21.847790 4859 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-migration-ssh-key-0\" (UniqueName: \"kubernetes.io/secret/4035f4f1-5231-4112-89f6-d9d1c7c7eeb1-nova-migration-ssh-key-0\") pod \"nova-edpm-deployment-openstack-edpm-ipam-sg6sh\" (UID: \"4035f4f1-5231-4112-89f6-d9d1c7c7eeb1\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-sg6sh" Oct 08 18:57:21 crc kubenswrapper[4859]: I1008 18:57:21.847810 4859 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/4035f4f1-5231-4112-89f6-d9d1c7c7eeb1-inventory\") pod \"nova-edpm-deployment-openstack-edpm-ipam-sg6sh\" (UID: \"4035f4f1-5231-4112-89f6-d9d1c7c7eeb1\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-sg6sh" Oct 08 18:57:21 crc kubenswrapper[4859]: I1008 18:57:21.847828 4859 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-extra-config-0\" (UniqueName: \"kubernetes.io/configmap/4035f4f1-5231-4112-89f6-d9d1c7c7eeb1-nova-extra-config-0\") pod \"nova-edpm-deployment-openstack-edpm-ipam-sg6sh\" (UID: \"4035f4f1-5231-4112-89f6-d9d1c7c7eeb1\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-sg6sh" Oct 08 18:57:21 crc kubenswrapper[4859]: I1008 18:57:21.848794 4859 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-extra-config-0\" (UniqueName: \"kubernetes.io/configmap/4035f4f1-5231-4112-89f6-d9d1c7c7eeb1-nova-extra-config-0\") pod \"nova-edpm-deployment-openstack-edpm-ipam-sg6sh\" (UID: \"4035f4f1-5231-4112-89f6-d9d1c7c7eeb1\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-sg6sh" Oct 08 18:57:21 crc kubenswrapper[4859]: I1008 18:57:21.852441 4859 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/4035f4f1-5231-4112-89f6-d9d1c7c7eeb1-inventory\") pod \"nova-edpm-deployment-openstack-edpm-ipam-sg6sh\" (UID: \"4035f4f1-5231-4112-89f6-d9d1c7c7eeb1\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-sg6sh" Oct 08 18:57:21 crc kubenswrapper[4859]: I1008 18:57:21.852537 4859 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-cell1-compute-config-1\" (UniqueName: \"kubernetes.io/secret/4035f4f1-5231-4112-89f6-d9d1c7c7eeb1-nova-cell1-compute-config-1\") pod \"nova-edpm-deployment-openstack-edpm-ipam-sg6sh\" (UID: \"4035f4f1-5231-4112-89f6-d9d1c7c7eeb1\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-sg6sh" Oct 08 18:57:21 crc kubenswrapper[4859]: I1008 18:57:21.853270 4859 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/4035f4f1-5231-4112-89f6-d9d1c7c7eeb1-ssh-key\") pod \"nova-edpm-deployment-openstack-edpm-ipam-sg6sh\" (UID: \"4035f4f1-5231-4112-89f6-d9d1c7c7eeb1\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-sg6sh" Oct 08 18:57:21 crc kubenswrapper[4859]: I1008 18:57:21.853386 4859 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4035f4f1-5231-4112-89f6-d9d1c7c7eeb1-nova-combined-ca-bundle\") pod \"nova-edpm-deployment-openstack-edpm-ipam-sg6sh\" (UID: \"4035f4f1-5231-4112-89f6-d9d1c7c7eeb1\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-sg6sh" Oct 08 18:57:21 crc kubenswrapper[4859]: I1008 18:57:21.856070 4859 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-cell1-compute-config-0\" (UniqueName: \"kubernetes.io/secret/4035f4f1-5231-4112-89f6-d9d1c7c7eeb1-nova-cell1-compute-config-0\") pod \"nova-edpm-deployment-openstack-edpm-ipam-sg6sh\" (UID: \"4035f4f1-5231-4112-89f6-d9d1c7c7eeb1\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-sg6sh" Oct 08 18:57:21 crc kubenswrapper[4859]: I1008 18:57:21.857339 4859 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-migration-ssh-key-0\" (UniqueName: \"kubernetes.io/secret/4035f4f1-5231-4112-89f6-d9d1c7c7eeb1-nova-migration-ssh-key-0\") pod \"nova-edpm-deployment-openstack-edpm-ipam-sg6sh\" (UID: \"4035f4f1-5231-4112-89f6-d9d1c7c7eeb1\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-sg6sh" Oct 08 18:57:21 crc kubenswrapper[4859]: I1008 18:57:21.869241 4859 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-migration-ssh-key-1\" (UniqueName: \"kubernetes.io/secret/4035f4f1-5231-4112-89f6-d9d1c7c7eeb1-nova-migration-ssh-key-1\") pod \"nova-edpm-deployment-openstack-edpm-ipam-sg6sh\" (UID: \"4035f4f1-5231-4112-89f6-d9d1c7c7eeb1\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-sg6sh" Oct 08 18:57:21 crc kubenswrapper[4859]: I1008 18:57:21.872173 4859 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-bgz5p\" (UniqueName: \"kubernetes.io/projected/4035f4f1-5231-4112-89f6-d9d1c7c7eeb1-kube-api-access-bgz5p\") pod \"nova-edpm-deployment-openstack-edpm-ipam-sg6sh\" (UID: \"4035f4f1-5231-4112-89f6-d9d1c7c7eeb1\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-sg6sh" Oct 08 18:57:22 crc kubenswrapper[4859]: I1008 18:57:22.002809 4859 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-sg6sh" Oct 08 18:57:22 crc kubenswrapper[4859]: I1008 18:57:22.565205 4859 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-edpm-deployment-openstack-edpm-ipam-sg6sh"] Oct 08 18:57:22 crc kubenswrapper[4859]: I1008 18:57:22.592804 4859 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-sg6sh" event={"ID":"4035f4f1-5231-4112-89f6-d9d1c7c7eeb1","Type":"ContainerStarted","Data":"bb11a7df0b2ad4ddfa91ba90319c698fb38f6d8c8d3efc114f2abf6a970cbae7"} Oct 08 18:57:23 crc kubenswrapper[4859]: I1008 18:57:23.610640 4859 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-sg6sh" event={"ID":"4035f4f1-5231-4112-89f6-d9d1c7c7eeb1","Type":"ContainerStarted","Data":"9bfcc3c84a379d54bf72f4a143a9e38414f0e9352c2a57cdf92cac39a56b7c00"} Oct 08 18:57:23 crc kubenswrapper[4859]: I1008 18:57:23.646299 4859 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-sg6sh" podStartSLOduration=2.22188462 podStartE2EDuration="2.646271199s" podCreationTimestamp="2025-10-08 18:57:21 +0000 UTC" firstStartedPulling="2025-10-08 18:57:22.577242019 +0000 UTC m=+2412.824081398" lastFinishedPulling="2025-10-08 18:57:23.001628578 +0000 UTC m=+2413.248467977" observedRunningTime="2025-10-08 18:57:23.632937846 +0000 UTC m=+2413.879777225" watchObservedRunningTime="2025-10-08 18:57:23.646271199 +0000 UTC m=+2413.893110618" Oct 08 18:57:29 crc kubenswrapper[4859]: I1008 18:57:29.471626 4859 scope.go:117] "RemoveContainer" containerID="265e61e43a0f4edee3adcaf87995c3c44fad0b135f61f14030c7ee297b24d748" Oct 08 18:57:29 crc kubenswrapper[4859]: E1008 18:57:29.472431 4859 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-82s52_openshift-machine-config-operator(b23a6a6c-9d92-4e7b-840e-55cfda873a2d)\"" pod="openshift-machine-config-operator/machine-config-daemon-82s52" podUID="b23a6a6c-9d92-4e7b-840e-55cfda873a2d" Oct 08 18:57:44 crc kubenswrapper[4859]: I1008 18:57:44.469459 4859 scope.go:117] "RemoveContainer" containerID="265e61e43a0f4edee3adcaf87995c3c44fad0b135f61f14030c7ee297b24d748" Oct 08 18:57:44 crc kubenswrapper[4859]: E1008 18:57:44.470179 4859 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-82s52_openshift-machine-config-operator(b23a6a6c-9d92-4e7b-840e-55cfda873a2d)\"" pod="openshift-machine-config-operator/machine-config-daemon-82s52" podUID="b23a6a6c-9d92-4e7b-840e-55cfda873a2d" Oct 08 18:57:56 crc kubenswrapper[4859]: I1008 18:57:56.470025 4859 scope.go:117] "RemoveContainer" containerID="265e61e43a0f4edee3adcaf87995c3c44fad0b135f61f14030c7ee297b24d748" Oct 08 18:57:56 crc kubenswrapper[4859]: E1008 18:57:56.470855 4859 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-82s52_openshift-machine-config-operator(b23a6a6c-9d92-4e7b-840e-55cfda873a2d)\"" pod="openshift-machine-config-operator/machine-config-daemon-82s52" podUID="b23a6a6c-9d92-4e7b-840e-55cfda873a2d" Oct 08 18:58:10 crc kubenswrapper[4859]: I1008 18:58:10.476370 4859 scope.go:117] "RemoveContainer" containerID="265e61e43a0f4edee3adcaf87995c3c44fad0b135f61f14030c7ee297b24d748" Oct 08 18:58:10 crc kubenswrapper[4859]: E1008 18:58:10.481403 4859 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-82s52_openshift-machine-config-operator(b23a6a6c-9d92-4e7b-840e-55cfda873a2d)\"" pod="openshift-machine-config-operator/machine-config-daemon-82s52" podUID="b23a6a6c-9d92-4e7b-840e-55cfda873a2d" Oct 08 18:58:25 crc kubenswrapper[4859]: I1008 18:58:25.470552 4859 scope.go:117] "RemoveContainer" containerID="265e61e43a0f4edee3adcaf87995c3c44fad0b135f61f14030c7ee297b24d748" Oct 08 18:58:25 crc kubenswrapper[4859]: E1008 18:58:25.471413 4859 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-82s52_openshift-machine-config-operator(b23a6a6c-9d92-4e7b-840e-55cfda873a2d)\"" pod="openshift-machine-config-operator/machine-config-daemon-82s52" podUID="b23a6a6c-9d92-4e7b-840e-55cfda873a2d" Oct 08 18:58:36 crc kubenswrapper[4859]: I1008 18:58:36.469948 4859 scope.go:117] "RemoveContainer" containerID="265e61e43a0f4edee3adcaf87995c3c44fad0b135f61f14030c7ee297b24d748" Oct 08 18:58:36 crc kubenswrapper[4859]: E1008 18:58:36.470715 4859 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-82s52_openshift-machine-config-operator(b23a6a6c-9d92-4e7b-840e-55cfda873a2d)\"" pod="openshift-machine-config-operator/machine-config-daemon-82s52" podUID="b23a6a6c-9d92-4e7b-840e-55cfda873a2d" Oct 08 18:58:49 crc kubenswrapper[4859]: I1008 18:58:49.470215 4859 scope.go:117] "RemoveContainer" containerID="265e61e43a0f4edee3adcaf87995c3c44fad0b135f61f14030c7ee297b24d748" Oct 08 18:58:49 crc kubenswrapper[4859]: E1008 18:58:49.471545 4859 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-82s52_openshift-machine-config-operator(b23a6a6c-9d92-4e7b-840e-55cfda873a2d)\"" pod="openshift-machine-config-operator/machine-config-daemon-82s52" podUID="b23a6a6c-9d92-4e7b-840e-55cfda873a2d" Oct 08 18:59:01 crc kubenswrapper[4859]: I1008 18:59:01.469274 4859 scope.go:117] "RemoveContainer" containerID="265e61e43a0f4edee3adcaf87995c3c44fad0b135f61f14030c7ee297b24d748" Oct 08 18:59:01 crc kubenswrapper[4859]: E1008 18:59:01.469998 4859 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-82s52_openshift-machine-config-operator(b23a6a6c-9d92-4e7b-840e-55cfda873a2d)\"" pod="openshift-machine-config-operator/machine-config-daemon-82s52" podUID="b23a6a6c-9d92-4e7b-840e-55cfda873a2d" Oct 08 18:59:16 crc kubenswrapper[4859]: I1008 18:59:16.470309 4859 scope.go:117] "RemoveContainer" containerID="265e61e43a0f4edee3adcaf87995c3c44fad0b135f61f14030c7ee297b24d748" Oct 08 18:59:16 crc kubenswrapper[4859]: E1008 18:59:16.471244 4859 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-82s52_openshift-machine-config-operator(b23a6a6c-9d92-4e7b-840e-55cfda873a2d)\"" pod="openshift-machine-config-operator/machine-config-daemon-82s52" podUID="b23a6a6c-9d92-4e7b-840e-55cfda873a2d" Oct 08 18:59:27 crc kubenswrapper[4859]: I1008 18:59:27.470724 4859 scope.go:117] "RemoveContainer" containerID="265e61e43a0f4edee3adcaf87995c3c44fad0b135f61f14030c7ee297b24d748" Oct 08 18:59:27 crc kubenswrapper[4859]: E1008 18:59:27.471931 4859 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-82s52_openshift-machine-config-operator(b23a6a6c-9d92-4e7b-840e-55cfda873a2d)\"" pod="openshift-machine-config-operator/machine-config-daemon-82s52" podUID="b23a6a6c-9d92-4e7b-840e-55cfda873a2d" Oct 08 18:59:42 crc kubenswrapper[4859]: I1008 18:59:42.470438 4859 scope.go:117] "RemoveContainer" containerID="265e61e43a0f4edee3adcaf87995c3c44fad0b135f61f14030c7ee297b24d748" Oct 08 18:59:42 crc kubenswrapper[4859]: E1008 18:59:42.472063 4859 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-82s52_openshift-machine-config-operator(b23a6a6c-9d92-4e7b-840e-55cfda873a2d)\"" pod="openshift-machine-config-operator/machine-config-daemon-82s52" podUID="b23a6a6c-9d92-4e7b-840e-55cfda873a2d" Oct 08 18:59:54 crc kubenswrapper[4859]: I1008 18:59:54.470949 4859 scope.go:117] "RemoveContainer" containerID="265e61e43a0f4edee3adcaf87995c3c44fad0b135f61f14030c7ee297b24d748" Oct 08 18:59:54 crc kubenswrapper[4859]: E1008 18:59:54.471973 4859 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-82s52_openshift-machine-config-operator(b23a6a6c-9d92-4e7b-840e-55cfda873a2d)\"" pod="openshift-machine-config-operator/machine-config-daemon-82s52" podUID="b23a6a6c-9d92-4e7b-840e-55cfda873a2d" Oct 08 19:00:00 crc kubenswrapper[4859]: I1008 19:00:00.162913 4859 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29332500-p86x4"] Oct 08 19:00:00 crc kubenswrapper[4859]: I1008 19:00:00.167511 4859 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29332500-p86x4" Oct 08 19:00:00 crc kubenswrapper[4859]: I1008 19:00:00.170300 4859 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"collect-profiles-config" Oct 08 19:00:00 crc kubenswrapper[4859]: I1008 19:00:00.170400 4859 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"collect-profiles-dockercfg-kzf4t" Oct 08 19:00:00 crc kubenswrapper[4859]: I1008 19:00:00.178569 4859 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29332500-p86x4"] Oct 08 19:00:00 crc kubenswrapper[4859]: I1008 19:00:00.249320 4859 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/37fd514d-24a1-4204-9125-578bb9c0779e-config-volume\") pod \"collect-profiles-29332500-p86x4\" (UID: \"37fd514d-24a1-4204-9125-578bb9c0779e\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29332500-p86x4" Oct 08 19:00:00 crc kubenswrapper[4859]: I1008 19:00:00.249978 4859 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/37fd514d-24a1-4204-9125-578bb9c0779e-secret-volume\") pod \"collect-profiles-29332500-p86x4\" (UID: \"37fd514d-24a1-4204-9125-578bb9c0779e\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29332500-p86x4" Oct 08 19:00:00 crc kubenswrapper[4859]: I1008 19:00:00.250098 4859 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-zgf6g\" (UniqueName: \"kubernetes.io/projected/37fd514d-24a1-4204-9125-578bb9c0779e-kube-api-access-zgf6g\") pod \"collect-profiles-29332500-p86x4\" (UID: \"37fd514d-24a1-4204-9125-578bb9c0779e\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29332500-p86x4" Oct 08 19:00:00 crc kubenswrapper[4859]: I1008 19:00:00.351698 4859 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/37fd514d-24a1-4204-9125-578bb9c0779e-secret-volume\") pod \"collect-profiles-29332500-p86x4\" (UID: \"37fd514d-24a1-4204-9125-578bb9c0779e\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29332500-p86x4" Oct 08 19:00:00 crc kubenswrapper[4859]: I1008 19:00:00.351760 4859 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-zgf6g\" (UniqueName: \"kubernetes.io/projected/37fd514d-24a1-4204-9125-578bb9c0779e-kube-api-access-zgf6g\") pod \"collect-profiles-29332500-p86x4\" (UID: \"37fd514d-24a1-4204-9125-578bb9c0779e\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29332500-p86x4" Oct 08 19:00:00 crc kubenswrapper[4859]: I1008 19:00:00.351805 4859 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/37fd514d-24a1-4204-9125-578bb9c0779e-config-volume\") pod \"collect-profiles-29332500-p86x4\" (UID: \"37fd514d-24a1-4204-9125-578bb9c0779e\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29332500-p86x4" Oct 08 19:00:00 crc kubenswrapper[4859]: I1008 19:00:00.353409 4859 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/37fd514d-24a1-4204-9125-578bb9c0779e-config-volume\") pod \"collect-profiles-29332500-p86x4\" (UID: \"37fd514d-24a1-4204-9125-578bb9c0779e\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29332500-p86x4" Oct 08 19:00:00 crc kubenswrapper[4859]: I1008 19:00:00.357661 4859 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/37fd514d-24a1-4204-9125-578bb9c0779e-secret-volume\") pod \"collect-profiles-29332500-p86x4\" (UID: \"37fd514d-24a1-4204-9125-578bb9c0779e\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29332500-p86x4" Oct 08 19:00:00 crc kubenswrapper[4859]: I1008 19:00:00.369157 4859 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-zgf6g\" (UniqueName: \"kubernetes.io/projected/37fd514d-24a1-4204-9125-578bb9c0779e-kube-api-access-zgf6g\") pod \"collect-profiles-29332500-p86x4\" (UID: \"37fd514d-24a1-4204-9125-578bb9c0779e\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29332500-p86x4" Oct 08 19:00:00 crc kubenswrapper[4859]: I1008 19:00:00.498651 4859 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29332500-p86x4" Oct 08 19:00:00 crc kubenswrapper[4859]: I1008 19:00:00.934394 4859 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29332500-p86x4"] Oct 08 19:00:00 crc kubenswrapper[4859]: W1008 19:00:00.947060 4859 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod37fd514d_24a1_4204_9125_578bb9c0779e.slice/crio-b66d4e5ab0a26bfe12b36228953b9b71b1e258aa0dd734e0c9a33f4781052a7b WatchSource:0}: Error finding container b66d4e5ab0a26bfe12b36228953b9b71b1e258aa0dd734e0c9a33f4781052a7b: Status 404 returned error can't find the container with id b66d4e5ab0a26bfe12b36228953b9b71b1e258aa0dd734e0c9a33f4781052a7b Oct 08 19:00:01 crc kubenswrapper[4859]: I1008 19:00:01.126346 4859 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29332500-p86x4" event={"ID":"37fd514d-24a1-4204-9125-578bb9c0779e","Type":"ContainerStarted","Data":"bb6a83dccccdb85d07f6c0caf85b3bc458ce5f450fbce482f8dd8554ce7b28a0"} Oct 08 19:00:01 crc kubenswrapper[4859]: I1008 19:00:01.126410 4859 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29332500-p86x4" event={"ID":"37fd514d-24a1-4204-9125-578bb9c0779e","Type":"ContainerStarted","Data":"b66d4e5ab0a26bfe12b36228953b9b71b1e258aa0dd734e0c9a33f4781052a7b"} Oct 08 19:00:01 crc kubenswrapper[4859]: I1008 19:00:01.147085 4859 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-operator-lifecycle-manager/collect-profiles-29332500-p86x4" podStartSLOduration=1.147068326 podStartE2EDuration="1.147068326s" podCreationTimestamp="2025-10-08 19:00:00 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-08 19:00:01.141580817 +0000 UTC m=+2571.388420196" watchObservedRunningTime="2025-10-08 19:00:01.147068326 +0000 UTC m=+2571.393907705" Oct 08 19:00:02 crc kubenswrapper[4859]: I1008 19:00:02.136605 4859 generic.go:334] "Generic (PLEG): container finished" podID="37fd514d-24a1-4204-9125-578bb9c0779e" containerID="bb6a83dccccdb85d07f6c0caf85b3bc458ce5f450fbce482f8dd8554ce7b28a0" exitCode=0 Oct 08 19:00:02 crc kubenswrapper[4859]: I1008 19:00:02.136671 4859 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29332500-p86x4" event={"ID":"37fd514d-24a1-4204-9125-578bb9c0779e","Type":"ContainerDied","Data":"bb6a83dccccdb85d07f6c0caf85b3bc458ce5f450fbce482f8dd8554ce7b28a0"} Oct 08 19:00:03 crc kubenswrapper[4859]: I1008 19:00:03.559121 4859 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29332500-p86x4" Oct 08 19:00:03 crc kubenswrapper[4859]: I1008 19:00:03.717074 4859 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-zgf6g\" (UniqueName: \"kubernetes.io/projected/37fd514d-24a1-4204-9125-578bb9c0779e-kube-api-access-zgf6g\") pod \"37fd514d-24a1-4204-9125-578bb9c0779e\" (UID: \"37fd514d-24a1-4204-9125-578bb9c0779e\") " Oct 08 19:00:03 crc kubenswrapper[4859]: I1008 19:00:03.717208 4859 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/37fd514d-24a1-4204-9125-578bb9c0779e-secret-volume\") pod \"37fd514d-24a1-4204-9125-578bb9c0779e\" (UID: \"37fd514d-24a1-4204-9125-578bb9c0779e\") " Oct 08 19:00:03 crc kubenswrapper[4859]: I1008 19:00:03.717324 4859 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/37fd514d-24a1-4204-9125-578bb9c0779e-config-volume\") pod \"37fd514d-24a1-4204-9125-578bb9c0779e\" (UID: \"37fd514d-24a1-4204-9125-578bb9c0779e\") " Oct 08 19:00:03 crc kubenswrapper[4859]: I1008 19:00:03.718454 4859 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/37fd514d-24a1-4204-9125-578bb9c0779e-config-volume" (OuterVolumeSpecName: "config-volume") pod "37fd514d-24a1-4204-9125-578bb9c0779e" (UID: "37fd514d-24a1-4204-9125-578bb9c0779e"). InnerVolumeSpecName "config-volume". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 08 19:00:03 crc kubenswrapper[4859]: I1008 19:00:03.718933 4859 reconciler_common.go:293] "Volume detached for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/37fd514d-24a1-4204-9125-578bb9c0779e-config-volume\") on node \"crc\" DevicePath \"\"" Oct 08 19:00:03 crc kubenswrapper[4859]: I1008 19:00:03.723872 4859 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/37fd514d-24a1-4204-9125-578bb9c0779e-secret-volume" (OuterVolumeSpecName: "secret-volume") pod "37fd514d-24a1-4204-9125-578bb9c0779e" (UID: "37fd514d-24a1-4204-9125-578bb9c0779e"). InnerVolumeSpecName "secret-volume". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 08 19:00:03 crc kubenswrapper[4859]: I1008 19:00:03.723978 4859 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/37fd514d-24a1-4204-9125-578bb9c0779e-kube-api-access-zgf6g" (OuterVolumeSpecName: "kube-api-access-zgf6g") pod "37fd514d-24a1-4204-9125-578bb9c0779e" (UID: "37fd514d-24a1-4204-9125-578bb9c0779e"). InnerVolumeSpecName "kube-api-access-zgf6g". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 08 19:00:03 crc kubenswrapper[4859]: I1008 19:00:03.822074 4859 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-zgf6g\" (UniqueName: \"kubernetes.io/projected/37fd514d-24a1-4204-9125-578bb9c0779e-kube-api-access-zgf6g\") on node \"crc\" DevicePath \"\"" Oct 08 19:00:03 crc kubenswrapper[4859]: I1008 19:00:03.822314 4859 reconciler_common.go:293] "Volume detached for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/37fd514d-24a1-4204-9125-578bb9c0779e-secret-volume\") on node \"crc\" DevicePath \"\"" Oct 08 19:00:04 crc kubenswrapper[4859]: I1008 19:00:04.157555 4859 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29332500-p86x4" event={"ID":"37fd514d-24a1-4204-9125-578bb9c0779e","Type":"ContainerDied","Data":"b66d4e5ab0a26bfe12b36228953b9b71b1e258aa0dd734e0c9a33f4781052a7b"} Oct 08 19:00:04 crc kubenswrapper[4859]: I1008 19:00:04.157602 4859 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="b66d4e5ab0a26bfe12b36228953b9b71b1e258aa0dd734e0c9a33f4781052a7b" Oct 08 19:00:04 crc kubenswrapper[4859]: I1008 19:00:04.157726 4859 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29332500-p86x4" Oct 08 19:00:04 crc kubenswrapper[4859]: I1008 19:00:04.630631 4859 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29332455-w9wt8"] Oct 08 19:00:04 crc kubenswrapper[4859]: I1008 19:00:04.639473 4859 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29332455-w9wt8"] Oct 08 19:00:05 crc kubenswrapper[4859]: I1008 19:00:05.471308 4859 scope.go:117] "RemoveContainer" containerID="265e61e43a0f4edee3adcaf87995c3c44fad0b135f61f14030c7ee297b24d748" Oct 08 19:00:05 crc kubenswrapper[4859]: E1008 19:00:05.472828 4859 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-82s52_openshift-machine-config-operator(b23a6a6c-9d92-4e7b-840e-55cfda873a2d)\"" pod="openshift-machine-config-operator/machine-config-daemon-82s52" podUID="b23a6a6c-9d92-4e7b-840e-55cfda873a2d" Oct 08 19:00:06 crc kubenswrapper[4859]: I1008 19:00:06.485183 4859 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="d403dae9-eb8c-4c8d-9daa-bf2a0c815652" path="/var/lib/kubelet/pods/d403dae9-eb8c-4c8d-9daa-bf2a0c815652/volumes" Oct 08 19:00:17 crc kubenswrapper[4859]: I1008 19:00:17.471211 4859 scope.go:117] "RemoveContainer" containerID="265e61e43a0f4edee3adcaf87995c3c44fad0b135f61f14030c7ee297b24d748" Oct 08 19:00:17 crc kubenswrapper[4859]: E1008 19:00:17.472105 4859 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-82s52_openshift-machine-config-operator(b23a6a6c-9d92-4e7b-840e-55cfda873a2d)\"" pod="openshift-machine-config-operator/machine-config-daemon-82s52" podUID="b23a6a6c-9d92-4e7b-840e-55cfda873a2d" Oct 08 19:00:24 crc kubenswrapper[4859]: I1008 19:00:24.278634 4859 scope.go:117] "RemoveContainer" containerID="562f5b1c2a23fe3445feaf2252f9159c42be63c462f3daaa37a9a6a551fc6acc" Oct 08 19:00:32 crc kubenswrapper[4859]: I1008 19:00:32.470026 4859 scope.go:117] "RemoveContainer" containerID="265e61e43a0f4edee3adcaf87995c3c44fad0b135f61f14030c7ee297b24d748" Oct 08 19:00:32 crc kubenswrapper[4859]: E1008 19:00:32.472207 4859 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-82s52_openshift-machine-config-operator(b23a6a6c-9d92-4e7b-840e-55cfda873a2d)\"" pod="openshift-machine-config-operator/machine-config-daemon-82s52" podUID="b23a6a6c-9d92-4e7b-840e-55cfda873a2d" Oct 08 19:00:44 crc kubenswrapper[4859]: I1008 19:00:44.471301 4859 scope.go:117] "RemoveContainer" containerID="265e61e43a0f4edee3adcaf87995c3c44fad0b135f61f14030c7ee297b24d748" Oct 08 19:00:44 crc kubenswrapper[4859]: E1008 19:00:44.472075 4859 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-82s52_openshift-machine-config-operator(b23a6a6c-9d92-4e7b-840e-55cfda873a2d)\"" pod="openshift-machine-config-operator/machine-config-daemon-82s52" podUID="b23a6a6c-9d92-4e7b-840e-55cfda873a2d" Oct 08 19:00:55 crc kubenswrapper[4859]: I1008 19:00:55.469709 4859 scope.go:117] "RemoveContainer" containerID="265e61e43a0f4edee3adcaf87995c3c44fad0b135f61f14030c7ee297b24d748" Oct 08 19:00:55 crc kubenswrapper[4859]: E1008 19:00:55.470648 4859 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-82s52_openshift-machine-config-operator(b23a6a6c-9d92-4e7b-840e-55cfda873a2d)\"" pod="openshift-machine-config-operator/machine-config-daemon-82s52" podUID="b23a6a6c-9d92-4e7b-840e-55cfda873a2d" Oct 08 19:00:58 crc kubenswrapper[4859]: I1008 19:00:58.696814 4859 generic.go:334] "Generic (PLEG): container finished" podID="4035f4f1-5231-4112-89f6-d9d1c7c7eeb1" containerID="9bfcc3c84a379d54bf72f4a143a9e38414f0e9352c2a57cdf92cac39a56b7c00" exitCode=0 Oct 08 19:00:58 crc kubenswrapper[4859]: I1008 19:00:58.696944 4859 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-sg6sh" event={"ID":"4035f4f1-5231-4112-89f6-d9d1c7c7eeb1","Type":"ContainerDied","Data":"9bfcc3c84a379d54bf72f4a143a9e38414f0e9352c2a57cdf92cac39a56b7c00"} Oct 08 19:01:00 crc kubenswrapper[4859]: I1008 19:01:00.151607 4859 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/keystone-cron-29332501-rxhgw"] Oct 08 19:01:00 crc kubenswrapper[4859]: E1008 19:01:00.152437 4859 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="37fd514d-24a1-4204-9125-578bb9c0779e" containerName="collect-profiles" Oct 08 19:01:00 crc kubenswrapper[4859]: I1008 19:01:00.152454 4859 state_mem.go:107] "Deleted CPUSet assignment" podUID="37fd514d-24a1-4204-9125-578bb9c0779e" containerName="collect-profiles" Oct 08 19:01:00 crc kubenswrapper[4859]: I1008 19:01:00.152656 4859 memory_manager.go:354] "RemoveStaleState removing state" podUID="37fd514d-24a1-4204-9125-578bb9c0779e" containerName="collect-profiles" Oct 08 19:01:00 crc kubenswrapper[4859]: I1008 19:01:00.154441 4859 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-cron-29332501-rxhgw" Oct 08 19:01:00 crc kubenswrapper[4859]: I1008 19:01:00.168077 4859 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-cron-29332501-rxhgw"] Oct 08 19:01:00 crc kubenswrapper[4859]: I1008 19:01:00.183547 4859 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-sg6sh" Oct 08 19:01:00 crc kubenswrapper[4859]: I1008 19:01:00.192093 4859 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/555ffac5-9b03-43eb-8610-fad69a164a42-combined-ca-bundle\") pod \"keystone-cron-29332501-rxhgw\" (UID: \"555ffac5-9b03-43eb-8610-fad69a164a42\") " pod="openstack/keystone-cron-29332501-rxhgw" Oct 08 19:01:00 crc kubenswrapper[4859]: I1008 19:01:00.192147 4859 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/555ffac5-9b03-43eb-8610-fad69a164a42-config-data\") pod \"keystone-cron-29332501-rxhgw\" (UID: \"555ffac5-9b03-43eb-8610-fad69a164a42\") " pod="openstack/keystone-cron-29332501-rxhgw" Oct 08 19:01:00 crc kubenswrapper[4859]: I1008 19:01:00.192426 4859 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-ngz8v\" (UniqueName: \"kubernetes.io/projected/555ffac5-9b03-43eb-8610-fad69a164a42-kube-api-access-ngz8v\") pod \"keystone-cron-29332501-rxhgw\" (UID: \"555ffac5-9b03-43eb-8610-fad69a164a42\") " pod="openstack/keystone-cron-29332501-rxhgw" Oct 08 19:01:00 crc kubenswrapper[4859]: I1008 19:01:00.192697 4859 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/555ffac5-9b03-43eb-8610-fad69a164a42-fernet-keys\") pod \"keystone-cron-29332501-rxhgw\" (UID: \"555ffac5-9b03-43eb-8610-fad69a164a42\") " pod="openstack/keystone-cron-29332501-rxhgw" Oct 08 19:01:00 crc kubenswrapper[4859]: I1008 19:01:00.294233 4859 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"nova-cell1-compute-config-1\" (UniqueName: \"kubernetes.io/secret/4035f4f1-5231-4112-89f6-d9d1c7c7eeb1-nova-cell1-compute-config-1\") pod \"4035f4f1-5231-4112-89f6-d9d1c7c7eeb1\" (UID: \"4035f4f1-5231-4112-89f6-d9d1c7c7eeb1\") " Oct 08 19:01:00 crc kubenswrapper[4859]: I1008 19:01:00.294714 4859 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"nova-cell1-compute-config-0\" (UniqueName: \"kubernetes.io/secret/4035f4f1-5231-4112-89f6-d9d1c7c7eeb1-nova-cell1-compute-config-0\") pod \"4035f4f1-5231-4112-89f6-d9d1c7c7eeb1\" (UID: \"4035f4f1-5231-4112-89f6-d9d1c7c7eeb1\") " Oct 08 19:01:00 crc kubenswrapper[4859]: I1008 19:01:00.294761 4859 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/4035f4f1-5231-4112-89f6-d9d1c7c7eeb1-inventory\") pod \"4035f4f1-5231-4112-89f6-d9d1c7c7eeb1\" (UID: \"4035f4f1-5231-4112-89f6-d9d1c7c7eeb1\") " Oct 08 19:01:00 crc kubenswrapper[4859]: I1008 19:01:00.294809 4859 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/4035f4f1-5231-4112-89f6-d9d1c7c7eeb1-ssh-key\") pod \"4035f4f1-5231-4112-89f6-d9d1c7c7eeb1\" (UID: \"4035f4f1-5231-4112-89f6-d9d1c7c7eeb1\") " Oct 08 19:01:00 crc kubenswrapper[4859]: I1008 19:01:00.294849 4859 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"nova-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4035f4f1-5231-4112-89f6-d9d1c7c7eeb1-nova-combined-ca-bundle\") pod \"4035f4f1-5231-4112-89f6-d9d1c7c7eeb1\" (UID: \"4035f4f1-5231-4112-89f6-d9d1c7c7eeb1\") " Oct 08 19:01:00 crc kubenswrapper[4859]: I1008 19:01:00.294885 4859 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"nova-migration-ssh-key-1\" (UniqueName: \"kubernetes.io/secret/4035f4f1-5231-4112-89f6-d9d1c7c7eeb1-nova-migration-ssh-key-1\") pod \"4035f4f1-5231-4112-89f6-d9d1c7c7eeb1\" (UID: \"4035f4f1-5231-4112-89f6-d9d1c7c7eeb1\") " Oct 08 19:01:00 crc kubenswrapper[4859]: I1008 19:01:00.294945 4859 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"nova-migration-ssh-key-0\" (UniqueName: \"kubernetes.io/secret/4035f4f1-5231-4112-89f6-d9d1c7c7eeb1-nova-migration-ssh-key-0\") pod \"4035f4f1-5231-4112-89f6-d9d1c7c7eeb1\" (UID: \"4035f4f1-5231-4112-89f6-d9d1c7c7eeb1\") " Oct 08 19:01:00 crc kubenswrapper[4859]: I1008 19:01:00.294974 4859 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-bgz5p\" (UniqueName: \"kubernetes.io/projected/4035f4f1-5231-4112-89f6-d9d1c7c7eeb1-kube-api-access-bgz5p\") pod \"4035f4f1-5231-4112-89f6-d9d1c7c7eeb1\" (UID: \"4035f4f1-5231-4112-89f6-d9d1c7c7eeb1\") " Oct 08 19:01:00 crc kubenswrapper[4859]: I1008 19:01:00.295048 4859 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"nova-extra-config-0\" (UniqueName: \"kubernetes.io/configmap/4035f4f1-5231-4112-89f6-d9d1c7c7eeb1-nova-extra-config-0\") pod \"4035f4f1-5231-4112-89f6-d9d1c7c7eeb1\" (UID: \"4035f4f1-5231-4112-89f6-d9d1c7c7eeb1\") " Oct 08 19:01:00 crc kubenswrapper[4859]: I1008 19:01:00.295338 4859 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-ngz8v\" (UniqueName: \"kubernetes.io/projected/555ffac5-9b03-43eb-8610-fad69a164a42-kube-api-access-ngz8v\") pod \"keystone-cron-29332501-rxhgw\" (UID: \"555ffac5-9b03-43eb-8610-fad69a164a42\") " pod="openstack/keystone-cron-29332501-rxhgw" Oct 08 19:01:00 crc kubenswrapper[4859]: I1008 19:01:00.295435 4859 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/555ffac5-9b03-43eb-8610-fad69a164a42-fernet-keys\") pod \"keystone-cron-29332501-rxhgw\" (UID: \"555ffac5-9b03-43eb-8610-fad69a164a42\") " pod="openstack/keystone-cron-29332501-rxhgw" Oct 08 19:01:00 crc kubenswrapper[4859]: I1008 19:01:00.296232 4859 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/555ffac5-9b03-43eb-8610-fad69a164a42-combined-ca-bundle\") pod \"keystone-cron-29332501-rxhgw\" (UID: \"555ffac5-9b03-43eb-8610-fad69a164a42\") " pod="openstack/keystone-cron-29332501-rxhgw" Oct 08 19:01:00 crc kubenswrapper[4859]: I1008 19:01:00.296314 4859 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/555ffac5-9b03-43eb-8610-fad69a164a42-config-data\") pod \"keystone-cron-29332501-rxhgw\" (UID: \"555ffac5-9b03-43eb-8610-fad69a164a42\") " pod="openstack/keystone-cron-29332501-rxhgw" Oct 08 19:01:00 crc kubenswrapper[4859]: I1008 19:01:00.301309 4859 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/4035f4f1-5231-4112-89f6-d9d1c7c7eeb1-nova-combined-ca-bundle" (OuterVolumeSpecName: "nova-combined-ca-bundle") pod "4035f4f1-5231-4112-89f6-d9d1c7c7eeb1" (UID: "4035f4f1-5231-4112-89f6-d9d1c7c7eeb1"). InnerVolumeSpecName "nova-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 08 19:01:00 crc kubenswrapper[4859]: I1008 19:01:00.301461 4859 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/555ffac5-9b03-43eb-8610-fad69a164a42-fernet-keys\") pod \"keystone-cron-29332501-rxhgw\" (UID: \"555ffac5-9b03-43eb-8610-fad69a164a42\") " pod="openstack/keystone-cron-29332501-rxhgw" Oct 08 19:01:00 crc kubenswrapper[4859]: I1008 19:01:00.302297 4859 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/555ffac5-9b03-43eb-8610-fad69a164a42-combined-ca-bundle\") pod \"keystone-cron-29332501-rxhgw\" (UID: \"555ffac5-9b03-43eb-8610-fad69a164a42\") " pod="openstack/keystone-cron-29332501-rxhgw" Oct 08 19:01:00 crc kubenswrapper[4859]: I1008 19:01:00.302724 4859 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/555ffac5-9b03-43eb-8610-fad69a164a42-config-data\") pod \"keystone-cron-29332501-rxhgw\" (UID: \"555ffac5-9b03-43eb-8610-fad69a164a42\") " pod="openstack/keystone-cron-29332501-rxhgw" Oct 08 19:01:00 crc kubenswrapper[4859]: I1008 19:01:00.313913 4859 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/4035f4f1-5231-4112-89f6-d9d1c7c7eeb1-kube-api-access-bgz5p" (OuterVolumeSpecName: "kube-api-access-bgz5p") pod "4035f4f1-5231-4112-89f6-d9d1c7c7eeb1" (UID: "4035f4f1-5231-4112-89f6-d9d1c7c7eeb1"). InnerVolumeSpecName "kube-api-access-bgz5p". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 08 19:01:00 crc kubenswrapper[4859]: I1008 19:01:00.314450 4859 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-ngz8v\" (UniqueName: \"kubernetes.io/projected/555ffac5-9b03-43eb-8610-fad69a164a42-kube-api-access-ngz8v\") pod \"keystone-cron-29332501-rxhgw\" (UID: \"555ffac5-9b03-43eb-8610-fad69a164a42\") " pod="openstack/keystone-cron-29332501-rxhgw" Oct 08 19:01:00 crc kubenswrapper[4859]: I1008 19:01:00.323919 4859 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/4035f4f1-5231-4112-89f6-d9d1c7c7eeb1-nova-cell1-compute-config-0" (OuterVolumeSpecName: "nova-cell1-compute-config-0") pod "4035f4f1-5231-4112-89f6-d9d1c7c7eeb1" (UID: "4035f4f1-5231-4112-89f6-d9d1c7c7eeb1"). InnerVolumeSpecName "nova-cell1-compute-config-0". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 08 19:01:00 crc kubenswrapper[4859]: I1008 19:01:00.326210 4859 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/4035f4f1-5231-4112-89f6-d9d1c7c7eeb1-nova-migration-ssh-key-1" (OuterVolumeSpecName: "nova-migration-ssh-key-1") pod "4035f4f1-5231-4112-89f6-d9d1c7c7eeb1" (UID: "4035f4f1-5231-4112-89f6-d9d1c7c7eeb1"). InnerVolumeSpecName "nova-migration-ssh-key-1". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 08 19:01:00 crc kubenswrapper[4859]: I1008 19:01:00.326819 4859 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/4035f4f1-5231-4112-89f6-d9d1c7c7eeb1-inventory" (OuterVolumeSpecName: "inventory") pod "4035f4f1-5231-4112-89f6-d9d1c7c7eeb1" (UID: "4035f4f1-5231-4112-89f6-d9d1c7c7eeb1"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 08 19:01:00 crc kubenswrapper[4859]: I1008 19:01:00.327901 4859 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/4035f4f1-5231-4112-89f6-d9d1c7c7eeb1-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "4035f4f1-5231-4112-89f6-d9d1c7c7eeb1" (UID: "4035f4f1-5231-4112-89f6-d9d1c7c7eeb1"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 08 19:01:00 crc kubenswrapper[4859]: I1008 19:01:00.328245 4859 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/4035f4f1-5231-4112-89f6-d9d1c7c7eeb1-nova-migration-ssh-key-0" (OuterVolumeSpecName: "nova-migration-ssh-key-0") pod "4035f4f1-5231-4112-89f6-d9d1c7c7eeb1" (UID: "4035f4f1-5231-4112-89f6-d9d1c7c7eeb1"). InnerVolumeSpecName "nova-migration-ssh-key-0". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 08 19:01:00 crc kubenswrapper[4859]: I1008 19:01:00.332626 4859 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/4035f4f1-5231-4112-89f6-d9d1c7c7eeb1-nova-cell1-compute-config-1" (OuterVolumeSpecName: "nova-cell1-compute-config-1") pod "4035f4f1-5231-4112-89f6-d9d1c7c7eeb1" (UID: "4035f4f1-5231-4112-89f6-d9d1c7c7eeb1"). InnerVolumeSpecName "nova-cell1-compute-config-1". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 08 19:01:00 crc kubenswrapper[4859]: I1008 19:01:00.334553 4859 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/4035f4f1-5231-4112-89f6-d9d1c7c7eeb1-nova-extra-config-0" (OuterVolumeSpecName: "nova-extra-config-0") pod "4035f4f1-5231-4112-89f6-d9d1c7c7eeb1" (UID: "4035f4f1-5231-4112-89f6-d9d1c7c7eeb1"). InnerVolumeSpecName "nova-extra-config-0". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 08 19:01:00 crc kubenswrapper[4859]: I1008 19:01:00.398665 4859 reconciler_common.go:293] "Volume detached for volume \"nova-migration-ssh-key-1\" (UniqueName: \"kubernetes.io/secret/4035f4f1-5231-4112-89f6-d9d1c7c7eeb1-nova-migration-ssh-key-1\") on node \"crc\" DevicePath \"\"" Oct 08 19:01:00 crc kubenswrapper[4859]: I1008 19:01:00.398718 4859 reconciler_common.go:293] "Volume detached for volume \"nova-migration-ssh-key-0\" (UniqueName: \"kubernetes.io/secret/4035f4f1-5231-4112-89f6-d9d1c7c7eeb1-nova-migration-ssh-key-0\") on node \"crc\" DevicePath \"\"" Oct 08 19:01:00 crc kubenswrapper[4859]: I1008 19:01:00.398736 4859 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-bgz5p\" (UniqueName: \"kubernetes.io/projected/4035f4f1-5231-4112-89f6-d9d1c7c7eeb1-kube-api-access-bgz5p\") on node \"crc\" DevicePath \"\"" Oct 08 19:01:00 crc kubenswrapper[4859]: I1008 19:01:00.398750 4859 reconciler_common.go:293] "Volume detached for volume \"nova-extra-config-0\" (UniqueName: \"kubernetes.io/configmap/4035f4f1-5231-4112-89f6-d9d1c7c7eeb1-nova-extra-config-0\") on node \"crc\" DevicePath \"\"" Oct 08 19:01:00 crc kubenswrapper[4859]: I1008 19:01:00.398761 4859 reconciler_common.go:293] "Volume detached for volume \"nova-cell1-compute-config-1\" (UniqueName: \"kubernetes.io/secret/4035f4f1-5231-4112-89f6-d9d1c7c7eeb1-nova-cell1-compute-config-1\") on node \"crc\" DevicePath \"\"" Oct 08 19:01:00 crc kubenswrapper[4859]: I1008 19:01:00.398770 4859 reconciler_common.go:293] "Volume detached for volume \"nova-cell1-compute-config-0\" (UniqueName: \"kubernetes.io/secret/4035f4f1-5231-4112-89f6-d9d1c7c7eeb1-nova-cell1-compute-config-0\") on node \"crc\" DevicePath \"\"" Oct 08 19:01:00 crc kubenswrapper[4859]: I1008 19:01:00.398779 4859 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/4035f4f1-5231-4112-89f6-d9d1c7c7eeb1-inventory\") on node \"crc\" DevicePath \"\"" Oct 08 19:01:00 crc kubenswrapper[4859]: I1008 19:01:00.398787 4859 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/4035f4f1-5231-4112-89f6-d9d1c7c7eeb1-ssh-key\") on node \"crc\" DevicePath \"\"" Oct 08 19:01:00 crc kubenswrapper[4859]: I1008 19:01:00.398797 4859 reconciler_common.go:293] "Volume detached for volume \"nova-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4035f4f1-5231-4112-89f6-d9d1c7c7eeb1-nova-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 08 19:01:00 crc kubenswrapper[4859]: I1008 19:01:00.495609 4859 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-cron-29332501-rxhgw" Oct 08 19:01:00 crc kubenswrapper[4859]: I1008 19:01:00.718393 4859 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-sg6sh" event={"ID":"4035f4f1-5231-4112-89f6-d9d1c7c7eeb1","Type":"ContainerDied","Data":"bb11a7df0b2ad4ddfa91ba90319c698fb38f6d8c8d3efc114f2abf6a970cbae7"} Oct 08 19:01:00 crc kubenswrapper[4859]: I1008 19:01:00.718613 4859 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="bb11a7df0b2ad4ddfa91ba90319c698fb38f6d8c8d3efc114f2abf6a970cbae7" Oct 08 19:01:00 crc kubenswrapper[4859]: I1008 19:01:00.718486 4859 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-sg6sh" Oct 08 19:01:00 crc kubenswrapper[4859]: I1008 19:01:00.806775 4859 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/telemetry-edpm-deployment-openstack-edpm-ipam-twhvj"] Oct 08 19:01:00 crc kubenswrapper[4859]: E1008 19:01:00.807263 4859 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="4035f4f1-5231-4112-89f6-d9d1c7c7eeb1" containerName="nova-edpm-deployment-openstack-edpm-ipam" Oct 08 19:01:00 crc kubenswrapper[4859]: I1008 19:01:00.807287 4859 state_mem.go:107] "Deleted CPUSet assignment" podUID="4035f4f1-5231-4112-89f6-d9d1c7c7eeb1" containerName="nova-edpm-deployment-openstack-edpm-ipam" Oct 08 19:01:00 crc kubenswrapper[4859]: I1008 19:01:00.807523 4859 memory_manager.go:354] "RemoveStaleState removing state" podUID="4035f4f1-5231-4112-89f6-d9d1c7c7eeb1" containerName="nova-edpm-deployment-openstack-edpm-ipam" Oct 08 19:01:00 crc kubenswrapper[4859]: I1008 19:01:00.808345 4859 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-twhvj" Oct 08 19:01:00 crc kubenswrapper[4859]: I1008 19:01:00.808937 4859 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/telemetry-edpm-deployment-openstack-edpm-ipam-twhvj"] Oct 08 19:01:00 crc kubenswrapper[4859]: I1008 19:01:00.811074 4859 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-gxqst" Oct 08 19:01:00 crc kubenswrapper[4859]: I1008 19:01:00.811268 4859 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Oct 08 19:01:00 crc kubenswrapper[4859]: I1008 19:01:00.811437 4859 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-compute-config-data" Oct 08 19:01:00 crc kubenswrapper[4859]: I1008 19:01:00.811548 4859 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Oct 08 19:01:00 crc kubenswrapper[4859]: I1008 19:01:00.811596 4859 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Oct 08 19:01:00 crc kubenswrapper[4859]: I1008 19:01:00.907292 4859 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/65244148-7971-47bf-b900-96932c6cd473-ssh-key\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-twhvj\" (UID: \"65244148-7971-47bf-b900-96932c6cd473\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-twhvj" Oct 08 19:01:00 crc kubenswrapper[4859]: I1008 19:01:00.907360 4859 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceilometer-compute-config-data-0\" (UniqueName: \"kubernetes.io/secret/65244148-7971-47bf-b900-96932c6cd473-ceilometer-compute-config-data-0\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-twhvj\" (UID: \"65244148-7971-47bf-b900-96932c6cd473\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-twhvj" Oct 08 19:01:00 crc kubenswrapper[4859]: I1008 19:01:00.907410 4859 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceilometer-compute-config-data-2\" (UniqueName: \"kubernetes.io/secret/65244148-7971-47bf-b900-96932c6cd473-ceilometer-compute-config-data-2\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-twhvj\" (UID: \"65244148-7971-47bf-b900-96932c6cd473\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-twhvj" Oct 08 19:01:00 crc kubenswrapper[4859]: I1008 19:01:00.907430 4859 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/65244148-7971-47bf-b900-96932c6cd473-inventory\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-twhvj\" (UID: \"65244148-7971-47bf-b900-96932c6cd473\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-twhvj" Oct 08 19:01:00 crc kubenswrapper[4859]: I1008 19:01:00.907516 4859 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-8ms4f\" (UniqueName: \"kubernetes.io/projected/65244148-7971-47bf-b900-96932c6cd473-kube-api-access-8ms4f\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-twhvj\" (UID: \"65244148-7971-47bf-b900-96932c6cd473\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-twhvj" Oct 08 19:01:00 crc kubenswrapper[4859]: I1008 19:01:00.907905 4859 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"telemetry-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/65244148-7971-47bf-b900-96932c6cd473-telemetry-combined-ca-bundle\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-twhvj\" (UID: \"65244148-7971-47bf-b900-96932c6cd473\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-twhvj" Oct 08 19:01:00 crc kubenswrapper[4859]: I1008 19:01:00.907947 4859 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceilometer-compute-config-data-1\" (UniqueName: \"kubernetes.io/secret/65244148-7971-47bf-b900-96932c6cd473-ceilometer-compute-config-data-1\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-twhvj\" (UID: \"65244148-7971-47bf-b900-96932c6cd473\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-twhvj" Oct 08 19:01:00 crc kubenswrapper[4859]: I1008 19:01:00.934580 4859 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-cron-29332501-rxhgw"] Oct 08 19:01:01 crc kubenswrapper[4859]: I1008 19:01:01.009288 4859 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceilometer-compute-config-data-2\" (UniqueName: \"kubernetes.io/secret/65244148-7971-47bf-b900-96932c6cd473-ceilometer-compute-config-data-2\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-twhvj\" (UID: \"65244148-7971-47bf-b900-96932c6cd473\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-twhvj" Oct 08 19:01:01 crc kubenswrapper[4859]: I1008 19:01:01.009350 4859 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/65244148-7971-47bf-b900-96932c6cd473-inventory\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-twhvj\" (UID: \"65244148-7971-47bf-b900-96932c6cd473\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-twhvj" Oct 08 19:01:01 crc kubenswrapper[4859]: I1008 19:01:01.009386 4859 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-8ms4f\" (UniqueName: \"kubernetes.io/projected/65244148-7971-47bf-b900-96932c6cd473-kube-api-access-8ms4f\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-twhvj\" (UID: \"65244148-7971-47bf-b900-96932c6cd473\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-twhvj" Oct 08 19:01:01 crc kubenswrapper[4859]: I1008 19:01:01.009522 4859 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"telemetry-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/65244148-7971-47bf-b900-96932c6cd473-telemetry-combined-ca-bundle\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-twhvj\" (UID: \"65244148-7971-47bf-b900-96932c6cd473\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-twhvj" Oct 08 19:01:01 crc kubenswrapper[4859]: I1008 19:01:01.009554 4859 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceilometer-compute-config-data-1\" (UniqueName: \"kubernetes.io/secret/65244148-7971-47bf-b900-96932c6cd473-ceilometer-compute-config-data-1\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-twhvj\" (UID: \"65244148-7971-47bf-b900-96932c6cd473\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-twhvj" Oct 08 19:01:01 crc kubenswrapper[4859]: I1008 19:01:01.009610 4859 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/65244148-7971-47bf-b900-96932c6cd473-ssh-key\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-twhvj\" (UID: \"65244148-7971-47bf-b900-96932c6cd473\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-twhvj" Oct 08 19:01:01 crc kubenswrapper[4859]: I1008 19:01:01.009654 4859 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceilometer-compute-config-data-0\" (UniqueName: \"kubernetes.io/secret/65244148-7971-47bf-b900-96932c6cd473-ceilometer-compute-config-data-0\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-twhvj\" (UID: \"65244148-7971-47bf-b900-96932c6cd473\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-twhvj" Oct 08 19:01:01 crc kubenswrapper[4859]: I1008 19:01:01.017640 4859 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceilometer-compute-config-data-0\" (UniqueName: \"kubernetes.io/secret/65244148-7971-47bf-b900-96932c6cd473-ceilometer-compute-config-data-0\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-twhvj\" (UID: \"65244148-7971-47bf-b900-96932c6cd473\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-twhvj" Oct 08 19:01:01 crc kubenswrapper[4859]: I1008 19:01:01.018113 4859 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/65244148-7971-47bf-b900-96932c6cd473-ssh-key\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-twhvj\" (UID: \"65244148-7971-47bf-b900-96932c6cd473\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-twhvj" Oct 08 19:01:01 crc kubenswrapper[4859]: I1008 19:01:01.019170 4859 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceilometer-compute-config-data-2\" (UniqueName: \"kubernetes.io/secret/65244148-7971-47bf-b900-96932c6cd473-ceilometer-compute-config-data-2\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-twhvj\" (UID: \"65244148-7971-47bf-b900-96932c6cd473\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-twhvj" Oct 08 19:01:01 crc kubenswrapper[4859]: I1008 19:01:01.023408 4859 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceilometer-compute-config-data-1\" (UniqueName: \"kubernetes.io/secret/65244148-7971-47bf-b900-96932c6cd473-ceilometer-compute-config-data-1\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-twhvj\" (UID: \"65244148-7971-47bf-b900-96932c6cd473\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-twhvj" Oct 08 19:01:01 crc kubenswrapper[4859]: I1008 19:01:01.023982 4859 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/65244148-7971-47bf-b900-96932c6cd473-inventory\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-twhvj\" (UID: \"65244148-7971-47bf-b900-96932c6cd473\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-twhvj" Oct 08 19:01:01 crc kubenswrapper[4859]: I1008 19:01:01.026798 4859 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"telemetry-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/65244148-7971-47bf-b900-96932c6cd473-telemetry-combined-ca-bundle\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-twhvj\" (UID: \"65244148-7971-47bf-b900-96932c6cd473\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-twhvj" Oct 08 19:01:01 crc kubenswrapper[4859]: I1008 19:01:01.069821 4859 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-8ms4f\" (UniqueName: \"kubernetes.io/projected/65244148-7971-47bf-b900-96932c6cd473-kube-api-access-8ms4f\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-twhvj\" (UID: \"65244148-7971-47bf-b900-96932c6cd473\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-twhvj" Oct 08 19:01:01 crc kubenswrapper[4859]: I1008 19:01:01.129282 4859 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-twhvj" Oct 08 19:01:01 crc kubenswrapper[4859]: I1008 19:01:01.636883 4859 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/telemetry-edpm-deployment-openstack-edpm-ipam-twhvj"] Oct 08 19:01:01 crc kubenswrapper[4859]: W1008 19:01:01.637552 4859 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod65244148_7971_47bf_b900_96932c6cd473.slice/crio-a54589f2121dfea6dd9ba650fa898645fad3cd095578c241c4a6e1e769db4e40 WatchSource:0}: Error finding container a54589f2121dfea6dd9ba650fa898645fad3cd095578c241c4a6e1e769db4e40: Status 404 returned error can't find the container with id a54589f2121dfea6dd9ba650fa898645fad3cd095578c241c4a6e1e769db4e40 Oct 08 19:01:01 crc kubenswrapper[4859]: I1008 19:01:01.640801 4859 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Oct 08 19:01:01 crc kubenswrapper[4859]: I1008 19:01:01.728419 4859 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-cron-29332501-rxhgw" event={"ID":"555ffac5-9b03-43eb-8610-fad69a164a42","Type":"ContainerStarted","Data":"b69813d06dd357b4470c1841f8c4f257140d2b3fd301a8a9fbbbc348f6715d77"} Oct 08 19:01:01 crc kubenswrapper[4859]: I1008 19:01:01.728466 4859 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-cron-29332501-rxhgw" event={"ID":"555ffac5-9b03-43eb-8610-fad69a164a42","Type":"ContainerStarted","Data":"4a0f8d433b67b1881fb1e5954f59fd509653e35091276fe6a078838c707a09d5"} Oct 08 19:01:01 crc kubenswrapper[4859]: I1008 19:01:01.730316 4859 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-twhvj" event={"ID":"65244148-7971-47bf-b900-96932c6cd473","Type":"ContainerStarted","Data":"a54589f2121dfea6dd9ba650fa898645fad3cd095578c241c4a6e1e769db4e40"} Oct 08 19:01:01 crc kubenswrapper[4859]: I1008 19:01:01.744169 4859 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/keystone-cron-29332501-rxhgw" podStartSLOduration=1.7441527639999999 podStartE2EDuration="1.744152764s" podCreationTimestamp="2025-10-08 19:01:00 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-08 19:01:01.74125523 +0000 UTC m=+2631.988094609" watchObservedRunningTime="2025-10-08 19:01:01.744152764 +0000 UTC m=+2631.990992133" Oct 08 19:01:02 crc kubenswrapper[4859]: I1008 19:01:02.742805 4859 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-twhvj" event={"ID":"65244148-7971-47bf-b900-96932c6cd473","Type":"ContainerStarted","Data":"52d2b1eeb68202f40996e1ab631e4a82637a852b77dd732d27c483a364a58b94"} Oct 08 19:01:02 crc kubenswrapper[4859]: I1008 19:01:02.762705 4859 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-twhvj" podStartSLOduration=2.205862801 podStartE2EDuration="2.762674034s" podCreationTimestamp="2025-10-08 19:01:00 +0000 UTC" firstStartedPulling="2025-10-08 19:01:01.64049574 +0000 UTC m=+2631.887335129" lastFinishedPulling="2025-10-08 19:01:02.197306983 +0000 UTC m=+2632.444146362" observedRunningTime="2025-10-08 19:01:02.761973264 +0000 UTC m=+2633.008812643" watchObservedRunningTime="2025-10-08 19:01:02.762674034 +0000 UTC m=+2633.009513413" Oct 08 19:01:03 crc kubenswrapper[4859]: I1008 19:01:03.752906 4859 generic.go:334] "Generic (PLEG): container finished" podID="555ffac5-9b03-43eb-8610-fad69a164a42" containerID="b69813d06dd357b4470c1841f8c4f257140d2b3fd301a8a9fbbbc348f6715d77" exitCode=0 Oct 08 19:01:03 crc kubenswrapper[4859]: I1008 19:01:03.753000 4859 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-cron-29332501-rxhgw" event={"ID":"555ffac5-9b03-43eb-8610-fad69a164a42","Type":"ContainerDied","Data":"b69813d06dd357b4470c1841f8c4f257140d2b3fd301a8a9fbbbc348f6715d77"} Oct 08 19:01:05 crc kubenswrapper[4859]: I1008 19:01:05.183946 4859 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-cron-29332501-rxhgw" Oct 08 19:01:05 crc kubenswrapper[4859]: I1008 19:01:05.300467 4859 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/555ffac5-9b03-43eb-8610-fad69a164a42-fernet-keys\") pod \"555ffac5-9b03-43eb-8610-fad69a164a42\" (UID: \"555ffac5-9b03-43eb-8610-fad69a164a42\") " Oct 08 19:01:05 crc kubenswrapper[4859]: I1008 19:01:05.300646 4859 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/555ffac5-9b03-43eb-8610-fad69a164a42-config-data\") pod \"555ffac5-9b03-43eb-8610-fad69a164a42\" (UID: \"555ffac5-9b03-43eb-8610-fad69a164a42\") " Oct 08 19:01:05 crc kubenswrapper[4859]: I1008 19:01:05.300724 4859 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-ngz8v\" (UniqueName: \"kubernetes.io/projected/555ffac5-9b03-43eb-8610-fad69a164a42-kube-api-access-ngz8v\") pod \"555ffac5-9b03-43eb-8610-fad69a164a42\" (UID: \"555ffac5-9b03-43eb-8610-fad69a164a42\") " Oct 08 19:01:05 crc kubenswrapper[4859]: I1008 19:01:05.300752 4859 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/555ffac5-9b03-43eb-8610-fad69a164a42-combined-ca-bundle\") pod \"555ffac5-9b03-43eb-8610-fad69a164a42\" (UID: \"555ffac5-9b03-43eb-8610-fad69a164a42\") " Oct 08 19:01:05 crc kubenswrapper[4859]: I1008 19:01:05.307141 4859 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/555ffac5-9b03-43eb-8610-fad69a164a42-kube-api-access-ngz8v" (OuterVolumeSpecName: "kube-api-access-ngz8v") pod "555ffac5-9b03-43eb-8610-fad69a164a42" (UID: "555ffac5-9b03-43eb-8610-fad69a164a42"). InnerVolumeSpecName "kube-api-access-ngz8v". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 08 19:01:05 crc kubenswrapper[4859]: I1008 19:01:05.307161 4859 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/555ffac5-9b03-43eb-8610-fad69a164a42-fernet-keys" (OuterVolumeSpecName: "fernet-keys") pod "555ffac5-9b03-43eb-8610-fad69a164a42" (UID: "555ffac5-9b03-43eb-8610-fad69a164a42"). InnerVolumeSpecName "fernet-keys". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 08 19:01:05 crc kubenswrapper[4859]: I1008 19:01:05.334120 4859 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/555ffac5-9b03-43eb-8610-fad69a164a42-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "555ffac5-9b03-43eb-8610-fad69a164a42" (UID: "555ffac5-9b03-43eb-8610-fad69a164a42"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 08 19:01:05 crc kubenswrapper[4859]: I1008 19:01:05.361818 4859 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/555ffac5-9b03-43eb-8610-fad69a164a42-config-data" (OuterVolumeSpecName: "config-data") pod "555ffac5-9b03-43eb-8610-fad69a164a42" (UID: "555ffac5-9b03-43eb-8610-fad69a164a42"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 08 19:01:05 crc kubenswrapper[4859]: I1008 19:01:05.404078 4859 reconciler_common.go:293] "Volume detached for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/555ffac5-9b03-43eb-8610-fad69a164a42-fernet-keys\") on node \"crc\" DevicePath \"\"" Oct 08 19:01:05 crc kubenswrapper[4859]: I1008 19:01:05.404139 4859 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/555ffac5-9b03-43eb-8610-fad69a164a42-config-data\") on node \"crc\" DevicePath \"\"" Oct 08 19:01:05 crc kubenswrapper[4859]: I1008 19:01:05.404153 4859 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-ngz8v\" (UniqueName: \"kubernetes.io/projected/555ffac5-9b03-43eb-8610-fad69a164a42-kube-api-access-ngz8v\") on node \"crc\" DevicePath \"\"" Oct 08 19:01:05 crc kubenswrapper[4859]: I1008 19:01:05.404169 4859 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/555ffac5-9b03-43eb-8610-fad69a164a42-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 08 19:01:05 crc kubenswrapper[4859]: I1008 19:01:05.768541 4859 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-cron-29332501-rxhgw" event={"ID":"555ffac5-9b03-43eb-8610-fad69a164a42","Type":"ContainerDied","Data":"4a0f8d433b67b1881fb1e5954f59fd509653e35091276fe6a078838c707a09d5"} Oct 08 19:01:05 crc kubenswrapper[4859]: I1008 19:01:05.768588 4859 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="4a0f8d433b67b1881fb1e5954f59fd509653e35091276fe6a078838c707a09d5" Oct 08 19:01:05 crc kubenswrapper[4859]: I1008 19:01:05.768653 4859 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-cron-29332501-rxhgw" Oct 08 19:01:09 crc kubenswrapper[4859]: I1008 19:01:09.470289 4859 scope.go:117] "RemoveContainer" containerID="265e61e43a0f4edee3adcaf87995c3c44fad0b135f61f14030c7ee297b24d748" Oct 08 19:01:09 crc kubenswrapper[4859]: E1008 19:01:09.470916 4859 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-82s52_openshift-machine-config-operator(b23a6a6c-9d92-4e7b-840e-55cfda873a2d)\"" pod="openshift-machine-config-operator/machine-config-daemon-82s52" podUID="b23a6a6c-9d92-4e7b-840e-55cfda873a2d" Oct 08 19:01:24 crc kubenswrapper[4859]: I1008 19:01:24.469982 4859 scope.go:117] "RemoveContainer" containerID="265e61e43a0f4edee3adcaf87995c3c44fad0b135f61f14030c7ee297b24d748" Oct 08 19:01:24 crc kubenswrapper[4859]: E1008 19:01:24.470753 4859 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-82s52_openshift-machine-config-operator(b23a6a6c-9d92-4e7b-840e-55cfda873a2d)\"" pod="openshift-machine-config-operator/machine-config-daemon-82s52" podUID="b23a6a6c-9d92-4e7b-840e-55cfda873a2d" Oct 08 19:01:37 crc kubenswrapper[4859]: I1008 19:01:37.323326 4859 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-24svc"] Oct 08 19:01:37 crc kubenswrapper[4859]: E1008 19:01:37.324303 4859 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="555ffac5-9b03-43eb-8610-fad69a164a42" containerName="keystone-cron" Oct 08 19:01:37 crc kubenswrapper[4859]: I1008 19:01:37.324318 4859 state_mem.go:107] "Deleted CPUSet assignment" podUID="555ffac5-9b03-43eb-8610-fad69a164a42" containerName="keystone-cron" Oct 08 19:01:37 crc kubenswrapper[4859]: I1008 19:01:37.324597 4859 memory_manager.go:354] "RemoveStaleState removing state" podUID="555ffac5-9b03-43eb-8610-fad69a164a42" containerName="keystone-cron" Oct 08 19:01:37 crc kubenswrapper[4859]: I1008 19:01:37.326332 4859 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-24svc" Oct 08 19:01:37 crc kubenswrapper[4859]: I1008 19:01:37.336340 4859 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-24svc"] Oct 08 19:01:37 crc kubenswrapper[4859]: I1008 19:01:37.399664 4859 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-vq78x\" (UniqueName: \"kubernetes.io/projected/997fd032-41f3-4d25-a8fb-c666c5891748-kube-api-access-vq78x\") pod \"community-operators-24svc\" (UID: \"997fd032-41f3-4d25-a8fb-c666c5891748\") " pod="openshift-marketplace/community-operators-24svc" Oct 08 19:01:37 crc kubenswrapper[4859]: I1008 19:01:37.399801 4859 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/997fd032-41f3-4d25-a8fb-c666c5891748-catalog-content\") pod \"community-operators-24svc\" (UID: \"997fd032-41f3-4d25-a8fb-c666c5891748\") " pod="openshift-marketplace/community-operators-24svc" Oct 08 19:01:37 crc kubenswrapper[4859]: I1008 19:01:37.399826 4859 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/997fd032-41f3-4d25-a8fb-c666c5891748-utilities\") pod \"community-operators-24svc\" (UID: \"997fd032-41f3-4d25-a8fb-c666c5891748\") " pod="openshift-marketplace/community-operators-24svc" Oct 08 19:01:37 crc kubenswrapper[4859]: I1008 19:01:37.501578 4859 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-vq78x\" (UniqueName: \"kubernetes.io/projected/997fd032-41f3-4d25-a8fb-c666c5891748-kube-api-access-vq78x\") pod \"community-operators-24svc\" (UID: \"997fd032-41f3-4d25-a8fb-c666c5891748\") " pod="openshift-marketplace/community-operators-24svc" Oct 08 19:01:37 crc kubenswrapper[4859]: I1008 19:01:37.501658 4859 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/997fd032-41f3-4d25-a8fb-c666c5891748-catalog-content\") pod \"community-operators-24svc\" (UID: \"997fd032-41f3-4d25-a8fb-c666c5891748\") " pod="openshift-marketplace/community-operators-24svc" Oct 08 19:01:37 crc kubenswrapper[4859]: I1008 19:01:37.501674 4859 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/997fd032-41f3-4d25-a8fb-c666c5891748-utilities\") pod \"community-operators-24svc\" (UID: \"997fd032-41f3-4d25-a8fb-c666c5891748\") " pod="openshift-marketplace/community-operators-24svc" Oct 08 19:01:37 crc kubenswrapper[4859]: I1008 19:01:37.502161 4859 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/997fd032-41f3-4d25-a8fb-c666c5891748-utilities\") pod \"community-operators-24svc\" (UID: \"997fd032-41f3-4d25-a8fb-c666c5891748\") " pod="openshift-marketplace/community-operators-24svc" Oct 08 19:01:37 crc kubenswrapper[4859]: I1008 19:01:37.502190 4859 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/997fd032-41f3-4d25-a8fb-c666c5891748-catalog-content\") pod \"community-operators-24svc\" (UID: \"997fd032-41f3-4d25-a8fb-c666c5891748\") " pod="openshift-marketplace/community-operators-24svc" Oct 08 19:01:37 crc kubenswrapper[4859]: I1008 19:01:37.525479 4859 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-vq78x\" (UniqueName: \"kubernetes.io/projected/997fd032-41f3-4d25-a8fb-c666c5891748-kube-api-access-vq78x\") pod \"community-operators-24svc\" (UID: \"997fd032-41f3-4d25-a8fb-c666c5891748\") " pod="openshift-marketplace/community-operators-24svc" Oct 08 19:01:37 crc kubenswrapper[4859]: I1008 19:01:37.662573 4859 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-24svc" Oct 08 19:01:38 crc kubenswrapper[4859]: I1008 19:01:38.171742 4859 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-24svc"] Oct 08 19:01:38 crc kubenswrapper[4859]: I1008 19:01:38.470135 4859 scope.go:117] "RemoveContainer" containerID="265e61e43a0f4edee3adcaf87995c3c44fad0b135f61f14030c7ee297b24d748" Oct 08 19:01:38 crc kubenswrapper[4859]: E1008 19:01:38.470963 4859 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-82s52_openshift-machine-config-operator(b23a6a6c-9d92-4e7b-840e-55cfda873a2d)\"" pod="openshift-machine-config-operator/machine-config-daemon-82s52" podUID="b23a6a6c-9d92-4e7b-840e-55cfda873a2d" Oct 08 19:01:39 crc kubenswrapper[4859]: I1008 19:01:39.084469 4859 generic.go:334] "Generic (PLEG): container finished" podID="997fd032-41f3-4d25-a8fb-c666c5891748" containerID="271a64f6f2c8e783f297af3e29599b321af389d17dc8916c0057d66e41d2e2d9" exitCode=0 Oct 08 19:01:39 crc kubenswrapper[4859]: I1008 19:01:39.084516 4859 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-24svc" event={"ID":"997fd032-41f3-4d25-a8fb-c666c5891748","Type":"ContainerDied","Data":"271a64f6f2c8e783f297af3e29599b321af389d17dc8916c0057d66e41d2e2d9"} Oct 08 19:01:39 crc kubenswrapper[4859]: I1008 19:01:39.084561 4859 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-24svc" event={"ID":"997fd032-41f3-4d25-a8fb-c666c5891748","Type":"ContainerStarted","Data":"c8f276c6475fe1d81ded5e9a338f7cb896ae477bcb3118ed26da820d98c436d7"} Oct 08 19:01:42 crc kubenswrapper[4859]: I1008 19:01:42.115074 4859 generic.go:334] "Generic (PLEG): container finished" podID="997fd032-41f3-4d25-a8fb-c666c5891748" containerID="0049642410a1d2f9a8c5adb62557418ed03e11d631aeb03734224999cac0650b" exitCode=0 Oct 08 19:01:42 crc kubenswrapper[4859]: I1008 19:01:42.115208 4859 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-24svc" event={"ID":"997fd032-41f3-4d25-a8fb-c666c5891748","Type":"ContainerDied","Data":"0049642410a1d2f9a8c5adb62557418ed03e11d631aeb03734224999cac0650b"} Oct 08 19:01:43 crc kubenswrapper[4859]: I1008 19:01:43.135908 4859 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-24svc" event={"ID":"997fd032-41f3-4d25-a8fb-c666c5891748","Type":"ContainerStarted","Data":"e0952d49e86ea007b2164ff75bdd494e510b13aad05f8770587f6c27356b05cf"} Oct 08 19:01:43 crc kubenswrapper[4859]: I1008 19:01:43.164954 4859 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-24svc" podStartSLOduration=2.716001828 podStartE2EDuration="6.164931575s" podCreationTimestamp="2025-10-08 19:01:37 +0000 UTC" firstStartedPulling="2025-10-08 19:01:39.086190123 +0000 UTC m=+2669.333029502" lastFinishedPulling="2025-10-08 19:01:42.53511983 +0000 UTC m=+2672.781959249" observedRunningTime="2025-10-08 19:01:43.1558138 +0000 UTC m=+2673.402653189" watchObservedRunningTime="2025-10-08 19:01:43.164931575 +0000 UTC m=+2673.411770964" Oct 08 19:01:47 crc kubenswrapper[4859]: I1008 19:01:47.663786 4859 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-24svc" Oct 08 19:01:47 crc kubenswrapper[4859]: I1008 19:01:47.664423 4859 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-24svc" Oct 08 19:01:47 crc kubenswrapper[4859]: I1008 19:01:47.725258 4859 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-24svc" Oct 08 19:01:48 crc kubenswrapper[4859]: I1008 19:01:48.221140 4859 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-24svc" Oct 08 19:01:48 crc kubenswrapper[4859]: I1008 19:01:48.268018 4859 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-24svc"] Oct 08 19:01:50 crc kubenswrapper[4859]: I1008 19:01:50.199273 4859 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/community-operators-24svc" podUID="997fd032-41f3-4d25-a8fb-c666c5891748" containerName="registry-server" containerID="cri-o://e0952d49e86ea007b2164ff75bdd494e510b13aad05f8770587f6c27356b05cf" gracePeriod=2 Oct 08 19:01:50 crc kubenswrapper[4859]: I1008 19:01:50.470007 4859 scope.go:117] "RemoveContainer" containerID="265e61e43a0f4edee3adcaf87995c3c44fad0b135f61f14030c7ee297b24d748" Oct 08 19:01:50 crc kubenswrapper[4859]: E1008 19:01:50.470828 4859 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-82s52_openshift-machine-config-operator(b23a6a6c-9d92-4e7b-840e-55cfda873a2d)\"" pod="openshift-machine-config-operator/machine-config-daemon-82s52" podUID="b23a6a6c-9d92-4e7b-840e-55cfda873a2d" Oct 08 19:01:50 crc kubenswrapper[4859]: I1008 19:01:50.631060 4859 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-24svc" Oct 08 19:01:50 crc kubenswrapper[4859]: I1008 19:01:50.749217 4859 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-vq78x\" (UniqueName: \"kubernetes.io/projected/997fd032-41f3-4d25-a8fb-c666c5891748-kube-api-access-vq78x\") pod \"997fd032-41f3-4d25-a8fb-c666c5891748\" (UID: \"997fd032-41f3-4d25-a8fb-c666c5891748\") " Oct 08 19:01:50 crc kubenswrapper[4859]: I1008 19:01:50.749262 4859 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/997fd032-41f3-4d25-a8fb-c666c5891748-utilities\") pod \"997fd032-41f3-4d25-a8fb-c666c5891748\" (UID: \"997fd032-41f3-4d25-a8fb-c666c5891748\") " Oct 08 19:01:50 crc kubenswrapper[4859]: I1008 19:01:50.749366 4859 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/997fd032-41f3-4d25-a8fb-c666c5891748-catalog-content\") pod \"997fd032-41f3-4d25-a8fb-c666c5891748\" (UID: \"997fd032-41f3-4d25-a8fb-c666c5891748\") " Oct 08 19:01:50 crc kubenswrapper[4859]: I1008 19:01:50.750606 4859 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/997fd032-41f3-4d25-a8fb-c666c5891748-utilities" (OuterVolumeSpecName: "utilities") pod "997fd032-41f3-4d25-a8fb-c666c5891748" (UID: "997fd032-41f3-4d25-a8fb-c666c5891748"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 08 19:01:50 crc kubenswrapper[4859]: I1008 19:01:50.755104 4859 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/997fd032-41f3-4d25-a8fb-c666c5891748-kube-api-access-vq78x" (OuterVolumeSpecName: "kube-api-access-vq78x") pod "997fd032-41f3-4d25-a8fb-c666c5891748" (UID: "997fd032-41f3-4d25-a8fb-c666c5891748"). InnerVolumeSpecName "kube-api-access-vq78x". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 08 19:01:50 crc kubenswrapper[4859]: I1008 19:01:50.801240 4859 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/997fd032-41f3-4d25-a8fb-c666c5891748-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "997fd032-41f3-4d25-a8fb-c666c5891748" (UID: "997fd032-41f3-4d25-a8fb-c666c5891748"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 08 19:01:50 crc kubenswrapper[4859]: I1008 19:01:50.851082 4859 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/997fd032-41f3-4d25-a8fb-c666c5891748-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 08 19:01:50 crc kubenswrapper[4859]: I1008 19:01:50.851113 4859 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-vq78x\" (UniqueName: \"kubernetes.io/projected/997fd032-41f3-4d25-a8fb-c666c5891748-kube-api-access-vq78x\") on node \"crc\" DevicePath \"\"" Oct 08 19:01:50 crc kubenswrapper[4859]: I1008 19:01:50.851125 4859 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/997fd032-41f3-4d25-a8fb-c666c5891748-utilities\") on node \"crc\" DevicePath \"\"" Oct 08 19:01:51 crc kubenswrapper[4859]: I1008 19:01:51.208790 4859 generic.go:334] "Generic (PLEG): container finished" podID="997fd032-41f3-4d25-a8fb-c666c5891748" containerID="e0952d49e86ea007b2164ff75bdd494e510b13aad05f8770587f6c27356b05cf" exitCode=0 Oct 08 19:01:51 crc kubenswrapper[4859]: I1008 19:01:51.208866 4859 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-24svc" event={"ID":"997fd032-41f3-4d25-a8fb-c666c5891748","Type":"ContainerDied","Data":"e0952d49e86ea007b2164ff75bdd494e510b13aad05f8770587f6c27356b05cf"} Oct 08 19:01:51 crc kubenswrapper[4859]: I1008 19:01:51.208927 4859 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-24svc" Oct 08 19:01:51 crc kubenswrapper[4859]: I1008 19:01:51.209175 4859 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-24svc" event={"ID":"997fd032-41f3-4d25-a8fb-c666c5891748","Type":"ContainerDied","Data":"c8f276c6475fe1d81ded5e9a338f7cb896ae477bcb3118ed26da820d98c436d7"} Oct 08 19:01:51 crc kubenswrapper[4859]: I1008 19:01:51.209198 4859 scope.go:117] "RemoveContainer" containerID="e0952d49e86ea007b2164ff75bdd494e510b13aad05f8770587f6c27356b05cf" Oct 08 19:01:51 crc kubenswrapper[4859]: I1008 19:01:51.236578 4859 scope.go:117] "RemoveContainer" containerID="0049642410a1d2f9a8c5adb62557418ed03e11d631aeb03734224999cac0650b" Oct 08 19:01:51 crc kubenswrapper[4859]: I1008 19:01:51.254417 4859 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-24svc"] Oct 08 19:01:51 crc kubenswrapper[4859]: I1008 19:01:51.265135 4859 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/community-operators-24svc"] Oct 08 19:01:51 crc kubenswrapper[4859]: I1008 19:01:51.287649 4859 scope.go:117] "RemoveContainer" containerID="271a64f6f2c8e783f297af3e29599b321af389d17dc8916c0057d66e41d2e2d9" Oct 08 19:01:51 crc kubenswrapper[4859]: I1008 19:01:51.306740 4859 scope.go:117] "RemoveContainer" containerID="e0952d49e86ea007b2164ff75bdd494e510b13aad05f8770587f6c27356b05cf" Oct 08 19:01:51 crc kubenswrapper[4859]: E1008 19:01:51.307137 4859 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"e0952d49e86ea007b2164ff75bdd494e510b13aad05f8770587f6c27356b05cf\": container with ID starting with e0952d49e86ea007b2164ff75bdd494e510b13aad05f8770587f6c27356b05cf not found: ID does not exist" containerID="e0952d49e86ea007b2164ff75bdd494e510b13aad05f8770587f6c27356b05cf" Oct 08 19:01:51 crc kubenswrapper[4859]: I1008 19:01:51.307195 4859 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"e0952d49e86ea007b2164ff75bdd494e510b13aad05f8770587f6c27356b05cf"} err="failed to get container status \"e0952d49e86ea007b2164ff75bdd494e510b13aad05f8770587f6c27356b05cf\": rpc error: code = NotFound desc = could not find container \"e0952d49e86ea007b2164ff75bdd494e510b13aad05f8770587f6c27356b05cf\": container with ID starting with e0952d49e86ea007b2164ff75bdd494e510b13aad05f8770587f6c27356b05cf not found: ID does not exist" Oct 08 19:01:51 crc kubenswrapper[4859]: I1008 19:01:51.307227 4859 scope.go:117] "RemoveContainer" containerID="0049642410a1d2f9a8c5adb62557418ed03e11d631aeb03734224999cac0650b" Oct 08 19:01:51 crc kubenswrapper[4859]: E1008 19:01:51.308025 4859 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"0049642410a1d2f9a8c5adb62557418ed03e11d631aeb03734224999cac0650b\": container with ID starting with 0049642410a1d2f9a8c5adb62557418ed03e11d631aeb03734224999cac0650b not found: ID does not exist" containerID="0049642410a1d2f9a8c5adb62557418ed03e11d631aeb03734224999cac0650b" Oct 08 19:01:51 crc kubenswrapper[4859]: I1008 19:01:51.308070 4859 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"0049642410a1d2f9a8c5adb62557418ed03e11d631aeb03734224999cac0650b"} err="failed to get container status \"0049642410a1d2f9a8c5adb62557418ed03e11d631aeb03734224999cac0650b\": rpc error: code = NotFound desc = could not find container \"0049642410a1d2f9a8c5adb62557418ed03e11d631aeb03734224999cac0650b\": container with ID starting with 0049642410a1d2f9a8c5adb62557418ed03e11d631aeb03734224999cac0650b not found: ID does not exist" Oct 08 19:01:51 crc kubenswrapper[4859]: I1008 19:01:51.308143 4859 scope.go:117] "RemoveContainer" containerID="271a64f6f2c8e783f297af3e29599b321af389d17dc8916c0057d66e41d2e2d9" Oct 08 19:01:51 crc kubenswrapper[4859]: E1008 19:01:51.308408 4859 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"271a64f6f2c8e783f297af3e29599b321af389d17dc8916c0057d66e41d2e2d9\": container with ID starting with 271a64f6f2c8e783f297af3e29599b321af389d17dc8916c0057d66e41d2e2d9 not found: ID does not exist" containerID="271a64f6f2c8e783f297af3e29599b321af389d17dc8916c0057d66e41d2e2d9" Oct 08 19:01:51 crc kubenswrapper[4859]: I1008 19:01:51.308431 4859 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"271a64f6f2c8e783f297af3e29599b321af389d17dc8916c0057d66e41d2e2d9"} err="failed to get container status \"271a64f6f2c8e783f297af3e29599b321af389d17dc8916c0057d66e41d2e2d9\": rpc error: code = NotFound desc = could not find container \"271a64f6f2c8e783f297af3e29599b321af389d17dc8916c0057d66e41d2e2d9\": container with ID starting with 271a64f6f2c8e783f297af3e29599b321af389d17dc8916c0057d66e41d2e2d9 not found: ID does not exist" Oct 08 19:01:52 crc kubenswrapper[4859]: I1008 19:01:52.493969 4859 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="997fd032-41f3-4d25-a8fb-c666c5891748" path="/var/lib/kubelet/pods/997fd032-41f3-4d25-a8fb-c666c5891748/volumes" Oct 08 19:02:03 crc kubenswrapper[4859]: I1008 19:02:03.470794 4859 scope.go:117] "RemoveContainer" containerID="265e61e43a0f4edee3adcaf87995c3c44fad0b135f61f14030c7ee297b24d748" Oct 08 19:02:03 crc kubenswrapper[4859]: E1008 19:02:03.471861 4859 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-82s52_openshift-machine-config-operator(b23a6a6c-9d92-4e7b-840e-55cfda873a2d)\"" pod="openshift-machine-config-operator/machine-config-daemon-82s52" podUID="b23a6a6c-9d92-4e7b-840e-55cfda873a2d" Oct 08 19:02:15 crc kubenswrapper[4859]: I1008 19:02:15.470206 4859 scope.go:117] "RemoveContainer" containerID="265e61e43a0f4edee3adcaf87995c3c44fad0b135f61f14030c7ee297b24d748" Oct 08 19:02:15 crc kubenswrapper[4859]: E1008 19:02:15.471036 4859 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-82s52_openshift-machine-config-operator(b23a6a6c-9d92-4e7b-840e-55cfda873a2d)\"" pod="openshift-machine-config-operator/machine-config-daemon-82s52" podUID="b23a6a6c-9d92-4e7b-840e-55cfda873a2d" Oct 08 19:02:27 crc kubenswrapper[4859]: I1008 19:02:27.035798 4859 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-cpddq"] Oct 08 19:02:27 crc kubenswrapper[4859]: E1008 19:02:27.036894 4859 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="997fd032-41f3-4d25-a8fb-c666c5891748" containerName="extract-content" Oct 08 19:02:27 crc kubenswrapper[4859]: I1008 19:02:27.036910 4859 state_mem.go:107] "Deleted CPUSet assignment" podUID="997fd032-41f3-4d25-a8fb-c666c5891748" containerName="extract-content" Oct 08 19:02:27 crc kubenswrapper[4859]: E1008 19:02:27.036929 4859 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="997fd032-41f3-4d25-a8fb-c666c5891748" containerName="extract-utilities" Oct 08 19:02:27 crc kubenswrapper[4859]: I1008 19:02:27.036937 4859 state_mem.go:107] "Deleted CPUSet assignment" podUID="997fd032-41f3-4d25-a8fb-c666c5891748" containerName="extract-utilities" Oct 08 19:02:27 crc kubenswrapper[4859]: E1008 19:02:27.036977 4859 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="997fd032-41f3-4d25-a8fb-c666c5891748" containerName="registry-server" Oct 08 19:02:27 crc kubenswrapper[4859]: I1008 19:02:27.036984 4859 state_mem.go:107] "Deleted CPUSet assignment" podUID="997fd032-41f3-4d25-a8fb-c666c5891748" containerName="registry-server" Oct 08 19:02:27 crc kubenswrapper[4859]: I1008 19:02:27.037158 4859 memory_manager.go:354] "RemoveStaleState removing state" podUID="997fd032-41f3-4d25-a8fb-c666c5891748" containerName="registry-server" Oct 08 19:02:27 crc kubenswrapper[4859]: I1008 19:02:27.038494 4859 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-cpddq" Oct 08 19:02:27 crc kubenswrapper[4859]: I1008 19:02:27.062389 4859 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-cpddq"] Oct 08 19:02:27 crc kubenswrapper[4859]: I1008 19:02:27.157461 4859 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/105f2c22-da1f-4ec4-b522-26043af2edca-utilities\") pod \"certified-operators-cpddq\" (UID: \"105f2c22-da1f-4ec4-b522-26043af2edca\") " pod="openshift-marketplace/certified-operators-cpddq" Oct 08 19:02:27 crc kubenswrapper[4859]: I1008 19:02:27.157572 4859 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/105f2c22-da1f-4ec4-b522-26043af2edca-catalog-content\") pod \"certified-operators-cpddq\" (UID: \"105f2c22-da1f-4ec4-b522-26043af2edca\") " pod="openshift-marketplace/certified-operators-cpddq" Oct 08 19:02:27 crc kubenswrapper[4859]: I1008 19:02:27.157935 4859 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-zlxhl\" (UniqueName: \"kubernetes.io/projected/105f2c22-da1f-4ec4-b522-26043af2edca-kube-api-access-zlxhl\") pod \"certified-operators-cpddq\" (UID: \"105f2c22-da1f-4ec4-b522-26043af2edca\") " pod="openshift-marketplace/certified-operators-cpddq" Oct 08 19:02:27 crc kubenswrapper[4859]: I1008 19:02:27.259515 4859 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/105f2c22-da1f-4ec4-b522-26043af2edca-catalog-content\") pod \"certified-operators-cpddq\" (UID: \"105f2c22-da1f-4ec4-b522-26043af2edca\") " pod="openshift-marketplace/certified-operators-cpddq" Oct 08 19:02:27 crc kubenswrapper[4859]: I1008 19:02:27.259892 4859 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-zlxhl\" (UniqueName: \"kubernetes.io/projected/105f2c22-da1f-4ec4-b522-26043af2edca-kube-api-access-zlxhl\") pod \"certified-operators-cpddq\" (UID: \"105f2c22-da1f-4ec4-b522-26043af2edca\") " pod="openshift-marketplace/certified-operators-cpddq" Oct 08 19:02:27 crc kubenswrapper[4859]: I1008 19:02:27.260031 4859 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/105f2c22-da1f-4ec4-b522-26043af2edca-utilities\") pod \"certified-operators-cpddq\" (UID: \"105f2c22-da1f-4ec4-b522-26043af2edca\") " pod="openshift-marketplace/certified-operators-cpddq" Oct 08 19:02:27 crc kubenswrapper[4859]: I1008 19:02:27.260060 4859 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/105f2c22-da1f-4ec4-b522-26043af2edca-catalog-content\") pod \"certified-operators-cpddq\" (UID: \"105f2c22-da1f-4ec4-b522-26043af2edca\") " pod="openshift-marketplace/certified-operators-cpddq" Oct 08 19:02:27 crc kubenswrapper[4859]: I1008 19:02:27.260271 4859 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/105f2c22-da1f-4ec4-b522-26043af2edca-utilities\") pod \"certified-operators-cpddq\" (UID: \"105f2c22-da1f-4ec4-b522-26043af2edca\") " pod="openshift-marketplace/certified-operators-cpddq" Oct 08 19:02:27 crc kubenswrapper[4859]: I1008 19:02:27.286585 4859 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-zlxhl\" (UniqueName: \"kubernetes.io/projected/105f2c22-da1f-4ec4-b522-26043af2edca-kube-api-access-zlxhl\") pod \"certified-operators-cpddq\" (UID: \"105f2c22-da1f-4ec4-b522-26043af2edca\") " pod="openshift-marketplace/certified-operators-cpddq" Oct 08 19:02:27 crc kubenswrapper[4859]: I1008 19:02:27.382862 4859 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-cpddq" Oct 08 19:02:27 crc kubenswrapper[4859]: I1008 19:02:27.909611 4859 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-cpddq"] Oct 08 19:02:28 crc kubenswrapper[4859]: I1008 19:02:28.593823 4859 generic.go:334] "Generic (PLEG): container finished" podID="105f2c22-da1f-4ec4-b522-26043af2edca" containerID="4a24e98502b548eb6e551a382af1c2e0028a29006074e5fc5f78d4fce9edf166" exitCode=0 Oct 08 19:02:28 crc kubenswrapper[4859]: I1008 19:02:28.593909 4859 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-cpddq" event={"ID":"105f2c22-da1f-4ec4-b522-26043af2edca","Type":"ContainerDied","Data":"4a24e98502b548eb6e551a382af1c2e0028a29006074e5fc5f78d4fce9edf166"} Oct 08 19:02:28 crc kubenswrapper[4859]: I1008 19:02:28.594159 4859 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-cpddq" event={"ID":"105f2c22-da1f-4ec4-b522-26043af2edca","Type":"ContainerStarted","Data":"9a23c99d44de8d23c8f2755bffac77605d92f1d678449c4be01987ce4e278731"} Oct 08 19:02:29 crc kubenswrapper[4859]: I1008 19:02:29.469851 4859 scope.go:117] "RemoveContainer" containerID="265e61e43a0f4edee3adcaf87995c3c44fad0b135f61f14030c7ee297b24d748" Oct 08 19:02:30 crc kubenswrapper[4859]: I1008 19:02:30.615074 4859 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-82s52" event={"ID":"b23a6a6c-9d92-4e7b-840e-55cfda873a2d","Type":"ContainerStarted","Data":"a3d24a0dfcbf833b7bcf1af5e19a45816904f485fab5db993921d32e47607790"} Oct 08 19:02:30 crc kubenswrapper[4859]: I1008 19:02:30.619093 4859 generic.go:334] "Generic (PLEG): container finished" podID="105f2c22-da1f-4ec4-b522-26043af2edca" containerID="baec2ef746f68fae84ce57207baf74bcb54426444511eeec57a5d82e507f0021" exitCode=0 Oct 08 19:02:30 crc kubenswrapper[4859]: I1008 19:02:30.619148 4859 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-cpddq" event={"ID":"105f2c22-da1f-4ec4-b522-26043af2edca","Type":"ContainerDied","Data":"baec2ef746f68fae84ce57207baf74bcb54426444511eeec57a5d82e507f0021"} Oct 08 19:02:31 crc kubenswrapper[4859]: I1008 19:02:31.630502 4859 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-cpddq" event={"ID":"105f2c22-da1f-4ec4-b522-26043af2edca","Type":"ContainerStarted","Data":"b3525d33870d1d59f4919166d11295dd36529d22a1461cea2953c90ab9d57b59"} Oct 08 19:02:31 crc kubenswrapper[4859]: I1008 19:02:31.672001 4859 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-cpddq" podStartSLOduration=2.267935149 podStartE2EDuration="4.671980565s" podCreationTimestamp="2025-10-08 19:02:27 +0000 UTC" firstStartedPulling="2025-10-08 19:02:28.59672194 +0000 UTC m=+2718.843561319" lastFinishedPulling="2025-10-08 19:02:31.000767356 +0000 UTC m=+2721.247606735" observedRunningTime="2025-10-08 19:02:31.659734431 +0000 UTC m=+2721.906573820" watchObservedRunningTime="2025-10-08 19:02:31.671980565 +0000 UTC m=+2721.918819954" Oct 08 19:02:37 crc kubenswrapper[4859]: I1008 19:02:37.383146 4859 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-cpddq" Oct 08 19:02:37 crc kubenswrapper[4859]: I1008 19:02:37.383981 4859 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-cpddq" Oct 08 19:02:37 crc kubenswrapper[4859]: I1008 19:02:37.455824 4859 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-cpddq" Oct 08 19:02:37 crc kubenswrapper[4859]: I1008 19:02:37.739643 4859 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-cpddq" Oct 08 19:02:37 crc kubenswrapper[4859]: I1008 19:02:37.797140 4859 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-cpddq"] Oct 08 19:02:39 crc kubenswrapper[4859]: I1008 19:02:39.706366 4859 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/certified-operators-cpddq" podUID="105f2c22-da1f-4ec4-b522-26043af2edca" containerName="registry-server" containerID="cri-o://b3525d33870d1d59f4919166d11295dd36529d22a1461cea2953c90ab9d57b59" gracePeriod=2 Oct 08 19:02:39 crc kubenswrapper[4859]: E1008 19:02:39.910033 4859 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod105f2c22_da1f_4ec4_b522_26043af2edca.slice/crio-conmon-b3525d33870d1d59f4919166d11295dd36529d22a1461cea2953c90ab9d57b59.scope\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod105f2c22_da1f_4ec4_b522_26043af2edca.slice/crio-b3525d33870d1d59f4919166d11295dd36529d22a1461cea2953c90ab9d57b59.scope\": RecentStats: unable to find data in memory cache]" Oct 08 19:02:40 crc kubenswrapper[4859]: I1008 19:02:40.207444 4859 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-cpddq" Oct 08 19:02:40 crc kubenswrapper[4859]: I1008 19:02:40.260838 4859 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/105f2c22-da1f-4ec4-b522-26043af2edca-catalog-content\") pod \"105f2c22-da1f-4ec4-b522-26043af2edca\" (UID: \"105f2c22-da1f-4ec4-b522-26043af2edca\") " Oct 08 19:02:40 crc kubenswrapper[4859]: I1008 19:02:40.261003 4859 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/105f2c22-da1f-4ec4-b522-26043af2edca-utilities\") pod \"105f2c22-da1f-4ec4-b522-26043af2edca\" (UID: \"105f2c22-da1f-4ec4-b522-26043af2edca\") " Oct 08 19:02:40 crc kubenswrapper[4859]: I1008 19:02:40.261140 4859 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-zlxhl\" (UniqueName: \"kubernetes.io/projected/105f2c22-da1f-4ec4-b522-26043af2edca-kube-api-access-zlxhl\") pod \"105f2c22-da1f-4ec4-b522-26043af2edca\" (UID: \"105f2c22-da1f-4ec4-b522-26043af2edca\") " Oct 08 19:02:40 crc kubenswrapper[4859]: I1008 19:02:40.262031 4859 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/105f2c22-da1f-4ec4-b522-26043af2edca-utilities" (OuterVolumeSpecName: "utilities") pod "105f2c22-da1f-4ec4-b522-26043af2edca" (UID: "105f2c22-da1f-4ec4-b522-26043af2edca"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 08 19:02:40 crc kubenswrapper[4859]: I1008 19:02:40.271933 4859 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/105f2c22-da1f-4ec4-b522-26043af2edca-kube-api-access-zlxhl" (OuterVolumeSpecName: "kube-api-access-zlxhl") pod "105f2c22-da1f-4ec4-b522-26043af2edca" (UID: "105f2c22-da1f-4ec4-b522-26043af2edca"). InnerVolumeSpecName "kube-api-access-zlxhl". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 08 19:02:40 crc kubenswrapper[4859]: I1008 19:02:40.314980 4859 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/105f2c22-da1f-4ec4-b522-26043af2edca-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "105f2c22-da1f-4ec4-b522-26043af2edca" (UID: "105f2c22-da1f-4ec4-b522-26043af2edca"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 08 19:02:40 crc kubenswrapper[4859]: I1008 19:02:40.362871 4859 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/105f2c22-da1f-4ec4-b522-26043af2edca-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 08 19:02:40 crc kubenswrapper[4859]: I1008 19:02:40.362901 4859 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/105f2c22-da1f-4ec4-b522-26043af2edca-utilities\") on node \"crc\" DevicePath \"\"" Oct 08 19:02:40 crc kubenswrapper[4859]: I1008 19:02:40.362913 4859 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-zlxhl\" (UniqueName: \"kubernetes.io/projected/105f2c22-da1f-4ec4-b522-26043af2edca-kube-api-access-zlxhl\") on node \"crc\" DevicePath \"\"" Oct 08 19:02:40 crc kubenswrapper[4859]: I1008 19:02:40.764201 4859 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-cpddq" event={"ID":"105f2c22-da1f-4ec4-b522-26043af2edca","Type":"ContainerDied","Data":"b3525d33870d1d59f4919166d11295dd36529d22a1461cea2953c90ab9d57b59"} Oct 08 19:02:40 crc kubenswrapper[4859]: I1008 19:02:40.764583 4859 scope.go:117] "RemoveContainer" containerID="b3525d33870d1d59f4919166d11295dd36529d22a1461cea2953c90ab9d57b59" Oct 08 19:02:40 crc kubenswrapper[4859]: I1008 19:02:40.764176 4859 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-cpddq" Oct 08 19:02:40 crc kubenswrapper[4859]: I1008 19:02:40.764056 4859 generic.go:334] "Generic (PLEG): container finished" podID="105f2c22-da1f-4ec4-b522-26043af2edca" containerID="b3525d33870d1d59f4919166d11295dd36529d22a1461cea2953c90ab9d57b59" exitCode=0 Oct 08 19:02:40 crc kubenswrapper[4859]: I1008 19:02:40.764714 4859 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-cpddq" event={"ID":"105f2c22-da1f-4ec4-b522-26043af2edca","Type":"ContainerDied","Data":"9a23c99d44de8d23c8f2755bffac77605d92f1d678449c4be01987ce4e278731"} Oct 08 19:02:40 crc kubenswrapper[4859]: I1008 19:02:40.789632 4859 scope.go:117] "RemoveContainer" containerID="baec2ef746f68fae84ce57207baf74bcb54426444511eeec57a5d82e507f0021" Oct 08 19:02:40 crc kubenswrapper[4859]: I1008 19:02:40.792646 4859 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-cpddq"] Oct 08 19:02:40 crc kubenswrapper[4859]: I1008 19:02:40.803856 4859 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/certified-operators-cpddq"] Oct 08 19:02:40 crc kubenswrapper[4859]: I1008 19:02:40.807378 4859 scope.go:117] "RemoveContainer" containerID="4a24e98502b548eb6e551a382af1c2e0028a29006074e5fc5f78d4fce9edf166" Oct 08 19:02:40 crc kubenswrapper[4859]: I1008 19:02:40.852821 4859 scope.go:117] "RemoveContainer" containerID="b3525d33870d1d59f4919166d11295dd36529d22a1461cea2953c90ab9d57b59" Oct 08 19:02:40 crc kubenswrapper[4859]: E1008 19:02:40.853272 4859 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"b3525d33870d1d59f4919166d11295dd36529d22a1461cea2953c90ab9d57b59\": container with ID starting with b3525d33870d1d59f4919166d11295dd36529d22a1461cea2953c90ab9d57b59 not found: ID does not exist" containerID="b3525d33870d1d59f4919166d11295dd36529d22a1461cea2953c90ab9d57b59" Oct 08 19:02:40 crc kubenswrapper[4859]: I1008 19:02:40.853303 4859 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"b3525d33870d1d59f4919166d11295dd36529d22a1461cea2953c90ab9d57b59"} err="failed to get container status \"b3525d33870d1d59f4919166d11295dd36529d22a1461cea2953c90ab9d57b59\": rpc error: code = NotFound desc = could not find container \"b3525d33870d1d59f4919166d11295dd36529d22a1461cea2953c90ab9d57b59\": container with ID starting with b3525d33870d1d59f4919166d11295dd36529d22a1461cea2953c90ab9d57b59 not found: ID does not exist" Oct 08 19:02:40 crc kubenswrapper[4859]: I1008 19:02:40.853323 4859 scope.go:117] "RemoveContainer" containerID="baec2ef746f68fae84ce57207baf74bcb54426444511eeec57a5d82e507f0021" Oct 08 19:02:40 crc kubenswrapper[4859]: E1008 19:02:40.853861 4859 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"baec2ef746f68fae84ce57207baf74bcb54426444511eeec57a5d82e507f0021\": container with ID starting with baec2ef746f68fae84ce57207baf74bcb54426444511eeec57a5d82e507f0021 not found: ID does not exist" containerID="baec2ef746f68fae84ce57207baf74bcb54426444511eeec57a5d82e507f0021" Oct 08 19:02:40 crc kubenswrapper[4859]: I1008 19:02:40.853910 4859 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"baec2ef746f68fae84ce57207baf74bcb54426444511eeec57a5d82e507f0021"} err="failed to get container status \"baec2ef746f68fae84ce57207baf74bcb54426444511eeec57a5d82e507f0021\": rpc error: code = NotFound desc = could not find container \"baec2ef746f68fae84ce57207baf74bcb54426444511eeec57a5d82e507f0021\": container with ID starting with baec2ef746f68fae84ce57207baf74bcb54426444511eeec57a5d82e507f0021 not found: ID does not exist" Oct 08 19:02:40 crc kubenswrapper[4859]: I1008 19:02:40.853936 4859 scope.go:117] "RemoveContainer" containerID="4a24e98502b548eb6e551a382af1c2e0028a29006074e5fc5f78d4fce9edf166" Oct 08 19:02:40 crc kubenswrapper[4859]: E1008 19:02:40.854224 4859 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"4a24e98502b548eb6e551a382af1c2e0028a29006074e5fc5f78d4fce9edf166\": container with ID starting with 4a24e98502b548eb6e551a382af1c2e0028a29006074e5fc5f78d4fce9edf166 not found: ID does not exist" containerID="4a24e98502b548eb6e551a382af1c2e0028a29006074e5fc5f78d4fce9edf166" Oct 08 19:02:40 crc kubenswrapper[4859]: I1008 19:02:40.854249 4859 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"4a24e98502b548eb6e551a382af1c2e0028a29006074e5fc5f78d4fce9edf166"} err="failed to get container status \"4a24e98502b548eb6e551a382af1c2e0028a29006074e5fc5f78d4fce9edf166\": rpc error: code = NotFound desc = could not find container \"4a24e98502b548eb6e551a382af1c2e0028a29006074e5fc5f78d4fce9edf166\": container with ID starting with 4a24e98502b548eb6e551a382af1c2e0028a29006074e5fc5f78d4fce9edf166 not found: ID does not exist" Oct 08 19:02:42 crc kubenswrapper[4859]: I1008 19:02:42.488635 4859 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="105f2c22-da1f-4ec4-b522-26043af2edca" path="/var/lib/kubelet/pods/105f2c22-da1f-4ec4-b522-26043af2edca/volumes" Oct 08 19:03:29 crc kubenswrapper[4859]: I1008 19:03:29.020949 4859 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-mtkqm"] Oct 08 19:03:29 crc kubenswrapper[4859]: E1008 19:03:29.023208 4859 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="105f2c22-da1f-4ec4-b522-26043af2edca" containerName="extract-content" Oct 08 19:03:29 crc kubenswrapper[4859]: I1008 19:03:29.023354 4859 state_mem.go:107] "Deleted CPUSet assignment" podUID="105f2c22-da1f-4ec4-b522-26043af2edca" containerName="extract-content" Oct 08 19:03:29 crc kubenswrapper[4859]: E1008 19:03:29.023445 4859 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="105f2c22-da1f-4ec4-b522-26043af2edca" containerName="registry-server" Oct 08 19:03:29 crc kubenswrapper[4859]: I1008 19:03:29.023523 4859 state_mem.go:107] "Deleted CPUSet assignment" podUID="105f2c22-da1f-4ec4-b522-26043af2edca" containerName="registry-server" Oct 08 19:03:29 crc kubenswrapper[4859]: E1008 19:03:29.023621 4859 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="105f2c22-da1f-4ec4-b522-26043af2edca" containerName="extract-utilities" Oct 08 19:03:29 crc kubenswrapper[4859]: I1008 19:03:29.023731 4859 state_mem.go:107] "Deleted CPUSet assignment" podUID="105f2c22-da1f-4ec4-b522-26043af2edca" containerName="extract-utilities" Oct 08 19:03:29 crc kubenswrapper[4859]: I1008 19:03:29.024051 4859 memory_manager.go:354] "RemoveStaleState removing state" podUID="105f2c22-da1f-4ec4-b522-26043af2edca" containerName="registry-server" Oct 08 19:03:29 crc kubenswrapper[4859]: I1008 19:03:29.027728 4859 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-mtkqm" Oct 08 19:03:29 crc kubenswrapper[4859]: I1008 19:03:29.038978 4859 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-mtkqm"] Oct 08 19:03:29 crc kubenswrapper[4859]: I1008 19:03:29.227998 4859 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-fwfd2\" (UniqueName: \"kubernetes.io/projected/f7931881-9829-40f5-9b68-c04957e42476-kube-api-access-fwfd2\") pod \"redhat-operators-mtkqm\" (UID: \"f7931881-9829-40f5-9b68-c04957e42476\") " pod="openshift-marketplace/redhat-operators-mtkqm" Oct 08 19:03:29 crc kubenswrapper[4859]: I1008 19:03:29.228413 4859 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/f7931881-9829-40f5-9b68-c04957e42476-catalog-content\") pod \"redhat-operators-mtkqm\" (UID: \"f7931881-9829-40f5-9b68-c04957e42476\") " pod="openshift-marketplace/redhat-operators-mtkqm" Oct 08 19:03:29 crc kubenswrapper[4859]: I1008 19:03:29.228464 4859 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/f7931881-9829-40f5-9b68-c04957e42476-utilities\") pod \"redhat-operators-mtkqm\" (UID: \"f7931881-9829-40f5-9b68-c04957e42476\") " pod="openshift-marketplace/redhat-operators-mtkqm" Oct 08 19:03:29 crc kubenswrapper[4859]: I1008 19:03:29.330028 4859 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-fwfd2\" (UniqueName: \"kubernetes.io/projected/f7931881-9829-40f5-9b68-c04957e42476-kube-api-access-fwfd2\") pod \"redhat-operators-mtkqm\" (UID: \"f7931881-9829-40f5-9b68-c04957e42476\") " pod="openshift-marketplace/redhat-operators-mtkqm" Oct 08 19:03:29 crc kubenswrapper[4859]: I1008 19:03:29.330186 4859 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/f7931881-9829-40f5-9b68-c04957e42476-catalog-content\") pod \"redhat-operators-mtkqm\" (UID: \"f7931881-9829-40f5-9b68-c04957e42476\") " pod="openshift-marketplace/redhat-operators-mtkqm" Oct 08 19:03:29 crc kubenswrapper[4859]: I1008 19:03:29.330231 4859 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/f7931881-9829-40f5-9b68-c04957e42476-utilities\") pod \"redhat-operators-mtkqm\" (UID: \"f7931881-9829-40f5-9b68-c04957e42476\") " pod="openshift-marketplace/redhat-operators-mtkqm" Oct 08 19:03:29 crc kubenswrapper[4859]: I1008 19:03:29.330679 4859 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/f7931881-9829-40f5-9b68-c04957e42476-catalog-content\") pod \"redhat-operators-mtkqm\" (UID: \"f7931881-9829-40f5-9b68-c04957e42476\") " pod="openshift-marketplace/redhat-operators-mtkqm" Oct 08 19:03:29 crc kubenswrapper[4859]: I1008 19:03:29.330757 4859 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/f7931881-9829-40f5-9b68-c04957e42476-utilities\") pod \"redhat-operators-mtkqm\" (UID: \"f7931881-9829-40f5-9b68-c04957e42476\") " pod="openshift-marketplace/redhat-operators-mtkqm" Oct 08 19:03:29 crc kubenswrapper[4859]: I1008 19:03:29.349093 4859 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-fwfd2\" (UniqueName: \"kubernetes.io/projected/f7931881-9829-40f5-9b68-c04957e42476-kube-api-access-fwfd2\") pod \"redhat-operators-mtkqm\" (UID: \"f7931881-9829-40f5-9b68-c04957e42476\") " pod="openshift-marketplace/redhat-operators-mtkqm" Oct 08 19:03:29 crc kubenswrapper[4859]: I1008 19:03:29.352903 4859 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-mtkqm" Oct 08 19:03:29 crc kubenswrapper[4859]: I1008 19:03:29.841074 4859 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-mtkqm"] Oct 08 19:03:30 crc kubenswrapper[4859]: I1008 19:03:30.256871 4859 generic.go:334] "Generic (PLEG): container finished" podID="f7931881-9829-40f5-9b68-c04957e42476" containerID="2e0d37cac2fd94803e0475b396e79a82f1d1ce81ec7fc4df61c7980b85bee134" exitCode=0 Oct 08 19:03:30 crc kubenswrapper[4859]: I1008 19:03:30.257089 4859 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-mtkqm" event={"ID":"f7931881-9829-40f5-9b68-c04957e42476","Type":"ContainerDied","Data":"2e0d37cac2fd94803e0475b396e79a82f1d1ce81ec7fc4df61c7980b85bee134"} Oct 08 19:03:30 crc kubenswrapper[4859]: I1008 19:03:30.257156 4859 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-mtkqm" event={"ID":"f7931881-9829-40f5-9b68-c04957e42476","Type":"ContainerStarted","Data":"40e475723c5d09346cc3f654a390a135127c3e5b660d1b6839ca83ada241cc4e"} Oct 08 19:03:32 crc kubenswrapper[4859]: I1008 19:03:32.282884 4859 generic.go:334] "Generic (PLEG): container finished" podID="f7931881-9829-40f5-9b68-c04957e42476" containerID="edc2df851fe6e07513152f866f5b777cc2382374f21b13b616ad749b21d917b9" exitCode=0 Oct 08 19:03:32 crc kubenswrapper[4859]: I1008 19:03:32.282998 4859 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-mtkqm" event={"ID":"f7931881-9829-40f5-9b68-c04957e42476","Type":"ContainerDied","Data":"edc2df851fe6e07513152f866f5b777cc2382374f21b13b616ad749b21d917b9"} Oct 08 19:03:33 crc kubenswrapper[4859]: I1008 19:03:33.298909 4859 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-mtkqm" event={"ID":"f7931881-9829-40f5-9b68-c04957e42476","Type":"ContainerStarted","Data":"280ce0e1e6978b689d83ef635869dfe05d88e8a1ecdc409f40db409565d6bc9f"} Oct 08 19:03:33 crc kubenswrapper[4859]: I1008 19:03:33.336903 4859 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-mtkqm" podStartSLOduration=2.863184745 podStartE2EDuration="5.336878798s" podCreationTimestamp="2025-10-08 19:03:28 +0000 UTC" firstStartedPulling="2025-10-08 19:03:30.259099489 +0000 UTC m=+2780.505938868" lastFinishedPulling="2025-10-08 19:03:32.732793542 +0000 UTC m=+2782.979632921" observedRunningTime="2025-10-08 19:03:33.325878099 +0000 UTC m=+2783.572717518" watchObservedRunningTime="2025-10-08 19:03:33.336878798 +0000 UTC m=+2783.583718177" Oct 08 19:03:36 crc kubenswrapper[4859]: I1008 19:03:36.325533 4859 generic.go:334] "Generic (PLEG): container finished" podID="65244148-7971-47bf-b900-96932c6cd473" containerID="52d2b1eeb68202f40996e1ab631e4a82637a852b77dd732d27c483a364a58b94" exitCode=0 Oct 08 19:03:36 crc kubenswrapper[4859]: I1008 19:03:36.325618 4859 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-twhvj" event={"ID":"65244148-7971-47bf-b900-96932c6cd473","Type":"ContainerDied","Data":"52d2b1eeb68202f40996e1ab631e4a82637a852b77dd732d27c483a364a58b94"} Oct 08 19:03:37 crc kubenswrapper[4859]: I1008 19:03:37.724540 4859 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-twhvj" Oct 08 19:03:37 crc kubenswrapper[4859]: I1008 19:03:37.896975 4859 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ceilometer-compute-config-data-0\" (UniqueName: \"kubernetes.io/secret/65244148-7971-47bf-b900-96932c6cd473-ceilometer-compute-config-data-0\") pod \"65244148-7971-47bf-b900-96932c6cd473\" (UID: \"65244148-7971-47bf-b900-96932c6cd473\") " Oct 08 19:03:37 crc kubenswrapper[4859]: I1008 19:03:37.897015 4859 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ceilometer-compute-config-data-2\" (UniqueName: \"kubernetes.io/secret/65244148-7971-47bf-b900-96932c6cd473-ceilometer-compute-config-data-2\") pod \"65244148-7971-47bf-b900-96932c6cd473\" (UID: \"65244148-7971-47bf-b900-96932c6cd473\") " Oct 08 19:03:37 crc kubenswrapper[4859]: I1008 19:03:37.897070 4859 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/65244148-7971-47bf-b900-96932c6cd473-ssh-key\") pod \"65244148-7971-47bf-b900-96932c6cd473\" (UID: \"65244148-7971-47bf-b900-96932c6cd473\") " Oct 08 19:03:37 crc kubenswrapper[4859]: I1008 19:03:37.897103 4859 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ceilometer-compute-config-data-1\" (UniqueName: \"kubernetes.io/secret/65244148-7971-47bf-b900-96932c6cd473-ceilometer-compute-config-data-1\") pod \"65244148-7971-47bf-b900-96932c6cd473\" (UID: \"65244148-7971-47bf-b900-96932c6cd473\") " Oct 08 19:03:37 crc kubenswrapper[4859]: I1008 19:03:37.897127 4859 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-8ms4f\" (UniqueName: \"kubernetes.io/projected/65244148-7971-47bf-b900-96932c6cd473-kube-api-access-8ms4f\") pod \"65244148-7971-47bf-b900-96932c6cd473\" (UID: \"65244148-7971-47bf-b900-96932c6cd473\") " Oct 08 19:03:37 crc kubenswrapper[4859]: I1008 19:03:37.897249 4859 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/65244148-7971-47bf-b900-96932c6cd473-inventory\") pod \"65244148-7971-47bf-b900-96932c6cd473\" (UID: \"65244148-7971-47bf-b900-96932c6cd473\") " Oct 08 19:03:37 crc kubenswrapper[4859]: I1008 19:03:37.897302 4859 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"telemetry-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/65244148-7971-47bf-b900-96932c6cd473-telemetry-combined-ca-bundle\") pod \"65244148-7971-47bf-b900-96932c6cd473\" (UID: \"65244148-7971-47bf-b900-96932c6cd473\") " Oct 08 19:03:37 crc kubenswrapper[4859]: I1008 19:03:37.910010 4859 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/65244148-7971-47bf-b900-96932c6cd473-kube-api-access-8ms4f" (OuterVolumeSpecName: "kube-api-access-8ms4f") pod "65244148-7971-47bf-b900-96932c6cd473" (UID: "65244148-7971-47bf-b900-96932c6cd473"). InnerVolumeSpecName "kube-api-access-8ms4f". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 08 19:03:37 crc kubenswrapper[4859]: I1008 19:03:37.913196 4859 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/65244148-7971-47bf-b900-96932c6cd473-telemetry-combined-ca-bundle" (OuterVolumeSpecName: "telemetry-combined-ca-bundle") pod "65244148-7971-47bf-b900-96932c6cd473" (UID: "65244148-7971-47bf-b900-96932c6cd473"). InnerVolumeSpecName "telemetry-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 08 19:03:37 crc kubenswrapper[4859]: I1008 19:03:37.928659 4859 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/65244148-7971-47bf-b900-96932c6cd473-ceilometer-compute-config-data-1" (OuterVolumeSpecName: "ceilometer-compute-config-data-1") pod "65244148-7971-47bf-b900-96932c6cd473" (UID: "65244148-7971-47bf-b900-96932c6cd473"). InnerVolumeSpecName "ceilometer-compute-config-data-1". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 08 19:03:37 crc kubenswrapper[4859]: I1008 19:03:37.929165 4859 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/65244148-7971-47bf-b900-96932c6cd473-ceilometer-compute-config-data-2" (OuterVolumeSpecName: "ceilometer-compute-config-data-2") pod "65244148-7971-47bf-b900-96932c6cd473" (UID: "65244148-7971-47bf-b900-96932c6cd473"). InnerVolumeSpecName "ceilometer-compute-config-data-2". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 08 19:03:37 crc kubenswrapper[4859]: I1008 19:03:37.929527 4859 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/65244148-7971-47bf-b900-96932c6cd473-ceilometer-compute-config-data-0" (OuterVolumeSpecName: "ceilometer-compute-config-data-0") pod "65244148-7971-47bf-b900-96932c6cd473" (UID: "65244148-7971-47bf-b900-96932c6cd473"). InnerVolumeSpecName "ceilometer-compute-config-data-0". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 08 19:03:37 crc kubenswrapper[4859]: I1008 19:03:37.933229 4859 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/65244148-7971-47bf-b900-96932c6cd473-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "65244148-7971-47bf-b900-96932c6cd473" (UID: "65244148-7971-47bf-b900-96932c6cd473"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 08 19:03:37 crc kubenswrapper[4859]: I1008 19:03:37.944283 4859 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/65244148-7971-47bf-b900-96932c6cd473-inventory" (OuterVolumeSpecName: "inventory") pod "65244148-7971-47bf-b900-96932c6cd473" (UID: "65244148-7971-47bf-b900-96932c6cd473"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 08 19:03:37 crc kubenswrapper[4859]: I1008 19:03:37.999955 4859 reconciler_common.go:293] "Volume detached for volume \"ceilometer-compute-config-data-0\" (UniqueName: \"kubernetes.io/secret/65244148-7971-47bf-b900-96932c6cd473-ceilometer-compute-config-data-0\") on node \"crc\" DevicePath \"\"" Oct 08 19:03:37 crc kubenswrapper[4859]: I1008 19:03:37.999985 4859 reconciler_common.go:293] "Volume detached for volume \"ceilometer-compute-config-data-2\" (UniqueName: \"kubernetes.io/secret/65244148-7971-47bf-b900-96932c6cd473-ceilometer-compute-config-data-2\") on node \"crc\" DevicePath \"\"" Oct 08 19:03:38 crc kubenswrapper[4859]: I1008 19:03:37.999995 4859 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/65244148-7971-47bf-b900-96932c6cd473-ssh-key\") on node \"crc\" DevicePath \"\"" Oct 08 19:03:38 crc kubenswrapper[4859]: I1008 19:03:38.000005 4859 reconciler_common.go:293] "Volume detached for volume \"ceilometer-compute-config-data-1\" (UniqueName: \"kubernetes.io/secret/65244148-7971-47bf-b900-96932c6cd473-ceilometer-compute-config-data-1\") on node \"crc\" DevicePath \"\"" Oct 08 19:03:38 crc kubenswrapper[4859]: I1008 19:03:38.000015 4859 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-8ms4f\" (UniqueName: \"kubernetes.io/projected/65244148-7971-47bf-b900-96932c6cd473-kube-api-access-8ms4f\") on node \"crc\" DevicePath \"\"" Oct 08 19:03:38 crc kubenswrapper[4859]: I1008 19:03:38.000025 4859 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/65244148-7971-47bf-b900-96932c6cd473-inventory\") on node \"crc\" DevicePath \"\"" Oct 08 19:03:38 crc kubenswrapper[4859]: I1008 19:03:38.000034 4859 reconciler_common.go:293] "Volume detached for volume \"telemetry-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/65244148-7971-47bf-b900-96932c6cd473-telemetry-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 08 19:03:38 crc kubenswrapper[4859]: I1008 19:03:38.351374 4859 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-twhvj" event={"ID":"65244148-7971-47bf-b900-96932c6cd473","Type":"ContainerDied","Data":"a54589f2121dfea6dd9ba650fa898645fad3cd095578c241c4a6e1e769db4e40"} Oct 08 19:03:38 crc kubenswrapper[4859]: I1008 19:03:38.351443 4859 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-twhvj" Oct 08 19:03:38 crc kubenswrapper[4859]: I1008 19:03:38.351459 4859 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="a54589f2121dfea6dd9ba650fa898645fad3cd095578c241c4a6e1e769db4e40" Oct 08 19:03:39 crc kubenswrapper[4859]: I1008 19:03:39.354203 4859 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-mtkqm" Oct 08 19:03:39 crc kubenswrapper[4859]: I1008 19:03:39.354577 4859 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-mtkqm" Oct 08 19:03:39 crc kubenswrapper[4859]: I1008 19:03:39.422464 4859 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-mtkqm" Oct 08 19:03:40 crc kubenswrapper[4859]: I1008 19:03:40.426475 4859 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-mtkqm" Oct 08 19:03:40 crc kubenswrapper[4859]: I1008 19:03:40.509904 4859 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-mtkqm"] Oct 08 19:03:42 crc kubenswrapper[4859]: I1008 19:03:42.390739 4859 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-operators-mtkqm" podUID="f7931881-9829-40f5-9b68-c04957e42476" containerName="registry-server" containerID="cri-o://280ce0e1e6978b689d83ef635869dfe05d88e8a1ecdc409f40db409565d6bc9f" gracePeriod=2 Oct 08 19:03:43 crc kubenswrapper[4859]: I1008 19:03:43.359803 4859 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-mtkqm" Oct 08 19:03:43 crc kubenswrapper[4859]: I1008 19:03:43.403104 4859 generic.go:334] "Generic (PLEG): container finished" podID="f7931881-9829-40f5-9b68-c04957e42476" containerID="280ce0e1e6978b689d83ef635869dfe05d88e8a1ecdc409f40db409565d6bc9f" exitCode=0 Oct 08 19:03:43 crc kubenswrapper[4859]: I1008 19:03:43.403161 4859 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-mtkqm" Oct 08 19:03:43 crc kubenswrapper[4859]: I1008 19:03:43.403163 4859 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-mtkqm" event={"ID":"f7931881-9829-40f5-9b68-c04957e42476","Type":"ContainerDied","Data":"280ce0e1e6978b689d83ef635869dfe05d88e8a1ecdc409f40db409565d6bc9f"} Oct 08 19:03:43 crc kubenswrapper[4859]: I1008 19:03:43.403240 4859 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-mtkqm" event={"ID":"f7931881-9829-40f5-9b68-c04957e42476","Type":"ContainerDied","Data":"40e475723c5d09346cc3f654a390a135127c3e5b660d1b6839ca83ada241cc4e"} Oct 08 19:03:43 crc kubenswrapper[4859]: I1008 19:03:43.403269 4859 scope.go:117] "RemoveContainer" containerID="280ce0e1e6978b689d83ef635869dfe05d88e8a1ecdc409f40db409565d6bc9f" Oct 08 19:03:43 crc kubenswrapper[4859]: I1008 19:03:43.438437 4859 scope.go:117] "RemoveContainer" containerID="edc2df851fe6e07513152f866f5b777cc2382374f21b13b616ad749b21d917b9" Oct 08 19:03:43 crc kubenswrapper[4859]: I1008 19:03:43.478521 4859 scope.go:117] "RemoveContainer" containerID="2e0d37cac2fd94803e0475b396e79a82f1d1ce81ec7fc4df61c7980b85bee134" Oct 08 19:03:43 crc kubenswrapper[4859]: I1008 19:03:43.505959 4859 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/f7931881-9829-40f5-9b68-c04957e42476-utilities\") pod \"f7931881-9829-40f5-9b68-c04957e42476\" (UID: \"f7931881-9829-40f5-9b68-c04957e42476\") " Oct 08 19:03:43 crc kubenswrapper[4859]: I1008 19:03:43.506058 4859 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-fwfd2\" (UniqueName: \"kubernetes.io/projected/f7931881-9829-40f5-9b68-c04957e42476-kube-api-access-fwfd2\") pod \"f7931881-9829-40f5-9b68-c04957e42476\" (UID: \"f7931881-9829-40f5-9b68-c04957e42476\") " Oct 08 19:03:43 crc kubenswrapper[4859]: I1008 19:03:43.506226 4859 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/f7931881-9829-40f5-9b68-c04957e42476-catalog-content\") pod \"f7931881-9829-40f5-9b68-c04957e42476\" (UID: \"f7931881-9829-40f5-9b68-c04957e42476\") " Oct 08 19:03:43 crc kubenswrapper[4859]: I1008 19:03:43.507164 4859 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/f7931881-9829-40f5-9b68-c04957e42476-utilities" (OuterVolumeSpecName: "utilities") pod "f7931881-9829-40f5-9b68-c04957e42476" (UID: "f7931881-9829-40f5-9b68-c04957e42476"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 08 19:03:43 crc kubenswrapper[4859]: I1008 19:03:43.512198 4859 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/f7931881-9829-40f5-9b68-c04957e42476-kube-api-access-fwfd2" (OuterVolumeSpecName: "kube-api-access-fwfd2") pod "f7931881-9829-40f5-9b68-c04957e42476" (UID: "f7931881-9829-40f5-9b68-c04957e42476"). InnerVolumeSpecName "kube-api-access-fwfd2". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 08 19:03:43 crc kubenswrapper[4859]: I1008 19:03:43.514961 4859 scope.go:117] "RemoveContainer" containerID="280ce0e1e6978b689d83ef635869dfe05d88e8a1ecdc409f40db409565d6bc9f" Oct 08 19:03:43 crc kubenswrapper[4859]: E1008 19:03:43.515449 4859 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"280ce0e1e6978b689d83ef635869dfe05d88e8a1ecdc409f40db409565d6bc9f\": container with ID starting with 280ce0e1e6978b689d83ef635869dfe05d88e8a1ecdc409f40db409565d6bc9f not found: ID does not exist" containerID="280ce0e1e6978b689d83ef635869dfe05d88e8a1ecdc409f40db409565d6bc9f" Oct 08 19:03:43 crc kubenswrapper[4859]: I1008 19:03:43.515573 4859 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"280ce0e1e6978b689d83ef635869dfe05d88e8a1ecdc409f40db409565d6bc9f"} err="failed to get container status \"280ce0e1e6978b689d83ef635869dfe05d88e8a1ecdc409f40db409565d6bc9f\": rpc error: code = NotFound desc = could not find container \"280ce0e1e6978b689d83ef635869dfe05d88e8a1ecdc409f40db409565d6bc9f\": container with ID starting with 280ce0e1e6978b689d83ef635869dfe05d88e8a1ecdc409f40db409565d6bc9f not found: ID does not exist" Oct 08 19:03:43 crc kubenswrapper[4859]: I1008 19:03:43.515613 4859 scope.go:117] "RemoveContainer" containerID="edc2df851fe6e07513152f866f5b777cc2382374f21b13b616ad749b21d917b9" Oct 08 19:03:43 crc kubenswrapper[4859]: E1008 19:03:43.516333 4859 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"edc2df851fe6e07513152f866f5b777cc2382374f21b13b616ad749b21d917b9\": container with ID starting with edc2df851fe6e07513152f866f5b777cc2382374f21b13b616ad749b21d917b9 not found: ID does not exist" containerID="edc2df851fe6e07513152f866f5b777cc2382374f21b13b616ad749b21d917b9" Oct 08 19:03:43 crc kubenswrapper[4859]: I1008 19:03:43.516363 4859 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"edc2df851fe6e07513152f866f5b777cc2382374f21b13b616ad749b21d917b9"} err="failed to get container status \"edc2df851fe6e07513152f866f5b777cc2382374f21b13b616ad749b21d917b9\": rpc error: code = NotFound desc = could not find container \"edc2df851fe6e07513152f866f5b777cc2382374f21b13b616ad749b21d917b9\": container with ID starting with edc2df851fe6e07513152f866f5b777cc2382374f21b13b616ad749b21d917b9 not found: ID does not exist" Oct 08 19:03:43 crc kubenswrapper[4859]: I1008 19:03:43.516382 4859 scope.go:117] "RemoveContainer" containerID="2e0d37cac2fd94803e0475b396e79a82f1d1ce81ec7fc4df61c7980b85bee134" Oct 08 19:03:43 crc kubenswrapper[4859]: E1008 19:03:43.516610 4859 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"2e0d37cac2fd94803e0475b396e79a82f1d1ce81ec7fc4df61c7980b85bee134\": container with ID starting with 2e0d37cac2fd94803e0475b396e79a82f1d1ce81ec7fc4df61c7980b85bee134 not found: ID does not exist" containerID="2e0d37cac2fd94803e0475b396e79a82f1d1ce81ec7fc4df61c7980b85bee134" Oct 08 19:03:43 crc kubenswrapper[4859]: I1008 19:03:43.516633 4859 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"2e0d37cac2fd94803e0475b396e79a82f1d1ce81ec7fc4df61c7980b85bee134"} err="failed to get container status \"2e0d37cac2fd94803e0475b396e79a82f1d1ce81ec7fc4df61c7980b85bee134\": rpc error: code = NotFound desc = could not find container \"2e0d37cac2fd94803e0475b396e79a82f1d1ce81ec7fc4df61c7980b85bee134\": container with ID starting with 2e0d37cac2fd94803e0475b396e79a82f1d1ce81ec7fc4df61c7980b85bee134 not found: ID does not exist" Oct 08 19:03:43 crc kubenswrapper[4859]: I1008 19:03:43.594244 4859 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/f7931881-9829-40f5-9b68-c04957e42476-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "f7931881-9829-40f5-9b68-c04957e42476" (UID: "f7931881-9829-40f5-9b68-c04957e42476"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 08 19:03:43 crc kubenswrapper[4859]: I1008 19:03:43.610157 4859 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/f7931881-9829-40f5-9b68-c04957e42476-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 08 19:03:43 crc kubenswrapper[4859]: I1008 19:03:43.610408 4859 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/f7931881-9829-40f5-9b68-c04957e42476-utilities\") on node \"crc\" DevicePath \"\"" Oct 08 19:03:43 crc kubenswrapper[4859]: I1008 19:03:43.610495 4859 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-fwfd2\" (UniqueName: \"kubernetes.io/projected/f7931881-9829-40f5-9b68-c04957e42476-kube-api-access-fwfd2\") on node \"crc\" DevicePath \"\"" Oct 08 19:03:43 crc kubenswrapper[4859]: I1008 19:03:43.757397 4859 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-mtkqm"] Oct 08 19:03:43 crc kubenswrapper[4859]: I1008 19:03:43.763763 4859 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-operators-mtkqm"] Oct 08 19:03:44 crc kubenswrapper[4859]: I1008 19:03:44.482518 4859 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="f7931881-9829-40f5-9b68-c04957e42476" path="/var/lib/kubelet/pods/f7931881-9829-40f5-9b68-c04957e42476/volumes" Oct 08 19:04:23 crc kubenswrapper[4859]: I1008 19:04:23.631920 4859 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/tempest-tests-tempest"] Oct 08 19:04:23 crc kubenswrapper[4859]: E1008 19:04:23.633015 4859 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="65244148-7971-47bf-b900-96932c6cd473" containerName="telemetry-edpm-deployment-openstack-edpm-ipam" Oct 08 19:04:23 crc kubenswrapper[4859]: I1008 19:04:23.633035 4859 state_mem.go:107] "Deleted CPUSet assignment" podUID="65244148-7971-47bf-b900-96932c6cd473" containerName="telemetry-edpm-deployment-openstack-edpm-ipam" Oct 08 19:04:23 crc kubenswrapper[4859]: E1008 19:04:23.633079 4859 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f7931881-9829-40f5-9b68-c04957e42476" containerName="registry-server" Oct 08 19:04:23 crc kubenswrapper[4859]: I1008 19:04:23.633087 4859 state_mem.go:107] "Deleted CPUSet assignment" podUID="f7931881-9829-40f5-9b68-c04957e42476" containerName="registry-server" Oct 08 19:04:23 crc kubenswrapper[4859]: E1008 19:04:23.633109 4859 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f7931881-9829-40f5-9b68-c04957e42476" containerName="extract-utilities" Oct 08 19:04:23 crc kubenswrapper[4859]: I1008 19:04:23.633116 4859 state_mem.go:107] "Deleted CPUSet assignment" podUID="f7931881-9829-40f5-9b68-c04957e42476" containerName="extract-utilities" Oct 08 19:04:23 crc kubenswrapper[4859]: E1008 19:04:23.633134 4859 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f7931881-9829-40f5-9b68-c04957e42476" containerName="extract-content" Oct 08 19:04:23 crc kubenswrapper[4859]: I1008 19:04:23.633142 4859 state_mem.go:107] "Deleted CPUSet assignment" podUID="f7931881-9829-40f5-9b68-c04957e42476" containerName="extract-content" Oct 08 19:04:23 crc kubenswrapper[4859]: I1008 19:04:23.633378 4859 memory_manager.go:354] "RemoveStaleState removing state" podUID="65244148-7971-47bf-b900-96932c6cd473" containerName="telemetry-edpm-deployment-openstack-edpm-ipam" Oct 08 19:04:23 crc kubenswrapper[4859]: I1008 19:04:23.633407 4859 memory_manager.go:354] "RemoveStaleState removing state" podUID="f7931881-9829-40f5-9b68-c04957e42476" containerName="registry-server" Oct 08 19:04:23 crc kubenswrapper[4859]: I1008 19:04:23.634237 4859 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/tempest-tests-tempest" Oct 08 19:04:23 crc kubenswrapper[4859]: I1008 19:04:23.639756 4859 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"default-dockercfg-mr7g8" Oct 08 19:04:23 crc kubenswrapper[4859]: I1008 19:04:23.639835 4859 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"test-operator-controller-priv-key" Oct 08 19:04:23 crc kubenswrapper[4859]: I1008 19:04:23.640347 4859 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"tempest-tests-tempest-env-vars-s0" Oct 08 19:04:23 crc kubenswrapper[4859]: I1008 19:04:23.640632 4859 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"tempest-tests-tempest-custom-data-s0" Oct 08 19:04:23 crc kubenswrapper[4859]: I1008 19:04:23.665433 4859 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/tempest-tests-tempest"] Oct 08 19:04:23 crc kubenswrapper[4859]: I1008 19:04:23.792927 4859 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/be909489-e6c5-4464-9292-aacd831885d7-openstack-config\") pod \"tempest-tests-tempest\" (UID: \"be909489-e6c5-4464-9292-aacd831885d7\") " pod="openstack/tempest-tests-tempest" Oct 08 19:04:23 crc kubenswrapper[4859]: I1008 19:04:23.792997 4859 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage11-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage11-crc\") pod \"tempest-tests-tempest\" (UID: \"be909489-e6c5-4464-9292-aacd831885d7\") " pod="openstack/tempest-tests-tempest" Oct 08 19:04:23 crc kubenswrapper[4859]: I1008 19:04:23.793020 4859 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ca-certs\" (UniqueName: \"kubernetes.io/secret/be909489-e6c5-4464-9292-aacd831885d7-ca-certs\") pod \"tempest-tests-tempest\" (UID: \"be909489-e6c5-4464-9292-aacd831885d7\") " pod="openstack/tempest-tests-tempest" Oct 08 19:04:23 crc kubenswrapper[4859]: I1008 19:04:23.793227 4859 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"test-operator-ephemeral-workdir\" (UniqueName: \"kubernetes.io/empty-dir/be909489-e6c5-4464-9292-aacd831885d7-test-operator-ephemeral-workdir\") pod \"tempest-tests-tempest\" (UID: \"be909489-e6c5-4464-9292-aacd831885d7\") " pod="openstack/tempest-tests-tempest" Oct 08 19:04:23 crc kubenswrapper[4859]: I1008 19:04:23.793361 4859 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-8psds\" (UniqueName: \"kubernetes.io/projected/be909489-e6c5-4464-9292-aacd831885d7-kube-api-access-8psds\") pod \"tempest-tests-tempest\" (UID: \"be909489-e6c5-4464-9292-aacd831885d7\") " pod="openstack/tempest-tests-tempest" Oct 08 19:04:23 crc kubenswrapper[4859]: I1008 19:04:23.793417 4859 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/be909489-e6c5-4464-9292-aacd831885d7-ssh-key\") pod \"tempest-tests-tempest\" (UID: \"be909489-e6c5-4464-9292-aacd831885d7\") " pod="openstack/tempest-tests-tempest" Oct 08 19:04:23 crc kubenswrapper[4859]: I1008 19:04:23.793761 4859 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"test-operator-ephemeral-temporary\" (UniqueName: \"kubernetes.io/empty-dir/be909489-e6c5-4464-9292-aacd831885d7-test-operator-ephemeral-temporary\") pod \"tempest-tests-tempest\" (UID: \"be909489-e6c5-4464-9292-aacd831885d7\") " pod="openstack/tempest-tests-tempest" Oct 08 19:04:23 crc kubenswrapper[4859]: I1008 19:04:23.793913 4859 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/be909489-e6c5-4464-9292-aacd831885d7-openstack-config-secret\") pod \"tempest-tests-tempest\" (UID: \"be909489-e6c5-4464-9292-aacd831885d7\") " pod="openstack/tempest-tests-tempest" Oct 08 19:04:23 crc kubenswrapper[4859]: I1008 19:04:23.793954 4859 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/be909489-e6c5-4464-9292-aacd831885d7-config-data\") pod \"tempest-tests-tempest\" (UID: \"be909489-e6c5-4464-9292-aacd831885d7\") " pod="openstack/tempest-tests-tempest" Oct 08 19:04:23 crc kubenswrapper[4859]: I1008 19:04:23.895622 4859 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"test-operator-ephemeral-temporary\" (UniqueName: \"kubernetes.io/empty-dir/be909489-e6c5-4464-9292-aacd831885d7-test-operator-ephemeral-temporary\") pod \"tempest-tests-tempest\" (UID: \"be909489-e6c5-4464-9292-aacd831885d7\") " pod="openstack/tempest-tests-tempest" Oct 08 19:04:23 crc kubenswrapper[4859]: I1008 19:04:23.895725 4859 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/be909489-e6c5-4464-9292-aacd831885d7-openstack-config-secret\") pod \"tempest-tests-tempest\" (UID: \"be909489-e6c5-4464-9292-aacd831885d7\") " pod="openstack/tempest-tests-tempest" Oct 08 19:04:23 crc kubenswrapper[4859]: I1008 19:04:23.895751 4859 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/be909489-e6c5-4464-9292-aacd831885d7-config-data\") pod \"tempest-tests-tempest\" (UID: \"be909489-e6c5-4464-9292-aacd831885d7\") " pod="openstack/tempest-tests-tempest" Oct 08 19:04:23 crc kubenswrapper[4859]: I1008 19:04:23.895810 4859 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/be909489-e6c5-4464-9292-aacd831885d7-openstack-config\") pod \"tempest-tests-tempest\" (UID: \"be909489-e6c5-4464-9292-aacd831885d7\") " pod="openstack/tempest-tests-tempest" Oct 08 19:04:23 crc kubenswrapper[4859]: I1008 19:04:23.895851 4859 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage11-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage11-crc\") pod \"tempest-tests-tempest\" (UID: \"be909489-e6c5-4464-9292-aacd831885d7\") " pod="openstack/tempest-tests-tempest" Oct 08 19:04:23 crc kubenswrapper[4859]: I1008 19:04:23.895875 4859 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ca-certs\" (UniqueName: \"kubernetes.io/secret/be909489-e6c5-4464-9292-aacd831885d7-ca-certs\") pod \"tempest-tests-tempest\" (UID: \"be909489-e6c5-4464-9292-aacd831885d7\") " pod="openstack/tempest-tests-tempest" Oct 08 19:04:23 crc kubenswrapper[4859]: I1008 19:04:23.895900 4859 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"test-operator-ephemeral-workdir\" (UniqueName: \"kubernetes.io/empty-dir/be909489-e6c5-4464-9292-aacd831885d7-test-operator-ephemeral-workdir\") pod \"tempest-tests-tempest\" (UID: \"be909489-e6c5-4464-9292-aacd831885d7\") " pod="openstack/tempest-tests-tempest" Oct 08 19:04:23 crc kubenswrapper[4859]: I1008 19:04:23.895930 4859 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-8psds\" (UniqueName: \"kubernetes.io/projected/be909489-e6c5-4464-9292-aacd831885d7-kube-api-access-8psds\") pod \"tempest-tests-tempest\" (UID: \"be909489-e6c5-4464-9292-aacd831885d7\") " pod="openstack/tempest-tests-tempest" Oct 08 19:04:23 crc kubenswrapper[4859]: I1008 19:04:23.895952 4859 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/be909489-e6c5-4464-9292-aacd831885d7-ssh-key\") pod \"tempest-tests-tempest\" (UID: \"be909489-e6c5-4464-9292-aacd831885d7\") " pod="openstack/tempest-tests-tempest" Oct 08 19:04:23 crc kubenswrapper[4859]: I1008 19:04:23.896342 4859 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage11-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage11-crc\") pod \"tempest-tests-tempest\" (UID: \"be909489-e6c5-4464-9292-aacd831885d7\") device mount path \"/mnt/openstack/pv11\"" pod="openstack/tempest-tests-tempest" Oct 08 19:04:23 crc kubenswrapper[4859]: I1008 19:04:23.897143 4859 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"test-operator-ephemeral-workdir\" (UniqueName: \"kubernetes.io/empty-dir/be909489-e6c5-4464-9292-aacd831885d7-test-operator-ephemeral-workdir\") pod \"tempest-tests-tempest\" (UID: \"be909489-e6c5-4464-9292-aacd831885d7\") " pod="openstack/tempest-tests-tempest" Oct 08 19:04:23 crc kubenswrapper[4859]: I1008 19:04:23.897671 4859 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"test-operator-ephemeral-temporary\" (UniqueName: \"kubernetes.io/empty-dir/be909489-e6c5-4464-9292-aacd831885d7-test-operator-ephemeral-temporary\") pod \"tempest-tests-tempest\" (UID: \"be909489-e6c5-4464-9292-aacd831885d7\") " pod="openstack/tempest-tests-tempest" Oct 08 19:04:23 crc kubenswrapper[4859]: I1008 19:04:23.898232 4859 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/be909489-e6c5-4464-9292-aacd831885d7-config-data\") pod \"tempest-tests-tempest\" (UID: \"be909489-e6c5-4464-9292-aacd831885d7\") " pod="openstack/tempest-tests-tempest" Oct 08 19:04:23 crc kubenswrapper[4859]: I1008 19:04:23.898339 4859 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/be909489-e6c5-4464-9292-aacd831885d7-openstack-config\") pod \"tempest-tests-tempest\" (UID: \"be909489-e6c5-4464-9292-aacd831885d7\") " pod="openstack/tempest-tests-tempest" Oct 08 19:04:23 crc kubenswrapper[4859]: I1008 19:04:23.904407 4859 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/be909489-e6c5-4464-9292-aacd831885d7-ssh-key\") pod \"tempest-tests-tempest\" (UID: \"be909489-e6c5-4464-9292-aacd831885d7\") " pod="openstack/tempest-tests-tempest" Oct 08 19:04:23 crc kubenswrapper[4859]: I1008 19:04:23.905599 4859 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/be909489-e6c5-4464-9292-aacd831885d7-openstack-config-secret\") pod \"tempest-tests-tempest\" (UID: \"be909489-e6c5-4464-9292-aacd831885d7\") " pod="openstack/tempest-tests-tempest" Oct 08 19:04:23 crc kubenswrapper[4859]: I1008 19:04:23.908393 4859 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ca-certs\" (UniqueName: \"kubernetes.io/secret/be909489-e6c5-4464-9292-aacd831885d7-ca-certs\") pod \"tempest-tests-tempest\" (UID: \"be909489-e6c5-4464-9292-aacd831885d7\") " pod="openstack/tempest-tests-tempest" Oct 08 19:04:23 crc kubenswrapper[4859]: I1008 19:04:23.926596 4859 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-8psds\" (UniqueName: \"kubernetes.io/projected/be909489-e6c5-4464-9292-aacd831885d7-kube-api-access-8psds\") pod \"tempest-tests-tempest\" (UID: \"be909489-e6c5-4464-9292-aacd831885d7\") " pod="openstack/tempest-tests-tempest" Oct 08 19:04:23 crc kubenswrapper[4859]: I1008 19:04:23.934907 4859 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage11-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage11-crc\") pod \"tempest-tests-tempest\" (UID: \"be909489-e6c5-4464-9292-aacd831885d7\") " pod="openstack/tempest-tests-tempest" Oct 08 19:04:23 crc kubenswrapper[4859]: I1008 19:04:23.974502 4859 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/tempest-tests-tempest" Oct 08 19:04:24 crc kubenswrapper[4859]: I1008 19:04:24.433679 4859 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/tempest-tests-tempest"] Oct 08 19:04:24 crc kubenswrapper[4859]: I1008 19:04:24.853174 4859 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/tempest-tests-tempest" event={"ID":"be909489-e6c5-4464-9292-aacd831885d7","Type":"ContainerStarted","Data":"2ec230b487b50f18cb8c914ed6316cb461e578fece9aa8b7a2ea76eaf15045de"} Oct 08 19:04:47 crc kubenswrapper[4859]: I1008 19:04:47.924521 4859 patch_prober.go:28] interesting pod/machine-config-daemon-82s52 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 08 19:04:47 crc kubenswrapper[4859]: I1008 19:04:47.925073 4859 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-82s52" podUID="b23a6a6c-9d92-4e7b-840e-55cfda873a2d" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 08 19:04:48 crc kubenswrapper[4859]: E1008 19:04:48.415549 4859 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/podified-antelope-centos9/openstack-tempest-all:current-podified" Oct 08 19:04:48 crc kubenswrapper[4859]: E1008 19:04:48.415873 4859 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:tempest-tests-tempest-tests-runner,Image:quay.io/podified-antelope-centos9/openstack-tempest-all:current-podified,Command:[],Args:[],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:test-operator-ephemeral-workdir,ReadOnly:false,MountPath:/var/lib/tempest,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:test-operator-ephemeral-temporary,ReadOnly:false,MountPath:/tmp,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:config-data,ReadOnly:false,MountPath:/etc/test_operator,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:test-operator-logs,ReadOnly:false,MountPath:/var/lib/tempest/external_files,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:openstack-config,ReadOnly:true,MountPath:/etc/openstack/clouds.yaml,SubPath:clouds.yaml,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:openstack-config,ReadOnly:true,MountPath:/var/lib/tempest/.config/openstack/clouds.yaml,SubPath:clouds.yaml,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:openstack-config-secret,ReadOnly:false,MountPath:/etc/openstack/secure.yaml,SubPath:secure.yaml,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:ca-certs,ReadOnly:true,MountPath:/etc/pki/ca-trust/extracted/pem/tls-ca-bundle.pem,SubPath:tls-ca-bundle.pem,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:ssh-key,ReadOnly:false,MountPath:/var/lib/tempest/id_ecdsa,SubPath:ssh_key,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-8psds,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*42480,RunAsNonRoot:*false,ReadOnlyRootFilesystem:*false,AllowPrivilegeEscalation:*true,RunAsGroup:*42480,ProcMount:nil,WindowsOptions:nil,SeccompProfile:&SeccompProfile{Type:RuntimeDefault,LocalhostProfile:nil,},AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{EnvFromSource{Prefix:,ConfigMapRef:&ConfigMapEnvSource{LocalObjectReference:LocalObjectReference{Name:tempest-tests-tempest-custom-data-s0,},Optional:nil,},SecretRef:nil,},EnvFromSource{Prefix:,ConfigMapRef:&ConfigMapEnvSource{LocalObjectReference:LocalObjectReference{Name:tempest-tests-tempest-env-vars-s0,},Optional:nil,},SecretRef:nil,},},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod tempest-tests-tempest_openstack(be909489-e6c5-4464-9292-aacd831885d7): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Oct 08 19:04:48 crc kubenswrapper[4859]: E1008 19:04:48.417076 4859 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"tempest-tests-tempest-tests-runner\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack/tempest-tests-tempest" podUID="be909489-e6c5-4464-9292-aacd831885d7" Oct 08 19:04:49 crc kubenswrapper[4859]: E1008 19:04:49.076212 4859 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"tempest-tests-tempest-tests-runner\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/podified-antelope-centos9/openstack-tempest-all:current-podified\\\"\"" pod="openstack/tempest-tests-tempest" podUID="be909489-e6c5-4464-9292-aacd831885d7" Oct 08 19:05:02 crc kubenswrapper[4859]: I1008 19:05:02.207234 4859 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/tempest-tests-tempest" event={"ID":"be909489-e6c5-4464-9292-aacd831885d7","Type":"ContainerStarted","Data":"139da404a051f6e011fc363bac25a65ad6b6fd39197c4b1e2be03afe722fa5ba"} Oct 08 19:05:02 crc kubenswrapper[4859]: I1008 19:05:02.230550 4859 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/tempest-tests-tempest" podStartSLOduration=3.695895195 podStartE2EDuration="40.230532618s" podCreationTimestamp="2025-10-08 19:04:22 +0000 UTC" firstStartedPulling="2025-10-08 19:04:24.46414485 +0000 UTC m=+2834.710984239" lastFinishedPulling="2025-10-08 19:05:00.998782283 +0000 UTC m=+2871.245621662" observedRunningTime="2025-10-08 19:05:02.223346909 +0000 UTC m=+2872.470186288" watchObservedRunningTime="2025-10-08 19:05:02.230532618 +0000 UTC m=+2872.477371997" Oct 08 19:05:17 crc kubenswrapper[4859]: I1008 19:05:17.924359 4859 patch_prober.go:28] interesting pod/machine-config-daemon-82s52 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 08 19:05:17 crc kubenswrapper[4859]: I1008 19:05:17.924931 4859 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-82s52" podUID="b23a6a6c-9d92-4e7b-840e-55cfda873a2d" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 08 19:05:47 crc kubenswrapper[4859]: I1008 19:05:47.924306 4859 patch_prober.go:28] interesting pod/machine-config-daemon-82s52 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 08 19:05:47 crc kubenswrapper[4859]: I1008 19:05:47.924884 4859 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-82s52" podUID="b23a6a6c-9d92-4e7b-840e-55cfda873a2d" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 08 19:05:47 crc kubenswrapper[4859]: I1008 19:05:47.924935 4859 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-82s52" Oct 08 19:05:47 crc kubenswrapper[4859]: I1008 19:05:47.925657 4859 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"a3d24a0dfcbf833b7bcf1af5e19a45816904f485fab5db993921d32e47607790"} pod="openshift-machine-config-operator/machine-config-daemon-82s52" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Oct 08 19:05:47 crc kubenswrapper[4859]: I1008 19:05:47.925740 4859 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-82s52" podUID="b23a6a6c-9d92-4e7b-840e-55cfda873a2d" containerName="machine-config-daemon" containerID="cri-o://a3d24a0dfcbf833b7bcf1af5e19a45816904f485fab5db993921d32e47607790" gracePeriod=600 Oct 08 19:05:48 crc kubenswrapper[4859]: I1008 19:05:48.669344 4859 generic.go:334] "Generic (PLEG): container finished" podID="b23a6a6c-9d92-4e7b-840e-55cfda873a2d" containerID="a3d24a0dfcbf833b7bcf1af5e19a45816904f485fab5db993921d32e47607790" exitCode=0 Oct 08 19:05:48 crc kubenswrapper[4859]: I1008 19:05:48.669416 4859 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-82s52" event={"ID":"b23a6a6c-9d92-4e7b-840e-55cfda873a2d","Type":"ContainerDied","Data":"a3d24a0dfcbf833b7bcf1af5e19a45816904f485fab5db993921d32e47607790"} Oct 08 19:05:48 crc kubenswrapper[4859]: I1008 19:05:48.669703 4859 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-82s52" event={"ID":"b23a6a6c-9d92-4e7b-840e-55cfda873a2d","Type":"ContainerStarted","Data":"cc836fd09088ba4df38f4588526d6fe50b7b131a7fb454b4fd3751d8b1b22560"} Oct 08 19:05:48 crc kubenswrapper[4859]: I1008 19:05:48.669728 4859 scope.go:117] "RemoveContainer" containerID="265e61e43a0f4edee3adcaf87995c3c44fad0b135f61f14030c7ee297b24d748" Oct 08 19:06:11 crc kubenswrapper[4859]: I1008 19:06:11.603073 4859 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-4z8nk"] Oct 08 19:06:11 crc kubenswrapper[4859]: I1008 19:06:11.607221 4859 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-4z8nk" Oct 08 19:06:11 crc kubenswrapper[4859]: I1008 19:06:11.626676 4859 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-4z8nk"] Oct 08 19:06:11 crc kubenswrapper[4859]: I1008 19:06:11.652202 4859 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-dl4xg\" (UniqueName: \"kubernetes.io/projected/42cbb501-291f-4dd2-8373-f80ebf57bdf2-kube-api-access-dl4xg\") pod \"redhat-marketplace-4z8nk\" (UID: \"42cbb501-291f-4dd2-8373-f80ebf57bdf2\") " pod="openshift-marketplace/redhat-marketplace-4z8nk" Oct 08 19:06:11 crc kubenswrapper[4859]: I1008 19:06:11.652414 4859 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/42cbb501-291f-4dd2-8373-f80ebf57bdf2-catalog-content\") pod \"redhat-marketplace-4z8nk\" (UID: \"42cbb501-291f-4dd2-8373-f80ebf57bdf2\") " pod="openshift-marketplace/redhat-marketplace-4z8nk" Oct 08 19:06:11 crc kubenswrapper[4859]: I1008 19:06:11.652538 4859 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/42cbb501-291f-4dd2-8373-f80ebf57bdf2-utilities\") pod \"redhat-marketplace-4z8nk\" (UID: \"42cbb501-291f-4dd2-8373-f80ebf57bdf2\") " pod="openshift-marketplace/redhat-marketplace-4z8nk" Oct 08 19:06:11 crc kubenswrapper[4859]: I1008 19:06:11.755067 4859 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/42cbb501-291f-4dd2-8373-f80ebf57bdf2-catalog-content\") pod \"redhat-marketplace-4z8nk\" (UID: \"42cbb501-291f-4dd2-8373-f80ebf57bdf2\") " pod="openshift-marketplace/redhat-marketplace-4z8nk" Oct 08 19:06:11 crc kubenswrapper[4859]: I1008 19:06:11.755420 4859 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/42cbb501-291f-4dd2-8373-f80ebf57bdf2-utilities\") pod \"redhat-marketplace-4z8nk\" (UID: \"42cbb501-291f-4dd2-8373-f80ebf57bdf2\") " pod="openshift-marketplace/redhat-marketplace-4z8nk" Oct 08 19:06:11 crc kubenswrapper[4859]: I1008 19:06:11.755550 4859 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-dl4xg\" (UniqueName: \"kubernetes.io/projected/42cbb501-291f-4dd2-8373-f80ebf57bdf2-kube-api-access-dl4xg\") pod \"redhat-marketplace-4z8nk\" (UID: \"42cbb501-291f-4dd2-8373-f80ebf57bdf2\") " pod="openshift-marketplace/redhat-marketplace-4z8nk" Oct 08 19:06:11 crc kubenswrapper[4859]: I1008 19:06:11.755619 4859 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/42cbb501-291f-4dd2-8373-f80ebf57bdf2-catalog-content\") pod \"redhat-marketplace-4z8nk\" (UID: \"42cbb501-291f-4dd2-8373-f80ebf57bdf2\") " pod="openshift-marketplace/redhat-marketplace-4z8nk" Oct 08 19:06:11 crc kubenswrapper[4859]: I1008 19:06:11.756029 4859 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/42cbb501-291f-4dd2-8373-f80ebf57bdf2-utilities\") pod \"redhat-marketplace-4z8nk\" (UID: \"42cbb501-291f-4dd2-8373-f80ebf57bdf2\") " pod="openshift-marketplace/redhat-marketplace-4z8nk" Oct 08 19:06:11 crc kubenswrapper[4859]: I1008 19:06:11.774631 4859 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-dl4xg\" (UniqueName: \"kubernetes.io/projected/42cbb501-291f-4dd2-8373-f80ebf57bdf2-kube-api-access-dl4xg\") pod \"redhat-marketplace-4z8nk\" (UID: \"42cbb501-291f-4dd2-8373-f80ebf57bdf2\") " pod="openshift-marketplace/redhat-marketplace-4z8nk" Oct 08 19:06:11 crc kubenswrapper[4859]: I1008 19:06:11.951329 4859 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-4z8nk" Oct 08 19:06:12 crc kubenswrapper[4859]: I1008 19:06:12.411421 4859 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-4z8nk"] Oct 08 19:06:12 crc kubenswrapper[4859]: I1008 19:06:12.927621 4859 generic.go:334] "Generic (PLEG): container finished" podID="42cbb501-291f-4dd2-8373-f80ebf57bdf2" containerID="27be93e86b3a217363cbba555c104e0efbc5c2a75aed4235788b5985048c3c76" exitCode=0 Oct 08 19:06:12 crc kubenswrapper[4859]: I1008 19:06:12.927992 4859 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-4z8nk" event={"ID":"42cbb501-291f-4dd2-8373-f80ebf57bdf2","Type":"ContainerDied","Data":"27be93e86b3a217363cbba555c104e0efbc5c2a75aed4235788b5985048c3c76"} Oct 08 19:06:12 crc kubenswrapper[4859]: I1008 19:06:12.928021 4859 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-4z8nk" event={"ID":"42cbb501-291f-4dd2-8373-f80ebf57bdf2","Type":"ContainerStarted","Data":"c92ae6b9863bac217baf03a6ad6624d19161f5478405de02b092c977d0ca044a"} Oct 08 19:06:12 crc kubenswrapper[4859]: I1008 19:06:12.929860 4859 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Oct 08 19:06:13 crc kubenswrapper[4859]: I1008 19:06:13.939086 4859 generic.go:334] "Generic (PLEG): container finished" podID="42cbb501-291f-4dd2-8373-f80ebf57bdf2" containerID="8740e1e27bb44aeec068579b52e081d0907e4f78e0a7af901e3689b603f55303" exitCode=0 Oct 08 19:06:13 crc kubenswrapper[4859]: I1008 19:06:13.939385 4859 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-4z8nk" event={"ID":"42cbb501-291f-4dd2-8373-f80ebf57bdf2","Type":"ContainerDied","Data":"8740e1e27bb44aeec068579b52e081d0907e4f78e0a7af901e3689b603f55303"} Oct 08 19:06:14 crc kubenswrapper[4859]: I1008 19:06:14.950416 4859 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-4z8nk" event={"ID":"42cbb501-291f-4dd2-8373-f80ebf57bdf2","Type":"ContainerStarted","Data":"44be98a1b71dab290021207cc3e4634b6a11ac84bdf55e804ffdf64699fb2ad8"} Oct 08 19:06:14 crc kubenswrapper[4859]: I1008 19:06:14.976428 4859 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-4z8nk" podStartSLOduration=2.444233411 podStartE2EDuration="3.976409196s" podCreationTimestamp="2025-10-08 19:06:11 +0000 UTC" firstStartedPulling="2025-10-08 19:06:12.929561165 +0000 UTC m=+2943.176400544" lastFinishedPulling="2025-10-08 19:06:14.46173691 +0000 UTC m=+2944.708576329" observedRunningTime="2025-10-08 19:06:14.974332655 +0000 UTC m=+2945.221172064" watchObservedRunningTime="2025-10-08 19:06:14.976409196 +0000 UTC m=+2945.223248585" Oct 08 19:06:21 crc kubenswrapper[4859]: I1008 19:06:21.952727 4859 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-4z8nk" Oct 08 19:06:21 crc kubenswrapper[4859]: I1008 19:06:21.955933 4859 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-4z8nk" Oct 08 19:06:22 crc kubenswrapper[4859]: I1008 19:06:22.023969 4859 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-4z8nk" Oct 08 19:06:22 crc kubenswrapper[4859]: I1008 19:06:22.102465 4859 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-4z8nk" Oct 08 19:06:22 crc kubenswrapper[4859]: I1008 19:06:22.264226 4859 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-4z8nk"] Oct 08 19:06:24 crc kubenswrapper[4859]: I1008 19:06:24.068531 4859 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-marketplace-4z8nk" podUID="42cbb501-291f-4dd2-8373-f80ebf57bdf2" containerName="registry-server" containerID="cri-o://44be98a1b71dab290021207cc3e4634b6a11ac84bdf55e804ffdf64699fb2ad8" gracePeriod=2 Oct 08 19:06:24 crc kubenswrapper[4859]: I1008 19:06:24.574376 4859 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-4z8nk" Oct 08 19:06:24 crc kubenswrapper[4859]: I1008 19:06:24.720675 4859 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-dl4xg\" (UniqueName: \"kubernetes.io/projected/42cbb501-291f-4dd2-8373-f80ebf57bdf2-kube-api-access-dl4xg\") pod \"42cbb501-291f-4dd2-8373-f80ebf57bdf2\" (UID: \"42cbb501-291f-4dd2-8373-f80ebf57bdf2\") " Oct 08 19:06:24 crc kubenswrapper[4859]: I1008 19:06:24.720739 4859 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/42cbb501-291f-4dd2-8373-f80ebf57bdf2-utilities\") pod \"42cbb501-291f-4dd2-8373-f80ebf57bdf2\" (UID: \"42cbb501-291f-4dd2-8373-f80ebf57bdf2\") " Oct 08 19:06:24 crc kubenswrapper[4859]: I1008 19:06:24.720890 4859 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/42cbb501-291f-4dd2-8373-f80ebf57bdf2-catalog-content\") pod \"42cbb501-291f-4dd2-8373-f80ebf57bdf2\" (UID: \"42cbb501-291f-4dd2-8373-f80ebf57bdf2\") " Oct 08 19:06:24 crc kubenswrapper[4859]: I1008 19:06:24.722539 4859 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/42cbb501-291f-4dd2-8373-f80ebf57bdf2-utilities" (OuterVolumeSpecName: "utilities") pod "42cbb501-291f-4dd2-8373-f80ebf57bdf2" (UID: "42cbb501-291f-4dd2-8373-f80ebf57bdf2"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 08 19:06:24 crc kubenswrapper[4859]: I1008 19:06:24.727248 4859 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/42cbb501-291f-4dd2-8373-f80ebf57bdf2-kube-api-access-dl4xg" (OuterVolumeSpecName: "kube-api-access-dl4xg") pod "42cbb501-291f-4dd2-8373-f80ebf57bdf2" (UID: "42cbb501-291f-4dd2-8373-f80ebf57bdf2"). InnerVolumeSpecName "kube-api-access-dl4xg". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 08 19:06:24 crc kubenswrapper[4859]: I1008 19:06:24.733654 4859 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/42cbb501-291f-4dd2-8373-f80ebf57bdf2-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "42cbb501-291f-4dd2-8373-f80ebf57bdf2" (UID: "42cbb501-291f-4dd2-8373-f80ebf57bdf2"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 08 19:06:24 crc kubenswrapper[4859]: I1008 19:06:24.823473 4859 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-dl4xg\" (UniqueName: \"kubernetes.io/projected/42cbb501-291f-4dd2-8373-f80ebf57bdf2-kube-api-access-dl4xg\") on node \"crc\" DevicePath \"\"" Oct 08 19:06:24 crc kubenswrapper[4859]: I1008 19:06:24.823519 4859 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/42cbb501-291f-4dd2-8373-f80ebf57bdf2-utilities\") on node \"crc\" DevicePath \"\"" Oct 08 19:06:24 crc kubenswrapper[4859]: I1008 19:06:24.823532 4859 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/42cbb501-291f-4dd2-8373-f80ebf57bdf2-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 08 19:06:25 crc kubenswrapper[4859]: I1008 19:06:25.078801 4859 generic.go:334] "Generic (PLEG): container finished" podID="42cbb501-291f-4dd2-8373-f80ebf57bdf2" containerID="44be98a1b71dab290021207cc3e4634b6a11ac84bdf55e804ffdf64699fb2ad8" exitCode=0 Oct 08 19:06:25 crc kubenswrapper[4859]: I1008 19:06:25.078870 4859 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-4z8nk" event={"ID":"42cbb501-291f-4dd2-8373-f80ebf57bdf2","Type":"ContainerDied","Data":"44be98a1b71dab290021207cc3e4634b6a11ac84bdf55e804ffdf64699fb2ad8"} Oct 08 19:06:25 crc kubenswrapper[4859]: I1008 19:06:25.078957 4859 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-4z8nk" event={"ID":"42cbb501-291f-4dd2-8373-f80ebf57bdf2","Type":"ContainerDied","Data":"c92ae6b9863bac217baf03a6ad6624d19161f5478405de02b092c977d0ca044a"} Oct 08 19:06:25 crc kubenswrapper[4859]: I1008 19:06:25.078987 4859 scope.go:117] "RemoveContainer" containerID="44be98a1b71dab290021207cc3e4634b6a11ac84bdf55e804ffdf64699fb2ad8" Oct 08 19:06:25 crc kubenswrapper[4859]: I1008 19:06:25.078896 4859 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-4z8nk" Oct 08 19:06:25 crc kubenswrapper[4859]: I1008 19:06:25.098513 4859 scope.go:117] "RemoveContainer" containerID="8740e1e27bb44aeec068579b52e081d0907e4f78e0a7af901e3689b603f55303" Oct 08 19:06:25 crc kubenswrapper[4859]: I1008 19:06:25.117222 4859 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-4z8nk"] Oct 08 19:06:25 crc kubenswrapper[4859]: I1008 19:06:25.128868 4859 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-marketplace-4z8nk"] Oct 08 19:06:25 crc kubenswrapper[4859]: I1008 19:06:25.141391 4859 scope.go:117] "RemoveContainer" containerID="27be93e86b3a217363cbba555c104e0efbc5c2a75aed4235788b5985048c3c76" Oct 08 19:06:25 crc kubenswrapper[4859]: I1008 19:06:25.175036 4859 scope.go:117] "RemoveContainer" containerID="44be98a1b71dab290021207cc3e4634b6a11ac84bdf55e804ffdf64699fb2ad8" Oct 08 19:06:25 crc kubenswrapper[4859]: E1008 19:06:25.175452 4859 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"44be98a1b71dab290021207cc3e4634b6a11ac84bdf55e804ffdf64699fb2ad8\": container with ID starting with 44be98a1b71dab290021207cc3e4634b6a11ac84bdf55e804ffdf64699fb2ad8 not found: ID does not exist" containerID="44be98a1b71dab290021207cc3e4634b6a11ac84bdf55e804ffdf64699fb2ad8" Oct 08 19:06:25 crc kubenswrapper[4859]: I1008 19:06:25.175498 4859 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"44be98a1b71dab290021207cc3e4634b6a11ac84bdf55e804ffdf64699fb2ad8"} err="failed to get container status \"44be98a1b71dab290021207cc3e4634b6a11ac84bdf55e804ffdf64699fb2ad8\": rpc error: code = NotFound desc = could not find container \"44be98a1b71dab290021207cc3e4634b6a11ac84bdf55e804ffdf64699fb2ad8\": container with ID starting with 44be98a1b71dab290021207cc3e4634b6a11ac84bdf55e804ffdf64699fb2ad8 not found: ID does not exist" Oct 08 19:06:25 crc kubenswrapper[4859]: I1008 19:06:25.175530 4859 scope.go:117] "RemoveContainer" containerID="8740e1e27bb44aeec068579b52e081d0907e4f78e0a7af901e3689b603f55303" Oct 08 19:06:25 crc kubenswrapper[4859]: E1008 19:06:25.175848 4859 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"8740e1e27bb44aeec068579b52e081d0907e4f78e0a7af901e3689b603f55303\": container with ID starting with 8740e1e27bb44aeec068579b52e081d0907e4f78e0a7af901e3689b603f55303 not found: ID does not exist" containerID="8740e1e27bb44aeec068579b52e081d0907e4f78e0a7af901e3689b603f55303" Oct 08 19:06:25 crc kubenswrapper[4859]: I1008 19:06:25.175870 4859 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"8740e1e27bb44aeec068579b52e081d0907e4f78e0a7af901e3689b603f55303"} err="failed to get container status \"8740e1e27bb44aeec068579b52e081d0907e4f78e0a7af901e3689b603f55303\": rpc error: code = NotFound desc = could not find container \"8740e1e27bb44aeec068579b52e081d0907e4f78e0a7af901e3689b603f55303\": container with ID starting with 8740e1e27bb44aeec068579b52e081d0907e4f78e0a7af901e3689b603f55303 not found: ID does not exist" Oct 08 19:06:25 crc kubenswrapper[4859]: I1008 19:06:25.175882 4859 scope.go:117] "RemoveContainer" containerID="27be93e86b3a217363cbba555c104e0efbc5c2a75aed4235788b5985048c3c76" Oct 08 19:06:25 crc kubenswrapper[4859]: E1008 19:06:25.176212 4859 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"27be93e86b3a217363cbba555c104e0efbc5c2a75aed4235788b5985048c3c76\": container with ID starting with 27be93e86b3a217363cbba555c104e0efbc5c2a75aed4235788b5985048c3c76 not found: ID does not exist" containerID="27be93e86b3a217363cbba555c104e0efbc5c2a75aed4235788b5985048c3c76" Oct 08 19:06:25 crc kubenswrapper[4859]: I1008 19:06:25.176249 4859 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"27be93e86b3a217363cbba555c104e0efbc5c2a75aed4235788b5985048c3c76"} err="failed to get container status \"27be93e86b3a217363cbba555c104e0efbc5c2a75aed4235788b5985048c3c76\": rpc error: code = NotFound desc = could not find container \"27be93e86b3a217363cbba555c104e0efbc5c2a75aed4235788b5985048c3c76\": container with ID starting with 27be93e86b3a217363cbba555c104e0efbc5c2a75aed4235788b5985048c3c76 not found: ID does not exist" Oct 08 19:06:26 crc kubenswrapper[4859]: I1008 19:06:26.483126 4859 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="42cbb501-291f-4dd2-8373-f80ebf57bdf2" path="/var/lib/kubelet/pods/42cbb501-291f-4dd2-8373-f80ebf57bdf2/volumes" Oct 08 19:08:17 crc kubenswrapper[4859]: I1008 19:08:17.925306 4859 patch_prober.go:28] interesting pod/machine-config-daemon-82s52 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 08 19:08:17 crc kubenswrapper[4859]: I1008 19:08:17.925914 4859 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-82s52" podUID="b23a6a6c-9d92-4e7b-840e-55cfda873a2d" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 08 19:08:47 crc kubenswrapper[4859]: I1008 19:08:47.925411 4859 patch_prober.go:28] interesting pod/machine-config-daemon-82s52 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 08 19:08:47 crc kubenswrapper[4859]: I1008 19:08:47.926199 4859 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-82s52" podUID="b23a6a6c-9d92-4e7b-840e-55cfda873a2d" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 08 19:09:17 crc kubenswrapper[4859]: I1008 19:09:17.925116 4859 patch_prober.go:28] interesting pod/machine-config-daemon-82s52 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 08 19:09:17 crc kubenswrapper[4859]: I1008 19:09:17.925709 4859 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-82s52" podUID="b23a6a6c-9d92-4e7b-840e-55cfda873a2d" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 08 19:09:17 crc kubenswrapper[4859]: I1008 19:09:17.925756 4859 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-82s52" Oct 08 19:09:17 crc kubenswrapper[4859]: I1008 19:09:17.926450 4859 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"cc836fd09088ba4df38f4588526d6fe50b7b131a7fb454b4fd3751d8b1b22560"} pod="openshift-machine-config-operator/machine-config-daemon-82s52" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Oct 08 19:09:17 crc kubenswrapper[4859]: I1008 19:09:17.926494 4859 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-82s52" podUID="b23a6a6c-9d92-4e7b-840e-55cfda873a2d" containerName="machine-config-daemon" containerID="cri-o://cc836fd09088ba4df38f4588526d6fe50b7b131a7fb454b4fd3751d8b1b22560" gracePeriod=600 Oct 08 19:09:18 crc kubenswrapper[4859]: E1008 19:09:18.073552 4859 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-82s52_openshift-machine-config-operator(b23a6a6c-9d92-4e7b-840e-55cfda873a2d)\"" pod="openshift-machine-config-operator/machine-config-daemon-82s52" podUID="b23a6a6c-9d92-4e7b-840e-55cfda873a2d" Oct 08 19:09:18 crc kubenswrapper[4859]: I1008 19:09:18.813824 4859 generic.go:334] "Generic (PLEG): container finished" podID="b23a6a6c-9d92-4e7b-840e-55cfda873a2d" containerID="cc836fd09088ba4df38f4588526d6fe50b7b131a7fb454b4fd3751d8b1b22560" exitCode=0 Oct 08 19:09:18 crc kubenswrapper[4859]: I1008 19:09:18.813884 4859 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-82s52" event={"ID":"b23a6a6c-9d92-4e7b-840e-55cfda873a2d","Type":"ContainerDied","Data":"cc836fd09088ba4df38f4588526d6fe50b7b131a7fb454b4fd3751d8b1b22560"} Oct 08 19:09:18 crc kubenswrapper[4859]: I1008 19:09:18.813925 4859 scope.go:117] "RemoveContainer" containerID="a3d24a0dfcbf833b7bcf1af5e19a45816904f485fab5db993921d32e47607790" Oct 08 19:09:18 crc kubenswrapper[4859]: I1008 19:09:18.814790 4859 scope.go:117] "RemoveContainer" containerID="cc836fd09088ba4df38f4588526d6fe50b7b131a7fb454b4fd3751d8b1b22560" Oct 08 19:09:18 crc kubenswrapper[4859]: E1008 19:09:18.815199 4859 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-82s52_openshift-machine-config-operator(b23a6a6c-9d92-4e7b-840e-55cfda873a2d)\"" pod="openshift-machine-config-operator/machine-config-daemon-82s52" podUID="b23a6a6c-9d92-4e7b-840e-55cfda873a2d" Oct 08 19:09:30 crc kubenswrapper[4859]: I1008 19:09:30.479794 4859 scope.go:117] "RemoveContainer" containerID="cc836fd09088ba4df38f4588526d6fe50b7b131a7fb454b4fd3751d8b1b22560" Oct 08 19:09:30 crc kubenswrapper[4859]: E1008 19:09:30.480822 4859 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-82s52_openshift-machine-config-operator(b23a6a6c-9d92-4e7b-840e-55cfda873a2d)\"" pod="openshift-machine-config-operator/machine-config-daemon-82s52" podUID="b23a6a6c-9d92-4e7b-840e-55cfda873a2d" Oct 08 19:09:43 crc kubenswrapper[4859]: I1008 19:09:43.470213 4859 scope.go:117] "RemoveContainer" containerID="cc836fd09088ba4df38f4588526d6fe50b7b131a7fb454b4fd3751d8b1b22560" Oct 08 19:09:43 crc kubenswrapper[4859]: E1008 19:09:43.471042 4859 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-82s52_openshift-machine-config-operator(b23a6a6c-9d92-4e7b-840e-55cfda873a2d)\"" pod="openshift-machine-config-operator/machine-config-daemon-82s52" podUID="b23a6a6c-9d92-4e7b-840e-55cfda873a2d" Oct 08 19:09:56 crc kubenswrapper[4859]: I1008 19:09:56.472412 4859 scope.go:117] "RemoveContainer" containerID="cc836fd09088ba4df38f4588526d6fe50b7b131a7fb454b4fd3751d8b1b22560" Oct 08 19:09:56 crc kubenswrapper[4859]: E1008 19:09:56.473206 4859 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-82s52_openshift-machine-config-operator(b23a6a6c-9d92-4e7b-840e-55cfda873a2d)\"" pod="openshift-machine-config-operator/machine-config-daemon-82s52" podUID="b23a6a6c-9d92-4e7b-840e-55cfda873a2d" Oct 08 19:10:10 crc kubenswrapper[4859]: I1008 19:10:10.488779 4859 scope.go:117] "RemoveContainer" containerID="cc836fd09088ba4df38f4588526d6fe50b7b131a7fb454b4fd3751d8b1b22560" Oct 08 19:10:10 crc kubenswrapper[4859]: E1008 19:10:10.489925 4859 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-82s52_openshift-machine-config-operator(b23a6a6c-9d92-4e7b-840e-55cfda873a2d)\"" pod="openshift-machine-config-operator/machine-config-daemon-82s52" podUID="b23a6a6c-9d92-4e7b-840e-55cfda873a2d" Oct 08 19:10:24 crc kubenswrapper[4859]: I1008 19:10:24.470645 4859 scope.go:117] "RemoveContainer" containerID="cc836fd09088ba4df38f4588526d6fe50b7b131a7fb454b4fd3751d8b1b22560" Oct 08 19:10:24 crc kubenswrapper[4859]: E1008 19:10:24.471587 4859 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-82s52_openshift-machine-config-operator(b23a6a6c-9d92-4e7b-840e-55cfda873a2d)\"" pod="openshift-machine-config-operator/machine-config-daemon-82s52" podUID="b23a6a6c-9d92-4e7b-840e-55cfda873a2d" Oct 08 19:10:39 crc kubenswrapper[4859]: I1008 19:10:39.470192 4859 scope.go:117] "RemoveContainer" containerID="cc836fd09088ba4df38f4588526d6fe50b7b131a7fb454b4fd3751d8b1b22560" Oct 08 19:10:39 crc kubenswrapper[4859]: E1008 19:10:39.471132 4859 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-82s52_openshift-machine-config-operator(b23a6a6c-9d92-4e7b-840e-55cfda873a2d)\"" pod="openshift-machine-config-operator/machine-config-daemon-82s52" podUID="b23a6a6c-9d92-4e7b-840e-55cfda873a2d" Oct 08 19:10:51 crc kubenswrapper[4859]: I1008 19:10:51.469742 4859 scope.go:117] "RemoveContainer" containerID="cc836fd09088ba4df38f4588526d6fe50b7b131a7fb454b4fd3751d8b1b22560" Oct 08 19:10:51 crc kubenswrapper[4859]: E1008 19:10:51.470602 4859 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-82s52_openshift-machine-config-operator(b23a6a6c-9d92-4e7b-840e-55cfda873a2d)\"" pod="openshift-machine-config-operator/machine-config-daemon-82s52" podUID="b23a6a6c-9d92-4e7b-840e-55cfda873a2d" Oct 08 19:11:04 crc kubenswrapper[4859]: I1008 19:11:04.470652 4859 scope.go:117] "RemoveContainer" containerID="cc836fd09088ba4df38f4588526d6fe50b7b131a7fb454b4fd3751d8b1b22560" Oct 08 19:11:04 crc kubenswrapper[4859]: E1008 19:11:04.472564 4859 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-82s52_openshift-machine-config-operator(b23a6a6c-9d92-4e7b-840e-55cfda873a2d)\"" pod="openshift-machine-config-operator/machine-config-daemon-82s52" podUID="b23a6a6c-9d92-4e7b-840e-55cfda873a2d" Oct 08 19:11:15 crc kubenswrapper[4859]: I1008 19:11:15.469954 4859 scope.go:117] "RemoveContainer" containerID="cc836fd09088ba4df38f4588526d6fe50b7b131a7fb454b4fd3751d8b1b22560" Oct 08 19:11:15 crc kubenswrapper[4859]: E1008 19:11:15.471027 4859 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-82s52_openshift-machine-config-operator(b23a6a6c-9d92-4e7b-840e-55cfda873a2d)\"" pod="openshift-machine-config-operator/machine-config-daemon-82s52" podUID="b23a6a6c-9d92-4e7b-840e-55cfda873a2d" Oct 08 19:11:26 crc kubenswrapper[4859]: I1008 19:11:26.470552 4859 scope.go:117] "RemoveContainer" containerID="cc836fd09088ba4df38f4588526d6fe50b7b131a7fb454b4fd3751d8b1b22560" Oct 08 19:11:26 crc kubenswrapper[4859]: E1008 19:11:26.471751 4859 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-82s52_openshift-machine-config-operator(b23a6a6c-9d92-4e7b-840e-55cfda873a2d)\"" pod="openshift-machine-config-operator/machine-config-daemon-82s52" podUID="b23a6a6c-9d92-4e7b-840e-55cfda873a2d" Oct 08 19:11:37 crc kubenswrapper[4859]: I1008 19:11:37.469907 4859 scope.go:117] "RemoveContainer" containerID="cc836fd09088ba4df38f4588526d6fe50b7b131a7fb454b4fd3751d8b1b22560" Oct 08 19:11:37 crc kubenswrapper[4859]: E1008 19:11:37.470798 4859 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-82s52_openshift-machine-config-operator(b23a6a6c-9d92-4e7b-840e-55cfda873a2d)\"" pod="openshift-machine-config-operator/machine-config-daemon-82s52" podUID="b23a6a6c-9d92-4e7b-840e-55cfda873a2d" Oct 08 19:11:48 crc kubenswrapper[4859]: I1008 19:11:48.470287 4859 scope.go:117] "RemoveContainer" containerID="cc836fd09088ba4df38f4588526d6fe50b7b131a7fb454b4fd3751d8b1b22560" Oct 08 19:11:48 crc kubenswrapper[4859]: E1008 19:11:48.471659 4859 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-82s52_openshift-machine-config-operator(b23a6a6c-9d92-4e7b-840e-55cfda873a2d)\"" pod="openshift-machine-config-operator/machine-config-daemon-82s52" podUID="b23a6a6c-9d92-4e7b-840e-55cfda873a2d" Oct 08 19:11:59 crc kubenswrapper[4859]: I1008 19:11:59.470860 4859 scope.go:117] "RemoveContainer" containerID="cc836fd09088ba4df38f4588526d6fe50b7b131a7fb454b4fd3751d8b1b22560" Oct 08 19:11:59 crc kubenswrapper[4859]: E1008 19:11:59.471856 4859 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-82s52_openshift-machine-config-operator(b23a6a6c-9d92-4e7b-840e-55cfda873a2d)\"" pod="openshift-machine-config-operator/machine-config-daemon-82s52" podUID="b23a6a6c-9d92-4e7b-840e-55cfda873a2d" Oct 08 19:12:14 crc kubenswrapper[4859]: I1008 19:12:14.471362 4859 scope.go:117] "RemoveContainer" containerID="cc836fd09088ba4df38f4588526d6fe50b7b131a7fb454b4fd3751d8b1b22560" Oct 08 19:12:14 crc kubenswrapper[4859]: E1008 19:12:14.472451 4859 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-82s52_openshift-machine-config-operator(b23a6a6c-9d92-4e7b-840e-55cfda873a2d)\"" pod="openshift-machine-config-operator/machine-config-daemon-82s52" podUID="b23a6a6c-9d92-4e7b-840e-55cfda873a2d" Oct 08 19:12:26 crc kubenswrapper[4859]: I1008 19:12:26.475316 4859 scope.go:117] "RemoveContainer" containerID="cc836fd09088ba4df38f4588526d6fe50b7b131a7fb454b4fd3751d8b1b22560" Oct 08 19:12:26 crc kubenswrapper[4859]: E1008 19:12:26.475991 4859 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-82s52_openshift-machine-config-operator(b23a6a6c-9d92-4e7b-840e-55cfda873a2d)\"" pod="openshift-machine-config-operator/machine-config-daemon-82s52" podUID="b23a6a6c-9d92-4e7b-840e-55cfda873a2d" Oct 08 19:12:40 crc kubenswrapper[4859]: I1008 19:12:40.470587 4859 scope.go:117] "RemoveContainer" containerID="cc836fd09088ba4df38f4588526d6fe50b7b131a7fb454b4fd3751d8b1b22560" Oct 08 19:12:40 crc kubenswrapper[4859]: E1008 19:12:40.471770 4859 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-82s52_openshift-machine-config-operator(b23a6a6c-9d92-4e7b-840e-55cfda873a2d)\"" pod="openshift-machine-config-operator/machine-config-daemon-82s52" podUID="b23a6a6c-9d92-4e7b-840e-55cfda873a2d" Oct 08 19:12:44 crc kubenswrapper[4859]: I1008 19:12:44.826941 4859 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-mds9x"] Oct 08 19:12:44 crc kubenswrapper[4859]: E1008 19:12:44.828102 4859 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="42cbb501-291f-4dd2-8373-f80ebf57bdf2" containerName="extract-content" Oct 08 19:12:44 crc kubenswrapper[4859]: I1008 19:12:44.828123 4859 state_mem.go:107] "Deleted CPUSet assignment" podUID="42cbb501-291f-4dd2-8373-f80ebf57bdf2" containerName="extract-content" Oct 08 19:12:44 crc kubenswrapper[4859]: E1008 19:12:44.828148 4859 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="42cbb501-291f-4dd2-8373-f80ebf57bdf2" containerName="registry-server" Oct 08 19:12:44 crc kubenswrapper[4859]: I1008 19:12:44.828164 4859 state_mem.go:107] "Deleted CPUSet assignment" podUID="42cbb501-291f-4dd2-8373-f80ebf57bdf2" containerName="registry-server" Oct 08 19:12:44 crc kubenswrapper[4859]: E1008 19:12:44.828238 4859 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="42cbb501-291f-4dd2-8373-f80ebf57bdf2" containerName="extract-utilities" Oct 08 19:12:44 crc kubenswrapper[4859]: I1008 19:12:44.828257 4859 state_mem.go:107] "Deleted CPUSet assignment" podUID="42cbb501-291f-4dd2-8373-f80ebf57bdf2" containerName="extract-utilities" Oct 08 19:12:44 crc kubenswrapper[4859]: I1008 19:12:44.828817 4859 memory_manager.go:354] "RemoveStaleState removing state" podUID="42cbb501-291f-4dd2-8373-f80ebf57bdf2" containerName="registry-server" Oct 08 19:12:44 crc kubenswrapper[4859]: I1008 19:12:44.837516 4859 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-mds9x" Oct 08 19:12:44 crc kubenswrapper[4859]: I1008 19:12:44.866504 4859 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-mds9x"] Oct 08 19:12:44 crc kubenswrapper[4859]: I1008 19:12:44.923010 4859 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-kc62x\" (UniqueName: \"kubernetes.io/projected/7c3480f8-6e54-47b3-a15d-82ecc7e1737a-kube-api-access-kc62x\") pod \"community-operators-mds9x\" (UID: \"7c3480f8-6e54-47b3-a15d-82ecc7e1737a\") " pod="openshift-marketplace/community-operators-mds9x" Oct 08 19:12:44 crc kubenswrapper[4859]: I1008 19:12:44.923257 4859 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/7c3480f8-6e54-47b3-a15d-82ecc7e1737a-utilities\") pod \"community-operators-mds9x\" (UID: \"7c3480f8-6e54-47b3-a15d-82ecc7e1737a\") " pod="openshift-marketplace/community-operators-mds9x" Oct 08 19:12:44 crc kubenswrapper[4859]: I1008 19:12:44.923296 4859 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/7c3480f8-6e54-47b3-a15d-82ecc7e1737a-catalog-content\") pod \"community-operators-mds9x\" (UID: \"7c3480f8-6e54-47b3-a15d-82ecc7e1737a\") " pod="openshift-marketplace/community-operators-mds9x" Oct 08 19:12:45 crc kubenswrapper[4859]: I1008 19:12:45.004568 4859 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-6gzxt"] Oct 08 19:12:45 crc kubenswrapper[4859]: I1008 19:12:45.009954 4859 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-6gzxt" Oct 08 19:12:45 crc kubenswrapper[4859]: I1008 19:12:45.017949 4859 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-6gzxt"] Oct 08 19:12:45 crc kubenswrapper[4859]: I1008 19:12:45.024987 4859 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/7c3480f8-6e54-47b3-a15d-82ecc7e1737a-utilities\") pod \"community-operators-mds9x\" (UID: \"7c3480f8-6e54-47b3-a15d-82ecc7e1737a\") " pod="openshift-marketplace/community-operators-mds9x" Oct 08 19:12:45 crc kubenswrapper[4859]: I1008 19:12:45.025037 4859 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/7c3480f8-6e54-47b3-a15d-82ecc7e1737a-catalog-content\") pod \"community-operators-mds9x\" (UID: \"7c3480f8-6e54-47b3-a15d-82ecc7e1737a\") " pod="openshift-marketplace/community-operators-mds9x" Oct 08 19:12:45 crc kubenswrapper[4859]: I1008 19:12:45.025063 4859 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-kc62x\" (UniqueName: \"kubernetes.io/projected/7c3480f8-6e54-47b3-a15d-82ecc7e1737a-kube-api-access-kc62x\") pod \"community-operators-mds9x\" (UID: \"7c3480f8-6e54-47b3-a15d-82ecc7e1737a\") " pod="openshift-marketplace/community-operators-mds9x" Oct 08 19:12:45 crc kubenswrapper[4859]: I1008 19:12:45.025528 4859 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/7c3480f8-6e54-47b3-a15d-82ecc7e1737a-utilities\") pod \"community-operators-mds9x\" (UID: \"7c3480f8-6e54-47b3-a15d-82ecc7e1737a\") " pod="openshift-marketplace/community-operators-mds9x" Oct 08 19:12:45 crc kubenswrapper[4859]: I1008 19:12:45.025538 4859 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/7c3480f8-6e54-47b3-a15d-82ecc7e1737a-catalog-content\") pod \"community-operators-mds9x\" (UID: \"7c3480f8-6e54-47b3-a15d-82ecc7e1737a\") " pod="openshift-marketplace/community-operators-mds9x" Oct 08 19:12:45 crc kubenswrapper[4859]: I1008 19:12:45.079528 4859 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-kc62x\" (UniqueName: \"kubernetes.io/projected/7c3480f8-6e54-47b3-a15d-82ecc7e1737a-kube-api-access-kc62x\") pod \"community-operators-mds9x\" (UID: \"7c3480f8-6e54-47b3-a15d-82ecc7e1737a\") " pod="openshift-marketplace/community-operators-mds9x" Oct 08 19:12:45 crc kubenswrapper[4859]: I1008 19:12:45.127077 4859 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/5e1f266e-f0c3-4020-937c-b30ea071a654-catalog-content\") pod \"certified-operators-6gzxt\" (UID: \"5e1f266e-f0c3-4020-937c-b30ea071a654\") " pod="openshift-marketplace/certified-operators-6gzxt" Oct 08 19:12:45 crc kubenswrapper[4859]: I1008 19:12:45.127188 4859 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-26znl\" (UniqueName: \"kubernetes.io/projected/5e1f266e-f0c3-4020-937c-b30ea071a654-kube-api-access-26znl\") pod \"certified-operators-6gzxt\" (UID: \"5e1f266e-f0c3-4020-937c-b30ea071a654\") " pod="openshift-marketplace/certified-operators-6gzxt" Oct 08 19:12:45 crc kubenswrapper[4859]: I1008 19:12:45.127215 4859 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/5e1f266e-f0c3-4020-937c-b30ea071a654-utilities\") pod \"certified-operators-6gzxt\" (UID: \"5e1f266e-f0c3-4020-937c-b30ea071a654\") " pod="openshift-marketplace/certified-operators-6gzxt" Oct 08 19:12:45 crc kubenswrapper[4859]: I1008 19:12:45.190522 4859 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-mds9x" Oct 08 19:12:45 crc kubenswrapper[4859]: I1008 19:12:45.228939 4859 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/5e1f266e-f0c3-4020-937c-b30ea071a654-utilities\") pod \"certified-operators-6gzxt\" (UID: \"5e1f266e-f0c3-4020-937c-b30ea071a654\") " pod="openshift-marketplace/certified-operators-6gzxt" Oct 08 19:12:45 crc kubenswrapper[4859]: I1008 19:12:45.229182 4859 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/5e1f266e-f0c3-4020-937c-b30ea071a654-catalog-content\") pod \"certified-operators-6gzxt\" (UID: \"5e1f266e-f0c3-4020-937c-b30ea071a654\") " pod="openshift-marketplace/certified-operators-6gzxt" Oct 08 19:12:45 crc kubenswrapper[4859]: I1008 19:12:45.229211 4859 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-26znl\" (UniqueName: \"kubernetes.io/projected/5e1f266e-f0c3-4020-937c-b30ea071a654-kube-api-access-26znl\") pod \"certified-operators-6gzxt\" (UID: \"5e1f266e-f0c3-4020-937c-b30ea071a654\") " pod="openshift-marketplace/certified-operators-6gzxt" Oct 08 19:12:45 crc kubenswrapper[4859]: I1008 19:12:45.230108 4859 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/5e1f266e-f0c3-4020-937c-b30ea071a654-catalog-content\") pod \"certified-operators-6gzxt\" (UID: \"5e1f266e-f0c3-4020-937c-b30ea071a654\") " pod="openshift-marketplace/certified-operators-6gzxt" Oct 08 19:12:45 crc kubenswrapper[4859]: I1008 19:12:45.230140 4859 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/5e1f266e-f0c3-4020-937c-b30ea071a654-utilities\") pod \"certified-operators-6gzxt\" (UID: \"5e1f266e-f0c3-4020-937c-b30ea071a654\") " pod="openshift-marketplace/certified-operators-6gzxt" Oct 08 19:12:45 crc kubenswrapper[4859]: I1008 19:12:45.252680 4859 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-26znl\" (UniqueName: \"kubernetes.io/projected/5e1f266e-f0c3-4020-937c-b30ea071a654-kube-api-access-26znl\") pod \"certified-operators-6gzxt\" (UID: \"5e1f266e-f0c3-4020-937c-b30ea071a654\") " pod="openshift-marketplace/certified-operators-6gzxt" Oct 08 19:12:45 crc kubenswrapper[4859]: I1008 19:12:45.328050 4859 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-6gzxt" Oct 08 19:12:45 crc kubenswrapper[4859]: I1008 19:12:45.726153 4859 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-mds9x"] Oct 08 19:12:45 crc kubenswrapper[4859]: I1008 19:12:45.889329 4859 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-mds9x" event={"ID":"7c3480f8-6e54-47b3-a15d-82ecc7e1737a","Type":"ContainerStarted","Data":"122b9c188e3fd5fbdc549b40c82a036cc190b3d5fa99270d0545916b179f54eb"} Oct 08 19:12:45 crc kubenswrapper[4859]: I1008 19:12:45.890509 4859 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-mds9x" event={"ID":"7c3480f8-6e54-47b3-a15d-82ecc7e1737a","Type":"ContainerStarted","Data":"b035dbe4357cc8bfc36a6feafdc65441756ec291c76ae1d84aa70ef9ea28a14a"} Oct 08 19:12:45 crc kubenswrapper[4859]: I1008 19:12:45.900312 4859 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-6gzxt"] Oct 08 19:12:46 crc kubenswrapper[4859]: I1008 19:12:46.898952 4859 generic.go:334] "Generic (PLEG): container finished" podID="5e1f266e-f0c3-4020-937c-b30ea071a654" containerID="4474089b1ae2e37c83ca57f4aedbb60cbff88ffef24a55572a15111d104fea3f" exitCode=0 Oct 08 19:12:46 crc kubenswrapper[4859]: I1008 19:12:46.899062 4859 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-6gzxt" event={"ID":"5e1f266e-f0c3-4020-937c-b30ea071a654","Type":"ContainerDied","Data":"4474089b1ae2e37c83ca57f4aedbb60cbff88ffef24a55572a15111d104fea3f"} Oct 08 19:12:46 crc kubenswrapper[4859]: I1008 19:12:46.899549 4859 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-6gzxt" event={"ID":"5e1f266e-f0c3-4020-937c-b30ea071a654","Type":"ContainerStarted","Data":"031386f3aaef44ccfb174915ba42452f4221a4b0cb2938ea935ce0d7c801b6d6"} Oct 08 19:12:46 crc kubenswrapper[4859]: I1008 19:12:46.901208 4859 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Oct 08 19:12:46 crc kubenswrapper[4859]: I1008 19:12:46.904655 4859 generic.go:334] "Generic (PLEG): container finished" podID="7c3480f8-6e54-47b3-a15d-82ecc7e1737a" containerID="122b9c188e3fd5fbdc549b40c82a036cc190b3d5fa99270d0545916b179f54eb" exitCode=0 Oct 08 19:12:46 crc kubenswrapper[4859]: I1008 19:12:46.905279 4859 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-mds9x" event={"ID":"7c3480f8-6e54-47b3-a15d-82ecc7e1737a","Type":"ContainerDied","Data":"122b9c188e3fd5fbdc549b40c82a036cc190b3d5fa99270d0545916b179f54eb"} Oct 08 19:12:47 crc kubenswrapper[4859]: I1008 19:12:47.918821 4859 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-mds9x" event={"ID":"7c3480f8-6e54-47b3-a15d-82ecc7e1737a","Type":"ContainerStarted","Data":"5d1ea73380cc55814ce8d036b21878d3bdfec5eae267674fdd2f54f57b24f9f6"} Oct 08 19:12:48 crc kubenswrapper[4859]: I1008 19:12:48.966189 4859 generic.go:334] "Generic (PLEG): container finished" podID="5e1f266e-f0c3-4020-937c-b30ea071a654" containerID="0c96317fc5b8f962c35ad2b7174b4d7403f324ab4fa141b0f64c59df51018a5b" exitCode=0 Oct 08 19:12:48 crc kubenswrapper[4859]: I1008 19:12:48.966407 4859 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-6gzxt" event={"ID":"5e1f266e-f0c3-4020-937c-b30ea071a654","Type":"ContainerDied","Data":"0c96317fc5b8f962c35ad2b7174b4d7403f324ab4fa141b0f64c59df51018a5b"} Oct 08 19:12:48 crc kubenswrapper[4859]: I1008 19:12:48.968651 4859 generic.go:334] "Generic (PLEG): container finished" podID="7c3480f8-6e54-47b3-a15d-82ecc7e1737a" containerID="5d1ea73380cc55814ce8d036b21878d3bdfec5eae267674fdd2f54f57b24f9f6" exitCode=0 Oct 08 19:12:48 crc kubenswrapper[4859]: I1008 19:12:48.969136 4859 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-mds9x" event={"ID":"7c3480f8-6e54-47b3-a15d-82ecc7e1737a","Type":"ContainerDied","Data":"5d1ea73380cc55814ce8d036b21878d3bdfec5eae267674fdd2f54f57b24f9f6"} Oct 08 19:12:49 crc kubenswrapper[4859]: I1008 19:12:49.984457 4859 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-mds9x" event={"ID":"7c3480f8-6e54-47b3-a15d-82ecc7e1737a","Type":"ContainerStarted","Data":"3b0f3f8fb0d39e0377749dfa7f65a3e6e8043eed009afdcf3394a5a10527cbb6"} Oct 08 19:12:49 crc kubenswrapper[4859]: I1008 19:12:49.987152 4859 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-6gzxt" event={"ID":"5e1f266e-f0c3-4020-937c-b30ea071a654","Type":"ContainerStarted","Data":"80d1d3d774dd5baf46c1afd7df49c1626780d4b8d0dfd7ff715a09f4b6649dc5"} Oct 08 19:12:50 crc kubenswrapper[4859]: I1008 19:12:50.010376 4859 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-mds9x" podStartSLOduration=3.486507707 podStartE2EDuration="6.010354102s" podCreationTimestamp="2025-10-08 19:12:44 +0000 UTC" firstStartedPulling="2025-10-08 19:12:46.906546698 +0000 UTC m=+3337.153386077" lastFinishedPulling="2025-10-08 19:12:49.430393083 +0000 UTC m=+3339.677232472" observedRunningTime="2025-10-08 19:12:50.010152846 +0000 UTC m=+3340.256992235" watchObservedRunningTime="2025-10-08 19:12:50.010354102 +0000 UTC m=+3340.257193481" Oct 08 19:12:50 crc kubenswrapper[4859]: I1008 19:12:50.039919 4859 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-6gzxt" podStartSLOduration=3.536440163 podStartE2EDuration="6.039896533s" podCreationTimestamp="2025-10-08 19:12:44 +0000 UTC" firstStartedPulling="2025-10-08 19:12:46.900821581 +0000 UTC m=+3337.147660960" lastFinishedPulling="2025-10-08 19:12:49.404277951 +0000 UTC m=+3339.651117330" observedRunningTime="2025-10-08 19:12:50.031818018 +0000 UTC m=+3340.278657407" watchObservedRunningTime="2025-10-08 19:12:50.039896533 +0000 UTC m=+3340.286735932" Oct 08 19:12:55 crc kubenswrapper[4859]: I1008 19:12:55.190674 4859 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-mds9x" Oct 08 19:12:55 crc kubenswrapper[4859]: I1008 19:12:55.192585 4859 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-mds9x" Oct 08 19:12:55 crc kubenswrapper[4859]: I1008 19:12:55.234442 4859 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-mds9x" Oct 08 19:12:55 crc kubenswrapper[4859]: I1008 19:12:55.328894 4859 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-6gzxt" Oct 08 19:12:55 crc kubenswrapper[4859]: I1008 19:12:55.328954 4859 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-6gzxt" Oct 08 19:12:55 crc kubenswrapper[4859]: I1008 19:12:55.379259 4859 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-6gzxt" Oct 08 19:12:55 crc kubenswrapper[4859]: I1008 19:12:55.469666 4859 scope.go:117] "RemoveContainer" containerID="cc836fd09088ba4df38f4588526d6fe50b7b131a7fb454b4fd3751d8b1b22560" Oct 08 19:12:55 crc kubenswrapper[4859]: E1008 19:12:55.469984 4859 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-82s52_openshift-machine-config-operator(b23a6a6c-9d92-4e7b-840e-55cfda873a2d)\"" pod="openshift-machine-config-operator/machine-config-daemon-82s52" podUID="b23a6a6c-9d92-4e7b-840e-55cfda873a2d" Oct 08 19:12:56 crc kubenswrapper[4859]: I1008 19:12:56.091496 4859 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-6gzxt" Oct 08 19:12:56 crc kubenswrapper[4859]: I1008 19:12:56.098950 4859 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-mds9x" Oct 08 19:12:57 crc kubenswrapper[4859]: I1008 19:12:57.000459 4859 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-6gzxt"] Oct 08 19:12:58 crc kubenswrapper[4859]: I1008 19:12:58.062040 4859 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/certified-operators-6gzxt" podUID="5e1f266e-f0c3-4020-937c-b30ea071a654" containerName="registry-server" containerID="cri-o://80d1d3d774dd5baf46c1afd7df49c1626780d4b8d0dfd7ff715a09f4b6649dc5" gracePeriod=2 Oct 08 19:12:58 crc kubenswrapper[4859]: I1008 19:12:58.406087 4859 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-mds9x"] Oct 08 19:12:58 crc kubenswrapper[4859]: I1008 19:12:58.592068 4859 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-6gzxt" Oct 08 19:12:58 crc kubenswrapper[4859]: I1008 19:12:58.702767 4859 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-26znl\" (UniqueName: \"kubernetes.io/projected/5e1f266e-f0c3-4020-937c-b30ea071a654-kube-api-access-26znl\") pod \"5e1f266e-f0c3-4020-937c-b30ea071a654\" (UID: \"5e1f266e-f0c3-4020-937c-b30ea071a654\") " Oct 08 19:12:58 crc kubenswrapper[4859]: I1008 19:12:58.702846 4859 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/5e1f266e-f0c3-4020-937c-b30ea071a654-utilities\") pod \"5e1f266e-f0c3-4020-937c-b30ea071a654\" (UID: \"5e1f266e-f0c3-4020-937c-b30ea071a654\") " Oct 08 19:12:58 crc kubenswrapper[4859]: I1008 19:12:58.702919 4859 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/5e1f266e-f0c3-4020-937c-b30ea071a654-catalog-content\") pod \"5e1f266e-f0c3-4020-937c-b30ea071a654\" (UID: \"5e1f266e-f0c3-4020-937c-b30ea071a654\") " Oct 08 19:12:58 crc kubenswrapper[4859]: I1008 19:12:58.703807 4859 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/5e1f266e-f0c3-4020-937c-b30ea071a654-utilities" (OuterVolumeSpecName: "utilities") pod "5e1f266e-f0c3-4020-937c-b30ea071a654" (UID: "5e1f266e-f0c3-4020-937c-b30ea071a654"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 08 19:12:58 crc kubenswrapper[4859]: I1008 19:12:58.710443 4859 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5e1f266e-f0c3-4020-937c-b30ea071a654-kube-api-access-26znl" (OuterVolumeSpecName: "kube-api-access-26znl") pod "5e1f266e-f0c3-4020-937c-b30ea071a654" (UID: "5e1f266e-f0c3-4020-937c-b30ea071a654"). InnerVolumeSpecName "kube-api-access-26znl". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 08 19:12:58 crc kubenswrapper[4859]: I1008 19:12:58.756806 4859 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/5e1f266e-f0c3-4020-937c-b30ea071a654-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "5e1f266e-f0c3-4020-937c-b30ea071a654" (UID: "5e1f266e-f0c3-4020-937c-b30ea071a654"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 08 19:12:58 crc kubenswrapper[4859]: I1008 19:12:58.805274 4859 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-26znl\" (UniqueName: \"kubernetes.io/projected/5e1f266e-f0c3-4020-937c-b30ea071a654-kube-api-access-26znl\") on node \"crc\" DevicePath \"\"" Oct 08 19:12:58 crc kubenswrapper[4859]: I1008 19:12:58.805322 4859 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/5e1f266e-f0c3-4020-937c-b30ea071a654-utilities\") on node \"crc\" DevicePath \"\"" Oct 08 19:12:58 crc kubenswrapper[4859]: I1008 19:12:58.805335 4859 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/5e1f266e-f0c3-4020-937c-b30ea071a654-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 08 19:12:59 crc kubenswrapper[4859]: I1008 19:12:59.075798 4859 generic.go:334] "Generic (PLEG): container finished" podID="5e1f266e-f0c3-4020-937c-b30ea071a654" containerID="80d1d3d774dd5baf46c1afd7df49c1626780d4b8d0dfd7ff715a09f4b6649dc5" exitCode=0 Oct 08 19:12:59 crc kubenswrapper[4859]: I1008 19:12:59.075917 4859 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-6gzxt" event={"ID":"5e1f266e-f0c3-4020-937c-b30ea071a654","Type":"ContainerDied","Data":"80d1d3d774dd5baf46c1afd7df49c1626780d4b8d0dfd7ff715a09f4b6649dc5"} Oct 08 19:12:59 crc kubenswrapper[4859]: I1008 19:12:59.075981 4859 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-6gzxt" event={"ID":"5e1f266e-f0c3-4020-937c-b30ea071a654","Type":"ContainerDied","Data":"031386f3aaef44ccfb174915ba42452f4221a4b0cb2938ea935ce0d7c801b6d6"} Oct 08 19:12:59 crc kubenswrapper[4859]: I1008 19:12:59.075933 4859 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-6gzxt" Oct 08 19:12:59 crc kubenswrapper[4859]: I1008 19:12:59.076009 4859 scope.go:117] "RemoveContainer" containerID="80d1d3d774dd5baf46c1afd7df49c1626780d4b8d0dfd7ff715a09f4b6649dc5" Oct 08 19:12:59 crc kubenswrapper[4859]: I1008 19:12:59.076027 4859 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/community-operators-mds9x" podUID="7c3480f8-6e54-47b3-a15d-82ecc7e1737a" containerName="registry-server" containerID="cri-o://3b0f3f8fb0d39e0377749dfa7f65a3e6e8043eed009afdcf3394a5a10527cbb6" gracePeriod=2 Oct 08 19:12:59 crc kubenswrapper[4859]: I1008 19:12:59.117239 4859 scope.go:117] "RemoveContainer" containerID="0c96317fc5b8f962c35ad2b7174b4d7403f324ab4fa141b0f64c59df51018a5b" Oct 08 19:12:59 crc kubenswrapper[4859]: I1008 19:12:59.128993 4859 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-6gzxt"] Oct 08 19:12:59 crc kubenswrapper[4859]: I1008 19:12:59.143434 4859 scope.go:117] "RemoveContainer" containerID="4474089b1ae2e37c83ca57f4aedbb60cbff88ffef24a55572a15111d104fea3f" Oct 08 19:12:59 crc kubenswrapper[4859]: I1008 19:12:59.145129 4859 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/certified-operators-6gzxt"] Oct 08 19:12:59 crc kubenswrapper[4859]: I1008 19:12:59.302122 4859 scope.go:117] "RemoveContainer" containerID="80d1d3d774dd5baf46c1afd7df49c1626780d4b8d0dfd7ff715a09f4b6649dc5" Oct 08 19:12:59 crc kubenswrapper[4859]: E1008 19:12:59.302917 4859 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"80d1d3d774dd5baf46c1afd7df49c1626780d4b8d0dfd7ff715a09f4b6649dc5\": container with ID starting with 80d1d3d774dd5baf46c1afd7df49c1626780d4b8d0dfd7ff715a09f4b6649dc5 not found: ID does not exist" containerID="80d1d3d774dd5baf46c1afd7df49c1626780d4b8d0dfd7ff715a09f4b6649dc5" Oct 08 19:12:59 crc kubenswrapper[4859]: I1008 19:12:59.302956 4859 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"80d1d3d774dd5baf46c1afd7df49c1626780d4b8d0dfd7ff715a09f4b6649dc5"} err="failed to get container status \"80d1d3d774dd5baf46c1afd7df49c1626780d4b8d0dfd7ff715a09f4b6649dc5\": rpc error: code = NotFound desc = could not find container \"80d1d3d774dd5baf46c1afd7df49c1626780d4b8d0dfd7ff715a09f4b6649dc5\": container with ID starting with 80d1d3d774dd5baf46c1afd7df49c1626780d4b8d0dfd7ff715a09f4b6649dc5 not found: ID does not exist" Oct 08 19:12:59 crc kubenswrapper[4859]: I1008 19:12:59.303002 4859 scope.go:117] "RemoveContainer" containerID="0c96317fc5b8f962c35ad2b7174b4d7403f324ab4fa141b0f64c59df51018a5b" Oct 08 19:12:59 crc kubenswrapper[4859]: E1008 19:12:59.303423 4859 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"0c96317fc5b8f962c35ad2b7174b4d7403f324ab4fa141b0f64c59df51018a5b\": container with ID starting with 0c96317fc5b8f962c35ad2b7174b4d7403f324ab4fa141b0f64c59df51018a5b not found: ID does not exist" containerID="0c96317fc5b8f962c35ad2b7174b4d7403f324ab4fa141b0f64c59df51018a5b" Oct 08 19:12:59 crc kubenswrapper[4859]: I1008 19:12:59.303452 4859 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"0c96317fc5b8f962c35ad2b7174b4d7403f324ab4fa141b0f64c59df51018a5b"} err="failed to get container status \"0c96317fc5b8f962c35ad2b7174b4d7403f324ab4fa141b0f64c59df51018a5b\": rpc error: code = NotFound desc = could not find container \"0c96317fc5b8f962c35ad2b7174b4d7403f324ab4fa141b0f64c59df51018a5b\": container with ID starting with 0c96317fc5b8f962c35ad2b7174b4d7403f324ab4fa141b0f64c59df51018a5b not found: ID does not exist" Oct 08 19:12:59 crc kubenswrapper[4859]: I1008 19:12:59.303472 4859 scope.go:117] "RemoveContainer" containerID="4474089b1ae2e37c83ca57f4aedbb60cbff88ffef24a55572a15111d104fea3f" Oct 08 19:12:59 crc kubenswrapper[4859]: E1008 19:12:59.303766 4859 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"4474089b1ae2e37c83ca57f4aedbb60cbff88ffef24a55572a15111d104fea3f\": container with ID starting with 4474089b1ae2e37c83ca57f4aedbb60cbff88ffef24a55572a15111d104fea3f not found: ID does not exist" containerID="4474089b1ae2e37c83ca57f4aedbb60cbff88ffef24a55572a15111d104fea3f" Oct 08 19:12:59 crc kubenswrapper[4859]: I1008 19:12:59.303800 4859 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"4474089b1ae2e37c83ca57f4aedbb60cbff88ffef24a55572a15111d104fea3f"} err="failed to get container status \"4474089b1ae2e37c83ca57f4aedbb60cbff88ffef24a55572a15111d104fea3f\": rpc error: code = NotFound desc = could not find container \"4474089b1ae2e37c83ca57f4aedbb60cbff88ffef24a55572a15111d104fea3f\": container with ID starting with 4474089b1ae2e37c83ca57f4aedbb60cbff88ffef24a55572a15111d104fea3f not found: ID does not exist" Oct 08 19:12:59 crc kubenswrapper[4859]: I1008 19:12:59.675679 4859 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-mds9x" Oct 08 19:12:59 crc kubenswrapper[4859]: I1008 19:12:59.826231 4859 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/7c3480f8-6e54-47b3-a15d-82ecc7e1737a-utilities\") pod \"7c3480f8-6e54-47b3-a15d-82ecc7e1737a\" (UID: \"7c3480f8-6e54-47b3-a15d-82ecc7e1737a\") " Oct 08 19:12:59 crc kubenswrapper[4859]: I1008 19:12:59.826624 4859 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/7c3480f8-6e54-47b3-a15d-82ecc7e1737a-catalog-content\") pod \"7c3480f8-6e54-47b3-a15d-82ecc7e1737a\" (UID: \"7c3480f8-6e54-47b3-a15d-82ecc7e1737a\") " Oct 08 19:12:59 crc kubenswrapper[4859]: I1008 19:12:59.826758 4859 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-kc62x\" (UniqueName: \"kubernetes.io/projected/7c3480f8-6e54-47b3-a15d-82ecc7e1737a-kube-api-access-kc62x\") pod \"7c3480f8-6e54-47b3-a15d-82ecc7e1737a\" (UID: \"7c3480f8-6e54-47b3-a15d-82ecc7e1737a\") " Oct 08 19:12:59 crc kubenswrapper[4859]: I1008 19:12:59.827465 4859 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/7c3480f8-6e54-47b3-a15d-82ecc7e1737a-utilities" (OuterVolumeSpecName: "utilities") pod "7c3480f8-6e54-47b3-a15d-82ecc7e1737a" (UID: "7c3480f8-6e54-47b3-a15d-82ecc7e1737a"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 08 19:12:59 crc kubenswrapper[4859]: I1008 19:12:59.832597 4859 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/7c3480f8-6e54-47b3-a15d-82ecc7e1737a-kube-api-access-kc62x" (OuterVolumeSpecName: "kube-api-access-kc62x") pod "7c3480f8-6e54-47b3-a15d-82ecc7e1737a" (UID: "7c3480f8-6e54-47b3-a15d-82ecc7e1737a"). InnerVolumeSpecName "kube-api-access-kc62x". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 08 19:12:59 crc kubenswrapper[4859]: I1008 19:12:59.873358 4859 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/7c3480f8-6e54-47b3-a15d-82ecc7e1737a-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "7c3480f8-6e54-47b3-a15d-82ecc7e1737a" (UID: "7c3480f8-6e54-47b3-a15d-82ecc7e1737a"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 08 19:12:59 crc kubenswrapper[4859]: I1008 19:12:59.929279 4859 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/7c3480f8-6e54-47b3-a15d-82ecc7e1737a-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 08 19:12:59 crc kubenswrapper[4859]: I1008 19:12:59.929327 4859 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-kc62x\" (UniqueName: \"kubernetes.io/projected/7c3480f8-6e54-47b3-a15d-82ecc7e1737a-kube-api-access-kc62x\") on node \"crc\" DevicePath \"\"" Oct 08 19:12:59 crc kubenswrapper[4859]: I1008 19:12:59.929340 4859 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/7c3480f8-6e54-47b3-a15d-82ecc7e1737a-utilities\") on node \"crc\" DevicePath \"\"" Oct 08 19:13:00 crc kubenswrapper[4859]: I1008 19:13:00.090416 4859 generic.go:334] "Generic (PLEG): container finished" podID="7c3480f8-6e54-47b3-a15d-82ecc7e1737a" containerID="3b0f3f8fb0d39e0377749dfa7f65a3e6e8043eed009afdcf3394a5a10527cbb6" exitCode=0 Oct 08 19:13:00 crc kubenswrapper[4859]: I1008 19:13:00.090475 4859 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-mds9x" event={"ID":"7c3480f8-6e54-47b3-a15d-82ecc7e1737a","Type":"ContainerDied","Data":"3b0f3f8fb0d39e0377749dfa7f65a3e6e8043eed009afdcf3394a5a10527cbb6"} Oct 08 19:13:00 crc kubenswrapper[4859]: I1008 19:13:00.090513 4859 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-mds9x" Oct 08 19:13:00 crc kubenswrapper[4859]: I1008 19:13:00.090536 4859 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-mds9x" event={"ID":"7c3480f8-6e54-47b3-a15d-82ecc7e1737a","Type":"ContainerDied","Data":"b035dbe4357cc8bfc36a6feafdc65441756ec291c76ae1d84aa70ef9ea28a14a"} Oct 08 19:13:00 crc kubenswrapper[4859]: I1008 19:13:00.090561 4859 scope.go:117] "RemoveContainer" containerID="3b0f3f8fb0d39e0377749dfa7f65a3e6e8043eed009afdcf3394a5a10527cbb6" Oct 08 19:13:00 crc kubenswrapper[4859]: I1008 19:13:00.131040 4859 scope.go:117] "RemoveContainer" containerID="5d1ea73380cc55814ce8d036b21878d3bdfec5eae267674fdd2f54f57b24f9f6" Oct 08 19:13:00 crc kubenswrapper[4859]: I1008 19:13:00.143831 4859 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-mds9x"] Oct 08 19:13:00 crc kubenswrapper[4859]: I1008 19:13:00.152908 4859 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/community-operators-mds9x"] Oct 08 19:13:00 crc kubenswrapper[4859]: I1008 19:13:00.240924 4859 scope.go:117] "RemoveContainer" containerID="122b9c188e3fd5fbdc549b40c82a036cc190b3d5fa99270d0545916b179f54eb" Oct 08 19:13:00 crc kubenswrapper[4859]: I1008 19:13:00.276396 4859 scope.go:117] "RemoveContainer" containerID="3b0f3f8fb0d39e0377749dfa7f65a3e6e8043eed009afdcf3394a5a10527cbb6" Oct 08 19:13:00 crc kubenswrapper[4859]: E1008 19:13:00.276848 4859 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"3b0f3f8fb0d39e0377749dfa7f65a3e6e8043eed009afdcf3394a5a10527cbb6\": container with ID starting with 3b0f3f8fb0d39e0377749dfa7f65a3e6e8043eed009afdcf3394a5a10527cbb6 not found: ID does not exist" containerID="3b0f3f8fb0d39e0377749dfa7f65a3e6e8043eed009afdcf3394a5a10527cbb6" Oct 08 19:13:00 crc kubenswrapper[4859]: I1008 19:13:00.276891 4859 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"3b0f3f8fb0d39e0377749dfa7f65a3e6e8043eed009afdcf3394a5a10527cbb6"} err="failed to get container status \"3b0f3f8fb0d39e0377749dfa7f65a3e6e8043eed009afdcf3394a5a10527cbb6\": rpc error: code = NotFound desc = could not find container \"3b0f3f8fb0d39e0377749dfa7f65a3e6e8043eed009afdcf3394a5a10527cbb6\": container with ID starting with 3b0f3f8fb0d39e0377749dfa7f65a3e6e8043eed009afdcf3394a5a10527cbb6 not found: ID does not exist" Oct 08 19:13:00 crc kubenswrapper[4859]: I1008 19:13:00.276913 4859 scope.go:117] "RemoveContainer" containerID="5d1ea73380cc55814ce8d036b21878d3bdfec5eae267674fdd2f54f57b24f9f6" Oct 08 19:13:00 crc kubenswrapper[4859]: E1008 19:13:00.277108 4859 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"5d1ea73380cc55814ce8d036b21878d3bdfec5eae267674fdd2f54f57b24f9f6\": container with ID starting with 5d1ea73380cc55814ce8d036b21878d3bdfec5eae267674fdd2f54f57b24f9f6 not found: ID does not exist" containerID="5d1ea73380cc55814ce8d036b21878d3bdfec5eae267674fdd2f54f57b24f9f6" Oct 08 19:13:00 crc kubenswrapper[4859]: I1008 19:13:00.277136 4859 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"5d1ea73380cc55814ce8d036b21878d3bdfec5eae267674fdd2f54f57b24f9f6"} err="failed to get container status \"5d1ea73380cc55814ce8d036b21878d3bdfec5eae267674fdd2f54f57b24f9f6\": rpc error: code = NotFound desc = could not find container \"5d1ea73380cc55814ce8d036b21878d3bdfec5eae267674fdd2f54f57b24f9f6\": container with ID starting with 5d1ea73380cc55814ce8d036b21878d3bdfec5eae267674fdd2f54f57b24f9f6 not found: ID does not exist" Oct 08 19:13:00 crc kubenswrapper[4859]: I1008 19:13:00.277149 4859 scope.go:117] "RemoveContainer" containerID="122b9c188e3fd5fbdc549b40c82a036cc190b3d5fa99270d0545916b179f54eb" Oct 08 19:13:00 crc kubenswrapper[4859]: E1008 19:13:00.277330 4859 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"122b9c188e3fd5fbdc549b40c82a036cc190b3d5fa99270d0545916b179f54eb\": container with ID starting with 122b9c188e3fd5fbdc549b40c82a036cc190b3d5fa99270d0545916b179f54eb not found: ID does not exist" containerID="122b9c188e3fd5fbdc549b40c82a036cc190b3d5fa99270d0545916b179f54eb" Oct 08 19:13:00 crc kubenswrapper[4859]: I1008 19:13:00.277352 4859 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"122b9c188e3fd5fbdc549b40c82a036cc190b3d5fa99270d0545916b179f54eb"} err="failed to get container status \"122b9c188e3fd5fbdc549b40c82a036cc190b3d5fa99270d0545916b179f54eb\": rpc error: code = NotFound desc = could not find container \"122b9c188e3fd5fbdc549b40c82a036cc190b3d5fa99270d0545916b179f54eb\": container with ID starting with 122b9c188e3fd5fbdc549b40c82a036cc190b3d5fa99270d0545916b179f54eb not found: ID does not exist" Oct 08 19:13:00 crc kubenswrapper[4859]: I1008 19:13:00.482569 4859 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="5e1f266e-f0c3-4020-937c-b30ea071a654" path="/var/lib/kubelet/pods/5e1f266e-f0c3-4020-937c-b30ea071a654/volumes" Oct 08 19:13:00 crc kubenswrapper[4859]: I1008 19:13:00.483528 4859 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="7c3480f8-6e54-47b3-a15d-82ecc7e1737a" path="/var/lib/kubelet/pods/7c3480f8-6e54-47b3-a15d-82ecc7e1737a/volumes" Oct 08 19:13:10 crc kubenswrapper[4859]: I1008 19:13:10.477459 4859 scope.go:117] "RemoveContainer" containerID="cc836fd09088ba4df38f4588526d6fe50b7b131a7fb454b4fd3751d8b1b22560" Oct 08 19:13:10 crc kubenswrapper[4859]: E1008 19:13:10.479651 4859 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-82s52_openshift-machine-config-operator(b23a6a6c-9d92-4e7b-840e-55cfda873a2d)\"" pod="openshift-machine-config-operator/machine-config-daemon-82s52" podUID="b23a6a6c-9d92-4e7b-840e-55cfda873a2d" Oct 08 19:13:21 crc kubenswrapper[4859]: I1008 19:13:21.469471 4859 scope.go:117] "RemoveContainer" containerID="cc836fd09088ba4df38f4588526d6fe50b7b131a7fb454b4fd3751d8b1b22560" Oct 08 19:13:21 crc kubenswrapper[4859]: E1008 19:13:21.470164 4859 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-82s52_openshift-machine-config-operator(b23a6a6c-9d92-4e7b-840e-55cfda873a2d)\"" pod="openshift-machine-config-operator/machine-config-daemon-82s52" podUID="b23a6a6c-9d92-4e7b-840e-55cfda873a2d" Oct 08 19:13:33 crc kubenswrapper[4859]: I1008 19:13:33.470784 4859 scope.go:117] "RemoveContainer" containerID="cc836fd09088ba4df38f4588526d6fe50b7b131a7fb454b4fd3751d8b1b22560" Oct 08 19:13:33 crc kubenswrapper[4859]: E1008 19:13:33.472050 4859 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-82s52_openshift-machine-config-operator(b23a6a6c-9d92-4e7b-840e-55cfda873a2d)\"" pod="openshift-machine-config-operator/machine-config-daemon-82s52" podUID="b23a6a6c-9d92-4e7b-840e-55cfda873a2d" Oct 08 19:13:44 crc kubenswrapper[4859]: I1008 19:13:44.186261 4859 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-kt7nx"] Oct 08 19:13:44 crc kubenswrapper[4859]: E1008 19:13:44.187266 4859 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7c3480f8-6e54-47b3-a15d-82ecc7e1737a" containerName="extract-utilities" Oct 08 19:13:44 crc kubenswrapper[4859]: I1008 19:13:44.187282 4859 state_mem.go:107] "Deleted CPUSet assignment" podUID="7c3480f8-6e54-47b3-a15d-82ecc7e1737a" containerName="extract-utilities" Oct 08 19:13:44 crc kubenswrapper[4859]: E1008 19:13:44.187296 4859 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5e1f266e-f0c3-4020-937c-b30ea071a654" containerName="extract-content" Oct 08 19:13:44 crc kubenswrapper[4859]: I1008 19:13:44.187304 4859 state_mem.go:107] "Deleted CPUSet assignment" podUID="5e1f266e-f0c3-4020-937c-b30ea071a654" containerName="extract-content" Oct 08 19:13:44 crc kubenswrapper[4859]: E1008 19:13:44.187320 4859 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7c3480f8-6e54-47b3-a15d-82ecc7e1737a" containerName="registry-server" Oct 08 19:13:44 crc kubenswrapper[4859]: I1008 19:13:44.187326 4859 state_mem.go:107] "Deleted CPUSet assignment" podUID="7c3480f8-6e54-47b3-a15d-82ecc7e1737a" containerName="registry-server" Oct 08 19:13:44 crc kubenswrapper[4859]: E1008 19:13:44.187354 4859 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7c3480f8-6e54-47b3-a15d-82ecc7e1737a" containerName="extract-content" Oct 08 19:13:44 crc kubenswrapper[4859]: I1008 19:13:44.187361 4859 state_mem.go:107] "Deleted CPUSet assignment" podUID="7c3480f8-6e54-47b3-a15d-82ecc7e1737a" containerName="extract-content" Oct 08 19:13:44 crc kubenswrapper[4859]: E1008 19:13:44.187376 4859 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5e1f266e-f0c3-4020-937c-b30ea071a654" containerName="extract-utilities" Oct 08 19:13:44 crc kubenswrapper[4859]: I1008 19:13:44.187382 4859 state_mem.go:107] "Deleted CPUSet assignment" podUID="5e1f266e-f0c3-4020-937c-b30ea071a654" containerName="extract-utilities" Oct 08 19:13:44 crc kubenswrapper[4859]: E1008 19:13:44.187395 4859 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5e1f266e-f0c3-4020-937c-b30ea071a654" containerName="registry-server" Oct 08 19:13:44 crc kubenswrapper[4859]: I1008 19:13:44.187401 4859 state_mem.go:107] "Deleted CPUSet assignment" podUID="5e1f266e-f0c3-4020-937c-b30ea071a654" containerName="registry-server" Oct 08 19:13:44 crc kubenswrapper[4859]: I1008 19:13:44.187575 4859 memory_manager.go:354] "RemoveStaleState removing state" podUID="7c3480f8-6e54-47b3-a15d-82ecc7e1737a" containerName="registry-server" Oct 08 19:13:44 crc kubenswrapper[4859]: I1008 19:13:44.187593 4859 memory_manager.go:354] "RemoveStaleState removing state" podUID="5e1f266e-f0c3-4020-937c-b30ea071a654" containerName="registry-server" Oct 08 19:13:44 crc kubenswrapper[4859]: I1008 19:13:44.191357 4859 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-kt7nx" Oct 08 19:13:44 crc kubenswrapper[4859]: I1008 19:13:44.207033 4859 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-kt7nx"] Oct 08 19:13:44 crc kubenswrapper[4859]: I1008 19:13:44.217164 4859 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-78vtx\" (UniqueName: \"kubernetes.io/projected/b3988f77-f1e7-4f1e-b4e5-fd1a0cdfefa8-kube-api-access-78vtx\") pod \"redhat-operators-kt7nx\" (UID: \"b3988f77-f1e7-4f1e-b4e5-fd1a0cdfefa8\") " pod="openshift-marketplace/redhat-operators-kt7nx" Oct 08 19:13:44 crc kubenswrapper[4859]: I1008 19:13:44.217406 4859 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b3988f77-f1e7-4f1e-b4e5-fd1a0cdfefa8-catalog-content\") pod \"redhat-operators-kt7nx\" (UID: \"b3988f77-f1e7-4f1e-b4e5-fd1a0cdfefa8\") " pod="openshift-marketplace/redhat-operators-kt7nx" Oct 08 19:13:44 crc kubenswrapper[4859]: I1008 19:13:44.217443 4859 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b3988f77-f1e7-4f1e-b4e5-fd1a0cdfefa8-utilities\") pod \"redhat-operators-kt7nx\" (UID: \"b3988f77-f1e7-4f1e-b4e5-fd1a0cdfefa8\") " pod="openshift-marketplace/redhat-operators-kt7nx" Oct 08 19:13:44 crc kubenswrapper[4859]: I1008 19:13:44.319670 4859 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-78vtx\" (UniqueName: \"kubernetes.io/projected/b3988f77-f1e7-4f1e-b4e5-fd1a0cdfefa8-kube-api-access-78vtx\") pod \"redhat-operators-kt7nx\" (UID: \"b3988f77-f1e7-4f1e-b4e5-fd1a0cdfefa8\") " pod="openshift-marketplace/redhat-operators-kt7nx" Oct 08 19:13:44 crc kubenswrapper[4859]: I1008 19:13:44.319856 4859 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b3988f77-f1e7-4f1e-b4e5-fd1a0cdfefa8-catalog-content\") pod \"redhat-operators-kt7nx\" (UID: \"b3988f77-f1e7-4f1e-b4e5-fd1a0cdfefa8\") " pod="openshift-marketplace/redhat-operators-kt7nx" Oct 08 19:13:44 crc kubenswrapper[4859]: I1008 19:13:44.319901 4859 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b3988f77-f1e7-4f1e-b4e5-fd1a0cdfefa8-utilities\") pod \"redhat-operators-kt7nx\" (UID: \"b3988f77-f1e7-4f1e-b4e5-fd1a0cdfefa8\") " pod="openshift-marketplace/redhat-operators-kt7nx" Oct 08 19:13:44 crc kubenswrapper[4859]: I1008 19:13:44.320398 4859 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b3988f77-f1e7-4f1e-b4e5-fd1a0cdfefa8-catalog-content\") pod \"redhat-operators-kt7nx\" (UID: \"b3988f77-f1e7-4f1e-b4e5-fd1a0cdfefa8\") " pod="openshift-marketplace/redhat-operators-kt7nx" Oct 08 19:13:44 crc kubenswrapper[4859]: I1008 19:13:44.320423 4859 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b3988f77-f1e7-4f1e-b4e5-fd1a0cdfefa8-utilities\") pod \"redhat-operators-kt7nx\" (UID: \"b3988f77-f1e7-4f1e-b4e5-fd1a0cdfefa8\") " pod="openshift-marketplace/redhat-operators-kt7nx" Oct 08 19:13:44 crc kubenswrapper[4859]: I1008 19:13:44.339505 4859 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-78vtx\" (UniqueName: \"kubernetes.io/projected/b3988f77-f1e7-4f1e-b4e5-fd1a0cdfefa8-kube-api-access-78vtx\") pod \"redhat-operators-kt7nx\" (UID: \"b3988f77-f1e7-4f1e-b4e5-fd1a0cdfefa8\") " pod="openshift-marketplace/redhat-operators-kt7nx" Oct 08 19:13:44 crc kubenswrapper[4859]: I1008 19:13:44.532574 4859 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-kt7nx" Oct 08 19:13:45 crc kubenswrapper[4859]: I1008 19:13:45.002256 4859 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-kt7nx"] Oct 08 19:13:45 crc kubenswrapper[4859]: W1008 19:13:45.050497 4859 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podb3988f77_f1e7_4f1e_b4e5_fd1a0cdfefa8.slice/crio-c4767bc532f19a159068ead4da4ea1d2372b5c49dbf1cf3071d212bc67130343 WatchSource:0}: Error finding container c4767bc532f19a159068ead4da4ea1d2372b5c49dbf1cf3071d212bc67130343: Status 404 returned error can't find the container with id c4767bc532f19a159068ead4da4ea1d2372b5c49dbf1cf3071d212bc67130343 Oct 08 19:13:45 crc kubenswrapper[4859]: I1008 19:13:45.592524 4859 generic.go:334] "Generic (PLEG): container finished" podID="b3988f77-f1e7-4f1e-b4e5-fd1a0cdfefa8" containerID="3a44b9a37d506efb5dd245323894ad5a5a794dfbf6ec2be9a0e4bc76fed1b75c" exitCode=0 Oct 08 19:13:45 crc kubenswrapper[4859]: I1008 19:13:45.592610 4859 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-kt7nx" event={"ID":"b3988f77-f1e7-4f1e-b4e5-fd1a0cdfefa8","Type":"ContainerDied","Data":"3a44b9a37d506efb5dd245323894ad5a5a794dfbf6ec2be9a0e4bc76fed1b75c"} Oct 08 19:13:45 crc kubenswrapper[4859]: I1008 19:13:45.593002 4859 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-kt7nx" event={"ID":"b3988f77-f1e7-4f1e-b4e5-fd1a0cdfefa8","Type":"ContainerStarted","Data":"c4767bc532f19a159068ead4da4ea1d2372b5c49dbf1cf3071d212bc67130343"} Oct 08 19:13:46 crc kubenswrapper[4859]: I1008 19:13:46.606717 4859 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-kt7nx" event={"ID":"b3988f77-f1e7-4f1e-b4e5-fd1a0cdfefa8","Type":"ContainerStarted","Data":"c2b4306f65814797b96b2752d18efcd2fdd3af2a2216d107677196e6d1899c7d"} Oct 08 19:13:47 crc kubenswrapper[4859]: I1008 19:13:47.470091 4859 scope.go:117] "RemoveContainer" containerID="cc836fd09088ba4df38f4588526d6fe50b7b131a7fb454b4fd3751d8b1b22560" Oct 08 19:13:47 crc kubenswrapper[4859]: E1008 19:13:47.470661 4859 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-82s52_openshift-machine-config-operator(b23a6a6c-9d92-4e7b-840e-55cfda873a2d)\"" pod="openshift-machine-config-operator/machine-config-daemon-82s52" podUID="b23a6a6c-9d92-4e7b-840e-55cfda873a2d" Oct 08 19:13:47 crc kubenswrapper[4859]: I1008 19:13:47.617843 4859 generic.go:334] "Generic (PLEG): container finished" podID="b3988f77-f1e7-4f1e-b4e5-fd1a0cdfefa8" containerID="c2b4306f65814797b96b2752d18efcd2fdd3af2a2216d107677196e6d1899c7d" exitCode=0 Oct 08 19:13:47 crc kubenswrapper[4859]: I1008 19:13:47.617890 4859 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-kt7nx" event={"ID":"b3988f77-f1e7-4f1e-b4e5-fd1a0cdfefa8","Type":"ContainerDied","Data":"c2b4306f65814797b96b2752d18efcd2fdd3af2a2216d107677196e6d1899c7d"} Oct 08 19:13:48 crc kubenswrapper[4859]: I1008 19:13:48.629604 4859 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-kt7nx" event={"ID":"b3988f77-f1e7-4f1e-b4e5-fd1a0cdfefa8","Type":"ContainerStarted","Data":"de6aa78b904ed133a0479183999ed33dff1679815f4e415b9f75c7f16d98d17a"} Oct 08 19:13:48 crc kubenswrapper[4859]: I1008 19:13:48.654434 4859 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-kt7nx" podStartSLOduration=2.175185871 podStartE2EDuration="4.654417993s" podCreationTimestamp="2025-10-08 19:13:44 +0000 UTC" firstStartedPulling="2025-10-08 19:13:45.595306285 +0000 UTC m=+3395.842145664" lastFinishedPulling="2025-10-08 19:13:48.074538407 +0000 UTC m=+3398.321377786" observedRunningTime="2025-10-08 19:13:48.651507518 +0000 UTC m=+3398.898346907" watchObservedRunningTime="2025-10-08 19:13:48.654417993 +0000 UTC m=+3398.901257362" Oct 08 19:13:54 crc kubenswrapper[4859]: I1008 19:13:54.532899 4859 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-kt7nx" Oct 08 19:13:54 crc kubenswrapper[4859]: I1008 19:13:54.533430 4859 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-kt7nx" Oct 08 19:13:54 crc kubenswrapper[4859]: I1008 19:13:54.580273 4859 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-kt7nx" Oct 08 19:13:54 crc kubenswrapper[4859]: I1008 19:13:54.751379 4859 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-kt7nx" Oct 08 19:13:54 crc kubenswrapper[4859]: I1008 19:13:54.821140 4859 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-kt7nx"] Oct 08 19:13:56 crc kubenswrapper[4859]: I1008 19:13:56.700670 4859 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-operators-kt7nx" podUID="b3988f77-f1e7-4f1e-b4e5-fd1a0cdfefa8" containerName="registry-server" containerID="cri-o://de6aa78b904ed133a0479183999ed33dff1679815f4e415b9f75c7f16d98d17a" gracePeriod=2 Oct 08 19:13:57 crc kubenswrapper[4859]: I1008 19:13:57.179635 4859 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-kt7nx" Oct 08 19:13:57 crc kubenswrapper[4859]: I1008 19:13:57.292946 4859 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-78vtx\" (UniqueName: \"kubernetes.io/projected/b3988f77-f1e7-4f1e-b4e5-fd1a0cdfefa8-kube-api-access-78vtx\") pod \"b3988f77-f1e7-4f1e-b4e5-fd1a0cdfefa8\" (UID: \"b3988f77-f1e7-4f1e-b4e5-fd1a0cdfefa8\") " Oct 08 19:13:57 crc kubenswrapper[4859]: I1008 19:13:57.293115 4859 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b3988f77-f1e7-4f1e-b4e5-fd1a0cdfefa8-catalog-content\") pod \"b3988f77-f1e7-4f1e-b4e5-fd1a0cdfefa8\" (UID: \"b3988f77-f1e7-4f1e-b4e5-fd1a0cdfefa8\") " Oct 08 19:13:57 crc kubenswrapper[4859]: I1008 19:13:57.293153 4859 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b3988f77-f1e7-4f1e-b4e5-fd1a0cdfefa8-utilities\") pod \"b3988f77-f1e7-4f1e-b4e5-fd1a0cdfefa8\" (UID: \"b3988f77-f1e7-4f1e-b4e5-fd1a0cdfefa8\") " Oct 08 19:13:57 crc kubenswrapper[4859]: I1008 19:13:57.294356 4859 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/b3988f77-f1e7-4f1e-b4e5-fd1a0cdfefa8-utilities" (OuterVolumeSpecName: "utilities") pod "b3988f77-f1e7-4f1e-b4e5-fd1a0cdfefa8" (UID: "b3988f77-f1e7-4f1e-b4e5-fd1a0cdfefa8"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 08 19:13:57 crc kubenswrapper[4859]: I1008 19:13:57.298526 4859 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b3988f77-f1e7-4f1e-b4e5-fd1a0cdfefa8-kube-api-access-78vtx" (OuterVolumeSpecName: "kube-api-access-78vtx") pod "b3988f77-f1e7-4f1e-b4e5-fd1a0cdfefa8" (UID: "b3988f77-f1e7-4f1e-b4e5-fd1a0cdfefa8"). InnerVolumeSpecName "kube-api-access-78vtx". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 08 19:13:57 crc kubenswrapper[4859]: I1008 19:13:57.395169 4859 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-78vtx\" (UniqueName: \"kubernetes.io/projected/b3988f77-f1e7-4f1e-b4e5-fd1a0cdfefa8-kube-api-access-78vtx\") on node \"crc\" DevicePath \"\"" Oct 08 19:13:57 crc kubenswrapper[4859]: I1008 19:13:57.395219 4859 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b3988f77-f1e7-4f1e-b4e5-fd1a0cdfefa8-utilities\") on node \"crc\" DevicePath \"\"" Oct 08 19:13:57 crc kubenswrapper[4859]: I1008 19:13:57.710810 4859 generic.go:334] "Generic (PLEG): container finished" podID="b3988f77-f1e7-4f1e-b4e5-fd1a0cdfefa8" containerID="de6aa78b904ed133a0479183999ed33dff1679815f4e415b9f75c7f16d98d17a" exitCode=0 Oct 08 19:13:57 crc kubenswrapper[4859]: I1008 19:13:57.710873 4859 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-kt7nx" event={"ID":"b3988f77-f1e7-4f1e-b4e5-fd1a0cdfefa8","Type":"ContainerDied","Data":"de6aa78b904ed133a0479183999ed33dff1679815f4e415b9f75c7f16d98d17a"} Oct 08 19:13:57 crc kubenswrapper[4859]: I1008 19:13:57.710911 4859 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-kt7nx" event={"ID":"b3988f77-f1e7-4f1e-b4e5-fd1a0cdfefa8","Type":"ContainerDied","Data":"c4767bc532f19a159068ead4da4ea1d2372b5c49dbf1cf3071d212bc67130343"} Oct 08 19:13:57 crc kubenswrapper[4859]: I1008 19:13:57.710937 4859 scope.go:117] "RemoveContainer" containerID="de6aa78b904ed133a0479183999ed33dff1679815f4e415b9f75c7f16d98d17a" Oct 08 19:13:57 crc kubenswrapper[4859]: I1008 19:13:57.711113 4859 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-kt7nx" Oct 08 19:13:57 crc kubenswrapper[4859]: I1008 19:13:57.740356 4859 scope.go:117] "RemoveContainer" containerID="c2b4306f65814797b96b2752d18efcd2fdd3af2a2216d107677196e6d1899c7d" Oct 08 19:13:57 crc kubenswrapper[4859]: I1008 19:13:57.774860 4859 scope.go:117] "RemoveContainer" containerID="3a44b9a37d506efb5dd245323894ad5a5a794dfbf6ec2be9a0e4bc76fed1b75c" Oct 08 19:13:57 crc kubenswrapper[4859]: I1008 19:13:57.820898 4859 scope.go:117] "RemoveContainer" containerID="de6aa78b904ed133a0479183999ed33dff1679815f4e415b9f75c7f16d98d17a" Oct 08 19:13:57 crc kubenswrapper[4859]: E1008 19:13:57.821351 4859 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"de6aa78b904ed133a0479183999ed33dff1679815f4e415b9f75c7f16d98d17a\": container with ID starting with de6aa78b904ed133a0479183999ed33dff1679815f4e415b9f75c7f16d98d17a not found: ID does not exist" containerID="de6aa78b904ed133a0479183999ed33dff1679815f4e415b9f75c7f16d98d17a" Oct 08 19:13:57 crc kubenswrapper[4859]: I1008 19:13:57.821390 4859 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"de6aa78b904ed133a0479183999ed33dff1679815f4e415b9f75c7f16d98d17a"} err="failed to get container status \"de6aa78b904ed133a0479183999ed33dff1679815f4e415b9f75c7f16d98d17a\": rpc error: code = NotFound desc = could not find container \"de6aa78b904ed133a0479183999ed33dff1679815f4e415b9f75c7f16d98d17a\": container with ID starting with de6aa78b904ed133a0479183999ed33dff1679815f4e415b9f75c7f16d98d17a not found: ID does not exist" Oct 08 19:13:57 crc kubenswrapper[4859]: I1008 19:13:57.821417 4859 scope.go:117] "RemoveContainer" containerID="c2b4306f65814797b96b2752d18efcd2fdd3af2a2216d107677196e6d1899c7d" Oct 08 19:13:57 crc kubenswrapper[4859]: E1008 19:13:57.821653 4859 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"c2b4306f65814797b96b2752d18efcd2fdd3af2a2216d107677196e6d1899c7d\": container with ID starting with c2b4306f65814797b96b2752d18efcd2fdd3af2a2216d107677196e6d1899c7d not found: ID does not exist" containerID="c2b4306f65814797b96b2752d18efcd2fdd3af2a2216d107677196e6d1899c7d" Oct 08 19:13:57 crc kubenswrapper[4859]: I1008 19:13:57.821675 4859 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"c2b4306f65814797b96b2752d18efcd2fdd3af2a2216d107677196e6d1899c7d"} err="failed to get container status \"c2b4306f65814797b96b2752d18efcd2fdd3af2a2216d107677196e6d1899c7d\": rpc error: code = NotFound desc = could not find container \"c2b4306f65814797b96b2752d18efcd2fdd3af2a2216d107677196e6d1899c7d\": container with ID starting with c2b4306f65814797b96b2752d18efcd2fdd3af2a2216d107677196e6d1899c7d not found: ID does not exist" Oct 08 19:13:57 crc kubenswrapper[4859]: I1008 19:13:57.821702 4859 scope.go:117] "RemoveContainer" containerID="3a44b9a37d506efb5dd245323894ad5a5a794dfbf6ec2be9a0e4bc76fed1b75c" Oct 08 19:13:57 crc kubenswrapper[4859]: E1008 19:13:57.821969 4859 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"3a44b9a37d506efb5dd245323894ad5a5a794dfbf6ec2be9a0e4bc76fed1b75c\": container with ID starting with 3a44b9a37d506efb5dd245323894ad5a5a794dfbf6ec2be9a0e4bc76fed1b75c not found: ID does not exist" containerID="3a44b9a37d506efb5dd245323894ad5a5a794dfbf6ec2be9a0e4bc76fed1b75c" Oct 08 19:13:57 crc kubenswrapper[4859]: I1008 19:13:57.821993 4859 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"3a44b9a37d506efb5dd245323894ad5a5a794dfbf6ec2be9a0e4bc76fed1b75c"} err="failed to get container status \"3a44b9a37d506efb5dd245323894ad5a5a794dfbf6ec2be9a0e4bc76fed1b75c\": rpc error: code = NotFound desc = could not find container \"3a44b9a37d506efb5dd245323894ad5a5a794dfbf6ec2be9a0e4bc76fed1b75c\": container with ID starting with 3a44b9a37d506efb5dd245323894ad5a5a794dfbf6ec2be9a0e4bc76fed1b75c not found: ID does not exist" Oct 08 19:13:58 crc kubenswrapper[4859]: I1008 19:13:58.462029 4859 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/b3988f77-f1e7-4f1e-b4e5-fd1a0cdfefa8-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "b3988f77-f1e7-4f1e-b4e5-fd1a0cdfefa8" (UID: "b3988f77-f1e7-4f1e-b4e5-fd1a0cdfefa8"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 08 19:13:58 crc kubenswrapper[4859]: I1008 19:13:58.518825 4859 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b3988f77-f1e7-4f1e-b4e5-fd1a0cdfefa8-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 08 19:13:58 crc kubenswrapper[4859]: I1008 19:13:58.659970 4859 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-kt7nx"] Oct 08 19:13:58 crc kubenswrapper[4859]: I1008 19:13:58.669117 4859 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-operators-kt7nx"] Oct 08 19:14:00 crc kubenswrapper[4859]: I1008 19:14:00.483158 4859 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="b3988f77-f1e7-4f1e-b4e5-fd1a0cdfefa8" path="/var/lib/kubelet/pods/b3988f77-f1e7-4f1e-b4e5-fd1a0cdfefa8/volumes" Oct 08 19:14:01 crc kubenswrapper[4859]: I1008 19:14:01.471501 4859 scope.go:117] "RemoveContainer" containerID="cc836fd09088ba4df38f4588526d6fe50b7b131a7fb454b4fd3751d8b1b22560" Oct 08 19:14:01 crc kubenswrapper[4859]: E1008 19:14:01.471882 4859 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-82s52_openshift-machine-config-operator(b23a6a6c-9d92-4e7b-840e-55cfda873a2d)\"" pod="openshift-machine-config-operator/machine-config-daemon-82s52" podUID="b23a6a6c-9d92-4e7b-840e-55cfda873a2d" Oct 08 19:14:12 crc kubenswrapper[4859]: I1008 19:14:12.469804 4859 scope.go:117] "RemoveContainer" containerID="cc836fd09088ba4df38f4588526d6fe50b7b131a7fb454b4fd3751d8b1b22560" Oct 08 19:14:12 crc kubenswrapper[4859]: E1008 19:14:12.470649 4859 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-82s52_openshift-machine-config-operator(b23a6a6c-9d92-4e7b-840e-55cfda873a2d)\"" pod="openshift-machine-config-operator/machine-config-daemon-82s52" podUID="b23a6a6c-9d92-4e7b-840e-55cfda873a2d" Oct 08 19:14:24 crc kubenswrapper[4859]: I1008 19:14:24.470576 4859 scope.go:117] "RemoveContainer" containerID="cc836fd09088ba4df38f4588526d6fe50b7b131a7fb454b4fd3751d8b1b22560" Oct 08 19:14:24 crc kubenswrapper[4859]: I1008 19:14:24.986651 4859 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-82s52" event={"ID":"b23a6a6c-9d92-4e7b-840e-55cfda873a2d","Type":"ContainerStarted","Data":"09fd446535f97e643340dafbb38e492a2f1d2902d0d749edf84136fd6fabb2ac"} Oct 08 19:15:00 crc kubenswrapper[4859]: I1008 19:15:00.186119 4859 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29332515-tr2n4"] Oct 08 19:15:00 crc kubenswrapper[4859]: E1008 19:15:00.187035 4859 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b3988f77-f1e7-4f1e-b4e5-fd1a0cdfefa8" containerName="extract-utilities" Oct 08 19:15:00 crc kubenswrapper[4859]: I1008 19:15:00.187049 4859 state_mem.go:107] "Deleted CPUSet assignment" podUID="b3988f77-f1e7-4f1e-b4e5-fd1a0cdfefa8" containerName="extract-utilities" Oct 08 19:15:00 crc kubenswrapper[4859]: E1008 19:15:00.187071 4859 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b3988f77-f1e7-4f1e-b4e5-fd1a0cdfefa8" containerName="extract-content" Oct 08 19:15:00 crc kubenswrapper[4859]: I1008 19:15:00.187078 4859 state_mem.go:107] "Deleted CPUSet assignment" podUID="b3988f77-f1e7-4f1e-b4e5-fd1a0cdfefa8" containerName="extract-content" Oct 08 19:15:00 crc kubenswrapper[4859]: E1008 19:15:00.187102 4859 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b3988f77-f1e7-4f1e-b4e5-fd1a0cdfefa8" containerName="registry-server" Oct 08 19:15:00 crc kubenswrapper[4859]: I1008 19:15:00.187109 4859 state_mem.go:107] "Deleted CPUSet assignment" podUID="b3988f77-f1e7-4f1e-b4e5-fd1a0cdfefa8" containerName="registry-server" Oct 08 19:15:00 crc kubenswrapper[4859]: I1008 19:15:00.187296 4859 memory_manager.go:354] "RemoveStaleState removing state" podUID="b3988f77-f1e7-4f1e-b4e5-fd1a0cdfefa8" containerName="registry-server" Oct 08 19:15:00 crc kubenswrapper[4859]: I1008 19:15:00.187900 4859 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29332515-tr2n4" Oct 08 19:15:00 crc kubenswrapper[4859]: I1008 19:15:00.191124 4859 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"collect-profiles-config" Oct 08 19:15:00 crc kubenswrapper[4859]: I1008 19:15:00.191137 4859 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"collect-profiles-dockercfg-kzf4t" Oct 08 19:15:00 crc kubenswrapper[4859]: I1008 19:15:00.198168 4859 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29332515-tr2n4"] Oct 08 19:15:00 crc kubenswrapper[4859]: I1008 19:15:00.321791 4859 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-dmqtd\" (UniqueName: \"kubernetes.io/projected/c91f6897-a7a6-48ef-9dc3-5e3d57378b08-kube-api-access-dmqtd\") pod \"collect-profiles-29332515-tr2n4\" (UID: \"c91f6897-a7a6-48ef-9dc3-5e3d57378b08\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29332515-tr2n4" Oct 08 19:15:00 crc kubenswrapper[4859]: I1008 19:15:00.321864 4859 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/c91f6897-a7a6-48ef-9dc3-5e3d57378b08-config-volume\") pod \"collect-profiles-29332515-tr2n4\" (UID: \"c91f6897-a7a6-48ef-9dc3-5e3d57378b08\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29332515-tr2n4" Oct 08 19:15:00 crc kubenswrapper[4859]: I1008 19:15:00.321889 4859 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/c91f6897-a7a6-48ef-9dc3-5e3d57378b08-secret-volume\") pod \"collect-profiles-29332515-tr2n4\" (UID: \"c91f6897-a7a6-48ef-9dc3-5e3d57378b08\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29332515-tr2n4" Oct 08 19:15:00 crc kubenswrapper[4859]: I1008 19:15:00.423851 4859 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-dmqtd\" (UniqueName: \"kubernetes.io/projected/c91f6897-a7a6-48ef-9dc3-5e3d57378b08-kube-api-access-dmqtd\") pod \"collect-profiles-29332515-tr2n4\" (UID: \"c91f6897-a7a6-48ef-9dc3-5e3d57378b08\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29332515-tr2n4" Oct 08 19:15:00 crc kubenswrapper[4859]: I1008 19:15:00.423940 4859 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/c91f6897-a7a6-48ef-9dc3-5e3d57378b08-config-volume\") pod \"collect-profiles-29332515-tr2n4\" (UID: \"c91f6897-a7a6-48ef-9dc3-5e3d57378b08\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29332515-tr2n4" Oct 08 19:15:00 crc kubenswrapper[4859]: I1008 19:15:00.423965 4859 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/c91f6897-a7a6-48ef-9dc3-5e3d57378b08-secret-volume\") pod \"collect-profiles-29332515-tr2n4\" (UID: \"c91f6897-a7a6-48ef-9dc3-5e3d57378b08\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29332515-tr2n4" Oct 08 19:15:00 crc kubenswrapper[4859]: I1008 19:15:00.424853 4859 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/c91f6897-a7a6-48ef-9dc3-5e3d57378b08-config-volume\") pod \"collect-profiles-29332515-tr2n4\" (UID: \"c91f6897-a7a6-48ef-9dc3-5e3d57378b08\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29332515-tr2n4" Oct 08 19:15:00 crc kubenswrapper[4859]: I1008 19:15:00.429750 4859 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/c91f6897-a7a6-48ef-9dc3-5e3d57378b08-secret-volume\") pod \"collect-profiles-29332515-tr2n4\" (UID: \"c91f6897-a7a6-48ef-9dc3-5e3d57378b08\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29332515-tr2n4" Oct 08 19:15:00 crc kubenswrapper[4859]: I1008 19:15:00.464982 4859 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-dmqtd\" (UniqueName: \"kubernetes.io/projected/c91f6897-a7a6-48ef-9dc3-5e3d57378b08-kube-api-access-dmqtd\") pod \"collect-profiles-29332515-tr2n4\" (UID: \"c91f6897-a7a6-48ef-9dc3-5e3d57378b08\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29332515-tr2n4" Oct 08 19:15:00 crc kubenswrapper[4859]: I1008 19:15:00.518366 4859 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29332515-tr2n4" Oct 08 19:15:00 crc kubenswrapper[4859]: I1008 19:15:00.971422 4859 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29332515-tr2n4"] Oct 08 19:15:01 crc kubenswrapper[4859]: I1008 19:15:01.329896 4859 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29332515-tr2n4" event={"ID":"c91f6897-a7a6-48ef-9dc3-5e3d57378b08","Type":"ContainerStarted","Data":"1af3ecfa705bdfc395c1df5b5d74279198c679fe155fb8ab968363ab4c7fbcab"} Oct 08 19:15:01 crc kubenswrapper[4859]: I1008 19:15:01.330182 4859 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29332515-tr2n4" event={"ID":"c91f6897-a7a6-48ef-9dc3-5e3d57378b08","Type":"ContainerStarted","Data":"30fc35a5d3258033af812bdd3aaa242d9b87811a3d9b6a53cdc2a0db8b589ac7"} Oct 08 19:15:02 crc kubenswrapper[4859]: I1008 19:15:02.348565 4859 generic.go:334] "Generic (PLEG): container finished" podID="c91f6897-a7a6-48ef-9dc3-5e3d57378b08" containerID="1af3ecfa705bdfc395c1df5b5d74279198c679fe155fb8ab968363ab4c7fbcab" exitCode=0 Oct 08 19:15:02 crc kubenswrapper[4859]: I1008 19:15:02.348619 4859 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29332515-tr2n4" event={"ID":"c91f6897-a7a6-48ef-9dc3-5e3d57378b08","Type":"ContainerDied","Data":"1af3ecfa705bdfc395c1df5b5d74279198c679fe155fb8ab968363ab4c7fbcab"} Oct 08 19:15:02 crc kubenswrapper[4859]: I1008 19:15:02.738982 4859 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29332515-tr2n4" Oct 08 19:15:02 crc kubenswrapper[4859]: I1008 19:15:02.873292 4859 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-dmqtd\" (UniqueName: \"kubernetes.io/projected/c91f6897-a7a6-48ef-9dc3-5e3d57378b08-kube-api-access-dmqtd\") pod \"c91f6897-a7a6-48ef-9dc3-5e3d57378b08\" (UID: \"c91f6897-a7a6-48ef-9dc3-5e3d57378b08\") " Oct 08 19:15:02 crc kubenswrapper[4859]: I1008 19:15:02.873411 4859 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/c91f6897-a7a6-48ef-9dc3-5e3d57378b08-config-volume\") pod \"c91f6897-a7a6-48ef-9dc3-5e3d57378b08\" (UID: \"c91f6897-a7a6-48ef-9dc3-5e3d57378b08\") " Oct 08 19:15:02 crc kubenswrapper[4859]: I1008 19:15:02.873445 4859 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/c91f6897-a7a6-48ef-9dc3-5e3d57378b08-secret-volume\") pod \"c91f6897-a7a6-48ef-9dc3-5e3d57378b08\" (UID: \"c91f6897-a7a6-48ef-9dc3-5e3d57378b08\") " Oct 08 19:15:02 crc kubenswrapper[4859]: I1008 19:15:02.874344 4859 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/c91f6897-a7a6-48ef-9dc3-5e3d57378b08-config-volume" (OuterVolumeSpecName: "config-volume") pod "c91f6897-a7a6-48ef-9dc3-5e3d57378b08" (UID: "c91f6897-a7a6-48ef-9dc3-5e3d57378b08"). InnerVolumeSpecName "config-volume". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 08 19:15:02 crc kubenswrapper[4859]: I1008 19:15:02.878442 4859 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/c91f6897-a7a6-48ef-9dc3-5e3d57378b08-kube-api-access-dmqtd" (OuterVolumeSpecName: "kube-api-access-dmqtd") pod "c91f6897-a7a6-48ef-9dc3-5e3d57378b08" (UID: "c91f6897-a7a6-48ef-9dc3-5e3d57378b08"). InnerVolumeSpecName "kube-api-access-dmqtd". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 08 19:15:02 crc kubenswrapper[4859]: I1008 19:15:02.879440 4859 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c91f6897-a7a6-48ef-9dc3-5e3d57378b08-secret-volume" (OuterVolumeSpecName: "secret-volume") pod "c91f6897-a7a6-48ef-9dc3-5e3d57378b08" (UID: "c91f6897-a7a6-48ef-9dc3-5e3d57378b08"). InnerVolumeSpecName "secret-volume". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 08 19:15:02 crc kubenswrapper[4859]: I1008 19:15:02.975834 4859 reconciler_common.go:293] "Volume detached for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/c91f6897-a7a6-48ef-9dc3-5e3d57378b08-config-volume\") on node \"crc\" DevicePath \"\"" Oct 08 19:15:02 crc kubenswrapper[4859]: I1008 19:15:02.975877 4859 reconciler_common.go:293] "Volume detached for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/c91f6897-a7a6-48ef-9dc3-5e3d57378b08-secret-volume\") on node \"crc\" DevicePath \"\"" Oct 08 19:15:02 crc kubenswrapper[4859]: I1008 19:15:02.975892 4859 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-dmqtd\" (UniqueName: \"kubernetes.io/projected/c91f6897-a7a6-48ef-9dc3-5e3d57378b08-kube-api-access-dmqtd\") on node \"crc\" DevicePath \"\"" Oct 08 19:15:03 crc kubenswrapper[4859]: I1008 19:15:03.358380 4859 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29332515-tr2n4" event={"ID":"c91f6897-a7a6-48ef-9dc3-5e3d57378b08","Type":"ContainerDied","Data":"30fc35a5d3258033af812bdd3aaa242d9b87811a3d9b6a53cdc2a0db8b589ac7"} Oct 08 19:15:03 crc kubenswrapper[4859]: I1008 19:15:03.358420 4859 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="30fc35a5d3258033af812bdd3aaa242d9b87811a3d9b6a53cdc2a0db8b589ac7" Oct 08 19:15:03 crc kubenswrapper[4859]: I1008 19:15:03.358433 4859 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29332515-tr2n4" Oct 08 19:15:03 crc kubenswrapper[4859]: I1008 19:15:03.813558 4859 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29332470-sc2qz"] Oct 08 19:15:03 crc kubenswrapper[4859]: I1008 19:15:03.820677 4859 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29332470-sc2qz"] Oct 08 19:15:04 crc kubenswrapper[4859]: I1008 19:15:04.500419 4859 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="af0b7fda-8561-4159-b38f-882401a5fc29" path="/var/lib/kubelet/pods/af0b7fda-8561-4159-b38f-882401a5fc29/volumes" Oct 08 19:15:24 crc kubenswrapper[4859]: I1008 19:15:24.727077 4859 scope.go:117] "RemoveContainer" containerID="562b43d35c2214b9dffd19a414e87c80d6a071857c70d6baefe20e372475003c" Oct 08 19:16:25 crc kubenswrapper[4859]: I1008 19:16:25.208086 4859 generic.go:334] "Generic (PLEG): container finished" podID="be909489-e6c5-4464-9292-aacd831885d7" containerID="139da404a051f6e011fc363bac25a65ad6b6fd39197c4b1e2be03afe722fa5ba" exitCode=0 Oct 08 19:16:25 crc kubenswrapper[4859]: I1008 19:16:25.208266 4859 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/tempest-tests-tempest" event={"ID":"be909489-e6c5-4464-9292-aacd831885d7","Type":"ContainerDied","Data":"139da404a051f6e011fc363bac25a65ad6b6fd39197c4b1e2be03afe722fa5ba"} Oct 08 19:16:26 crc kubenswrapper[4859]: I1008 19:16:26.585936 4859 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/tempest-tests-tempest" Oct 08 19:16:26 crc kubenswrapper[4859]: I1008 19:16:26.711959 4859 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/be909489-e6c5-4464-9292-aacd831885d7-openstack-config\") pod \"be909489-e6c5-4464-9292-aacd831885d7\" (UID: \"be909489-e6c5-4464-9292-aacd831885d7\") " Oct 08 19:16:26 crc kubenswrapper[4859]: I1008 19:16:26.712034 4859 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ca-certs\" (UniqueName: \"kubernetes.io/secret/be909489-e6c5-4464-9292-aacd831885d7-ca-certs\") pod \"be909489-e6c5-4464-9292-aacd831885d7\" (UID: \"be909489-e6c5-4464-9292-aacd831885d7\") " Oct 08 19:16:26 crc kubenswrapper[4859]: I1008 19:16:26.712052 4859 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"test-operator-logs\" (UniqueName: \"kubernetes.io/local-volume/local-storage11-crc\") pod \"be909489-e6c5-4464-9292-aacd831885d7\" (UID: \"be909489-e6c5-4464-9292-aacd831885d7\") " Oct 08 19:16:26 crc kubenswrapper[4859]: I1008 19:16:26.712083 4859 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/be909489-e6c5-4464-9292-aacd831885d7-openstack-config-secret\") pod \"be909489-e6c5-4464-9292-aacd831885d7\" (UID: \"be909489-e6c5-4464-9292-aacd831885d7\") " Oct 08 19:16:26 crc kubenswrapper[4859]: I1008 19:16:26.712117 4859 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"test-operator-ephemeral-temporary\" (UniqueName: \"kubernetes.io/empty-dir/be909489-e6c5-4464-9292-aacd831885d7-test-operator-ephemeral-temporary\") pod \"be909489-e6c5-4464-9292-aacd831885d7\" (UID: \"be909489-e6c5-4464-9292-aacd831885d7\") " Oct 08 19:16:26 crc kubenswrapper[4859]: I1008 19:16:26.712223 4859 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/be909489-e6c5-4464-9292-aacd831885d7-ssh-key\") pod \"be909489-e6c5-4464-9292-aacd831885d7\" (UID: \"be909489-e6c5-4464-9292-aacd831885d7\") " Oct 08 19:16:26 crc kubenswrapper[4859]: I1008 19:16:26.712242 4859 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-8psds\" (UniqueName: \"kubernetes.io/projected/be909489-e6c5-4464-9292-aacd831885d7-kube-api-access-8psds\") pod \"be909489-e6c5-4464-9292-aacd831885d7\" (UID: \"be909489-e6c5-4464-9292-aacd831885d7\") " Oct 08 19:16:26 crc kubenswrapper[4859]: I1008 19:16:26.712262 4859 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/be909489-e6c5-4464-9292-aacd831885d7-config-data\") pod \"be909489-e6c5-4464-9292-aacd831885d7\" (UID: \"be909489-e6c5-4464-9292-aacd831885d7\") " Oct 08 19:16:26 crc kubenswrapper[4859]: I1008 19:16:26.712294 4859 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"test-operator-ephemeral-workdir\" (UniqueName: \"kubernetes.io/empty-dir/be909489-e6c5-4464-9292-aacd831885d7-test-operator-ephemeral-workdir\") pod \"be909489-e6c5-4464-9292-aacd831885d7\" (UID: \"be909489-e6c5-4464-9292-aacd831885d7\") " Oct 08 19:16:26 crc kubenswrapper[4859]: I1008 19:16:26.713151 4859 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/be909489-e6c5-4464-9292-aacd831885d7-test-operator-ephemeral-temporary" (OuterVolumeSpecName: "test-operator-ephemeral-temporary") pod "be909489-e6c5-4464-9292-aacd831885d7" (UID: "be909489-e6c5-4464-9292-aacd831885d7"). InnerVolumeSpecName "test-operator-ephemeral-temporary". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 08 19:16:26 crc kubenswrapper[4859]: I1008 19:16:26.713356 4859 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/be909489-e6c5-4464-9292-aacd831885d7-config-data" (OuterVolumeSpecName: "config-data") pod "be909489-e6c5-4464-9292-aacd831885d7" (UID: "be909489-e6c5-4464-9292-aacd831885d7"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 08 19:16:26 crc kubenswrapper[4859]: I1008 19:16:26.717920 4859 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/local-volume/local-storage11-crc" (OuterVolumeSpecName: "test-operator-logs") pod "be909489-e6c5-4464-9292-aacd831885d7" (UID: "be909489-e6c5-4464-9292-aacd831885d7"). InnerVolumeSpecName "local-storage11-crc". PluginName "kubernetes.io/local-volume", VolumeGidValue "" Oct 08 19:16:26 crc kubenswrapper[4859]: I1008 19:16:26.717994 4859 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/be909489-e6c5-4464-9292-aacd831885d7-kube-api-access-8psds" (OuterVolumeSpecName: "kube-api-access-8psds") pod "be909489-e6c5-4464-9292-aacd831885d7" (UID: "be909489-e6c5-4464-9292-aacd831885d7"). InnerVolumeSpecName "kube-api-access-8psds". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 08 19:16:26 crc kubenswrapper[4859]: I1008 19:16:26.722448 4859 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/be909489-e6c5-4464-9292-aacd831885d7-test-operator-ephemeral-workdir" (OuterVolumeSpecName: "test-operator-ephemeral-workdir") pod "be909489-e6c5-4464-9292-aacd831885d7" (UID: "be909489-e6c5-4464-9292-aacd831885d7"). InnerVolumeSpecName "test-operator-ephemeral-workdir". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 08 19:16:26 crc kubenswrapper[4859]: I1008 19:16:26.741077 4859 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/be909489-e6c5-4464-9292-aacd831885d7-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "be909489-e6c5-4464-9292-aacd831885d7" (UID: "be909489-e6c5-4464-9292-aacd831885d7"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 08 19:16:26 crc kubenswrapper[4859]: I1008 19:16:26.743479 4859 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/be909489-e6c5-4464-9292-aacd831885d7-openstack-config-secret" (OuterVolumeSpecName: "openstack-config-secret") pod "be909489-e6c5-4464-9292-aacd831885d7" (UID: "be909489-e6c5-4464-9292-aacd831885d7"). InnerVolumeSpecName "openstack-config-secret". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 08 19:16:26 crc kubenswrapper[4859]: I1008 19:16:26.747903 4859 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/be909489-e6c5-4464-9292-aacd831885d7-ca-certs" (OuterVolumeSpecName: "ca-certs") pod "be909489-e6c5-4464-9292-aacd831885d7" (UID: "be909489-e6c5-4464-9292-aacd831885d7"). InnerVolumeSpecName "ca-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 08 19:16:26 crc kubenswrapper[4859]: I1008 19:16:26.781196 4859 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/be909489-e6c5-4464-9292-aacd831885d7-openstack-config" (OuterVolumeSpecName: "openstack-config") pod "be909489-e6c5-4464-9292-aacd831885d7" (UID: "be909489-e6c5-4464-9292-aacd831885d7"). InnerVolumeSpecName "openstack-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 08 19:16:26 crc kubenswrapper[4859]: I1008 19:16:26.815227 4859 reconciler_common.go:293] "Volume detached for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/be909489-e6c5-4464-9292-aacd831885d7-openstack-config\") on node \"crc\" DevicePath \"\"" Oct 08 19:16:26 crc kubenswrapper[4859]: I1008 19:16:26.815276 4859 reconciler_common.go:293] "Volume detached for volume \"ca-certs\" (UniqueName: \"kubernetes.io/secret/be909489-e6c5-4464-9292-aacd831885d7-ca-certs\") on node \"crc\" DevicePath \"\"" Oct 08 19:16:26 crc kubenswrapper[4859]: I1008 19:16:26.815326 4859 reconciler_common.go:286] "operationExecutor.UnmountDevice started for volume \"local-storage11-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage11-crc\") on node \"crc\" " Oct 08 19:16:26 crc kubenswrapper[4859]: I1008 19:16:26.815344 4859 reconciler_common.go:293] "Volume detached for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/be909489-e6c5-4464-9292-aacd831885d7-openstack-config-secret\") on node \"crc\" DevicePath \"\"" Oct 08 19:16:26 crc kubenswrapper[4859]: I1008 19:16:26.815364 4859 reconciler_common.go:293] "Volume detached for volume \"test-operator-ephemeral-temporary\" (UniqueName: \"kubernetes.io/empty-dir/be909489-e6c5-4464-9292-aacd831885d7-test-operator-ephemeral-temporary\") on node \"crc\" DevicePath \"\"" Oct 08 19:16:26 crc kubenswrapper[4859]: I1008 19:16:26.815382 4859 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/be909489-e6c5-4464-9292-aacd831885d7-ssh-key\") on node \"crc\" DevicePath \"\"" Oct 08 19:16:26 crc kubenswrapper[4859]: I1008 19:16:26.815398 4859 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-8psds\" (UniqueName: \"kubernetes.io/projected/be909489-e6c5-4464-9292-aacd831885d7-kube-api-access-8psds\") on node \"crc\" DevicePath \"\"" Oct 08 19:16:26 crc kubenswrapper[4859]: I1008 19:16:26.815412 4859 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/be909489-e6c5-4464-9292-aacd831885d7-config-data\") on node \"crc\" DevicePath \"\"" Oct 08 19:16:26 crc kubenswrapper[4859]: I1008 19:16:26.815427 4859 reconciler_common.go:293] "Volume detached for volume \"test-operator-ephemeral-workdir\" (UniqueName: \"kubernetes.io/empty-dir/be909489-e6c5-4464-9292-aacd831885d7-test-operator-ephemeral-workdir\") on node \"crc\" DevicePath \"\"" Oct 08 19:16:26 crc kubenswrapper[4859]: I1008 19:16:26.841454 4859 operation_generator.go:917] UnmountDevice succeeded for volume "local-storage11-crc" (UniqueName: "kubernetes.io/local-volume/local-storage11-crc") on node "crc" Oct 08 19:16:26 crc kubenswrapper[4859]: I1008 19:16:26.916661 4859 reconciler_common.go:293] "Volume detached for volume \"local-storage11-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage11-crc\") on node \"crc\" DevicePath \"\"" Oct 08 19:16:27 crc kubenswrapper[4859]: I1008 19:16:27.229761 4859 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/tempest-tests-tempest" event={"ID":"be909489-e6c5-4464-9292-aacd831885d7","Type":"ContainerDied","Data":"2ec230b487b50f18cb8c914ed6316cb461e578fece9aa8b7a2ea76eaf15045de"} Oct 08 19:16:27 crc kubenswrapper[4859]: I1008 19:16:27.229815 4859 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="2ec230b487b50f18cb8c914ed6316cb461e578fece9aa8b7a2ea76eaf15045de" Oct 08 19:16:27 crc kubenswrapper[4859]: I1008 19:16:27.229860 4859 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/tempest-tests-tempest" Oct 08 19:16:35 crc kubenswrapper[4859]: I1008 19:16:35.809291 4859 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/test-operator-logs-pod-tempest-tempest-tests-tempest"] Oct 08 19:16:35 crc kubenswrapper[4859]: E1008 19:16:35.810251 4859 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c91f6897-a7a6-48ef-9dc3-5e3d57378b08" containerName="collect-profiles" Oct 08 19:16:35 crc kubenswrapper[4859]: I1008 19:16:35.810267 4859 state_mem.go:107] "Deleted CPUSet assignment" podUID="c91f6897-a7a6-48ef-9dc3-5e3d57378b08" containerName="collect-profiles" Oct 08 19:16:35 crc kubenswrapper[4859]: E1008 19:16:35.810304 4859 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="be909489-e6c5-4464-9292-aacd831885d7" containerName="tempest-tests-tempest-tests-runner" Oct 08 19:16:35 crc kubenswrapper[4859]: I1008 19:16:35.810313 4859 state_mem.go:107] "Deleted CPUSet assignment" podUID="be909489-e6c5-4464-9292-aacd831885d7" containerName="tempest-tests-tempest-tests-runner" Oct 08 19:16:35 crc kubenswrapper[4859]: I1008 19:16:35.810533 4859 memory_manager.go:354] "RemoveStaleState removing state" podUID="be909489-e6c5-4464-9292-aacd831885d7" containerName="tempest-tests-tempest-tests-runner" Oct 08 19:16:35 crc kubenswrapper[4859]: I1008 19:16:35.810567 4859 memory_manager.go:354] "RemoveStaleState removing state" podUID="c91f6897-a7a6-48ef-9dc3-5e3d57378b08" containerName="collect-profiles" Oct 08 19:16:35 crc kubenswrapper[4859]: I1008 19:16:35.811374 4859 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/test-operator-logs-pod-tempest-tempest-tests-tempest" Oct 08 19:16:35 crc kubenswrapper[4859]: I1008 19:16:35.813807 4859 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"default-dockercfg-mr7g8" Oct 08 19:16:35 crc kubenswrapper[4859]: I1008 19:16:35.832375 4859 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/test-operator-logs-pod-tempest-tempest-tests-tempest"] Oct 08 19:16:35 crc kubenswrapper[4859]: I1008 19:16:35.886598 4859 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage11-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage11-crc\") pod \"test-operator-logs-pod-tempest-tempest-tests-tempest\" (UID: \"f9597f73-13f2-4547-862b-fa23bba1c625\") " pod="openstack/test-operator-logs-pod-tempest-tempest-tests-tempest" Oct 08 19:16:35 crc kubenswrapper[4859]: I1008 19:16:35.886771 4859 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-89vc9\" (UniqueName: \"kubernetes.io/projected/f9597f73-13f2-4547-862b-fa23bba1c625-kube-api-access-89vc9\") pod \"test-operator-logs-pod-tempest-tempest-tests-tempest\" (UID: \"f9597f73-13f2-4547-862b-fa23bba1c625\") " pod="openstack/test-operator-logs-pod-tempest-tempest-tests-tempest" Oct 08 19:16:35 crc kubenswrapper[4859]: I1008 19:16:35.988312 4859 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage11-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage11-crc\") pod \"test-operator-logs-pod-tempest-tempest-tests-tempest\" (UID: \"f9597f73-13f2-4547-862b-fa23bba1c625\") " pod="openstack/test-operator-logs-pod-tempest-tempest-tests-tempest" Oct 08 19:16:35 crc kubenswrapper[4859]: I1008 19:16:35.988391 4859 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-89vc9\" (UniqueName: \"kubernetes.io/projected/f9597f73-13f2-4547-862b-fa23bba1c625-kube-api-access-89vc9\") pod \"test-operator-logs-pod-tempest-tempest-tests-tempest\" (UID: \"f9597f73-13f2-4547-862b-fa23bba1c625\") " pod="openstack/test-operator-logs-pod-tempest-tempest-tests-tempest" Oct 08 19:16:35 crc kubenswrapper[4859]: I1008 19:16:35.989470 4859 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage11-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage11-crc\") pod \"test-operator-logs-pod-tempest-tempest-tests-tempest\" (UID: \"f9597f73-13f2-4547-862b-fa23bba1c625\") device mount path \"/mnt/openstack/pv11\"" pod="openstack/test-operator-logs-pod-tempest-tempest-tests-tempest" Oct 08 19:16:36 crc kubenswrapper[4859]: I1008 19:16:36.017509 4859 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-89vc9\" (UniqueName: \"kubernetes.io/projected/f9597f73-13f2-4547-862b-fa23bba1c625-kube-api-access-89vc9\") pod \"test-operator-logs-pod-tempest-tempest-tests-tempest\" (UID: \"f9597f73-13f2-4547-862b-fa23bba1c625\") " pod="openstack/test-operator-logs-pod-tempest-tempest-tests-tempest" Oct 08 19:16:36 crc kubenswrapper[4859]: I1008 19:16:36.018040 4859 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage11-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage11-crc\") pod \"test-operator-logs-pod-tempest-tempest-tests-tempest\" (UID: \"f9597f73-13f2-4547-862b-fa23bba1c625\") " pod="openstack/test-operator-logs-pod-tempest-tempest-tests-tempest" Oct 08 19:16:36 crc kubenswrapper[4859]: I1008 19:16:36.142903 4859 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/test-operator-logs-pod-tempest-tempest-tests-tempest" Oct 08 19:16:36 crc kubenswrapper[4859]: I1008 19:16:36.618652 4859 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/test-operator-logs-pod-tempest-tempest-tests-tempest"] Oct 08 19:16:37 crc kubenswrapper[4859]: I1008 19:16:37.326149 4859 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/test-operator-logs-pod-tempest-tempest-tests-tempest" event={"ID":"f9597f73-13f2-4547-862b-fa23bba1c625","Type":"ContainerStarted","Data":"a99e629495cc882a066b02ce611cf16960b350c1cccbd8c1271d673c3443d7d2"} Oct 08 19:16:38 crc kubenswrapper[4859]: I1008 19:16:38.337851 4859 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/test-operator-logs-pod-tempest-tempest-tests-tempest" event={"ID":"f9597f73-13f2-4547-862b-fa23bba1c625","Type":"ContainerStarted","Data":"516e50250ce3d3c78bc50920bb7127675acebd5a7d3ea98c7714b988041dd959"} Oct 08 19:16:38 crc kubenswrapper[4859]: I1008 19:16:38.369313 4859 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/test-operator-logs-pod-tempest-tempest-tests-tempest" podStartSLOduration=2.319955297 podStartE2EDuration="3.369282916s" podCreationTimestamp="2025-10-08 19:16:35 +0000 UTC" firstStartedPulling="2025-10-08 19:16:36.626875448 +0000 UTC m=+3566.873714867" lastFinishedPulling="2025-10-08 19:16:37.676203107 +0000 UTC m=+3567.923042486" observedRunningTime="2025-10-08 19:16:38.351768235 +0000 UTC m=+3568.598607624" watchObservedRunningTime="2025-10-08 19:16:38.369282916 +0000 UTC m=+3568.616122335" Oct 08 19:16:47 crc kubenswrapper[4859]: I1008 19:16:47.925114 4859 patch_prober.go:28] interesting pod/machine-config-daemon-82s52 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 08 19:16:47 crc kubenswrapper[4859]: I1008 19:16:47.925627 4859 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-82s52" podUID="b23a6a6c-9d92-4e7b-840e-55cfda873a2d" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 08 19:16:54 crc kubenswrapper[4859]: I1008 19:16:54.763575 4859 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-must-gather-f7r6n/must-gather-9cn9k"] Oct 08 19:16:54 crc kubenswrapper[4859]: I1008 19:16:54.766184 4859 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-f7r6n/must-gather-9cn9k" Oct 08 19:16:54 crc kubenswrapper[4859]: I1008 19:16:54.767966 4859 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-must-gather-f7r6n"/"default-dockercfg-dklcr" Oct 08 19:16:54 crc kubenswrapper[4859]: I1008 19:16:54.770429 4859 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-must-gather-f7r6n"/"kube-root-ca.crt" Oct 08 19:16:54 crc kubenswrapper[4859]: I1008 19:16:54.771844 4859 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-must-gather-f7r6n"/"openshift-service-ca.crt" Oct 08 19:16:54 crc kubenswrapper[4859]: I1008 19:16:54.773516 4859 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-must-gather-f7r6n/must-gather-9cn9k"] Oct 08 19:16:54 crc kubenswrapper[4859]: I1008 19:16:54.882842 4859 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"must-gather-output\" (UniqueName: \"kubernetes.io/empty-dir/4dea6b3c-d8e9-4086-bfd0-46970f43f282-must-gather-output\") pod \"must-gather-9cn9k\" (UID: \"4dea6b3c-d8e9-4086-bfd0-46970f43f282\") " pod="openshift-must-gather-f7r6n/must-gather-9cn9k" Oct 08 19:16:54 crc kubenswrapper[4859]: I1008 19:16:54.882978 4859 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-zq5mf\" (UniqueName: \"kubernetes.io/projected/4dea6b3c-d8e9-4086-bfd0-46970f43f282-kube-api-access-zq5mf\") pod \"must-gather-9cn9k\" (UID: \"4dea6b3c-d8e9-4086-bfd0-46970f43f282\") " pod="openshift-must-gather-f7r6n/must-gather-9cn9k" Oct 08 19:16:54 crc kubenswrapper[4859]: I1008 19:16:54.985659 4859 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-zq5mf\" (UniqueName: \"kubernetes.io/projected/4dea6b3c-d8e9-4086-bfd0-46970f43f282-kube-api-access-zq5mf\") pod \"must-gather-9cn9k\" (UID: \"4dea6b3c-d8e9-4086-bfd0-46970f43f282\") " pod="openshift-must-gather-f7r6n/must-gather-9cn9k" Oct 08 19:16:54 crc kubenswrapper[4859]: I1008 19:16:54.985845 4859 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"must-gather-output\" (UniqueName: \"kubernetes.io/empty-dir/4dea6b3c-d8e9-4086-bfd0-46970f43f282-must-gather-output\") pod \"must-gather-9cn9k\" (UID: \"4dea6b3c-d8e9-4086-bfd0-46970f43f282\") " pod="openshift-must-gather-f7r6n/must-gather-9cn9k" Oct 08 19:16:54 crc kubenswrapper[4859]: I1008 19:16:54.986415 4859 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"must-gather-output\" (UniqueName: \"kubernetes.io/empty-dir/4dea6b3c-d8e9-4086-bfd0-46970f43f282-must-gather-output\") pod \"must-gather-9cn9k\" (UID: \"4dea6b3c-d8e9-4086-bfd0-46970f43f282\") " pod="openshift-must-gather-f7r6n/must-gather-9cn9k" Oct 08 19:16:55 crc kubenswrapper[4859]: I1008 19:16:55.013856 4859 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-zq5mf\" (UniqueName: \"kubernetes.io/projected/4dea6b3c-d8e9-4086-bfd0-46970f43f282-kube-api-access-zq5mf\") pod \"must-gather-9cn9k\" (UID: \"4dea6b3c-d8e9-4086-bfd0-46970f43f282\") " pod="openshift-must-gather-f7r6n/must-gather-9cn9k" Oct 08 19:16:55 crc kubenswrapper[4859]: I1008 19:16:55.084772 4859 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-f7r6n/must-gather-9cn9k" Oct 08 19:16:55 crc kubenswrapper[4859]: I1008 19:16:55.595139 4859 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-must-gather-f7r6n/must-gather-9cn9k"] Oct 08 19:16:56 crc kubenswrapper[4859]: I1008 19:16:56.531109 4859 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-f7r6n/must-gather-9cn9k" event={"ID":"4dea6b3c-d8e9-4086-bfd0-46970f43f282","Type":"ContainerStarted","Data":"95bde17856b28b9eb55b60fa2aed35175cce19931af1a5791966a99c01632964"} Oct 08 19:17:00 crc kubenswrapper[4859]: I1008 19:17:00.568614 4859 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-f7r6n/must-gather-9cn9k" event={"ID":"4dea6b3c-d8e9-4086-bfd0-46970f43f282","Type":"ContainerStarted","Data":"add7a3369622385591a2aacdd64391cb595efae300e79eadd5957a4f4eefd41a"} Oct 08 19:17:00 crc kubenswrapper[4859]: I1008 19:17:00.569029 4859 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-f7r6n/must-gather-9cn9k" event={"ID":"4dea6b3c-d8e9-4086-bfd0-46970f43f282","Type":"ContainerStarted","Data":"9caf66c4d88250d731a03fcb5e38db07b33c7bf6f56030210c181852d886f771"} Oct 08 19:17:00 crc kubenswrapper[4859]: I1008 19:17:00.592778 4859 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-must-gather-f7r6n/must-gather-9cn9k" podStartSLOduration=2.921029163 podStartE2EDuration="6.592758724s" podCreationTimestamp="2025-10-08 19:16:54 +0000 UTC" firstStartedPulling="2025-10-08 19:16:55.603992972 +0000 UTC m=+3585.850832361" lastFinishedPulling="2025-10-08 19:16:59.275722503 +0000 UTC m=+3589.522561922" observedRunningTime="2025-10-08 19:17:00.590029624 +0000 UTC m=+3590.836869003" watchObservedRunningTime="2025-10-08 19:17:00.592758724 +0000 UTC m=+3590.839598103" Oct 08 19:17:03 crc kubenswrapper[4859]: I1008 19:17:03.374827 4859 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-must-gather-f7r6n/crc-debug-829bd"] Oct 08 19:17:03 crc kubenswrapper[4859]: I1008 19:17:03.376661 4859 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-f7r6n/crc-debug-829bd" Oct 08 19:17:03 crc kubenswrapper[4859]: I1008 19:17:03.456637 4859 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/ca27de4c-c560-4441-99df-f9c49c3d97e0-host\") pod \"crc-debug-829bd\" (UID: \"ca27de4c-c560-4441-99df-f9c49c3d97e0\") " pod="openshift-must-gather-f7r6n/crc-debug-829bd" Oct 08 19:17:03 crc kubenswrapper[4859]: I1008 19:17:03.457055 4859 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-f2tpn\" (UniqueName: \"kubernetes.io/projected/ca27de4c-c560-4441-99df-f9c49c3d97e0-kube-api-access-f2tpn\") pod \"crc-debug-829bd\" (UID: \"ca27de4c-c560-4441-99df-f9c49c3d97e0\") " pod="openshift-must-gather-f7r6n/crc-debug-829bd" Oct 08 19:17:03 crc kubenswrapper[4859]: I1008 19:17:03.558665 4859 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-f2tpn\" (UniqueName: \"kubernetes.io/projected/ca27de4c-c560-4441-99df-f9c49c3d97e0-kube-api-access-f2tpn\") pod \"crc-debug-829bd\" (UID: \"ca27de4c-c560-4441-99df-f9c49c3d97e0\") " pod="openshift-must-gather-f7r6n/crc-debug-829bd" Oct 08 19:17:03 crc kubenswrapper[4859]: I1008 19:17:03.558797 4859 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/ca27de4c-c560-4441-99df-f9c49c3d97e0-host\") pod \"crc-debug-829bd\" (UID: \"ca27de4c-c560-4441-99df-f9c49c3d97e0\") " pod="openshift-must-gather-f7r6n/crc-debug-829bd" Oct 08 19:17:03 crc kubenswrapper[4859]: I1008 19:17:03.558911 4859 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host\" (UniqueName: \"kubernetes.io/host-path/ca27de4c-c560-4441-99df-f9c49c3d97e0-host\") pod \"crc-debug-829bd\" (UID: \"ca27de4c-c560-4441-99df-f9c49c3d97e0\") " pod="openshift-must-gather-f7r6n/crc-debug-829bd" Oct 08 19:17:03 crc kubenswrapper[4859]: I1008 19:17:03.585487 4859 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-f2tpn\" (UniqueName: \"kubernetes.io/projected/ca27de4c-c560-4441-99df-f9c49c3d97e0-kube-api-access-f2tpn\") pod \"crc-debug-829bd\" (UID: \"ca27de4c-c560-4441-99df-f9c49c3d97e0\") " pod="openshift-must-gather-f7r6n/crc-debug-829bd" Oct 08 19:17:03 crc kubenswrapper[4859]: I1008 19:17:03.695619 4859 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-f7r6n/crc-debug-829bd" Oct 08 19:17:03 crc kubenswrapper[4859]: W1008 19:17:03.741863 4859 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podca27de4c_c560_4441_99df_f9c49c3d97e0.slice/crio-8c9ea630b2c2c397af7f079555b8799c93a69ae3cca6820976e55c77418db6ea WatchSource:0}: Error finding container 8c9ea630b2c2c397af7f079555b8799c93a69ae3cca6820976e55c77418db6ea: Status 404 returned error can't find the container with id 8c9ea630b2c2c397af7f079555b8799c93a69ae3cca6820976e55c77418db6ea Oct 08 19:17:04 crc kubenswrapper[4859]: I1008 19:17:04.612552 4859 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-f7r6n/crc-debug-829bd" event={"ID":"ca27de4c-c560-4441-99df-f9c49c3d97e0","Type":"ContainerStarted","Data":"8c9ea630b2c2c397af7f079555b8799c93a69ae3cca6820976e55c77418db6ea"} Oct 08 19:17:11 crc kubenswrapper[4859]: I1008 19:17:11.163392 4859 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-wlkbs"] Oct 08 19:17:11 crc kubenswrapper[4859]: I1008 19:17:11.166091 4859 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-wlkbs" Oct 08 19:17:11 crc kubenswrapper[4859]: I1008 19:17:11.183830 4859 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-wlkbs"] Oct 08 19:17:11 crc kubenswrapper[4859]: I1008 19:17:11.324980 4859 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/640efdd5-e75b-4957-8aaa-e7a9a5828962-catalog-content\") pod \"redhat-marketplace-wlkbs\" (UID: \"640efdd5-e75b-4957-8aaa-e7a9a5828962\") " pod="openshift-marketplace/redhat-marketplace-wlkbs" Oct 08 19:17:11 crc kubenswrapper[4859]: I1008 19:17:11.325344 4859 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/640efdd5-e75b-4957-8aaa-e7a9a5828962-utilities\") pod \"redhat-marketplace-wlkbs\" (UID: \"640efdd5-e75b-4957-8aaa-e7a9a5828962\") " pod="openshift-marketplace/redhat-marketplace-wlkbs" Oct 08 19:17:11 crc kubenswrapper[4859]: I1008 19:17:11.325446 4859 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-8w2wg\" (UniqueName: \"kubernetes.io/projected/640efdd5-e75b-4957-8aaa-e7a9a5828962-kube-api-access-8w2wg\") pod \"redhat-marketplace-wlkbs\" (UID: \"640efdd5-e75b-4957-8aaa-e7a9a5828962\") " pod="openshift-marketplace/redhat-marketplace-wlkbs" Oct 08 19:17:11 crc kubenswrapper[4859]: I1008 19:17:11.428142 4859 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/640efdd5-e75b-4957-8aaa-e7a9a5828962-catalog-content\") pod \"redhat-marketplace-wlkbs\" (UID: \"640efdd5-e75b-4957-8aaa-e7a9a5828962\") " pod="openshift-marketplace/redhat-marketplace-wlkbs" Oct 08 19:17:11 crc kubenswrapper[4859]: I1008 19:17:11.428368 4859 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/640efdd5-e75b-4957-8aaa-e7a9a5828962-utilities\") pod \"redhat-marketplace-wlkbs\" (UID: \"640efdd5-e75b-4957-8aaa-e7a9a5828962\") " pod="openshift-marketplace/redhat-marketplace-wlkbs" Oct 08 19:17:11 crc kubenswrapper[4859]: I1008 19:17:11.428430 4859 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-8w2wg\" (UniqueName: \"kubernetes.io/projected/640efdd5-e75b-4957-8aaa-e7a9a5828962-kube-api-access-8w2wg\") pod \"redhat-marketplace-wlkbs\" (UID: \"640efdd5-e75b-4957-8aaa-e7a9a5828962\") " pod="openshift-marketplace/redhat-marketplace-wlkbs" Oct 08 19:17:11 crc kubenswrapper[4859]: I1008 19:17:11.428948 4859 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/640efdd5-e75b-4957-8aaa-e7a9a5828962-catalog-content\") pod \"redhat-marketplace-wlkbs\" (UID: \"640efdd5-e75b-4957-8aaa-e7a9a5828962\") " pod="openshift-marketplace/redhat-marketplace-wlkbs" Oct 08 19:17:11 crc kubenswrapper[4859]: I1008 19:17:11.428974 4859 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/640efdd5-e75b-4957-8aaa-e7a9a5828962-utilities\") pod \"redhat-marketplace-wlkbs\" (UID: \"640efdd5-e75b-4957-8aaa-e7a9a5828962\") " pod="openshift-marketplace/redhat-marketplace-wlkbs" Oct 08 19:17:11 crc kubenswrapper[4859]: I1008 19:17:11.479103 4859 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-8w2wg\" (UniqueName: \"kubernetes.io/projected/640efdd5-e75b-4957-8aaa-e7a9a5828962-kube-api-access-8w2wg\") pod \"redhat-marketplace-wlkbs\" (UID: \"640efdd5-e75b-4957-8aaa-e7a9a5828962\") " pod="openshift-marketplace/redhat-marketplace-wlkbs" Oct 08 19:17:11 crc kubenswrapper[4859]: I1008 19:17:11.490489 4859 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-wlkbs" Oct 08 19:17:15 crc kubenswrapper[4859]: I1008 19:17:15.077423 4859 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-wlkbs"] Oct 08 19:17:15 crc kubenswrapper[4859]: W1008 19:17:15.096856 4859 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod640efdd5_e75b_4957_8aaa_e7a9a5828962.slice/crio-d2e8513b6a2809a0dd89b68b62eb44aa392931cb5c5c8f2b3c8c1546efc2a309 WatchSource:0}: Error finding container d2e8513b6a2809a0dd89b68b62eb44aa392931cb5c5c8f2b3c8c1546efc2a309: Status 404 returned error can't find the container with id d2e8513b6a2809a0dd89b68b62eb44aa392931cb5c5c8f2b3c8c1546efc2a309 Oct 08 19:17:15 crc kubenswrapper[4859]: I1008 19:17:15.706128 4859 generic.go:334] "Generic (PLEG): container finished" podID="640efdd5-e75b-4957-8aaa-e7a9a5828962" containerID="86d88bb3225d5f86cc0da842cc8ebd679177ac82f9680687685c58fec781bcb7" exitCode=0 Oct 08 19:17:15 crc kubenswrapper[4859]: I1008 19:17:15.706220 4859 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-wlkbs" event={"ID":"640efdd5-e75b-4957-8aaa-e7a9a5828962","Type":"ContainerDied","Data":"86d88bb3225d5f86cc0da842cc8ebd679177ac82f9680687685c58fec781bcb7"} Oct 08 19:17:15 crc kubenswrapper[4859]: I1008 19:17:15.706433 4859 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-wlkbs" event={"ID":"640efdd5-e75b-4957-8aaa-e7a9a5828962","Type":"ContainerStarted","Data":"d2e8513b6a2809a0dd89b68b62eb44aa392931cb5c5c8f2b3c8c1546efc2a309"} Oct 08 19:17:15 crc kubenswrapper[4859]: I1008 19:17:15.711172 4859 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-f7r6n/crc-debug-829bd" event={"ID":"ca27de4c-c560-4441-99df-f9c49c3d97e0","Type":"ContainerStarted","Data":"335899cba18961c57684d6717a63018db36b992f2dec8df2b77f0e7340d13582"} Oct 08 19:17:15 crc kubenswrapper[4859]: I1008 19:17:15.744592 4859 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-must-gather-f7r6n/crc-debug-829bd" podStartSLOduration=1.835476382 podStartE2EDuration="12.744572879s" podCreationTimestamp="2025-10-08 19:17:03 +0000 UTC" firstStartedPulling="2025-10-08 19:17:03.74497705 +0000 UTC m=+3593.991816429" lastFinishedPulling="2025-10-08 19:17:14.654073547 +0000 UTC m=+3604.900912926" observedRunningTime="2025-10-08 19:17:15.740464269 +0000 UTC m=+3605.987303668" watchObservedRunningTime="2025-10-08 19:17:15.744572879 +0000 UTC m=+3605.991412258" Oct 08 19:17:16 crc kubenswrapper[4859]: I1008 19:17:16.721598 4859 generic.go:334] "Generic (PLEG): container finished" podID="640efdd5-e75b-4957-8aaa-e7a9a5828962" containerID="9d5d53ae3a9fdad175a075d874f4e7d1d6cbd1fa73b99e1a52eead1d2f5ed97c" exitCode=0 Oct 08 19:17:16 crc kubenswrapper[4859]: I1008 19:17:16.721665 4859 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-wlkbs" event={"ID":"640efdd5-e75b-4957-8aaa-e7a9a5828962","Type":"ContainerDied","Data":"9d5d53ae3a9fdad175a075d874f4e7d1d6cbd1fa73b99e1a52eead1d2f5ed97c"} Oct 08 19:17:17 crc kubenswrapper[4859]: I1008 19:17:17.732054 4859 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-wlkbs" event={"ID":"640efdd5-e75b-4957-8aaa-e7a9a5828962","Type":"ContainerStarted","Data":"e41df4a54deb68f74ab738560f4bd5afa0a6227d2c0af6111748597918d6859c"} Oct 08 19:17:17 crc kubenswrapper[4859]: I1008 19:17:17.757150 4859 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-wlkbs" podStartSLOduration=5.362965359 podStartE2EDuration="6.757128619s" podCreationTimestamp="2025-10-08 19:17:11 +0000 UTC" firstStartedPulling="2025-10-08 19:17:15.709576568 +0000 UTC m=+3605.956415947" lastFinishedPulling="2025-10-08 19:17:17.103739818 +0000 UTC m=+3607.350579207" observedRunningTime="2025-10-08 19:17:17.75133971 +0000 UTC m=+3607.998179099" watchObservedRunningTime="2025-10-08 19:17:17.757128619 +0000 UTC m=+3608.003967998" Oct 08 19:17:17 crc kubenswrapper[4859]: I1008 19:17:17.924917 4859 patch_prober.go:28] interesting pod/machine-config-daemon-82s52 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 08 19:17:17 crc kubenswrapper[4859]: I1008 19:17:17.924987 4859 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-82s52" podUID="b23a6a6c-9d92-4e7b-840e-55cfda873a2d" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 08 19:17:21 crc kubenswrapper[4859]: I1008 19:17:21.491392 4859 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-wlkbs" Oct 08 19:17:21 crc kubenswrapper[4859]: I1008 19:17:21.491934 4859 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-wlkbs" Oct 08 19:17:21 crc kubenswrapper[4859]: I1008 19:17:21.546084 4859 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-wlkbs" Oct 08 19:17:31 crc kubenswrapper[4859]: I1008 19:17:31.558080 4859 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-wlkbs" Oct 08 19:17:31 crc kubenswrapper[4859]: I1008 19:17:31.603516 4859 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-wlkbs"] Oct 08 19:17:31 crc kubenswrapper[4859]: I1008 19:17:31.847196 4859 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-marketplace-wlkbs" podUID="640efdd5-e75b-4957-8aaa-e7a9a5828962" containerName="registry-server" containerID="cri-o://e41df4a54deb68f74ab738560f4bd5afa0a6227d2c0af6111748597918d6859c" gracePeriod=2 Oct 08 19:17:32 crc kubenswrapper[4859]: I1008 19:17:32.363263 4859 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-wlkbs" Oct 08 19:17:32 crc kubenswrapper[4859]: I1008 19:17:32.548845 4859 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/640efdd5-e75b-4957-8aaa-e7a9a5828962-utilities\") pod \"640efdd5-e75b-4957-8aaa-e7a9a5828962\" (UID: \"640efdd5-e75b-4957-8aaa-e7a9a5828962\") " Oct 08 19:17:32 crc kubenswrapper[4859]: I1008 19:17:32.549087 4859 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/640efdd5-e75b-4957-8aaa-e7a9a5828962-catalog-content\") pod \"640efdd5-e75b-4957-8aaa-e7a9a5828962\" (UID: \"640efdd5-e75b-4957-8aaa-e7a9a5828962\") " Oct 08 19:17:32 crc kubenswrapper[4859]: I1008 19:17:32.549189 4859 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-8w2wg\" (UniqueName: \"kubernetes.io/projected/640efdd5-e75b-4957-8aaa-e7a9a5828962-kube-api-access-8w2wg\") pod \"640efdd5-e75b-4957-8aaa-e7a9a5828962\" (UID: \"640efdd5-e75b-4957-8aaa-e7a9a5828962\") " Oct 08 19:17:32 crc kubenswrapper[4859]: I1008 19:17:32.550517 4859 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/640efdd5-e75b-4957-8aaa-e7a9a5828962-utilities" (OuterVolumeSpecName: "utilities") pod "640efdd5-e75b-4957-8aaa-e7a9a5828962" (UID: "640efdd5-e75b-4957-8aaa-e7a9a5828962"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 08 19:17:32 crc kubenswrapper[4859]: I1008 19:17:32.572252 4859 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/640efdd5-e75b-4957-8aaa-e7a9a5828962-kube-api-access-8w2wg" (OuterVolumeSpecName: "kube-api-access-8w2wg") pod "640efdd5-e75b-4957-8aaa-e7a9a5828962" (UID: "640efdd5-e75b-4957-8aaa-e7a9a5828962"). InnerVolumeSpecName "kube-api-access-8w2wg". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 08 19:17:32 crc kubenswrapper[4859]: I1008 19:17:32.619995 4859 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/640efdd5-e75b-4957-8aaa-e7a9a5828962-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "640efdd5-e75b-4957-8aaa-e7a9a5828962" (UID: "640efdd5-e75b-4957-8aaa-e7a9a5828962"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 08 19:17:32 crc kubenswrapper[4859]: I1008 19:17:32.652107 4859 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/640efdd5-e75b-4957-8aaa-e7a9a5828962-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 08 19:17:32 crc kubenswrapper[4859]: I1008 19:17:32.652143 4859 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-8w2wg\" (UniqueName: \"kubernetes.io/projected/640efdd5-e75b-4957-8aaa-e7a9a5828962-kube-api-access-8w2wg\") on node \"crc\" DevicePath \"\"" Oct 08 19:17:32 crc kubenswrapper[4859]: I1008 19:17:32.652154 4859 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/640efdd5-e75b-4957-8aaa-e7a9a5828962-utilities\") on node \"crc\" DevicePath \"\"" Oct 08 19:17:32 crc kubenswrapper[4859]: I1008 19:17:32.860861 4859 generic.go:334] "Generic (PLEG): container finished" podID="640efdd5-e75b-4957-8aaa-e7a9a5828962" containerID="e41df4a54deb68f74ab738560f4bd5afa0a6227d2c0af6111748597918d6859c" exitCode=0 Oct 08 19:17:32 crc kubenswrapper[4859]: I1008 19:17:32.860933 4859 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-wlkbs" Oct 08 19:17:32 crc kubenswrapper[4859]: I1008 19:17:32.860946 4859 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-wlkbs" event={"ID":"640efdd5-e75b-4957-8aaa-e7a9a5828962","Type":"ContainerDied","Data":"e41df4a54deb68f74ab738560f4bd5afa0a6227d2c0af6111748597918d6859c"} Oct 08 19:17:32 crc kubenswrapper[4859]: I1008 19:17:32.861382 4859 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-wlkbs" event={"ID":"640efdd5-e75b-4957-8aaa-e7a9a5828962","Type":"ContainerDied","Data":"d2e8513b6a2809a0dd89b68b62eb44aa392931cb5c5c8f2b3c8c1546efc2a309"} Oct 08 19:17:32 crc kubenswrapper[4859]: I1008 19:17:32.861409 4859 scope.go:117] "RemoveContainer" containerID="e41df4a54deb68f74ab738560f4bd5afa0a6227d2c0af6111748597918d6859c" Oct 08 19:17:32 crc kubenswrapper[4859]: I1008 19:17:32.882267 4859 scope.go:117] "RemoveContainer" containerID="9d5d53ae3a9fdad175a075d874f4e7d1d6cbd1fa73b99e1a52eead1d2f5ed97c" Oct 08 19:17:32 crc kubenswrapper[4859]: I1008 19:17:32.906339 4859 scope.go:117] "RemoveContainer" containerID="86d88bb3225d5f86cc0da842cc8ebd679177ac82f9680687685c58fec781bcb7" Oct 08 19:17:32 crc kubenswrapper[4859]: I1008 19:17:32.908134 4859 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-wlkbs"] Oct 08 19:17:32 crc kubenswrapper[4859]: I1008 19:17:32.919001 4859 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-marketplace-wlkbs"] Oct 08 19:17:32 crc kubenswrapper[4859]: I1008 19:17:32.965506 4859 scope.go:117] "RemoveContainer" containerID="e41df4a54deb68f74ab738560f4bd5afa0a6227d2c0af6111748597918d6859c" Oct 08 19:17:32 crc kubenswrapper[4859]: E1008 19:17:32.986990 4859 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"e41df4a54deb68f74ab738560f4bd5afa0a6227d2c0af6111748597918d6859c\": container with ID starting with e41df4a54deb68f74ab738560f4bd5afa0a6227d2c0af6111748597918d6859c not found: ID does not exist" containerID="e41df4a54deb68f74ab738560f4bd5afa0a6227d2c0af6111748597918d6859c" Oct 08 19:17:32 crc kubenswrapper[4859]: I1008 19:17:32.987040 4859 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"e41df4a54deb68f74ab738560f4bd5afa0a6227d2c0af6111748597918d6859c"} err="failed to get container status \"e41df4a54deb68f74ab738560f4bd5afa0a6227d2c0af6111748597918d6859c\": rpc error: code = NotFound desc = could not find container \"e41df4a54deb68f74ab738560f4bd5afa0a6227d2c0af6111748597918d6859c\": container with ID starting with e41df4a54deb68f74ab738560f4bd5afa0a6227d2c0af6111748597918d6859c not found: ID does not exist" Oct 08 19:17:32 crc kubenswrapper[4859]: I1008 19:17:32.987071 4859 scope.go:117] "RemoveContainer" containerID="9d5d53ae3a9fdad175a075d874f4e7d1d6cbd1fa73b99e1a52eead1d2f5ed97c" Oct 08 19:17:32 crc kubenswrapper[4859]: E1008 19:17:32.988655 4859 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"9d5d53ae3a9fdad175a075d874f4e7d1d6cbd1fa73b99e1a52eead1d2f5ed97c\": container with ID starting with 9d5d53ae3a9fdad175a075d874f4e7d1d6cbd1fa73b99e1a52eead1d2f5ed97c not found: ID does not exist" containerID="9d5d53ae3a9fdad175a075d874f4e7d1d6cbd1fa73b99e1a52eead1d2f5ed97c" Oct 08 19:17:32 crc kubenswrapper[4859]: I1008 19:17:32.988698 4859 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"9d5d53ae3a9fdad175a075d874f4e7d1d6cbd1fa73b99e1a52eead1d2f5ed97c"} err="failed to get container status \"9d5d53ae3a9fdad175a075d874f4e7d1d6cbd1fa73b99e1a52eead1d2f5ed97c\": rpc error: code = NotFound desc = could not find container \"9d5d53ae3a9fdad175a075d874f4e7d1d6cbd1fa73b99e1a52eead1d2f5ed97c\": container with ID starting with 9d5d53ae3a9fdad175a075d874f4e7d1d6cbd1fa73b99e1a52eead1d2f5ed97c not found: ID does not exist" Oct 08 19:17:32 crc kubenswrapper[4859]: I1008 19:17:32.988727 4859 scope.go:117] "RemoveContainer" containerID="86d88bb3225d5f86cc0da842cc8ebd679177ac82f9680687685c58fec781bcb7" Oct 08 19:17:32 crc kubenswrapper[4859]: E1008 19:17:32.989157 4859 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"86d88bb3225d5f86cc0da842cc8ebd679177ac82f9680687685c58fec781bcb7\": container with ID starting with 86d88bb3225d5f86cc0da842cc8ebd679177ac82f9680687685c58fec781bcb7 not found: ID does not exist" containerID="86d88bb3225d5f86cc0da842cc8ebd679177ac82f9680687685c58fec781bcb7" Oct 08 19:17:32 crc kubenswrapper[4859]: I1008 19:17:32.989184 4859 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"86d88bb3225d5f86cc0da842cc8ebd679177ac82f9680687685c58fec781bcb7"} err="failed to get container status \"86d88bb3225d5f86cc0da842cc8ebd679177ac82f9680687685c58fec781bcb7\": rpc error: code = NotFound desc = could not find container \"86d88bb3225d5f86cc0da842cc8ebd679177ac82f9680687685c58fec781bcb7\": container with ID starting with 86d88bb3225d5f86cc0da842cc8ebd679177ac82f9680687685c58fec781bcb7 not found: ID does not exist" Oct 08 19:17:34 crc kubenswrapper[4859]: I1008 19:17:34.479828 4859 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="640efdd5-e75b-4957-8aaa-e7a9a5828962" path="/var/lib/kubelet/pods/640efdd5-e75b-4957-8aaa-e7a9a5828962/volumes" Oct 08 19:17:47 crc kubenswrapper[4859]: I1008 19:17:47.924327 4859 patch_prober.go:28] interesting pod/machine-config-daemon-82s52 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 08 19:17:47 crc kubenswrapper[4859]: I1008 19:17:47.924951 4859 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-82s52" podUID="b23a6a6c-9d92-4e7b-840e-55cfda873a2d" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 08 19:17:47 crc kubenswrapper[4859]: I1008 19:17:47.925006 4859 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-82s52" Oct 08 19:17:47 crc kubenswrapper[4859]: I1008 19:17:47.925964 4859 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"09fd446535f97e643340dafbb38e492a2f1d2902d0d749edf84136fd6fabb2ac"} pod="openshift-machine-config-operator/machine-config-daemon-82s52" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Oct 08 19:17:47 crc kubenswrapper[4859]: I1008 19:17:47.926044 4859 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-82s52" podUID="b23a6a6c-9d92-4e7b-840e-55cfda873a2d" containerName="machine-config-daemon" containerID="cri-o://09fd446535f97e643340dafbb38e492a2f1d2902d0d749edf84136fd6fabb2ac" gracePeriod=600 Oct 08 19:17:49 crc kubenswrapper[4859]: I1008 19:17:49.010108 4859 generic.go:334] "Generic (PLEG): container finished" podID="b23a6a6c-9d92-4e7b-840e-55cfda873a2d" containerID="09fd446535f97e643340dafbb38e492a2f1d2902d0d749edf84136fd6fabb2ac" exitCode=0 Oct 08 19:17:49 crc kubenswrapper[4859]: I1008 19:17:49.010196 4859 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-82s52" event={"ID":"b23a6a6c-9d92-4e7b-840e-55cfda873a2d","Type":"ContainerDied","Data":"09fd446535f97e643340dafbb38e492a2f1d2902d0d749edf84136fd6fabb2ac"} Oct 08 19:17:49 crc kubenswrapper[4859]: I1008 19:17:49.010513 4859 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-82s52" event={"ID":"b23a6a6c-9d92-4e7b-840e-55cfda873a2d","Type":"ContainerStarted","Data":"de71f0b717129dba48f3f0bdd90240a8e45404a17f0df79d6812d58c9209feff"} Oct 08 19:17:49 crc kubenswrapper[4859]: I1008 19:17:49.010533 4859 scope.go:117] "RemoveContainer" containerID="cc836fd09088ba4df38f4588526d6fe50b7b131a7fb454b4fd3751d8b1b22560" Oct 08 19:18:04 crc kubenswrapper[4859]: I1008 19:18:04.665939 4859 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_barbican-api-588f98c7dd-c7l8h_3ecae548-eaa6-4c24-a538-b2a34fa0308f/barbican-api-log/0.log" Oct 08 19:18:04 crc kubenswrapper[4859]: I1008 19:18:04.683027 4859 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_barbican-api-588f98c7dd-c7l8h_3ecae548-eaa6-4c24-a538-b2a34fa0308f/barbican-api/0.log" Oct 08 19:18:04 crc kubenswrapper[4859]: I1008 19:18:04.872602 4859 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_barbican-keystone-listener-d798f6bcd-5zw8p_80660054-bd62-4742-a13d-5193b317b036/barbican-keystone-listener/0.log" Oct 08 19:18:04 crc kubenswrapper[4859]: I1008 19:18:04.972060 4859 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_barbican-keystone-listener-d798f6bcd-5zw8p_80660054-bd62-4742-a13d-5193b317b036/barbican-keystone-listener-log/0.log" Oct 08 19:18:05 crc kubenswrapper[4859]: I1008 19:18:05.165575 4859 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_barbican-worker-7849cf5875-6srbt_578aad66-818f-412f-9023-97ed6ae95de3/barbican-worker/0.log" Oct 08 19:18:05 crc kubenswrapper[4859]: I1008 19:18:05.175352 4859 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_barbican-worker-7849cf5875-6srbt_578aad66-818f-412f-9023-97ed6ae95de3/barbican-worker-log/0.log" Oct 08 19:18:05 crc kubenswrapper[4859]: I1008 19:18:05.344722 4859 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_bootstrap-edpm-deployment-openstack-edpm-ipam-nl8fp_e0bf2482-ef25-48bb-9491-282a3eabfec0/bootstrap-edpm-deployment-openstack-edpm-ipam/0.log" Oct 08 19:18:05 crc kubenswrapper[4859]: I1008 19:18:05.573393 4859 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ceilometer-0_53819555-42d9-40cb-9897-94c5496c8da9/proxy-httpd/0.log" Oct 08 19:18:05 crc kubenswrapper[4859]: I1008 19:18:05.577537 4859 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ceilometer-0_53819555-42d9-40cb-9897-94c5496c8da9/ceilometer-notification-agent/0.log" Oct 08 19:18:05 crc kubenswrapper[4859]: I1008 19:18:05.601897 4859 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ceilometer-0_53819555-42d9-40cb-9897-94c5496c8da9/ceilometer-central-agent/0.log" Oct 08 19:18:05 crc kubenswrapper[4859]: I1008 19:18:05.794820 4859 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ceilometer-0_53819555-42d9-40cb-9897-94c5496c8da9/sg-core/0.log" Oct 08 19:18:05 crc kubenswrapper[4859]: I1008 19:18:05.868879 4859 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_cinder-api-0_f8eeb421-2f6d-428f-93d5-5aca080785cc/cinder-api/0.log" Oct 08 19:18:05 crc kubenswrapper[4859]: I1008 19:18:05.976476 4859 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_cinder-api-0_f8eeb421-2f6d-428f-93d5-5aca080785cc/cinder-api-log/0.log" Oct 08 19:18:06 crc kubenswrapper[4859]: I1008 19:18:06.163789 4859 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_cinder-scheduler-0_f20f7045-6768-4824-9295-6e5ba74da387/cinder-scheduler/0.log" Oct 08 19:18:06 crc kubenswrapper[4859]: I1008 19:18:06.284332 4859 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_cinder-scheduler-0_f20f7045-6768-4824-9295-6e5ba74da387/probe/0.log" Oct 08 19:18:06 crc kubenswrapper[4859]: I1008 19:18:06.347036 4859 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_configure-network-edpm-deployment-openstack-edpm-ipam-6n7d4_70979534-d410-4768-95a9-5c9ff4ae1583/configure-network-edpm-deployment-openstack-edpm-ipam/0.log" Oct 08 19:18:06 crc kubenswrapper[4859]: I1008 19:18:06.581912 4859 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_configure-os-edpm-deployment-openstack-edpm-ipam-5dz4q_ff5a590b-fac3-4cf3-b472-db2f43030032/configure-os-edpm-deployment-openstack-edpm-ipam/0.log" Oct 08 19:18:06 crc kubenswrapper[4859]: I1008 19:18:06.726016 4859 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_configure-os-edpm-deployment-openstack-edpm-ipam-dnz8t_28feb067-adfe-4b48-aef3-321e18ac0e37/configure-os-edpm-deployment-openstack-edpm-ipam/0.log" Oct 08 19:18:06 crc kubenswrapper[4859]: I1008 19:18:06.824239 4859 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_dnsmasq-dns-6f586dc697-ncnrs_7f4d68a5-ccfe-48f2-a84f-e800d6718498/init/0.log" Oct 08 19:18:07 crc kubenswrapper[4859]: I1008 19:18:07.016377 4859 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_dnsmasq-dns-6f586dc697-ncnrs_7f4d68a5-ccfe-48f2-a84f-e800d6718498/init/0.log" Oct 08 19:18:07 crc kubenswrapper[4859]: I1008 19:18:07.074546 4859 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_dnsmasq-dns-6f586dc697-ncnrs_7f4d68a5-ccfe-48f2-a84f-e800d6718498/dnsmasq-dns/0.log" Oct 08 19:18:07 crc kubenswrapper[4859]: I1008 19:18:07.231803 4859 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_download-cache-edpm-deployment-openstack-edpm-ipam-hb4tw_2db15fd4-e39e-474c-b274-53262e92bccb/download-cache-edpm-deployment-openstack-edpm-ipam/0.log" Oct 08 19:18:07 crc kubenswrapper[4859]: I1008 19:18:07.396844 4859 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_glance-default-external-api-0_21a00512-7854-48c3-a040-795eb7d79744/glance-httpd/0.log" Oct 08 19:18:07 crc kubenswrapper[4859]: I1008 19:18:07.451887 4859 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_glance-default-external-api-0_21a00512-7854-48c3-a040-795eb7d79744/glance-log/0.log" Oct 08 19:18:07 crc kubenswrapper[4859]: I1008 19:18:07.629155 4859 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_glance-default-internal-api-0_d14936ce-b30c-44e7-95bc-9a4d516e71b1/glance-httpd/0.log" Oct 08 19:18:07 crc kubenswrapper[4859]: I1008 19:18:07.672446 4859 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_glance-default-internal-api-0_d14936ce-b30c-44e7-95bc-9a4d516e71b1/glance-log/0.log" Oct 08 19:18:07 crc kubenswrapper[4859]: I1008 19:18:07.907302 4859 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_horizon-6f569d5894-mx8v5_50b5de31-90f6-43cf-9e16-7b89b09f1e57/horizon/0.log" Oct 08 19:18:08 crc kubenswrapper[4859]: I1008 19:18:08.189418 4859 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_install-certs-edpm-deployment-openstack-edpm-ipam-w95ff_3b4b0004-0407-4671-a67e-0c14ef73e6ff/install-certs-edpm-deployment-openstack-edpm-ipam/0.log" Oct 08 19:18:08 crc kubenswrapper[4859]: I1008 19:18:08.231805 4859 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_horizon-6f569d5894-mx8v5_50b5de31-90f6-43cf-9e16-7b89b09f1e57/horizon-log/0.log" Oct 08 19:18:08 crc kubenswrapper[4859]: I1008 19:18:08.305732 4859 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_install-os-edpm-deployment-openstack-edpm-ipam-twjcj_5d2f49c5-b079-4409-9ad7-e7c6b419e843/install-os-edpm-deployment-openstack-edpm-ipam/0.log" Oct 08 19:18:08 crc kubenswrapper[4859]: I1008 19:18:08.555796 4859 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_keystone-cron-29332501-rxhgw_555ffac5-9b03-43eb-8610-fad69a164a42/keystone-cron/0.log" Oct 08 19:18:08 crc kubenswrapper[4859]: I1008 19:18:08.613227 4859 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_keystone-758ff8b98b-p6wm9_285bdd26-6109-4191-870a-2fde949a6a28/keystone-api/0.log" Oct 08 19:18:08 crc kubenswrapper[4859]: I1008 19:18:08.734360 4859 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_kube-state-metrics-0_3652de4f-2bd2-4fe2-b138-63dd20b42a61/kube-state-metrics/0.log" Oct 08 19:18:08 crc kubenswrapper[4859]: I1008 19:18:08.961305 4859 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_libvirt-edpm-deployment-openstack-edpm-ipam-rrrnw_a7cd7054-3a66-49ef-bcd5-405c815f8bfb/libvirt-edpm-deployment-openstack-edpm-ipam/0.log" Oct 08 19:18:09 crc kubenswrapper[4859]: I1008 19:18:09.378443 4859 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_neutron-5f9b9b6f45-wr5fx_f64840cf-331e-4d8d-b102-49225713df66/neutron-api/0.log" Oct 08 19:18:09 crc kubenswrapper[4859]: I1008 19:18:09.481676 4859 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_neutron-5f9b9b6f45-wr5fx_f64840cf-331e-4d8d-b102-49225713df66/neutron-httpd/0.log" Oct 08 19:18:09 crc kubenswrapper[4859]: I1008 19:18:09.624953 4859 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_neutron-metadata-edpm-deployment-openstack-edpm-ipam-fnxfs_3f1e49a1-2a8d-4d9e-9fde-529f8832d0dc/neutron-metadata-edpm-deployment-openstack-edpm-ipam/0.log" Oct 08 19:18:10 crc kubenswrapper[4859]: I1008 19:18:10.224443 4859 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-api-0_e56f2ed9-f7a3-4157-bb50-88fbd86d9c60/nova-api-log/0.log" Oct 08 19:18:10 crc kubenswrapper[4859]: I1008 19:18:10.332368 4859 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-cell0-conductor-0_8c9e1318-b507-4caa-832f-ea7c02f7c199/nova-cell0-conductor-conductor/0.log" Oct 08 19:18:10 crc kubenswrapper[4859]: I1008 19:18:10.474731 4859 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-api-0_e56f2ed9-f7a3-4157-bb50-88fbd86d9c60/nova-api-api/0.log" Oct 08 19:18:10 crc kubenswrapper[4859]: I1008 19:18:10.658552 4859 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-cell1-conductor-0_994d31f2-ad55-4817-a2e5-4e3743462cae/nova-cell1-conductor-conductor/0.log" Oct 08 19:18:10 crc kubenswrapper[4859]: I1008 19:18:10.925229 4859 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-cell1-novncproxy-0_f2fffadc-cd8a-4481-aafe-8f7e2dd338cf/nova-cell1-novncproxy-novncproxy/0.log" Oct 08 19:18:10 crc kubenswrapper[4859]: I1008 19:18:10.991283 4859 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-edpm-deployment-openstack-edpm-ipam-sg6sh_4035f4f1-5231-4112-89f6-d9d1c7c7eeb1/nova-edpm-deployment-openstack-edpm-ipam/0.log" Oct 08 19:18:11 crc kubenswrapper[4859]: I1008 19:18:11.225941 4859 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-metadata-0_6188ecfc-2b44-443d-a966-8b91ae535533/nova-metadata-log/0.log" Oct 08 19:18:11 crc kubenswrapper[4859]: I1008 19:18:11.689059 4859 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-scheduler-0_70906bc9-3b52-4436-8d68-cdb37f0d2478/nova-scheduler-scheduler/0.log" Oct 08 19:18:11 crc kubenswrapper[4859]: I1008 19:18:11.914307 4859 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_openstack-cell1-galera-0_a2c834dc-2b23-4081-8c71-e7ae462ca063/mysql-bootstrap/0.log" Oct 08 19:18:12 crc kubenswrapper[4859]: I1008 19:18:12.114290 4859 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_openstack-cell1-galera-0_a2c834dc-2b23-4081-8c71-e7ae462ca063/galera/0.log" Oct 08 19:18:12 crc kubenswrapper[4859]: I1008 19:18:12.146079 4859 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_openstack-cell1-galera-0_a2c834dc-2b23-4081-8c71-e7ae462ca063/mysql-bootstrap/0.log" Oct 08 19:18:12 crc kubenswrapper[4859]: I1008 19:18:12.342032 4859 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_openstack-galera-0_f1828698-e403-4643-a9ab-aedf302aae70/mysql-bootstrap/0.log" Oct 08 19:18:12 crc kubenswrapper[4859]: I1008 19:18:12.566588 4859 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_openstack-galera-0_f1828698-e403-4643-a9ab-aedf302aae70/mysql-bootstrap/0.log" Oct 08 19:18:12 crc kubenswrapper[4859]: I1008 19:18:12.649799 4859 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-metadata-0_6188ecfc-2b44-443d-a966-8b91ae535533/nova-metadata-metadata/0.log" Oct 08 19:18:12 crc kubenswrapper[4859]: I1008 19:18:12.682940 4859 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_openstack-galera-0_f1828698-e403-4643-a9ab-aedf302aae70/galera/0.log" Oct 08 19:18:12 crc kubenswrapper[4859]: I1008 19:18:12.901992 4859 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_openstackclient_1a631d33-4c0c-46c9-a3d5-4d9dc9f6de84/openstackclient/0.log" Oct 08 19:18:13 crc kubenswrapper[4859]: I1008 19:18:13.008540 4859 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-controller-b89lv_ec20820a-066c-4606-8f1b-a26b172b7720/ovn-controller/0.log" Oct 08 19:18:13 crc kubenswrapper[4859]: I1008 19:18:13.144791 4859 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-controller-metrics-bjb7d_f29e1ba4-3f25-45b7-ba3a-3592ce322840/openstack-network-exporter/0.log" Oct 08 19:18:13 crc kubenswrapper[4859]: I1008 19:18:13.354644 4859 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-controller-ovs-vthdk_eca01cc4-b7ef-429c-984d-1ccd1d95bcd2/ovsdb-server-init/0.log" Oct 08 19:18:13 crc kubenswrapper[4859]: I1008 19:18:13.558681 4859 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-controller-ovs-vthdk_eca01cc4-b7ef-429c-984d-1ccd1d95bcd2/ovsdb-server-init/0.log" Oct 08 19:18:13 crc kubenswrapper[4859]: I1008 19:18:13.569348 4859 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-controller-ovs-vthdk_eca01cc4-b7ef-429c-984d-1ccd1d95bcd2/ovs-vswitchd/0.log" Oct 08 19:18:13 crc kubenswrapper[4859]: I1008 19:18:13.621054 4859 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-controller-ovs-vthdk_eca01cc4-b7ef-429c-984d-1ccd1d95bcd2/ovsdb-server/0.log" Oct 08 19:18:13 crc kubenswrapper[4859]: I1008 19:18:13.868216 4859 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-edpm-deployment-openstack-edpm-ipam-gtbjg_4bbd12de-a0ea-4741-b6af-dc19464d2161/ovn-edpm-deployment-openstack-edpm-ipam/0.log" Oct 08 19:18:14 crc kubenswrapper[4859]: I1008 19:18:14.020732 4859 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-northd-0_3d31b043-3252-458e-8f23-afd68b624849/openstack-network-exporter/0.log" Oct 08 19:18:14 crc kubenswrapper[4859]: I1008 19:18:14.121394 4859 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-northd-0_3d31b043-3252-458e-8f23-afd68b624849/ovn-northd/0.log" Oct 08 19:18:14 crc kubenswrapper[4859]: I1008 19:18:14.242922 4859 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovsdbserver-nb-0_9fc7c8d2-a890-43c0-8081-6049a57c7474/openstack-network-exporter/0.log" Oct 08 19:18:14 crc kubenswrapper[4859]: I1008 19:18:14.365227 4859 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovsdbserver-nb-0_9fc7c8d2-a890-43c0-8081-6049a57c7474/ovsdbserver-nb/0.log" Oct 08 19:18:14 crc kubenswrapper[4859]: I1008 19:18:14.493411 4859 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovsdbserver-sb-0_fb46e63a-80c0-4141-8e6a-ad49c95bc924/openstack-network-exporter/0.log" Oct 08 19:18:14 crc kubenswrapper[4859]: I1008 19:18:14.579911 4859 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovsdbserver-sb-0_fb46e63a-80c0-4141-8e6a-ad49c95bc924/ovsdbserver-sb/0.log" Oct 08 19:18:14 crc kubenswrapper[4859]: I1008 19:18:14.839720 4859 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_placement-58c5f664f4-5lvrs_01262af8-82d8-4771-9fe8-472dc77e4b60/placement-api/0.log" Oct 08 19:18:14 crc kubenswrapper[4859]: I1008 19:18:14.893244 4859 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_placement-58c5f664f4-5lvrs_01262af8-82d8-4771-9fe8-472dc77e4b60/placement-log/0.log" Oct 08 19:18:15 crc kubenswrapper[4859]: I1008 19:18:15.050349 4859 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_rabbitmq-cell1-server-0_2c6db854-3378-4bc4-b622-480bc8c42bf5/setup-container/0.log" Oct 08 19:18:15 crc kubenswrapper[4859]: I1008 19:18:15.252651 4859 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_rabbitmq-cell1-server-0_2c6db854-3378-4bc4-b622-480bc8c42bf5/rabbitmq/0.log" Oct 08 19:18:15 crc kubenswrapper[4859]: I1008 19:18:15.290608 4859 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_rabbitmq-cell1-server-0_2c6db854-3378-4bc4-b622-480bc8c42bf5/setup-container/0.log" Oct 08 19:18:15 crc kubenswrapper[4859]: I1008 19:18:15.544785 4859 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_rabbitmq-server-0_94450380-649a-4476-b646-018ed24aa703/setup-container/0.log" Oct 08 19:18:15 crc kubenswrapper[4859]: I1008 19:18:15.678861 4859 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_rabbitmq-server-0_94450380-649a-4476-b646-018ed24aa703/rabbitmq/0.log" Oct 08 19:18:15 crc kubenswrapper[4859]: I1008 19:18:15.720356 4859 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_rabbitmq-server-0_94450380-649a-4476-b646-018ed24aa703/setup-container/0.log" Oct 08 19:18:15 crc kubenswrapper[4859]: I1008 19:18:15.975539 4859 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_reboot-os-edpm-deployment-openstack-edpm-ipam-cjdnl_5dc6ec8a-d555-43e6-b295-821ae62d0eb1/reboot-os-edpm-deployment-openstack-edpm-ipam/0.log" Oct 08 19:18:15 crc kubenswrapper[4859]: I1008 19:18:15.992502 4859 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_redhat-edpm-deployment-openstack-edpm-ipam-fz6lm_02a2d3c9-f778-4465-b0aa-62c051299f2b/redhat-edpm-deployment-openstack-edpm-ipam/0.log" Oct 08 19:18:16 crc kubenswrapper[4859]: I1008 19:18:16.208916 4859 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_repo-setup-edpm-deployment-openstack-edpm-ipam-n8rrq_53d801a5-a432-44f3-970a-23519033d1e3/repo-setup-edpm-deployment-openstack-edpm-ipam/0.log" Oct 08 19:18:16 crc kubenswrapper[4859]: I1008 19:18:16.384090 4859 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_run-os-edpm-deployment-openstack-edpm-ipam-746dn_e37b6e90-4b9a-41e8-9a83-db72281cc7f5/run-os-edpm-deployment-openstack-edpm-ipam/0.log" Oct 08 19:18:16 crc kubenswrapper[4859]: I1008 19:18:16.549254 4859 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ssh-known-hosts-edpm-deployment-f546c_252ee072-dfc8-4b4b-a8dc-c695f2be633a/ssh-known-hosts-edpm-deployment/0.log" Oct 08 19:18:16 crc kubenswrapper[4859]: I1008 19:18:16.802781 4859 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-proxy-776bbffb49-bdzz5_465dc520-34ae-41b8-869c-5d4ead3b16ab/proxy-server/0.log" Oct 08 19:18:16 crc kubenswrapper[4859]: I1008 19:18:16.824459 4859 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-proxy-776bbffb49-bdzz5_465dc520-34ae-41b8-869c-5d4ead3b16ab/proxy-httpd/0.log" Oct 08 19:18:16 crc kubenswrapper[4859]: I1008 19:18:16.987994 4859 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-ring-rebalance-qrpd6_431b4e6e-415d-4c14-8783-f94409691ae3/swift-ring-rebalance/0.log" Oct 08 19:18:17 crc kubenswrapper[4859]: I1008 19:18:17.143485 4859 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_c860866f-d309-4824-a31c-e88c5d820712/account-auditor/0.log" Oct 08 19:18:17 crc kubenswrapper[4859]: I1008 19:18:17.241197 4859 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_c860866f-d309-4824-a31c-e88c5d820712/account-reaper/0.log" Oct 08 19:18:17 crc kubenswrapper[4859]: I1008 19:18:17.332092 4859 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_c860866f-d309-4824-a31c-e88c5d820712/account-replicator/0.log" Oct 08 19:18:17 crc kubenswrapper[4859]: I1008 19:18:17.360504 4859 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_c860866f-d309-4824-a31c-e88c5d820712/account-server/0.log" Oct 08 19:18:17 crc kubenswrapper[4859]: I1008 19:18:17.452796 4859 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_c860866f-d309-4824-a31c-e88c5d820712/container-auditor/0.log" Oct 08 19:18:17 crc kubenswrapper[4859]: I1008 19:18:17.573181 4859 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_c860866f-d309-4824-a31c-e88c5d820712/container-replicator/0.log" Oct 08 19:18:17 crc kubenswrapper[4859]: I1008 19:18:17.602059 4859 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_c860866f-d309-4824-a31c-e88c5d820712/container-server/0.log" Oct 08 19:18:17 crc kubenswrapper[4859]: I1008 19:18:17.676652 4859 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_c860866f-d309-4824-a31c-e88c5d820712/container-updater/0.log" Oct 08 19:18:17 crc kubenswrapper[4859]: I1008 19:18:17.797854 4859 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_c860866f-d309-4824-a31c-e88c5d820712/object-expirer/0.log" Oct 08 19:18:17 crc kubenswrapper[4859]: I1008 19:18:17.812913 4859 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_c860866f-d309-4824-a31c-e88c5d820712/object-auditor/0.log" Oct 08 19:18:17 crc kubenswrapper[4859]: I1008 19:18:17.951320 4859 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_c860866f-d309-4824-a31c-e88c5d820712/object-replicator/0.log" Oct 08 19:18:18 crc kubenswrapper[4859]: I1008 19:18:18.008017 4859 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_c860866f-d309-4824-a31c-e88c5d820712/object-server/0.log" Oct 08 19:18:18 crc kubenswrapper[4859]: I1008 19:18:18.028121 4859 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_c860866f-d309-4824-a31c-e88c5d820712/object-updater/0.log" Oct 08 19:18:18 crc kubenswrapper[4859]: I1008 19:18:18.214230 4859 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_c860866f-d309-4824-a31c-e88c5d820712/rsync/0.log" Oct 08 19:18:18 crc kubenswrapper[4859]: I1008 19:18:18.289497 4859 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_c860866f-d309-4824-a31c-e88c5d820712/swift-recon-cron/0.log" Oct 08 19:18:18 crc kubenswrapper[4859]: I1008 19:18:18.492652 4859 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_telemetry-edpm-deployment-openstack-edpm-ipam-twhvj_65244148-7971-47bf-b900-96932c6cd473/telemetry-edpm-deployment-openstack-edpm-ipam/0.log" Oct 08 19:18:18 crc kubenswrapper[4859]: I1008 19:18:18.614037 4859 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_tempest-tests-tempest_be909489-e6c5-4464-9292-aacd831885d7/tempest-tests-tempest-tests-runner/0.log" Oct 08 19:18:18 crc kubenswrapper[4859]: I1008 19:18:18.730186 4859 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_test-operator-logs-pod-tempest-tempest-tests-tempest_f9597f73-13f2-4547-862b-fa23bba1c625/test-operator-logs-container/0.log" Oct 08 19:18:18 crc kubenswrapper[4859]: I1008 19:18:18.970937 4859 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_validate-network-edpm-deployment-openstack-edpm-ipam-2lvj9_67e6fffe-7caa-42a6-812c-efa6a8f2162b/validate-network-edpm-deployment-openstack-edpm-ipam/0.log" Oct 08 19:18:25 crc kubenswrapper[4859]: I1008 19:18:25.242196 4859 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_memcached-0_6dc3fa5d-109b-4140-93aa-61b855c46df2/memcached/0.log" Oct 08 19:19:04 crc kubenswrapper[4859]: I1008 19:19:04.815487 4859 generic.go:334] "Generic (PLEG): container finished" podID="ca27de4c-c560-4441-99df-f9c49c3d97e0" containerID="335899cba18961c57684d6717a63018db36b992f2dec8df2b77f0e7340d13582" exitCode=0 Oct 08 19:19:04 crc kubenswrapper[4859]: I1008 19:19:04.815620 4859 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-f7r6n/crc-debug-829bd" event={"ID":"ca27de4c-c560-4441-99df-f9c49c3d97e0","Type":"ContainerDied","Data":"335899cba18961c57684d6717a63018db36b992f2dec8df2b77f0e7340d13582"} Oct 08 19:19:05 crc kubenswrapper[4859]: I1008 19:19:05.920999 4859 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-f7r6n/crc-debug-829bd" Oct 08 19:19:05 crc kubenswrapper[4859]: I1008 19:19:05.965240 4859 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-must-gather-f7r6n/crc-debug-829bd"] Oct 08 19:19:05 crc kubenswrapper[4859]: I1008 19:19:05.973205 4859 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-must-gather-f7r6n/crc-debug-829bd"] Oct 08 19:19:06 crc kubenswrapper[4859]: I1008 19:19:06.023886 4859 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/ca27de4c-c560-4441-99df-f9c49c3d97e0-host\") pod \"ca27de4c-c560-4441-99df-f9c49c3d97e0\" (UID: \"ca27de4c-c560-4441-99df-f9c49c3d97e0\") " Oct 08 19:19:06 crc kubenswrapper[4859]: I1008 19:19:06.024269 4859 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-f2tpn\" (UniqueName: \"kubernetes.io/projected/ca27de4c-c560-4441-99df-f9c49c3d97e0-kube-api-access-f2tpn\") pod \"ca27de4c-c560-4441-99df-f9c49c3d97e0\" (UID: \"ca27de4c-c560-4441-99df-f9c49c3d97e0\") " Oct 08 19:19:06 crc kubenswrapper[4859]: I1008 19:19:06.024324 4859 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/ca27de4c-c560-4441-99df-f9c49c3d97e0-host" (OuterVolumeSpecName: "host") pod "ca27de4c-c560-4441-99df-f9c49c3d97e0" (UID: "ca27de4c-c560-4441-99df-f9c49c3d97e0"). InnerVolumeSpecName "host". PluginName "kubernetes.io/host-path", VolumeGidValue "" Oct 08 19:19:06 crc kubenswrapper[4859]: I1008 19:19:06.024924 4859 reconciler_common.go:293] "Volume detached for volume \"host\" (UniqueName: \"kubernetes.io/host-path/ca27de4c-c560-4441-99df-f9c49c3d97e0-host\") on node \"crc\" DevicePath \"\"" Oct 08 19:19:06 crc kubenswrapper[4859]: I1008 19:19:06.030472 4859 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/ca27de4c-c560-4441-99df-f9c49c3d97e0-kube-api-access-f2tpn" (OuterVolumeSpecName: "kube-api-access-f2tpn") pod "ca27de4c-c560-4441-99df-f9c49c3d97e0" (UID: "ca27de4c-c560-4441-99df-f9c49c3d97e0"). InnerVolumeSpecName "kube-api-access-f2tpn". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 08 19:19:06 crc kubenswrapper[4859]: I1008 19:19:06.126962 4859 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-f2tpn\" (UniqueName: \"kubernetes.io/projected/ca27de4c-c560-4441-99df-f9c49c3d97e0-kube-api-access-f2tpn\") on node \"crc\" DevicePath \"\"" Oct 08 19:19:06 crc kubenswrapper[4859]: I1008 19:19:06.490921 4859 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="ca27de4c-c560-4441-99df-f9c49c3d97e0" path="/var/lib/kubelet/pods/ca27de4c-c560-4441-99df-f9c49c3d97e0/volumes" Oct 08 19:19:06 crc kubenswrapper[4859]: I1008 19:19:06.838903 4859 scope.go:117] "RemoveContainer" containerID="335899cba18961c57684d6717a63018db36b992f2dec8df2b77f0e7340d13582" Oct 08 19:19:06 crc kubenswrapper[4859]: I1008 19:19:06.838961 4859 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-f7r6n/crc-debug-829bd" Oct 08 19:19:07 crc kubenswrapper[4859]: I1008 19:19:07.192506 4859 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-must-gather-f7r6n/crc-debug-8cwvm"] Oct 08 19:19:07 crc kubenswrapper[4859]: E1008 19:19:07.193622 4859 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="640efdd5-e75b-4957-8aaa-e7a9a5828962" containerName="registry-server" Oct 08 19:19:07 crc kubenswrapper[4859]: I1008 19:19:07.193757 4859 state_mem.go:107] "Deleted CPUSet assignment" podUID="640efdd5-e75b-4957-8aaa-e7a9a5828962" containerName="registry-server" Oct 08 19:19:07 crc kubenswrapper[4859]: E1008 19:19:07.193863 4859 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="640efdd5-e75b-4957-8aaa-e7a9a5828962" containerName="extract-content" Oct 08 19:19:07 crc kubenswrapper[4859]: I1008 19:19:07.193934 4859 state_mem.go:107] "Deleted CPUSet assignment" podUID="640efdd5-e75b-4957-8aaa-e7a9a5828962" containerName="extract-content" Oct 08 19:19:07 crc kubenswrapper[4859]: E1008 19:19:07.194003 4859 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ca27de4c-c560-4441-99df-f9c49c3d97e0" containerName="container-00" Oct 08 19:19:07 crc kubenswrapper[4859]: I1008 19:19:07.194061 4859 state_mem.go:107] "Deleted CPUSet assignment" podUID="ca27de4c-c560-4441-99df-f9c49c3d97e0" containerName="container-00" Oct 08 19:19:07 crc kubenswrapper[4859]: E1008 19:19:07.194125 4859 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="640efdd5-e75b-4957-8aaa-e7a9a5828962" containerName="extract-utilities" Oct 08 19:19:07 crc kubenswrapper[4859]: I1008 19:19:07.194182 4859 state_mem.go:107] "Deleted CPUSet assignment" podUID="640efdd5-e75b-4957-8aaa-e7a9a5828962" containerName="extract-utilities" Oct 08 19:19:07 crc kubenswrapper[4859]: I1008 19:19:07.194432 4859 memory_manager.go:354] "RemoveStaleState removing state" podUID="640efdd5-e75b-4957-8aaa-e7a9a5828962" containerName="registry-server" Oct 08 19:19:07 crc kubenswrapper[4859]: I1008 19:19:07.194510 4859 memory_manager.go:354] "RemoveStaleState removing state" podUID="ca27de4c-c560-4441-99df-f9c49c3d97e0" containerName="container-00" Oct 08 19:19:07 crc kubenswrapper[4859]: I1008 19:19:07.195238 4859 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-f7r6n/crc-debug-8cwvm" Oct 08 19:19:07 crc kubenswrapper[4859]: I1008 19:19:07.251848 4859 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/9e607837-1f09-480d-a057-33add773636e-host\") pod \"crc-debug-8cwvm\" (UID: \"9e607837-1f09-480d-a057-33add773636e\") " pod="openshift-must-gather-f7r6n/crc-debug-8cwvm" Oct 08 19:19:07 crc kubenswrapper[4859]: I1008 19:19:07.252146 4859 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-qh765\" (UniqueName: \"kubernetes.io/projected/9e607837-1f09-480d-a057-33add773636e-kube-api-access-qh765\") pod \"crc-debug-8cwvm\" (UID: \"9e607837-1f09-480d-a057-33add773636e\") " pod="openshift-must-gather-f7r6n/crc-debug-8cwvm" Oct 08 19:19:07 crc kubenswrapper[4859]: I1008 19:19:07.354410 4859 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-qh765\" (UniqueName: \"kubernetes.io/projected/9e607837-1f09-480d-a057-33add773636e-kube-api-access-qh765\") pod \"crc-debug-8cwvm\" (UID: \"9e607837-1f09-480d-a057-33add773636e\") " pod="openshift-must-gather-f7r6n/crc-debug-8cwvm" Oct 08 19:19:07 crc kubenswrapper[4859]: I1008 19:19:07.354609 4859 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/9e607837-1f09-480d-a057-33add773636e-host\") pod \"crc-debug-8cwvm\" (UID: \"9e607837-1f09-480d-a057-33add773636e\") " pod="openshift-must-gather-f7r6n/crc-debug-8cwvm" Oct 08 19:19:07 crc kubenswrapper[4859]: I1008 19:19:07.354829 4859 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host\" (UniqueName: \"kubernetes.io/host-path/9e607837-1f09-480d-a057-33add773636e-host\") pod \"crc-debug-8cwvm\" (UID: \"9e607837-1f09-480d-a057-33add773636e\") " pod="openshift-must-gather-f7r6n/crc-debug-8cwvm" Oct 08 19:19:07 crc kubenswrapper[4859]: I1008 19:19:07.375752 4859 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-qh765\" (UniqueName: \"kubernetes.io/projected/9e607837-1f09-480d-a057-33add773636e-kube-api-access-qh765\") pod \"crc-debug-8cwvm\" (UID: \"9e607837-1f09-480d-a057-33add773636e\") " pod="openshift-must-gather-f7r6n/crc-debug-8cwvm" Oct 08 19:19:07 crc kubenswrapper[4859]: I1008 19:19:07.517044 4859 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-f7r6n/crc-debug-8cwvm" Oct 08 19:19:07 crc kubenswrapper[4859]: I1008 19:19:07.849437 4859 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-f7r6n/crc-debug-8cwvm" event={"ID":"9e607837-1f09-480d-a057-33add773636e","Type":"ContainerStarted","Data":"aaf7a34a7ba622ed67c2d3b385e9a8219112acf5608aeea4d037e11fd5595b1f"} Oct 08 19:19:07 crc kubenswrapper[4859]: I1008 19:19:07.849917 4859 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-f7r6n/crc-debug-8cwvm" event={"ID":"9e607837-1f09-480d-a057-33add773636e","Type":"ContainerStarted","Data":"5b958248258ddcf0998a052454d8df28ef5309ffd57edb66175254e66227981c"} Oct 08 19:19:07 crc kubenswrapper[4859]: I1008 19:19:07.875564 4859 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-must-gather-f7r6n/crc-debug-8cwvm" podStartSLOduration=0.87554047 podStartE2EDuration="875.54047ms" podCreationTimestamp="2025-10-08 19:19:07 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-08 19:19:07.862760287 +0000 UTC m=+3718.109599686" watchObservedRunningTime="2025-10-08 19:19:07.87554047 +0000 UTC m=+3718.122379859" Oct 08 19:19:08 crc kubenswrapper[4859]: I1008 19:19:08.875591 4859 generic.go:334] "Generic (PLEG): container finished" podID="9e607837-1f09-480d-a057-33add773636e" containerID="aaf7a34a7ba622ed67c2d3b385e9a8219112acf5608aeea4d037e11fd5595b1f" exitCode=0 Oct 08 19:19:08 crc kubenswrapper[4859]: I1008 19:19:08.875660 4859 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-f7r6n/crc-debug-8cwvm" event={"ID":"9e607837-1f09-480d-a057-33add773636e","Type":"ContainerDied","Data":"aaf7a34a7ba622ed67c2d3b385e9a8219112acf5608aeea4d037e11fd5595b1f"} Oct 08 19:19:09 crc kubenswrapper[4859]: I1008 19:19:09.980859 4859 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-f7r6n/crc-debug-8cwvm" Oct 08 19:19:10 crc kubenswrapper[4859]: I1008 19:19:10.100388 4859 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-qh765\" (UniqueName: \"kubernetes.io/projected/9e607837-1f09-480d-a057-33add773636e-kube-api-access-qh765\") pod \"9e607837-1f09-480d-a057-33add773636e\" (UID: \"9e607837-1f09-480d-a057-33add773636e\") " Oct 08 19:19:10 crc kubenswrapper[4859]: I1008 19:19:10.100493 4859 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/9e607837-1f09-480d-a057-33add773636e-host\") pod \"9e607837-1f09-480d-a057-33add773636e\" (UID: \"9e607837-1f09-480d-a057-33add773636e\") " Oct 08 19:19:10 crc kubenswrapper[4859]: I1008 19:19:10.101233 4859 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/9e607837-1f09-480d-a057-33add773636e-host" (OuterVolumeSpecName: "host") pod "9e607837-1f09-480d-a057-33add773636e" (UID: "9e607837-1f09-480d-a057-33add773636e"). InnerVolumeSpecName "host". PluginName "kubernetes.io/host-path", VolumeGidValue "" Oct 08 19:19:10 crc kubenswrapper[4859]: I1008 19:19:10.111907 4859 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/9e607837-1f09-480d-a057-33add773636e-kube-api-access-qh765" (OuterVolumeSpecName: "kube-api-access-qh765") pod "9e607837-1f09-480d-a057-33add773636e" (UID: "9e607837-1f09-480d-a057-33add773636e"). InnerVolumeSpecName "kube-api-access-qh765". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 08 19:19:10 crc kubenswrapper[4859]: I1008 19:19:10.204729 4859 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-qh765\" (UniqueName: \"kubernetes.io/projected/9e607837-1f09-480d-a057-33add773636e-kube-api-access-qh765\") on node \"crc\" DevicePath \"\"" Oct 08 19:19:10 crc kubenswrapper[4859]: I1008 19:19:10.205051 4859 reconciler_common.go:293] "Volume detached for volume \"host\" (UniqueName: \"kubernetes.io/host-path/9e607837-1f09-480d-a057-33add773636e-host\") on node \"crc\" DevicePath \"\"" Oct 08 19:19:10 crc kubenswrapper[4859]: I1008 19:19:10.899960 4859 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-f7r6n/crc-debug-8cwvm" event={"ID":"9e607837-1f09-480d-a057-33add773636e","Type":"ContainerDied","Data":"5b958248258ddcf0998a052454d8df28ef5309ffd57edb66175254e66227981c"} Oct 08 19:19:10 crc kubenswrapper[4859]: I1008 19:19:10.900016 4859 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="5b958248258ddcf0998a052454d8df28ef5309ffd57edb66175254e66227981c" Oct 08 19:19:10 crc kubenswrapper[4859]: I1008 19:19:10.900094 4859 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-f7r6n/crc-debug-8cwvm" Oct 08 19:19:14 crc kubenswrapper[4859]: I1008 19:19:14.665196 4859 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-must-gather-f7r6n/crc-debug-8cwvm"] Oct 08 19:19:14 crc kubenswrapper[4859]: I1008 19:19:14.673101 4859 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-must-gather-f7r6n/crc-debug-8cwvm"] Oct 08 19:19:15 crc kubenswrapper[4859]: I1008 19:19:15.900601 4859 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-must-gather-f7r6n/crc-debug-c5n28"] Oct 08 19:19:15 crc kubenswrapper[4859]: E1008 19:19:15.901885 4859 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9e607837-1f09-480d-a057-33add773636e" containerName="container-00" Oct 08 19:19:15 crc kubenswrapper[4859]: I1008 19:19:15.901913 4859 state_mem.go:107] "Deleted CPUSet assignment" podUID="9e607837-1f09-480d-a057-33add773636e" containerName="container-00" Oct 08 19:19:15 crc kubenswrapper[4859]: I1008 19:19:15.902291 4859 memory_manager.go:354] "RemoveStaleState removing state" podUID="9e607837-1f09-480d-a057-33add773636e" containerName="container-00" Oct 08 19:19:15 crc kubenswrapper[4859]: I1008 19:19:15.903349 4859 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-f7r6n/crc-debug-c5n28" Oct 08 19:19:16 crc kubenswrapper[4859]: I1008 19:19:16.011551 4859 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/b6ec1f7a-65b5-4979-8f2c-3b77d03bb399-host\") pod \"crc-debug-c5n28\" (UID: \"b6ec1f7a-65b5-4979-8f2c-3b77d03bb399\") " pod="openshift-must-gather-f7r6n/crc-debug-c5n28" Oct 08 19:19:16 crc kubenswrapper[4859]: I1008 19:19:16.011642 4859 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-d7vx6\" (UniqueName: \"kubernetes.io/projected/b6ec1f7a-65b5-4979-8f2c-3b77d03bb399-kube-api-access-d7vx6\") pod \"crc-debug-c5n28\" (UID: \"b6ec1f7a-65b5-4979-8f2c-3b77d03bb399\") " pod="openshift-must-gather-f7r6n/crc-debug-c5n28" Oct 08 19:19:16 crc kubenswrapper[4859]: I1008 19:19:16.114172 4859 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-d7vx6\" (UniqueName: \"kubernetes.io/projected/b6ec1f7a-65b5-4979-8f2c-3b77d03bb399-kube-api-access-d7vx6\") pod \"crc-debug-c5n28\" (UID: \"b6ec1f7a-65b5-4979-8f2c-3b77d03bb399\") " pod="openshift-must-gather-f7r6n/crc-debug-c5n28" Oct 08 19:19:16 crc kubenswrapper[4859]: I1008 19:19:16.114233 4859 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/b6ec1f7a-65b5-4979-8f2c-3b77d03bb399-host\") pod \"crc-debug-c5n28\" (UID: \"b6ec1f7a-65b5-4979-8f2c-3b77d03bb399\") " pod="openshift-must-gather-f7r6n/crc-debug-c5n28" Oct 08 19:19:16 crc kubenswrapper[4859]: I1008 19:19:16.114539 4859 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host\" (UniqueName: \"kubernetes.io/host-path/b6ec1f7a-65b5-4979-8f2c-3b77d03bb399-host\") pod \"crc-debug-c5n28\" (UID: \"b6ec1f7a-65b5-4979-8f2c-3b77d03bb399\") " pod="openshift-must-gather-f7r6n/crc-debug-c5n28" Oct 08 19:19:16 crc kubenswrapper[4859]: I1008 19:19:16.159367 4859 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-d7vx6\" (UniqueName: \"kubernetes.io/projected/b6ec1f7a-65b5-4979-8f2c-3b77d03bb399-kube-api-access-d7vx6\") pod \"crc-debug-c5n28\" (UID: \"b6ec1f7a-65b5-4979-8f2c-3b77d03bb399\") " pod="openshift-must-gather-f7r6n/crc-debug-c5n28" Oct 08 19:19:16 crc kubenswrapper[4859]: I1008 19:19:16.239774 4859 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-f7r6n/crc-debug-c5n28" Oct 08 19:19:16 crc kubenswrapper[4859]: I1008 19:19:16.487328 4859 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="9e607837-1f09-480d-a057-33add773636e" path="/var/lib/kubelet/pods/9e607837-1f09-480d-a057-33add773636e/volumes" Oct 08 19:19:16 crc kubenswrapper[4859]: I1008 19:19:16.987850 4859 generic.go:334] "Generic (PLEG): container finished" podID="b6ec1f7a-65b5-4979-8f2c-3b77d03bb399" containerID="0e7169303b1f6d525281d8c48e1488b08668fc51e577b30c6f9ef77604f1f9d9" exitCode=0 Oct 08 19:19:16 crc kubenswrapper[4859]: I1008 19:19:16.987919 4859 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-f7r6n/crc-debug-c5n28" event={"ID":"b6ec1f7a-65b5-4979-8f2c-3b77d03bb399","Type":"ContainerDied","Data":"0e7169303b1f6d525281d8c48e1488b08668fc51e577b30c6f9ef77604f1f9d9"} Oct 08 19:19:16 crc kubenswrapper[4859]: I1008 19:19:16.988001 4859 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-f7r6n/crc-debug-c5n28" event={"ID":"b6ec1f7a-65b5-4979-8f2c-3b77d03bb399","Type":"ContainerStarted","Data":"b2c6f46055bf87ac47d75021d00ca2e9da1fa4efdc3994a1eb70f5ef7d5e111b"} Oct 08 19:19:17 crc kubenswrapper[4859]: I1008 19:19:17.036392 4859 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-must-gather-f7r6n/crc-debug-c5n28"] Oct 08 19:19:17 crc kubenswrapper[4859]: I1008 19:19:17.046552 4859 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-must-gather-f7r6n/crc-debug-c5n28"] Oct 08 19:19:18 crc kubenswrapper[4859]: I1008 19:19:18.099668 4859 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-f7r6n/crc-debug-c5n28" Oct 08 19:19:18 crc kubenswrapper[4859]: I1008 19:19:18.161508 4859 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/b6ec1f7a-65b5-4979-8f2c-3b77d03bb399-host\") pod \"b6ec1f7a-65b5-4979-8f2c-3b77d03bb399\" (UID: \"b6ec1f7a-65b5-4979-8f2c-3b77d03bb399\") " Oct 08 19:19:18 crc kubenswrapper[4859]: I1008 19:19:18.161589 4859 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/b6ec1f7a-65b5-4979-8f2c-3b77d03bb399-host" (OuterVolumeSpecName: "host") pod "b6ec1f7a-65b5-4979-8f2c-3b77d03bb399" (UID: "b6ec1f7a-65b5-4979-8f2c-3b77d03bb399"). InnerVolumeSpecName "host". PluginName "kubernetes.io/host-path", VolumeGidValue "" Oct 08 19:19:18 crc kubenswrapper[4859]: I1008 19:19:18.161850 4859 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-d7vx6\" (UniqueName: \"kubernetes.io/projected/b6ec1f7a-65b5-4979-8f2c-3b77d03bb399-kube-api-access-d7vx6\") pod \"b6ec1f7a-65b5-4979-8f2c-3b77d03bb399\" (UID: \"b6ec1f7a-65b5-4979-8f2c-3b77d03bb399\") " Oct 08 19:19:18 crc kubenswrapper[4859]: I1008 19:19:18.162333 4859 reconciler_common.go:293] "Volume detached for volume \"host\" (UniqueName: \"kubernetes.io/host-path/b6ec1f7a-65b5-4979-8f2c-3b77d03bb399-host\") on node \"crc\" DevicePath \"\"" Oct 08 19:19:18 crc kubenswrapper[4859]: I1008 19:19:18.168443 4859 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b6ec1f7a-65b5-4979-8f2c-3b77d03bb399-kube-api-access-d7vx6" (OuterVolumeSpecName: "kube-api-access-d7vx6") pod "b6ec1f7a-65b5-4979-8f2c-3b77d03bb399" (UID: "b6ec1f7a-65b5-4979-8f2c-3b77d03bb399"). InnerVolumeSpecName "kube-api-access-d7vx6". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 08 19:19:18 crc kubenswrapper[4859]: I1008 19:19:18.264181 4859 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-d7vx6\" (UniqueName: \"kubernetes.io/projected/b6ec1f7a-65b5-4979-8f2c-3b77d03bb399-kube-api-access-d7vx6\") on node \"crc\" DevicePath \"\"" Oct 08 19:19:18 crc kubenswrapper[4859]: I1008 19:19:18.533953 4859 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="b6ec1f7a-65b5-4979-8f2c-3b77d03bb399" path="/var/lib/kubelet/pods/b6ec1f7a-65b5-4979-8f2c-3b77d03bb399/volumes" Oct 08 19:19:18 crc kubenswrapper[4859]: I1008 19:19:18.552105 4859 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_1bfea55cf6f540853d0c0d919aed51b5fd6e2d105cc44a68cb64399f8bxp4tb_70914ba1-396a-4131-835f-817debc89338/util/0.log" Oct 08 19:19:18 crc kubenswrapper[4859]: I1008 19:19:18.754308 4859 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_1bfea55cf6f540853d0c0d919aed51b5fd6e2d105cc44a68cb64399f8bxp4tb_70914ba1-396a-4131-835f-817debc89338/util/0.log" Oct 08 19:19:18 crc kubenswrapper[4859]: I1008 19:19:18.775036 4859 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_1bfea55cf6f540853d0c0d919aed51b5fd6e2d105cc44a68cb64399f8bxp4tb_70914ba1-396a-4131-835f-817debc89338/pull/0.log" Oct 08 19:19:18 crc kubenswrapper[4859]: I1008 19:19:18.807351 4859 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_1bfea55cf6f540853d0c0d919aed51b5fd6e2d105cc44a68cb64399f8bxp4tb_70914ba1-396a-4131-835f-817debc89338/pull/0.log" Oct 08 19:19:18 crc kubenswrapper[4859]: I1008 19:19:18.978203 4859 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_1bfea55cf6f540853d0c0d919aed51b5fd6e2d105cc44a68cb64399f8bxp4tb_70914ba1-396a-4131-835f-817debc89338/util/0.log" Oct 08 19:19:19 crc kubenswrapper[4859]: I1008 19:19:19.001272 4859 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_1bfea55cf6f540853d0c0d919aed51b5fd6e2d105cc44a68cb64399f8bxp4tb_70914ba1-396a-4131-835f-817debc89338/pull/0.log" Oct 08 19:19:19 crc kubenswrapper[4859]: I1008 19:19:19.004407 4859 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_1bfea55cf6f540853d0c0d919aed51b5fd6e2d105cc44a68cb64399f8bxp4tb_70914ba1-396a-4131-835f-817debc89338/extract/0.log" Oct 08 19:19:19 crc kubenswrapper[4859]: I1008 19:19:19.004661 4859 scope.go:117] "RemoveContainer" containerID="0e7169303b1f6d525281d8c48e1488b08668fc51e577b30c6f9ef77604f1f9d9" Oct 08 19:19:19 crc kubenswrapper[4859]: I1008 19:19:19.004802 4859 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-f7r6n/crc-debug-c5n28" Oct 08 19:19:19 crc kubenswrapper[4859]: I1008 19:19:19.205179 4859 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_barbican-operator-controller-manager-658bdf4b74-8bfst_d097cd11-b23b-4d38-80d7-5d85c257cef4/kube-rbac-proxy/0.log" Oct 08 19:19:19 crc kubenswrapper[4859]: I1008 19:19:19.205666 4859 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_barbican-operator-controller-manager-658bdf4b74-8bfst_d097cd11-b23b-4d38-80d7-5d85c257cef4/manager/0.log" Oct 08 19:19:19 crc kubenswrapper[4859]: I1008 19:19:19.214854 4859 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_cinder-operator-controller-manager-7b7fb68549-lbw64_e2b104df-4109-45ab-8a5a-3569da53caa0/kube-rbac-proxy/0.log" Oct 08 19:19:19 crc kubenswrapper[4859]: I1008 19:19:19.438754 4859 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_cinder-operator-controller-manager-7b7fb68549-lbw64_e2b104df-4109-45ab-8a5a-3569da53caa0/manager/0.log" Oct 08 19:19:19 crc kubenswrapper[4859]: I1008 19:19:19.452020 4859 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_designate-operator-controller-manager-85d5d9dd78-nhqkh_d9a50bbe-f28f-46d8-9266-fd54e44388a3/manager/0.log" Oct 08 19:19:19 crc kubenswrapper[4859]: I1008 19:19:19.457132 4859 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_designate-operator-controller-manager-85d5d9dd78-nhqkh_d9a50bbe-f28f-46d8-9266-fd54e44388a3/kube-rbac-proxy/0.log" Oct 08 19:19:19 crc kubenswrapper[4859]: I1008 19:19:19.674197 4859 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_glance-operator-controller-manager-84b9b84486-d9gmv_0b068d16-0072-41f8-ad01-4194fd428014/kube-rbac-proxy/0.log" Oct 08 19:19:19 crc kubenswrapper[4859]: I1008 19:19:19.706166 4859 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_glance-operator-controller-manager-84b9b84486-d9gmv_0b068d16-0072-41f8-ad01-4194fd428014/manager/0.log" Oct 08 19:19:19 crc kubenswrapper[4859]: I1008 19:19:19.829858 4859 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_heat-operator-controller-manager-858f76bbdd-b5bfz_19fa1d1d-0731-4c21-9faa-b1665ec4af36/kube-rbac-proxy/0.log" Oct 08 19:19:19 crc kubenswrapper[4859]: I1008 19:19:19.849987 4859 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_heat-operator-controller-manager-858f76bbdd-b5bfz_19fa1d1d-0731-4c21-9faa-b1665ec4af36/manager/0.log" Oct 08 19:19:19 crc kubenswrapper[4859]: I1008 19:19:19.923174 4859 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_horizon-operator-controller-manager-7ffbcb7588-86z4n_ca71782e-3569-4ac2-a178-239f0ae200d0/kube-rbac-proxy/0.log" Oct 08 19:19:20 crc kubenswrapper[4859]: I1008 19:19:20.012373 4859 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_horizon-operator-controller-manager-7ffbcb7588-86z4n_ca71782e-3569-4ac2-a178-239f0ae200d0/manager/0.log" Oct 08 19:19:20 crc kubenswrapper[4859]: I1008 19:19:20.085058 4859 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_infra-operator-controller-manager-656bcbd775-ctjz9_1a1cfd66-310a-4077-8020-df7e6e069c53/kube-rbac-proxy/0.log" Oct 08 19:19:20 crc kubenswrapper[4859]: I1008 19:19:20.255094 4859 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_infra-operator-controller-manager-656bcbd775-ctjz9_1a1cfd66-310a-4077-8020-df7e6e069c53/manager/0.log" Oct 08 19:19:20 crc kubenswrapper[4859]: I1008 19:19:20.291993 4859 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_ironic-operator-controller-manager-9c5c78d49-z9952_2b67febf-c4c8-4e48-9350-29dd496439b4/kube-rbac-proxy/0.log" Oct 08 19:19:20 crc kubenswrapper[4859]: I1008 19:19:20.294864 4859 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_ironic-operator-controller-manager-9c5c78d49-z9952_2b67febf-c4c8-4e48-9350-29dd496439b4/manager/0.log" Oct 08 19:19:20 crc kubenswrapper[4859]: I1008 19:19:20.425661 4859 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_keystone-operator-controller-manager-55b6b7c7b8-rsbgr_0228033e-c835-4618-9603-2c67cf9ce57d/kube-rbac-proxy/0.log" Oct 08 19:19:20 crc kubenswrapper[4859]: I1008 19:19:20.562398 4859 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_keystone-operator-controller-manager-55b6b7c7b8-rsbgr_0228033e-c835-4618-9603-2c67cf9ce57d/manager/0.log" Oct 08 19:19:20 crc kubenswrapper[4859]: I1008 19:19:20.617706 4859 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_manila-operator-controller-manager-5f67fbc655-xg74g_fbce123a-545f-4c4c-86ca-2d68f81aa3f3/kube-rbac-proxy/0.log" Oct 08 19:19:20 crc kubenswrapper[4859]: I1008 19:19:20.796018 4859 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_manila-operator-controller-manager-5f67fbc655-xg74g_fbce123a-545f-4c4c-86ca-2d68f81aa3f3/manager/0.log" Oct 08 19:19:20 crc kubenswrapper[4859]: I1008 19:19:20.913028 4859 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_mariadb-operator-controller-manager-f9fb45f8f-ptpqs_b6a2cffa-8421-43c4-bad8-5d4d341a40cc/kube-rbac-proxy/0.log" Oct 08 19:19:20 crc kubenswrapper[4859]: I1008 19:19:20.927180 4859 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_mariadb-operator-controller-manager-f9fb45f8f-ptpqs_b6a2cffa-8421-43c4-bad8-5d4d341a40cc/manager/0.log" Oct 08 19:19:20 crc kubenswrapper[4859]: I1008 19:19:20.994054 4859 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_neutron-operator-controller-manager-79d585cb66-zn65q_e69529d1-1a2a-433c-a16b-3c6de62a0587/kube-rbac-proxy/0.log" Oct 08 19:19:21 crc kubenswrapper[4859]: I1008 19:19:21.119858 4859 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_neutron-operator-controller-manager-79d585cb66-zn65q_e69529d1-1a2a-433c-a16b-3c6de62a0587/manager/0.log" Oct 08 19:19:21 crc kubenswrapper[4859]: I1008 19:19:21.181451 4859 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_nova-operator-controller-manager-5df598886f-p99nw_e2889a4a-89ec-47de-99df-7e749c3e5514/kube-rbac-proxy/0.log" Oct 08 19:19:21 crc kubenswrapper[4859]: I1008 19:19:21.320210 4859 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_nova-operator-controller-manager-5df598886f-p99nw_e2889a4a-89ec-47de-99df-7e749c3e5514/manager/0.log" Oct 08 19:19:21 crc kubenswrapper[4859]: I1008 19:19:21.337651 4859 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_octavia-operator-controller-manager-69fdcfc5f5-lsl8l_550bb7bb-6f9c-4b20-9d92-2c85773e05a2/manager/0.log" Oct 08 19:19:21 crc kubenswrapper[4859]: I1008 19:19:21.396676 4859 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_octavia-operator-controller-manager-69fdcfc5f5-lsl8l_550bb7bb-6f9c-4b20-9d92-2c85773e05a2/kube-rbac-proxy/0.log" Oct 08 19:19:21 crc kubenswrapper[4859]: I1008 19:19:21.532096 4859 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_openstack-baremetal-operator-controller-manager-677c5f5bff5pg7p_dd924000-5690-490e-a4db-e29ed7f821ec/kube-rbac-proxy/0.log" Oct 08 19:19:21 crc kubenswrapper[4859]: I1008 19:19:21.609110 4859 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_openstack-baremetal-operator-controller-manager-677c5f5bff5pg7p_dd924000-5690-490e-a4db-e29ed7f821ec/manager/0.log" Oct 08 19:19:21 crc kubenswrapper[4859]: I1008 19:19:21.723768 4859 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_openstack-operator-controller-manager-598c4c5b5-6l55k_182ef4ca-109f-46bd-a935-091d05275271/kube-rbac-proxy/0.log" Oct 08 19:19:21 crc kubenswrapper[4859]: I1008 19:19:21.835248 4859 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_openstack-operator-controller-operator-bd6bc67fb-jkvgb_f60d2d10-2f0f-464f-9d1e-e63a5ef48972/kube-rbac-proxy/0.log" Oct 08 19:19:22 crc kubenswrapper[4859]: I1008 19:19:22.084708 4859 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_openstack-operator-controller-operator-bd6bc67fb-jkvgb_f60d2d10-2f0f-464f-9d1e-e63a5ef48972/operator/0.log" Oct 08 19:19:22 crc kubenswrapper[4859]: I1008 19:19:22.135649 4859 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_openstack-operator-index-z8f5z_6c1463a8-c973-42db-8671-f274a513230a/registry-server/0.log" Oct 08 19:19:22 crc kubenswrapper[4859]: I1008 19:19:22.269595 4859 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_ovn-operator-controller-manager-79db49b9fb-hhgrl_e2675d43-51ec-4a8c-aa2d-5aedc14f649a/kube-rbac-proxy/0.log" Oct 08 19:19:22 crc kubenswrapper[4859]: I1008 19:19:22.340640 4859 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_ovn-operator-controller-manager-79db49b9fb-hhgrl_e2675d43-51ec-4a8c-aa2d-5aedc14f649a/manager/0.log" Oct 08 19:19:22 crc kubenswrapper[4859]: I1008 19:19:22.345593 4859 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_placement-operator-controller-manager-68b6c87b68-xgjpk_f56e7a0c-7b3a-4dc8-9367-cfeef5c5cef9/kube-rbac-proxy/0.log" Oct 08 19:19:22 crc kubenswrapper[4859]: I1008 19:19:22.491347 4859 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_placement-operator-controller-manager-68b6c87b68-xgjpk_f56e7a0c-7b3a-4dc8-9367-cfeef5c5cef9/manager/0.log" Oct 08 19:19:22 crc kubenswrapper[4859]: I1008 19:19:22.567873 4859 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_rabbitmq-cluster-operator-manager-5f97d8c699-dxrl8_e1dbe8d7-f993-42b1-9a53-7d0c342780c1/operator/0.log" Oct 08 19:19:22 crc kubenswrapper[4859]: I1008 19:19:22.796252 4859 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_swift-operator-controller-manager-db6d7f97b-zjkwm_44980d79-924e-436f-950d-8fcdb4c42af1/kube-rbac-proxy/0.log" Oct 08 19:19:22 crc kubenswrapper[4859]: I1008 19:19:22.801580 4859 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_swift-operator-controller-manager-db6d7f97b-zjkwm_44980d79-924e-436f-950d-8fcdb4c42af1/manager/0.log" Oct 08 19:19:22 crc kubenswrapper[4859]: I1008 19:19:22.865484 4859 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_openstack-operator-controller-manager-598c4c5b5-6l55k_182ef4ca-109f-46bd-a935-091d05275271/manager/0.log" Oct 08 19:19:22 crc kubenswrapper[4859]: I1008 19:19:22.895026 4859 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_telemetry-operator-controller-manager-76796d4c6b-kjbfp_ffa2914a-717c-4763-8b1c-4a1ff18858f5/kube-rbac-proxy/0.log" Oct 08 19:19:23 crc kubenswrapper[4859]: I1008 19:19:23.060895 4859 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_test-operator-controller-manager-56c698c775-d2g8b_2d6cf85d-c5cc-4466-a471-bb17aa8fca9d/manager/0.log" Oct 08 19:19:23 crc kubenswrapper[4859]: I1008 19:19:23.065557 4859 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_test-operator-controller-manager-56c698c775-d2g8b_2d6cf85d-c5cc-4466-a471-bb17aa8fca9d/kube-rbac-proxy/0.log" Oct 08 19:19:23 crc kubenswrapper[4859]: I1008 19:19:23.071431 4859 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_telemetry-operator-controller-manager-76796d4c6b-kjbfp_ffa2914a-717c-4763-8b1c-4a1ff18858f5/manager/0.log" Oct 08 19:19:23 crc kubenswrapper[4859]: I1008 19:19:23.229194 4859 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_watcher-operator-controller-manager-7794bc6bd-2v65r_cf5b6741-336e-435c-8209-82b8e19f4896/kube-rbac-proxy/0.log" Oct 08 19:19:23 crc kubenswrapper[4859]: I1008 19:19:23.232534 4859 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_watcher-operator-controller-manager-7794bc6bd-2v65r_cf5b6741-336e-435c-8209-82b8e19f4896/manager/0.log" Oct 08 19:19:38 crc kubenswrapper[4859]: I1008 19:19:38.320102 4859 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-machine-api_control-plane-machine-set-operator-78cbb6b69f-q9nqj_492ef569-6d45-4dd0-89e5-506172689768/control-plane-machine-set-operator/0.log" Oct 08 19:19:38 crc kubenswrapper[4859]: I1008 19:19:38.478282 4859 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-machine-api_machine-api-operator-5694c8668f-gmbk9_a52b9536-1013-482b-9261-e024377687d2/kube-rbac-proxy/0.log" Oct 08 19:19:38 crc kubenswrapper[4859]: I1008 19:19:38.549242 4859 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-machine-api_machine-api-operator-5694c8668f-gmbk9_a52b9536-1013-482b-9261-e024377687d2/machine-api-operator/0.log" Oct 08 19:19:51 crc kubenswrapper[4859]: I1008 19:19:51.746315 4859 log.go:25] "Finished parsing log file" path="/var/log/pods/cert-manager_cert-manager-5b446d88c5-w95bk_2fa6fe28-fd3a-4611-ad50-2a4fc67ca563/cert-manager-controller/0.log" Oct 08 19:19:51 crc kubenswrapper[4859]: I1008 19:19:51.884765 4859 log.go:25] "Finished parsing log file" path="/var/log/pods/cert-manager_cert-manager-cainjector-7f985d654d-lqldn_ccaaab95-a4e1-43b8-a962-0b833beec038/cert-manager-cainjector/0.log" Oct 08 19:19:51 crc kubenswrapper[4859]: I1008 19:19:51.942530 4859 log.go:25] "Finished parsing log file" path="/var/log/pods/cert-manager_cert-manager-webhook-5655c58dd6-jkfxx_24b1f415-e420-420a-b6c1-7fde05b080b6/cert-manager-webhook/0.log" Oct 08 19:20:05 crc kubenswrapper[4859]: I1008 19:20:05.415349 4859 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-nmstate_nmstate-console-plugin-6b874cbd85-r5l52_a8be1e40-aedc-40bd-b47d-e381b5746080/nmstate-console-plugin/0.log" Oct 08 19:20:05 crc kubenswrapper[4859]: I1008 19:20:05.604578 4859 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-nmstate_nmstate-handler-htrgb_692be5e4-be66-4010-9547-610a2d3d85fd/nmstate-handler/0.log" Oct 08 19:20:06 crc kubenswrapper[4859]: I1008 19:20:06.214026 4859 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-nmstate_nmstate-metrics-fdff9cb8d-8vdx7_0beb8934-6616-4771-b93d-5b833f87600d/kube-rbac-proxy/0.log" Oct 08 19:20:06 crc kubenswrapper[4859]: I1008 19:20:06.260071 4859 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-nmstate_nmstate-operator-858ddd8f98-vbswp_3a114f4c-72fa-4640-86e4-cb4326069481/nmstate-operator/0.log" Oct 08 19:20:06 crc kubenswrapper[4859]: I1008 19:20:06.269443 4859 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-nmstate_nmstate-metrics-fdff9cb8d-8vdx7_0beb8934-6616-4771-b93d-5b833f87600d/nmstate-metrics/0.log" Oct 08 19:20:06 crc kubenswrapper[4859]: I1008 19:20:06.439782 4859 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-nmstate_nmstate-webhook-6cdbc54649-728jz_5af71d73-77d0-409b-8b48-cca9164c529e/nmstate-webhook/0.log" Oct 08 19:20:17 crc kubenswrapper[4859]: I1008 19:20:17.924644 4859 patch_prober.go:28] interesting pod/machine-config-daemon-82s52 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 08 19:20:17 crc kubenswrapper[4859]: I1008 19:20:17.925148 4859 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-82s52" podUID="b23a6a6c-9d92-4e7b-840e-55cfda873a2d" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 08 19:20:19 crc kubenswrapper[4859]: I1008 19:20:19.750567 4859 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_controller-68d546b9d8-cr5dk_7f1dbe0e-c03a-4ec2-ba3c-7179d2c67e41/kube-rbac-proxy/0.log" Oct 08 19:20:19 crc kubenswrapper[4859]: I1008 19:20:19.804032 4859 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_controller-68d546b9d8-cr5dk_7f1dbe0e-c03a-4ec2-ba3c-7179d2c67e41/controller/0.log" Oct 08 19:20:19 crc kubenswrapper[4859]: I1008 19:20:19.962547 4859 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-mfcd8_69cc95cc-1a58-40c2-9c84-4808c217c2e5/cp-frr-files/0.log" Oct 08 19:20:20 crc kubenswrapper[4859]: I1008 19:20:20.096274 4859 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-mfcd8_69cc95cc-1a58-40c2-9c84-4808c217c2e5/cp-frr-files/0.log" Oct 08 19:20:20 crc kubenswrapper[4859]: I1008 19:20:20.141377 4859 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-mfcd8_69cc95cc-1a58-40c2-9c84-4808c217c2e5/cp-reloader/0.log" Oct 08 19:20:20 crc kubenswrapper[4859]: I1008 19:20:20.179671 4859 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-mfcd8_69cc95cc-1a58-40c2-9c84-4808c217c2e5/cp-metrics/0.log" Oct 08 19:20:20 crc kubenswrapper[4859]: I1008 19:20:20.194932 4859 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-mfcd8_69cc95cc-1a58-40c2-9c84-4808c217c2e5/cp-reloader/0.log" Oct 08 19:20:20 crc kubenswrapper[4859]: I1008 19:20:20.302241 4859 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-mfcd8_69cc95cc-1a58-40c2-9c84-4808c217c2e5/cp-frr-files/0.log" Oct 08 19:20:20 crc kubenswrapper[4859]: I1008 19:20:20.340583 4859 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-mfcd8_69cc95cc-1a58-40c2-9c84-4808c217c2e5/cp-metrics/0.log" Oct 08 19:20:20 crc kubenswrapper[4859]: I1008 19:20:20.363737 4859 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-mfcd8_69cc95cc-1a58-40c2-9c84-4808c217c2e5/cp-metrics/0.log" Oct 08 19:20:20 crc kubenswrapper[4859]: I1008 19:20:20.380754 4859 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-mfcd8_69cc95cc-1a58-40c2-9c84-4808c217c2e5/cp-reloader/0.log" Oct 08 19:20:20 crc kubenswrapper[4859]: I1008 19:20:20.574530 4859 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-mfcd8_69cc95cc-1a58-40c2-9c84-4808c217c2e5/cp-reloader/0.log" Oct 08 19:20:20 crc kubenswrapper[4859]: I1008 19:20:20.586403 4859 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-mfcd8_69cc95cc-1a58-40c2-9c84-4808c217c2e5/cp-frr-files/0.log" Oct 08 19:20:20 crc kubenswrapper[4859]: I1008 19:20:20.586638 4859 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-mfcd8_69cc95cc-1a58-40c2-9c84-4808c217c2e5/cp-metrics/0.log" Oct 08 19:20:20 crc kubenswrapper[4859]: I1008 19:20:20.631568 4859 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-mfcd8_69cc95cc-1a58-40c2-9c84-4808c217c2e5/controller/0.log" Oct 08 19:20:20 crc kubenswrapper[4859]: I1008 19:20:20.738310 4859 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-mfcd8_69cc95cc-1a58-40c2-9c84-4808c217c2e5/frr-metrics/0.log" Oct 08 19:20:20 crc kubenswrapper[4859]: I1008 19:20:20.788194 4859 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-mfcd8_69cc95cc-1a58-40c2-9c84-4808c217c2e5/kube-rbac-proxy/0.log" Oct 08 19:20:20 crc kubenswrapper[4859]: I1008 19:20:20.794910 4859 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-mfcd8_69cc95cc-1a58-40c2-9c84-4808c217c2e5/kube-rbac-proxy-frr/0.log" Oct 08 19:20:20 crc kubenswrapper[4859]: I1008 19:20:20.936069 4859 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-mfcd8_69cc95cc-1a58-40c2-9c84-4808c217c2e5/reloader/0.log" Oct 08 19:20:20 crc kubenswrapper[4859]: I1008 19:20:20.988946 4859 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-webhook-server-64bf5d555-77xwq_4d682ed0-e5d8-48ad-8e52-cd29b2f65bd2/frr-k8s-webhook-server/0.log" Oct 08 19:20:21 crc kubenswrapper[4859]: I1008 19:20:21.192959 4859 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_metallb-operator-controller-manager-6789c5bc7c-k7fl4_47c9b7b8-561d-489b-a5d2-34ef1f6f4be3/manager/0.log" Oct 08 19:20:21 crc kubenswrapper[4859]: I1008 19:20:21.403105 4859 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_metallb-operator-webhook-server-c4d794b9c-6lnpl_c5192e5b-dc81-4df1-8ba2-f103a6b869eb/webhook-server/0.log" Oct 08 19:20:21 crc kubenswrapper[4859]: I1008 19:20:21.476570 4859 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_speaker-47524_7557fe67-9517-441e-944c-9d5286f5735c/kube-rbac-proxy/0.log" Oct 08 19:20:22 crc kubenswrapper[4859]: I1008 19:20:22.075884 4859 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_speaker-47524_7557fe67-9517-441e-944c-9d5286f5735c/speaker/0.log" Oct 08 19:20:22 crc kubenswrapper[4859]: I1008 19:20:22.218488 4859 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-mfcd8_69cc95cc-1a58-40c2-9c84-4808c217c2e5/frr/0.log" Oct 08 19:20:34 crc kubenswrapper[4859]: I1008 19:20:34.001477 4859 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_8f2f4ee801e5826a37d84a7b1fc4ccbf6b79de668302737d0f1152d8d2lfzpq_a200c80e-a0f7-40ee-b697-acf6d429d14f/util/0.log" Oct 08 19:20:34 crc kubenswrapper[4859]: I1008 19:20:34.170072 4859 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_8f2f4ee801e5826a37d84a7b1fc4ccbf6b79de668302737d0f1152d8d2lfzpq_a200c80e-a0f7-40ee-b697-acf6d429d14f/util/0.log" Oct 08 19:20:34 crc kubenswrapper[4859]: I1008 19:20:34.197266 4859 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_8f2f4ee801e5826a37d84a7b1fc4ccbf6b79de668302737d0f1152d8d2lfzpq_a200c80e-a0f7-40ee-b697-acf6d429d14f/pull/0.log" Oct 08 19:20:34 crc kubenswrapper[4859]: I1008 19:20:34.201192 4859 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_8f2f4ee801e5826a37d84a7b1fc4ccbf6b79de668302737d0f1152d8d2lfzpq_a200c80e-a0f7-40ee-b697-acf6d429d14f/pull/0.log" Oct 08 19:20:34 crc kubenswrapper[4859]: I1008 19:20:34.408520 4859 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_8f2f4ee801e5826a37d84a7b1fc4ccbf6b79de668302737d0f1152d8d2lfzpq_a200c80e-a0f7-40ee-b697-acf6d429d14f/util/0.log" Oct 08 19:20:34 crc kubenswrapper[4859]: I1008 19:20:34.432436 4859 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_8f2f4ee801e5826a37d84a7b1fc4ccbf6b79de668302737d0f1152d8d2lfzpq_a200c80e-a0f7-40ee-b697-acf6d429d14f/pull/0.log" Oct 08 19:20:34 crc kubenswrapper[4859]: I1008 19:20:34.433076 4859 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_8f2f4ee801e5826a37d84a7b1fc4ccbf6b79de668302737d0f1152d8d2lfzpq_a200c80e-a0f7-40ee-b697-acf6d429d14f/extract/0.log" Oct 08 19:20:34 crc kubenswrapper[4859]: I1008 19:20:34.585226 4859 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-d8bkm_6e7e5636-231d-4788-9277-83549fc6b404/extract-utilities/0.log" Oct 08 19:20:34 crc kubenswrapper[4859]: I1008 19:20:34.791829 4859 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-d8bkm_6e7e5636-231d-4788-9277-83549fc6b404/extract-utilities/0.log" Oct 08 19:20:34 crc kubenswrapper[4859]: I1008 19:20:34.816617 4859 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-d8bkm_6e7e5636-231d-4788-9277-83549fc6b404/extract-content/0.log" Oct 08 19:20:34 crc kubenswrapper[4859]: I1008 19:20:34.841220 4859 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-d8bkm_6e7e5636-231d-4788-9277-83549fc6b404/extract-content/0.log" Oct 08 19:20:35 crc kubenswrapper[4859]: I1008 19:20:35.189200 4859 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-d8bkm_6e7e5636-231d-4788-9277-83549fc6b404/extract-content/0.log" Oct 08 19:20:35 crc kubenswrapper[4859]: I1008 19:20:35.234227 4859 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-d8bkm_6e7e5636-231d-4788-9277-83549fc6b404/extract-utilities/0.log" Oct 08 19:20:35 crc kubenswrapper[4859]: I1008 19:20:35.431562 4859 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-p7bgm_ad354068-21f9-40aa-a2ae-739f84e32469/extract-utilities/0.log" Oct 08 19:20:35 crc kubenswrapper[4859]: I1008 19:20:35.670098 4859 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-p7bgm_ad354068-21f9-40aa-a2ae-739f84e32469/extract-content/0.log" Oct 08 19:20:35 crc kubenswrapper[4859]: I1008 19:20:35.678383 4859 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-d8bkm_6e7e5636-231d-4788-9277-83549fc6b404/registry-server/0.log" Oct 08 19:20:35 crc kubenswrapper[4859]: I1008 19:20:35.701456 4859 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-p7bgm_ad354068-21f9-40aa-a2ae-739f84e32469/extract-utilities/0.log" Oct 08 19:20:35 crc kubenswrapper[4859]: I1008 19:20:35.714362 4859 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-p7bgm_ad354068-21f9-40aa-a2ae-739f84e32469/extract-content/0.log" Oct 08 19:20:35 crc kubenswrapper[4859]: I1008 19:20:35.917125 4859 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-p7bgm_ad354068-21f9-40aa-a2ae-739f84e32469/extract-content/0.log" Oct 08 19:20:35 crc kubenswrapper[4859]: I1008 19:20:35.934014 4859 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-p7bgm_ad354068-21f9-40aa-a2ae-739f84e32469/extract-utilities/0.log" Oct 08 19:20:36 crc kubenswrapper[4859]: I1008 19:20:36.168084 4859 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_fa9831ede5d93c33d525b70ce6ddf94e500d80992af75a3305fe98835cdxmbx_cc21f1af-ca05-4cf4-b66f-f7fcde5026ee/util/0.log" Oct 08 19:20:36 crc kubenswrapper[4859]: I1008 19:20:36.413966 4859 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_fa9831ede5d93c33d525b70ce6ddf94e500d80992af75a3305fe98835cdxmbx_cc21f1af-ca05-4cf4-b66f-f7fcde5026ee/pull/0.log" Oct 08 19:20:36 crc kubenswrapper[4859]: I1008 19:20:36.432872 4859 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_fa9831ede5d93c33d525b70ce6ddf94e500d80992af75a3305fe98835cdxmbx_cc21f1af-ca05-4cf4-b66f-f7fcde5026ee/util/0.log" Oct 08 19:20:36 crc kubenswrapper[4859]: I1008 19:20:36.496734 4859 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_fa9831ede5d93c33d525b70ce6ddf94e500d80992af75a3305fe98835cdxmbx_cc21f1af-ca05-4cf4-b66f-f7fcde5026ee/pull/0.log" Oct 08 19:20:36 crc kubenswrapper[4859]: I1008 19:20:36.525397 4859 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-p7bgm_ad354068-21f9-40aa-a2ae-739f84e32469/registry-server/0.log" Oct 08 19:20:36 crc kubenswrapper[4859]: I1008 19:20:36.621357 4859 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_fa9831ede5d93c33d525b70ce6ddf94e500d80992af75a3305fe98835cdxmbx_cc21f1af-ca05-4cf4-b66f-f7fcde5026ee/pull/0.log" Oct 08 19:20:36 crc kubenswrapper[4859]: I1008 19:20:36.667056 4859 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_fa9831ede5d93c33d525b70ce6ddf94e500d80992af75a3305fe98835cdxmbx_cc21f1af-ca05-4cf4-b66f-f7fcde5026ee/util/0.log" Oct 08 19:20:36 crc kubenswrapper[4859]: I1008 19:20:36.709847 4859 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_fa9831ede5d93c33d525b70ce6ddf94e500d80992af75a3305fe98835cdxmbx_cc21f1af-ca05-4cf4-b66f-f7fcde5026ee/extract/0.log" Oct 08 19:20:36 crc kubenswrapper[4859]: I1008 19:20:36.820650 4859 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_marketplace-operator-79b997595-h4tnd_8fdf9d2f-74e1-46e7-89ed-1947c3104386/marketplace-operator/0.log" Oct 08 19:20:36 crc kubenswrapper[4859]: I1008 19:20:36.917720 4859 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-wlwhh_997d41aa-a006-46d1-b488-b6d29186874d/extract-utilities/0.log" Oct 08 19:20:37 crc kubenswrapper[4859]: I1008 19:20:37.117620 4859 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-wlwhh_997d41aa-a006-46d1-b488-b6d29186874d/extract-utilities/0.log" Oct 08 19:20:37 crc kubenswrapper[4859]: I1008 19:20:37.139420 4859 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-wlwhh_997d41aa-a006-46d1-b488-b6d29186874d/extract-content/0.log" Oct 08 19:20:37 crc kubenswrapper[4859]: I1008 19:20:37.155571 4859 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-wlwhh_997d41aa-a006-46d1-b488-b6d29186874d/extract-content/0.log" Oct 08 19:20:37 crc kubenswrapper[4859]: I1008 19:20:37.321573 4859 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-wlwhh_997d41aa-a006-46d1-b488-b6d29186874d/extract-content/0.log" Oct 08 19:20:37 crc kubenswrapper[4859]: I1008 19:20:37.335926 4859 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-wlwhh_997d41aa-a006-46d1-b488-b6d29186874d/extract-utilities/0.log" Oct 08 19:20:37 crc kubenswrapper[4859]: I1008 19:20:37.512137 4859 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-wlwhh_997d41aa-a006-46d1-b488-b6d29186874d/registry-server/0.log" Oct 08 19:20:37 crc kubenswrapper[4859]: I1008 19:20:37.569288 4859 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-hlqjs_748a4b84-e115-444a-b0f8-24029e8d1896/extract-utilities/0.log" Oct 08 19:20:37 crc kubenswrapper[4859]: I1008 19:20:37.733425 4859 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-hlqjs_748a4b84-e115-444a-b0f8-24029e8d1896/extract-utilities/0.log" Oct 08 19:20:37 crc kubenswrapper[4859]: I1008 19:20:37.749324 4859 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-hlqjs_748a4b84-e115-444a-b0f8-24029e8d1896/extract-content/0.log" Oct 08 19:20:37 crc kubenswrapper[4859]: I1008 19:20:37.773408 4859 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-hlqjs_748a4b84-e115-444a-b0f8-24029e8d1896/extract-content/0.log" Oct 08 19:20:37 crc kubenswrapper[4859]: I1008 19:20:37.933921 4859 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-hlqjs_748a4b84-e115-444a-b0f8-24029e8d1896/extract-utilities/0.log" Oct 08 19:20:37 crc kubenswrapper[4859]: I1008 19:20:37.984511 4859 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-hlqjs_748a4b84-e115-444a-b0f8-24029e8d1896/extract-content/0.log" Oct 08 19:20:38 crc kubenswrapper[4859]: I1008 19:20:38.449489 4859 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-hlqjs_748a4b84-e115-444a-b0f8-24029e8d1896/registry-server/0.log" Oct 08 19:20:47 crc kubenswrapper[4859]: I1008 19:20:47.925346 4859 patch_prober.go:28] interesting pod/machine-config-daemon-82s52 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 08 19:20:47 crc kubenswrapper[4859]: I1008 19:20:47.926146 4859 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-82s52" podUID="b23a6a6c-9d92-4e7b-840e-55cfda873a2d" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 08 19:21:17 crc kubenswrapper[4859]: I1008 19:21:17.925089 4859 patch_prober.go:28] interesting pod/machine-config-daemon-82s52 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 08 19:21:17 crc kubenswrapper[4859]: I1008 19:21:17.925672 4859 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-82s52" podUID="b23a6a6c-9d92-4e7b-840e-55cfda873a2d" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 08 19:21:17 crc kubenswrapper[4859]: I1008 19:21:17.925770 4859 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-82s52" Oct 08 19:21:17 crc kubenswrapper[4859]: I1008 19:21:17.926866 4859 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"de71f0b717129dba48f3f0bdd90240a8e45404a17f0df79d6812d58c9209feff"} pod="openshift-machine-config-operator/machine-config-daemon-82s52" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Oct 08 19:21:17 crc kubenswrapper[4859]: I1008 19:21:17.926970 4859 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-82s52" podUID="b23a6a6c-9d92-4e7b-840e-55cfda873a2d" containerName="machine-config-daemon" containerID="cri-o://de71f0b717129dba48f3f0bdd90240a8e45404a17f0df79d6812d58c9209feff" gracePeriod=600 Oct 08 19:21:18 crc kubenswrapper[4859]: E1008 19:21:18.079923 4859 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-82s52_openshift-machine-config-operator(b23a6a6c-9d92-4e7b-840e-55cfda873a2d)\"" pod="openshift-machine-config-operator/machine-config-daemon-82s52" podUID="b23a6a6c-9d92-4e7b-840e-55cfda873a2d" Oct 08 19:21:18 crc kubenswrapper[4859]: I1008 19:21:18.155245 4859 generic.go:334] "Generic (PLEG): container finished" podID="b23a6a6c-9d92-4e7b-840e-55cfda873a2d" containerID="de71f0b717129dba48f3f0bdd90240a8e45404a17f0df79d6812d58c9209feff" exitCode=0 Oct 08 19:21:18 crc kubenswrapper[4859]: I1008 19:21:18.155311 4859 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-82s52" event={"ID":"b23a6a6c-9d92-4e7b-840e-55cfda873a2d","Type":"ContainerDied","Data":"de71f0b717129dba48f3f0bdd90240a8e45404a17f0df79d6812d58c9209feff"} Oct 08 19:21:18 crc kubenswrapper[4859]: I1008 19:21:18.155372 4859 scope.go:117] "RemoveContainer" containerID="09fd446535f97e643340dafbb38e492a2f1d2902d0d749edf84136fd6fabb2ac" Oct 08 19:21:18 crc kubenswrapper[4859]: I1008 19:21:18.156086 4859 scope.go:117] "RemoveContainer" containerID="de71f0b717129dba48f3f0bdd90240a8e45404a17f0df79d6812d58c9209feff" Oct 08 19:21:18 crc kubenswrapper[4859]: E1008 19:21:18.156366 4859 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-82s52_openshift-machine-config-operator(b23a6a6c-9d92-4e7b-840e-55cfda873a2d)\"" pod="openshift-machine-config-operator/machine-config-daemon-82s52" podUID="b23a6a6c-9d92-4e7b-840e-55cfda873a2d" Oct 08 19:21:30 crc kubenswrapper[4859]: I1008 19:21:30.476181 4859 scope.go:117] "RemoveContainer" containerID="de71f0b717129dba48f3f0bdd90240a8e45404a17f0df79d6812d58c9209feff" Oct 08 19:21:30 crc kubenswrapper[4859]: E1008 19:21:30.477154 4859 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-82s52_openshift-machine-config-operator(b23a6a6c-9d92-4e7b-840e-55cfda873a2d)\"" pod="openshift-machine-config-operator/machine-config-daemon-82s52" podUID="b23a6a6c-9d92-4e7b-840e-55cfda873a2d" Oct 08 19:21:45 crc kubenswrapper[4859]: I1008 19:21:45.469939 4859 scope.go:117] "RemoveContainer" containerID="de71f0b717129dba48f3f0bdd90240a8e45404a17f0df79d6812d58c9209feff" Oct 08 19:21:45 crc kubenswrapper[4859]: E1008 19:21:45.470945 4859 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-82s52_openshift-machine-config-operator(b23a6a6c-9d92-4e7b-840e-55cfda873a2d)\"" pod="openshift-machine-config-operator/machine-config-daemon-82s52" podUID="b23a6a6c-9d92-4e7b-840e-55cfda873a2d" Oct 08 19:21:58 crc kubenswrapper[4859]: I1008 19:21:58.470325 4859 scope.go:117] "RemoveContainer" containerID="de71f0b717129dba48f3f0bdd90240a8e45404a17f0df79d6812d58c9209feff" Oct 08 19:21:58 crc kubenswrapper[4859]: E1008 19:21:58.471249 4859 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-82s52_openshift-machine-config-operator(b23a6a6c-9d92-4e7b-840e-55cfda873a2d)\"" pod="openshift-machine-config-operator/machine-config-daemon-82s52" podUID="b23a6a6c-9d92-4e7b-840e-55cfda873a2d" Oct 08 19:22:09 crc kubenswrapper[4859]: I1008 19:22:09.473435 4859 scope.go:117] "RemoveContainer" containerID="de71f0b717129dba48f3f0bdd90240a8e45404a17f0df79d6812d58c9209feff" Oct 08 19:22:09 crc kubenswrapper[4859]: E1008 19:22:09.476162 4859 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-82s52_openshift-machine-config-operator(b23a6a6c-9d92-4e7b-840e-55cfda873a2d)\"" pod="openshift-machine-config-operator/machine-config-daemon-82s52" podUID="b23a6a6c-9d92-4e7b-840e-55cfda873a2d" Oct 08 19:22:23 crc kubenswrapper[4859]: I1008 19:22:23.471247 4859 scope.go:117] "RemoveContainer" containerID="de71f0b717129dba48f3f0bdd90240a8e45404a17f0df79d6812d58c9209feff" Oct 08 19:22:23 crc kubenswrapper[4859]: E1008 19:22:23.472039 4859 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-82s52_openshift-machine-config-operator(b23a6a6c-9d92-4e7b-840e-55cfda873a2d)\"" pod="openshift-machine-config-operator/machine-config-daemon-82s52" podUID="b23a6a6c-9d92-4e7b-840e-55cfda873a2d" Oct 08 19:22:33 crc kubenswrapper[4859]: I1008 19:22:33.009914 4859 generic.go:334] "Generic (PLEG): container finished" podID="4dea6b3c-d8e9-4086-bfd0-46970f43f282" containerID="9caf66c4d88250d731a03fcb5e38db07b33c7bf6f56030210c181852d886f771" exitCode=0 Oct 08 19:22:33 crc kubenswrapper[4859]: I1008 19:22:33.010022 4859 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-f7r6n/must-gather-9cn9k" event={"ID":"4dea6b3c-d8e9-4086-bfd0-46970f43f282","Type":"ContainerDied","Data":"9caf66c4d88250d731a03fcb5e38db07b33c7bf6f56030210c181852d886f771"} Oct 08 19:22:33 crc kubenswrapper[4859]: I1008 19:22:33.012030 4859 scope.go:117] "RemoveContainer" containerID="9caf66c4d88250d731a03fcb5e38db07b33c7bf6f56030210c181852d886f771" Oct 08 19:22:33 crc kubenswrapper[4859]: I1008 19:22:33.326592 4859 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-must-gather-f7r6n_must-gather-9cn9k_4dea6b3c-d8e9-4086-bfd0-46970f43f282/gather/0.log" Oct 08 19:22:38 crc kubenswrapper[4859]: I1008 19:22:38.471135 4859 scope.go:117] "RemoveContainer" containerID="de71f0b717129dba48f3f0bdd90240a8e45404a17f0df79d6812d58c9209feff" Oct 08 19:22:38 crc kubenswrapper[4859]: E1008 19:22:38.475276 4859 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-82s52_openshift-machine-config-operator(b23a6a6c-9d92-4e7b-840e-55cfda873a2d)\"" pod="openshift-machine-config-operator/machine-config-daemon-82s52" podUID="b23a6a6c-9d92-4e7b-840e-55cfda873a2d" Oct 08 19:22:41 crc kubenswrapper[4859]: I1008 19:22:41.193601 4859 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-must-gather-f7r6n/must-gather-9cn9k"] Oct 08 19:22:41 crc kubenswrapper[4859]: I1008 19:22:41.194600 4859 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-must-gather-f7r6n/must-gather-9cn9k" podUID="4dea6b3c-d8e9-4086-bfd0-46970f43f282" containerName="copy" containerID="cri-o://add7a3369622385591a2aacdd64391cb595efae300e79eadd5957a4f4eefd41a" gracePeriod=2 Oct 08 19:22:41 crc kubenswrapper[4859]: I1008 19:22:41.204149 4859 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-must-gather-f7r6n/must-gather-9cn9k"] Oct 08 19:22:41 crc kubenswrapper[4859]: I1008 19:22:41.685186 4859 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-must-gather-f7r6n_must-gather-9cn9k_4dea6b3c-d8e9-4086-bfd0-46970f43f282/copy/0.log" Oct 08 19:22:41 crc kubenswrapper[4859]: I1008 19:22:41.686053 4859 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-f7r6n/must-gather-9cn9k" Oct 08 19:22:41 crc kubenswrapper[4859]: I1008 19:22:41.857601 4859 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-zq5mf\" (UniqueName: \"kubernetes.io/projected/4dea6b3c-d8e9-4086-bfd0-46970f43f282-kube-api-access-zq5mf\") pod \"4dea6b3c-d8e9-4086-bfd0-46970f43f282\" (UID: \"4dea6b3c-d8e9-4086-bfd0-46970f43f282\") " Oct 08 19:22:41 crc kubenswrapper[4859]: I1008 19:22:41.857790 4859 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"must-gather-output\" (UniqueName: \"kubernetes.io/empty-dir/4dea6b3c-d8e9-4086-bfd0-46970f43f282-must-gather-output\") pod \"4dea6b3c-d8e9-4086-bfd0-46970f43f282\" (UID: \"4dea6b3c-d8e9-4086-bfd0-46970f43f282\") " Oct 08 19:22:41 crc kubenswrapper[4859]: I1008 19:22:41.866112 4859 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/4dea6b3c-d8e9-4086-bfd0-46970f43f282-kube-api-access-zq5mf" (OuterVolumeSpecName: "kube-api-access-zq5mf") pod "4dea6b3c-d8e9-4086-bfd0-46970f43f282" (UID: "4dea6b3c-d8e9-4086-bfd0-46970f43f282"). InnerVolumeSpecName "kube-api-access-zq5mf". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 08 19:22:41 crc kubenswrapper[4859]: I1008 19:22:41.960948 4859 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-zq5mf\" (UniqueName: \"kubernetes.io/projected/4dea6b3c-d8e9-4086-bfd0-46970f43f282-kube-api-access-zq5mf\") on node \"crc\" DevicePath \"\"" Oct 08 19:22:42 crc kubenswrapper[4859]: I1008 19:22:42.004170 4859 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/4dea6b3c-d8e9-4086-bfd0-46970f43f282-must-gather-output" (OuterVolumeSpecName: "must-gather-output") pod "4dea6b3c-d8e9-4086-bfd0-46970f43f282" (UID: "4dea6b3c-d8e9-4086-bfd0-46970f43f282"). InnerVolumeSpecName "must-gather-output". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 08 19:22:42 crc kubenswrapper[4859]: I1008 19:22:42.063396 4859 reconciler_common.go:293] "Volume detached for volume \"must-gather-output\" (UniqueName: \"kubernetes.io/empty-dir/4dea6b3c-d8e9-4086-bfd0-46970f43f282-must-gather-output\") on node \"crc\" DevicePath \"\"" Oct 08 19:22:42 crc kubenswrapper[4859]: I1008 19:22:42.105854 4859 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-must-gather-f7r6n_must-gather-9cn9k_4dea6b3c-d8e9-4086-bfd0-46970f43f282/copy/0.log" Oct 08 19:22:42 crc kubenswrapper[4859]: I1008 19:22:42.106257 4859 generic.go:334] "Generic (PLEG): container finished" podID="4dea6b3c-d8e9-4086-bfd0-46970f43f282" containerID="add7a3369622385591a2aacdd64391cb595efae300e79eadd5957a4f4eefd41a" exitCode=143 Oct 08 19:22:42 crc kubenswrapper[4859]: I1008 19:22:42.106314 4859 scope.go:117] "RemoveContainer" containerID="add7a3369622385591a2aacdd64391cb595efae300e79eadd5957a4f4eefd41a" Oct 08 19:22:42 crc kubenswrapper[4859]: I1008 19:22:42.106496 4859 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-f7r6n/must-gather-9cn9k" Oct 08 19:22:42 crc kubenswrapper[4859]: I1008 19:22:42.139377 4859 scope.go:117] "RemoveContainer" containerID="9caf66c4d88250d731a03fcb5e38db07b33c7bf6f56030210c181852d886f771" Oct 08 19:22:42 crc kubenswrapper[4859]: I1008 19:22:42.200147 4859 scope.go:117] "RemoveContainer" containerID="add7a3369622385591a2aacdd64391cb595efae300e79eadd5957a4f4eefd41a" Oct 08 19:22:42 crc kubenswrapper[4859]: E1008 19:22:42.200608 4859 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"add7a3369622385591a2aacdd64391cb595efae300e79eadd5957a4f4eefd41a\": container with ID starting with add7a3369622385591a2aacdd64391cb595efae300e79eadd5957a4f4eefd41a not found: ID does not exist" containerID="add7a3369622385591a2aacdd64391cb595efae300e79eadd5957a4f4eefd41a" Oct 08 19:22:42 crc kubenswrapper[4859]: I1008 19:22:42.200843 4859 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"add7a3369622385591a2aacdd64391cb595efae300e79eadd5957a4f4eefd41a"} err="failed to get container status \"add7a3369622385591a2aacdd64391cb595efae300e79eadd5957a4f4eefd41a\": rpc error: code = NotFound desc = could not find container \"add7a3369622385591a2aacdd64391cb595efae300e79eadd5957a4f4eefd41a\": container with ID starting with add7a3369622385591a2aacdd64391cb595efae300e79eadd5957a4f4eefd41a not found: ID does not exist" Oct 08 19:22:42 crc kubenswrapper[4859]: I1008 19:22:42.200870 4859 scope.go:117] "RemoveContainer" containerID="9caf66c4d88250d731a03fcb5e38db07b33c7bf6f56030210c181852d886f771" Oct 08 19:22:42 crc kubenswrapper[4859]: E1008 19:22:42.201132 4859 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"9caf66c4d88250d731a03fcb5e38db07b33c7bf6f56030210c181852d886f771\": container with ID starting with 9caf66c4d88250d731a03fcb5e38db07b33c7bf6f56030210c181852d886f771 not found: ID does not exist" containerID="9caf66c4d88250d731a03fcb5e38db07b33c7bf6f56030210c181852d886f771" Oct 08 19:22:42 crc kubenswrapper[4859]: I1008 19:22:42.201174 4859 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"9caf66c4d88250d731a03fcb5e38db07b33c7bf6f56030210c181852d886f771"} err="failed to get container status \"9caf66c4d88250d731a03fcb5e38db07b33c7bf6f56030210c181852d886f771\": rpc error: code = NotFound desc = could not find container \"9caf66c4d88250d731a03fcb5e38db07b33c7bf6f56030210c181852d886f771\": container with ID starting with 9caf66c4d88250d731a03fcb5e38db07b33c7bf6f56030210c181852d886f771 not found: ID does not exist" Oct 08 19:22:42 crc kubenswrapper[4859]: I1008 19:22:42.482542 4859 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="4dea6b3c-d8e9-4086-bfd0-46970f43f282" path="/var/lib/kubelet/pods/4dea6b3c-d8e9-4086-bfd0-46970f43f282/volumes" Oct 08 19:22:49 crc kubenswrapper[4859]: I1008 19:22:49.470660 4859 scope.go:117] "RemoveContainer" containerID="de71f0b717129dba48f3f0bdd90240a8e45404a17f0df79d6812d58c9209feff" Oct 08 19:22:49 crc kubenswrapper[4859]: E1008 19:22:49.471570 4859 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-82s52_openshift-machine-config-operator(b23a6a6c-9d92-4e7b-840e-55cfda873a2d)\"" pod="openshift-machine-config-operator/machine-config-daemon-82s52" podUID="b23a6a6c-9d92-4e7b-840e-55cfda873a2d" Oct 08 19:22:53 crc kubenswrapper[4859]: I1008 19:22:53.722349 4859 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-zltrt"] Oct 08 19:22:53 crc kubenswrapper[4859]: E1008 19:22:53.723393 4859 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="4dea6b3c-d8e9-4086-bfd0-46970f43f282" containerName="copy" Oct 08 19:22:53 crc kubenswrapper[4859]: I1008 19:22:53.723412 4859 state_mem.go:107] "Deleted CPUSet assignment" podUID="4dea6b3c-d8e9-4086-bfd0-46970f43f282" containerName="copy" Oct 08 19:22:53 crc kubenswrapper[4859]: E1008 19:22:53.723439 4859 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b6ec1f7a-65b5-4979-8f2c-3b77d03bb399" containerName="container-00" Oct 08 19:22:53 crc kubenswrapper[4859]: I1008 19:22:53.723449 4859 state_mem.go:107] "Deleted CPUSet assignment" podUID="b6ec1f7a-65b5-4979-8f2c-3b77d03bb399" containerName="container-00" Oct 08 19:22:53 crc kubenswrapper[4859]: E1008 19:22:53.723474 4859 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="4dea6b3c-d8e9-4086-bfd0-46970f43f282" containerName="gather" Oct 08 19:22:53 crc kubenswrapper[4859]: I1008 19:22:53.723484 4859 state_mem.go:107] "Deleted CPUSet assignment" podUID="4dea6b3c-d8e9-4086-bfd0-46970f43f282" containerName="gather" Oct 08 19:22:53 crc kubenswrapper[4859]: I1008 19:22:53.723754 4859 memory_manager.go:354] "RemoveStaleState removing state" podUID="4dea6b3c-d8e9-4086-bfd0-46970f43f282" containerName="gather" Oct 08 19:22:53 crc kubenswrapper[4859]: I1008 19:22:53.723775 4859 memory_manager.go:354] "RemoveStaleState removing state" podUID="b6ec1f7a-65b5-4979-8f2c-3b77d03bb399" containerName="container-00" Oct 08 19:22:53 crc kubenswrapper[4859]: I1008 19:22:53.723788 4859 memory_manager.go:354] "RemoveStaleState removing state" podUID="4dea6b3c-d8e9-4086-bfd0-46970f43f282" containerName="copy" Oct 08 19:22:53 crc kubenswrapper[4859]: I1008 19:22:53.725677 4859 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-zltrt" Oct 08 19:22:53 crc kubenswrapper[4859]: I1008 19:22:53.734634 4859 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-zltrt"] Oct 08 19:22:53 crc kubenswrapper[4859]: I1008 19:22:53.895937 4859 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/03b56ca9-6072-4f44-b144-2e8654fc26af-catalog-content\") pod \"certified-operators-zltrt\" (UID: \"03b56ca9-6072-4f44-b144-2e8654fc26af\") " pod="openshift-marketplace/certified-operators-zltrt" Oct 08 19:22:53 crc kubenswrapper[4859]: I1008 19:22:53.896068 4859 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-pbzqj\" (UniqueName: \"kubernetes.io/projected/03b56ca9-6072-4f44-b144-2e8654fc26af-kube-api-access-pbzqj\") pod \"certified-operators-zltrt\" (UID: \"03b56ca9-6072-4f44-b144-2e8654fc26af\") " pod="openshift-marketplace/certified-operators-zltrt" Oct 08 19:22:53 crc kubenswrapper[4859]: I1008 19:22:53.896104 4859 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/03b56ca9-6072-4f44-b144-2e8654fc26af-utilities\") pod \"certified-operators-zltrt\" (UID: \"03b56ca9-6072-4f44-b144-2e8654fc26af\") " pod="openshift-marketplace/certified-operators-zltrt" Oct 08 19:22:53 crc kubenswrapper[4859]: I1008 19:22:53.912661 4859 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-sjt8j"] Oct 08 19:22:53 crc kubenswrapper[4859]: I1008 19:22:53.915254 4859 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-sjt8j" Oct 08 19:22:53 crc kubenswrapper[4859]: I1008 19:22:53.923553 4859 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-sjt8j"] Oct 08 19:22:53 crc kubenswrapper[4859]: I1008 19:22:53.997866 4859 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/f1c2e055-2fe8-40c5-a896-801d2b6d02bd-catalog-content\") pod \"community-operators-sjt8j\" (UID: \"f1c2e055-2fe8-40c5-a896-801d2b6d02bd\") " pod="openshift-marketplace/community-operators-sjt8j" Oct 08 19:22:53 crc kubenswrapper[4859]: I1008 19:22:53.997970 4859 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/f1c2e055-2fe8-40c5-a896-801d2b6d02bd-utilities\") pod \"community-operators-sjt8j\" (UID: \"f1c2e055-2fe8-40c5-a896-801d2b6d02bd\") " pod="openshift-marketplace/community-operators-sjt8j" Oct 08 19:22:53 crc kubenswrapper[4859]: I1008 19:22:53.998015 4859 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-pbzqj\" (UniqueName: \"kubernetes.io/projected/03b56ca9-6072-4f44-b144-2e8654fc26af-kube-api-access-pbzqj\") pod \"certified-operators-zltrt\" (UID: \"03b56ca9-6072-4f44-b144-2e8654fc26af\") " pod="openshift-marketplace/certified-operators-zltrt" Oct 08 19:22:53 crc kubenswrapper[4859]: I1008 19:22:53.998049 4859 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/03b56ca9-6072-4f44-b144-2e8654fc26af-utilities\") pod \"certified-operators-zltrt\" (UID: \"03b56ca9-6072-4f44-b144-2e8654fc26af\") " pod="openshift-marketplace/certified-operators-zltrt" Oct 08 19:22:53 crc kubenswrapper[4859]: I1008 19:22:53.998406 4859 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-5kc4g\" (UniqueName: \"kubernetes.io/projected/f1c2e055-2fe8-40c5-a896-801d2b6d02bd-kube-api-access-5kc4g\") pod \"community-operators-sjt8j\" (UID: \"f1c2e055-2fe8-40c5-a896-801d2b6d02bd\") " pod="openshift-marketplace/community-operators-sjt8j" Oct 08 19:22:53 crc kubenswrapper[4859]: I1008 19:22:53.998466 4859 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/03b56ca9-6072-4f44-b144-2e8654fc26af-utilities\") pod \"certified-operators-zltrt\" (UID: \"03b56ca9-6072-4f44-b144-2e8654fc26af\") " pod="openshift-marketplace/certified-operators-zltrt" Oct 08 19:22:53 crc kubenswrapper[4859]: I1008 19:22:53.998565 4859 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/03b56ca9-6072-4f44-b144-2e8654fc26af-catalog-content\") pod \"certified-operators-zltrt\" (UID: \"03b56ca9-6072-4f44-b144-2e8654fc26af\") " pod="openshift-marketplace/certified-operators-zltrt" Oct 08 19:22:53 crc kubenswrapper[4859]: I1008 19:22:53.998895 4859 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/03b56ca9-6072-4f44-b144-2e8654fc26af-catalog-content\") pod \"certified-operators-zltrt\" (UID: \"03b56ca9-6072-4f44-b144-2e8654fc26af\") " pod="openshift-marketplace/certified-operators-zltrt" Oct 08 19:22:54 crc kubenswrapper[4859]: I1008 19:22:54.019264 4859 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-pbzqj\" (UniqueName: \"kubernetes.io/projected/03b56ca9-6072-4f44-b144-2e8654fc26af-kube-api-access-pbzqj\") pod \"certified-operators-zltrt\" (UID: \"03b56ca9-6072-4f44-b144-2e8654fc26af\") " pod="openshift-marketplace/certified-operators-zltrt" Oct 08 19:22:54 crc kubenswrapper[4859]: I1008 19:22:54.046147 4859 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-zltrt" Oct 08 19:22:54 crc kubenswrapper[4859]: I1008 19:22:54.099983 4859 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-5kc4g\" (UniqueName: \"kubernetes.io/projected/f1c2e055-2fe8-40c5-a896-801d2b6d02bd-kube-api-access-5kc4g\") pod \"community-operators-sjt8j\" (UID: \"f1c2e055-2fe8-40c5-a896-801d2b6d02bd\") " pod="openshift-marketplace/community-operators-sjt8j" Oct 08 19:22:54 crc kubenswrapper[4859]: I1008 19:22:54.100348 4859 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/f1c2e055-2fe8-40c5-a896-801d2b6d02bd-catalog-content\") pod \"community-operators-sjt8j\" (UID: \"f1c2e055-2fe8-40c5-a896-801d2b6d02bd\") " pod="openshift-marketplace/community-operators-sjt8j" Oct 08 19:22:54 crc kubenswrapper[4859]: I1008 19:22:54.100387 4859 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/f1c2e055-2fe8-40c5-a896-801d2b6d02bd-utilities\") pod \"community-operators-sjt8j\" (UID: \"f1c2e055-2fe8-40c5-a896-801d2b6d02bd\") " pod="openshift-marketplace/community-operators-sjt8j" Oct 08 19:22:54 crc kubenswrapper[4859]: I1008 19:22:54.100832 4859 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/f1c2e055-2fe8-40c5-a896-801d2b6d02bd-catalog-content\") pod \"community-operators-sjt8j\" (UID: \"f1c2e055-2fe8-40c5-a896-801d2b6d02bd\") " pod="openshift-marketplace/community-operators-sjt8j" Oct 08 19:22:54 crc kubenswrapper[4859]: I1008 19:22:54.100844 4859 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/f1c2e055-2fe8-40c5-a896-801d2b6d02bd-utilities\") pod \"community-operators-sjt8j\" (UID: \"f1c2e055-2fe8-40c5-a896-801d2b6d02bd\") " pod="openshift-marketplace/community-operators-sjt8j" Oct 08 19:22:54 crc kubenswrapper[4859]: I1008 19:22:54.120407 4859 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-5kc4g\" (UniqueName: \"kubernetes.io/projected/f1c2e055-2fe8-40c5-a896-801d2b6d02bd-kube-api-access-5kc4g\") pod \"community-operators-sjt8j\" (UID: \"f1c2e055-2fe8-40c5-a896-801d2b6d02bd\") " pod="openshift-marketplace/community-operators-sjt8j" Oct 08 19:22:54 crc kubenswrapper[4859]: I1008 19:22:54.237197 4859 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-sjt8j" Oct 08 19:22:54 crc kubenswrapper[4859]: I1008 19:22:54.639865 4859 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-zltrt"] Oct 08 19:22:55 crc kubenswrapper[4859]: I1008 19:22:55.285511 4859 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-zltrt" event={"ID":"03b56ca9-6072-4f44-b144-2e8654fc26af","Type":"ContainerDied","Data":"6a807051dcf3a501a57a6451e555c30692c2d0ea55d98f63864f27651ddb4f56"} Oct 08 19:22:55 crc kubenswrapper[4859]: I1008 19:22:55.285468 4859 generic.go:334] "Generic (PLEG): container finished" podID="03b56ca9-6072-4f44-b144-2e8654fc26af" containerID="6a807051dcf3a501a57a6451e555c30692c2d0ea55d98f63864f27651ddb4f56" exitCode=0 Oct 08 19:22:55 crc kubenswrapper[4859]: I1008 19:22:55.285978 4859 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-zltrt" event={"ID":"03b56ca9-6072-4f44-b144-2e8654fc26af","Type":"ContainerStarted","Data":"047573012e432c99be0766c928ccc7732df45f912e17c107bc989562fec733a8"} Oct 08 19:22:55 crc kubenswrapper[4859]: I1008 19:22:55.292742 4859 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Oct 08 19:22:55 crc kubenswrapper[4859]: W1008 19:22:55.329897 4859 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podf1c2e055_2fe8_40c5_a896_801d2b6d02bd.slice/crio-49c44829cd84aacc7edf3ad6c4b5fa9cab5f44589800d45615d76def28948679 WatchSource:0}: Error finding container 49c44829cd84aacc7edf3ad6c4b5fa9cab5f44589800d45615d76def28948679: Status 404 returned error can't find the container with id 49c44829cd84aacc7edf3ad6c4b5fa9cab5f44589800d45615d76def28948679 Oct 08 19:22:55 crc kubenswrapper[4859]: I1008 19:22:55.347569 4859 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-sjt8j"] Oct 08 19:22:56 crc kubenswrapper[4859]: I1008 19:22:56.300518 4859 generic.go:334] "Generic (PLEG): container finished" podID="f1c2e055-2fe8-40c5-a896-801d2b6d02bd" containerID="50b23eb2303d4c5a18f4e6de83b53565b9e347da4a46dee75bc9bc512d3ad9a8" exitCode=0 Oct 08 19:22:56 crc kubenswrapper[4859]: I1008 19:22:56.300726 4859 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-sjt8j" event={"ID":"f1c2e055-2fe8-40c5-a896-801d2b6d02bd","Type":"ContainerDied","Data":"50b23eb2303d4c5a18f4e6de83b53565b9e347da4a46dee75bc9bc512d3ad9a8"} Oct 08 19:22:56 crc kubenswrapper[4859]: I1008 19:22:56.301289 4859 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-sjt8j" event={"ID":"f1c2e055-2fe8-40c5-a896-801d2b6d02bd","Type":"ContainerStarted","Data":"49c44829cd84aacc7edf3ad6c4b5fa9cab5f44589800d45615d76def28948679"} Oct 08 19:22:56 crc kubenswrapper[4859]: I1008 19:22:56.306447 4859 generic.go:334] "Generic (PLEG): container finished" podID="03b56ca9-6072-4f44-b144-2e8654fc26af" containerID="bb99e93a03ec37774df06deef51aa5ae1dff8f4a7d8c8a6b1e6a269df29aab88" exitCode=0 Oct 08 19:22:56 crc kubenswrapper[4859]: I1008 19:22:56.306482 4859 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-zltrt" event={"ID":"03b56ca9-6072-4f44-b144-2e8654fc26af","Type":"ContainerDied","Data":"bb99e93a03ec37774df06deef51aa5ae1dff8f4a7d8c8a6b1e6a269df29aab88"} Oct 08 19:22:57 crc kubenswrapper[4859]: I1008 19:22:57.320786 4859 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-zltrt" event={"ID":"03b56ca9-6072-4f44-b144-2e8654fc26af","Type":"ContainerStarted","Data":"3bafd7fad7bba7601dad2292419fd75fb4073cc5ae13ec8f6b640f2f87db5e72"} Oct 08 19:22:57 crc kubenswrapper[4859]: I1008 19:22:57.380295 4859 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-zltrt" podStartSLOduration=2.957384589 podStartE2EDuration="4.380274749s" podCreationTimestamp="2025-10-08 19:22:53 +0000 UTC" firstStartedPulling="2025-10-08 19:22:55.292312171 +0000 UTC m=+3945.539151590" lastFinishedPulling="2025-10-08 19:22:56.715202361 +0000 UTC m=+3946.962041750" observedRunningTime="2025-10-08 19:22:57.367559478 +0000 UTC m=+3947.614398857" watchObservedRunningTime="2025-10-08 19:22:57.380274749 +0000 UTC m=+3947.627114128" Oct 08 19:22:58 crc kubenswrapper[4859]: I1008 19:22:58.334875 4859 generic.go:334] "Generic (PLEG): container finished" podID="f1c2e055-2fe8-40c5-a896-801d2b6d02bd" containerID="1ea6a10e96ba2d3597e999b2d473cb709bde8c27ec749e59d8d22ba6caca3780" exitCode=0 Oct 08 19:22:58 crc kubenswrapper[4859]: I1008 19:22:58.334968 4859 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-sjt8j" event={"ID":"f1c2e055-2fe8-40c5-a896-801d2b6d02bd","Type":"ContainerDied","Data":"1ea6a10e96ba2d3597e999b2d473cb709bde8c27ec749e59d8d22ba6caca3780"} Oct 08 19:22:59 crc kubenswrapper[4859]: I1008 19:22:59.347681 4859 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-sjt8j" event={"ID":"f1c2e055-2fe8-40c5-a896-801d2b6d02bd","Type":"ContainerStarted","Data":"9b161947202d188bb35358ea28abbdd6c799ebfa0ad5227cffe9d251c045d89b"} Oct 08 19:22:59 crc kubenswrapper[4859]: I1008 19:22:59.371551 4859 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-sjt8j" podStartSLOduration=3.8100309660000002 podStartE2EDuration="6.371531558s" podCreationTimestamp="2025-10-08 19:22:53 +0000 UTC" firstStartedPulling="2025-10-08 19:22:56.304414506 +0000 UTC m=+3946.551253925" lastFinishedPulling="2025-10-08 19:22:58.865915118 +0000 UTC m=+3949.112754517" observedRunningTime="2025-10-08 19:22:59.363194365 +0000 UTC m=+3949.610033744" watchObservedRunningTime="2025-10-08 19:22:59.371531558 +0000 UTC m=+3949.618370937" Oct 08 19:23:00 crc kubenswrapper[4859]: I1008 19:23:00.479291 4859 scope.go:117] "RemoveContainer" containerID="de71f0b717129dba48f3f0bdd90240a8e45404a17f0df79d6812d58c9209feff" Oct 08 19:23:00 crc kubenswrapper[4859]: E1008 19:23:00.479915 4859 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-82s52_openshift-machine-config-operator(b23a6a6c-9d92-4e7b-840e-55cfda873a2d)\"" pod="openshift-machine-config-operator/machine-config-daemon-82s52" podUID="b23a6a6c-9d92-4e7b-840e-55cfda873a2d" Oct 08 19:23:04 crc kubenswrapper[4859]: I1008 19:23:04.046977 4859 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-zltrt" Oct 08 19:23:04 crc kubenswrapper[4859]: I1008 19:23:04.047664 4859 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-zltrt" Oct 08 19:23:04 crc kubenswrapper[4859]: I1008 19:23:04.135053 4859 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-zltrt" Oct 08 19:23:04 crc kubenswrapper[4859]: I1008 19:23:04.238604 4859 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-sjt8j" Oct 08 19:23:04 crc kubenswrapper[4859]: I1008 19:23:04.238716 4859 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-sjt8j" Oct 08 19:23:04 crc kubenswrapper[4859]: I1008 19:23:04.302982 4859 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-sjt8j" Oct 08 19:23:04 crc kubenswrapper[4859]: I1008 19:23:04.489672 4859 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-zltrt" Oct 08 19:23:04 crc kubenswrapper[4859]: I1008 19:23:04.496102 4859 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-sjt8j" Oct 08 19:23:06 crc kubenswrapper[4859]: I1008 19:23:06.510231 4859 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-zltrt"] Oct 08 19:23:06 crc kubenswrapper[4859]: I1008 19:23:06.510903 4859 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/certified-operators-zltrt" podUID="03b56ca9-6072-4f44-b144-2e8654fc26af" containerName="registry-server" containerID="cri-o://3bafd7fad7bba7601dad2292419fd75fb4073cc5ae13ec8f6b640f2f87db5e72" gracePeriod=2 Oct 08 19:23:06 crc kubenswrapper[4859]: I1008 19:23:06.705829 4859 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-sjt8j"] Oct 08 19:23:06 crc kubenswrapper[4859]: I1008 19:23:06.706114 4859 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/community-operators-sjt8j" podUID="f1c2e055-2fe8-40c5-a896-801d2b6d02bd" containerName="registry-server" containerID="cri-o://9b161947202d188bb35358ea28abbdd6c799ebfa0ad5227cffe9d251c045d89b" gracePeriod=2 Oct 08 19:23:07 crc kubenswrapper[4859]: I1008 19:23:07.265297 4859 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-sjt8j" Oct 08 19:23:07 crc kubenswrapper[4859]: I1008 19:23:07.404426 4859 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/f1c2e055-2fe8-40c5-a896-801d2b6d02bd-catalog-content\") pod \"f1c2e055-2fe8-40c5-a896-801d2b6d02bd\" (UID: \"f1c2e055-2fe8-40c5-a896-801d2b6d02bd\") " Oct 08 19:23:07 crc kubenswrapper[4859]: I1008 19:23:07.404666 4859 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/f1c2e055-2fe8-40c5-a896-801d2b6d02bd-utilities\") pod \"f1c2e055-2fe8-40c5-a896-801d2b6d02bd\" (UID: \"f1c2e055-2fe8-40c5-a896-801d2b6d02bd\") " Oct 08 19:23:07 crc kubenswrapper[4859]: I1008 19:23:07.404700 4859 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-5kc4g\" (UniqueName: \"kubernetes.io/projected/f1c2e055-2fe8-40c5-a896-801d2b6d02bd-kube-api-access-5kc4g\") pod \"f1c2e055-2fe8-40c5-a896-801d2b6d02bd\" (UID: \"f1c2e055-2fe8-40c5-a896-801d2b6d02bd\") " Oct 08 19:23:07 crc kubenswrapper[4859]: I1008 19:23:07.407364 4859 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/f1c2e055-2fe8-40c5-a896-801d2b6d02bd-utilities" (OuterVolumeSpecName: "utilities") pod "f1c2e055-2fe8-40c5-a896-801d2b6d02bd" (UID: "f1c2e055-2fe8-40c5-a896-801d2b6d02bd"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 08 19:23:07 crc kubenswrapper[4859]: I1008 19:23:07.408154 4859 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/f1c2e055-2fe8-40c5-a896-801d2b6d02bd-utilities\") on node \"crc\" DevicePath \"\"" Oct 08 19:23:07 crc kubenswrapper[4859]: I1008 19:23:07.427051 4859 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/f1c2e055-2fe8-40c5-a896-801d2b6d02bd-kube-api-access-5kc4g" (OuterVolumeSpecName: "kube-api-access-5kc4g") pod "f1c2e055-2fe8-40c5-a896-801d2b6d02bd" (UID: "f1c2e055-2fe8-40c5-a896-801d2b6d02bd"). InnerVolumeSpecName "kube-api-access-5kc4g". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 08 19:23:07 crc kubenswrapper[4859]: I1008 19:23:07.441238 4859 generic.go:334] "Generic (PLEG): container finished" podID="03b56ca9-6072-4f44-b144-2e8654fc26af" containerID="3bafd7fad7bba7601dad2292419fd75fb4073cc5ae13ec8f6b640f2f87db5e72" exitCode=0 Oct 08 19:23:07 crc kubenswrapper[4859]: I1008 19:23:07.441302 4859 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-zltrt" event={"ID":"03b56ca9-6072-4f44-b144-2e8654fc26af","Type":"ContainerDied","Data":"3bafd7fad7bba7601dad2292419fd75fb4073cc5ae13ec8f6b640f2f87db5e72"} Oct 08 19:23:07 crc kubenswrapper[4859]: I1008 19:23:07.441329 4859 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-zltrt" event={"ID":"03b56ca9-6072-4f44-b144-2e8654fc26af","Type":"ContainerDied","Data":"047573012e432c99be0766c928ccc7732df45f912e17c107bc989562fec733a8"} Oct 08 19:23:07 crc kubenswrapper[4859]: I1008 19:23:07.441338 4859 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="047573012e432c99be0766c928ccc7732df45f912e17c107bc989562fec733a8" Oct 08 19:23:07 crc kubenswrapper[4859]: I1008 19:23:07.447231 4859 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-zltrt" Oct 08 19:23:07 crc kubenswrapper[4859]: I1008 19:23:07.447703 4859 generic.go:334] "Generic (PLEG): container finished" podID="f1c2e055-2fe8-40c5-a896-801d2b6d02bd" containerID="9b161947202d188bb35358ea28abbdd6c799ebfa0ad5227cffe9d251c045d89b" exitCode=0 Oct 08 19:23:07 crc kubenswrapper[4859]: I1008 19:23:07.447744 4859 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-sjt8j" event={"ID":"f1c2e055-2fe8-40c5-a896-801d2b6d02bd","Type":"ContainerDied","Data":"9b161947202d188bb35358ea28abbdd6c799ebfa0ad5227cffe9d251c045d89b"} Oct 08 19:23:07 crc kubenswrapper[4859]: I1008 19:23:07.447772 4859 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-sjt8j" event={"ID":"f1c2e055-2fe8-40c5-a896-801d2b6d02bd","Type":"ContainerDied","Data":"49c44829cd84aacc7edf3ad6c4b5fa9cab5f44589800d45615d76def28948679"} Oct 08 19:23:07 crc kubenswrapper[4859]: I1008 19:23:07.447791 4859 scope.go:117] "RemoveContainer" containerID="9b161947202d188bb35358ea28abbdd6c799ebfa0ad5227cffe9d251c045d89b" Oct 08 19:23:07 crc kubenswrapper[4859]: I1008 19:23:07.447938 4859 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-sjt8j" Oct 08 19:23:07 crc kubenswrapper[4859]: I1008 19:23:07.502935 4859 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-must-gather-4fzzd/must-gather-rf7m5"] Oct 08 19:23:07 crc kubenswrapper[4859]: E1008 19:23:07.503438 4859 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f1c2e055-2fe8-40c5-a896-801d2b6d02bd" containerName="registry-server" Oct 08 19:23:07 crc kubenswrapper[4859]: I1008 19:23:07.503458 4859 state_mem.go:107] "Deleted CPUSet assignment" podUID="f1c2e055-2fe8-40c5-a896-801d2b6d02bd" containerName="registry-server" Oct 08 19:23:07 crc kubenswrapper[4859]: E1008 19:23:07.503480 4859 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="03b56ca9-6072-4f44-b144-2e8654fc26af" containerName="extract-utilities" Oct 08 19:23:07 crc kubenswrapper[4859]: I1008 19:23:07.503490 4859 state_mem.go:107] "Deleted CPUSet assignment" podUID="03b56ca9-6072-4f44-b144-2e8654fc26af" containerName="extract-utilities" Oct 08 19:23:07 crc kubenswrapper[4859]: E1008 19:23:07.503502 4859 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f1c2e055-2fe8-40c5-a896-801d2b6d02bd" containerName="extract-utilities" Oct 08 19:23:07 crc kubenswrapper[4859]: I1008 19:23:07.503510 4859 state_mem.go:107] "Deleted CPUSet assignment" podUID="f1c2e055-2fe8-40c5-a896-801d2b6d02bd" containerName="extract-utilities" Oct 08 19:23:07 crc kubenswrapper[4859]: E1008 19:23:07.503537 4859 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f1c2e055-2fe8-40c5-a896-801d2b6d02bd" containerName="extract-content" Oct 08 19:23:07 crc kubenswrapper[4859]: I1008 19:23:07.503544 4859 state_mem.go:107] "Deleted CPUSet assignment" podUID="f1c2e055-2fe8-40c5-a896-801d2b6d02bd" containerName="extract-content" Oct 08 19:23:07 crc kubenswrapper[4859]: E1008 19:23:07.503566 4859 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="03b56ca9-6072-4f44-b144-2e8654fc26af" containerName="extract-content" Oct 08 19:23:07 crc kubenswrapper[4859]: I1008 19:23:07.503573 4859 state_mem.go:107] "Deleted CPUSet assignment" podUID="03b56ca9-6072-4f44-b144-2e8654fc26af" containerName="extract-content" Oct 08 19:23:07 crc kubenswrapper[4859]: E1008 19:23:07.503584 4859 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="03b56ca9-6072-4f44-b144-2e8654fc26af" containerName="registry-server" Oct 08 19:23:07 crc kubenswrapper[4859]: I1008 19:23:07.503591 4859 state_mem.go:107] "Deleted CPUSet assignment" podUID="03b56ca9-6072-4f44-b144-2e8654fc26af" containerName="registry-server" Oct 08 19:23:07 crc kubenswrapper[4859]: I1008 19:23:07.503816 4859 memory_manager.go:354] "RemoveStaleState removing state" podUID="f1c2e055-2fe8-40c5-a896-801d2b6d02bd" containerName="registry-server" Oct 08 19:23:07 crc kubenswrapper[4859]: I1008 19:23:07.503872 4859 memory_manager.go:354] "RemoveStaleState removing state" podUID="03b56ca9-6072-4f44-b144-2e8654fc26af" containerName="registry-server" Oct 08 19:23:07 crc kubenswrapper[4859]: I1008 19:23:07.505098 4859 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-4fzzd/must-gather-rf7m5" Oct 08 19:23:07 crc kubenswrapper[4859]: I1008 19:23:07.509667 4859 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-5kc4g\" (UniqueName: \"kubernetes.io/projected/f1c2e055-2fe8-40c5-a896-801d2b6d02bd-kube-api-access-5kc4g\") on node \"crc\" DevicePath \"\"" Oct 08 19:23:07 crc kubenswrapper[4859]: I1008 19:23:07.509804 4859 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-must-gather-4fzzd"/"openshift-service-ca.crt" Oct 08 19:23:07 crc kubenswrapper[4859]: I1008 19:23:07.509806 4859 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-must-gather-4fzzd"/"default-dockercfg-5lsvc" Oct 08 19:23:07 crc kubenswrapper[4859]: I1008 19:23:07.517587 4859 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-must-gather-4fzzd"/"kube-root-ca.crt" Oct 08 19:23:07 crc kubenswrapper[4859]: I1008 19:23:07.518841 4859 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/f1c2e055-2fe8-40c5-a896-801d2b6d02bd-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "f1c2e055-2fe8-40c5-a896-801d2b6d02bd" (UID: "f1c2e055-2fe8-40c5-a896-801d2b6d02bd"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 08 19:23:07 crc kubenswrapper[4859]: I1008 19:23:07.523849 4859 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-must-gather-4fzzd/must-gather-rf7m5"] Oct 08 19:23:07 crc kubenswrapper[4859]: I1008 19:23:07.536038 4859 scope.go:117] "RemoveContainer" containerID="1ea6a10e96ba2d3597e999b2d473cb709bde8c27ec749e59d8d22ba6caca3780" Oct 08 19:23:07 crc kubenswrapper[4859]: I1008 19:23:07.577401 4859 scope.go:117] "RemoveContainer" containerID="50b23eb2303d4c5a18f4e6de83b53565b9e347da4a46dee75bc9bc512d3ad9a8" Oct 08 19:23:07 crc kubenswrapper[4859]: I1008 19:23:07.611504 4859 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/03b56ca9-6072-4f44-b144-2e8654fc26af-catalog-content\") pod \"03b56ca9-6072-4f44-b144-2e8654fc26af\" (UID: \"03b56ca9-6072-4f44-b144-2e8654fc26af\") " Oct 08 19:23:07 crc kubenswrapper[4859]: I1008 19:23:07.611560 4859 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/03b56ca9-6072-4f44-b144-2e8654fc26af-utilities\") pod \"03b56ca9-6072-4f44-b144-2e8654fc26af\" (UID: \"03b56ca9-6072-4f44-b144-2e8654fc26af\") " Oct 08 19:23:07 crc kubenswrapper[4859]: I1008 19:23:07.611637 4859 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-pbzqj\" (UniqueName: \"kubernetes.io/projected/03b56ca9-6072-4f44-b144-2e8654fc26af-kube-api-access-pbzqj\") pod \"03b56ca9-6072-4f44-b144-2e8654fc26af\" (UID: \"03b56ca9-6072-4f44-b144-2e8654fc26af\") " Oct 08 19:23:07 crc kubenswrapper[4859]: I1008 19:23:07.611943 4859 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"must-gather-output\" (UniqueName: \"kubernetes.io/empty-dir/23aa205e-275b-4feb-8b07-537fc2834e00-must-gather-output\") pod \"must-gather-rf7m5\" (UID: \"23aa205e-275b-4feb-8b07-537fc2834e00\") " pod="openshift-must-gather-4fzzd/must-gather-rf7m5" Oct 08 19:23:07 crc kubenswrapper[4859]: I1008 19:23:07.612195 4859 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-w2rq7\" (UniqueName: \"kubernetes.io/projected/23aa205e-275b-4feb-8b07-537fc2834e00-kube-api-access-w2rq7\") pod \"must-gather-rf7m5\" (UID: \"23aa205e-275b-4feb-8b07-537fc2834e00\") " pod="openshift-must-gather-4fzzd/must-gather-rf7m5" Oct 08 19:23:07 crc kubenswrapper[4859]: I1008 19:23:07.612425 4859 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/f1c2e055-2fe8-40c5-a896-801d2b6d02bd-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 08 19:23:07 crc kubenswrapper[4859]: I1008 19:23:07.614214 4859 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/03b56ca9-6072-4f44-b144-2e8654fc26af-utilities" (OuterVolumeSpecName: "utilities") pod "03b56ca9-6072-4f44-b144-2e8654fc26af" (UID: "03b56ca9-6072-4f44-b144-2e8654fc26af"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 08 19:23:07 crc kubenswrapper[4859]: I1008 19:23:07.618611 4859 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/03b56ca9-6072-4f44-b144-2e8654fc26af-kube-api-access-pbzqj" (OuterVolumeSpecName: "kube-api-access-pbzqj") pod "03b56ca9-6072-4f44-b144-2e8654fc26af" (UID: "03b56ca9-6072-4f44-b144-2e8654fc26af"). InnerVolumeSpecName "kube-api-access-pbzqj". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 08 19:23:07 crc kubenswrapper[4859]: I1008 19:23:07.620121 4859 scope.go:117] "RemoveContainer" containerID="9b161947202d188bb35358ea28abbdd6c799ebfa0ad5227cffe9d251c045d89b" Oct 08 19:23:07 crc kubenswrapper[4859]: E1008 19:23:07.620817 4859 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"9b161947202d188bb35358ea28abbdd6c799ebfa0ad5227cffe9d251c045d89b\": container with ID starting with 9b161947202d188bb35358ea28abbdd6c799ebfa0ad5227cffe9d251c045d89b not found: ID does not exist" containerID="9b161947202d188bb35358ea28abbdd6c799ebfa0ad5227cffe9d251c045d89b" Oct 08 19:23:07 crc kubenswrapper[4859]: I1008 19:23:07.620852 4859 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"9b161947202d188bb35358ea28abbdd6c799ebfa0ad5227cffe9d251c045d89b"} err="failed to get container status \"9b161947202d188bb35358ea28abbdd6c799ebfa0ad5227cffe9d251c045d89b\": rpc error: code = NotFound desc = could not find container \"9b161947202d188bb35358ea28abbdd6c799ebfa0ad5227cffe9d251c045d89b\": container with ID starting with 9b161947202d188bb35358ea28abbdd6c799ebfa0ad5227cffe9d251c045d89b not found: ID does not exist" Oct 08 19:23:07 crc kubenswrapper[4859]: I1008 19:23:07.620875 4859 scope.go:117] "RemoveContainer" containerID="1ea6a10e96ba2d3597e999b2d473cb709bde8c27ec749e59d8d22ba6caca3780" Oct 08 19:23:07 crc kubenswrapper[4859]: E1008 19:23:07.621292 4859 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"1ea6a10e96ba2d3597e999b2d473cb709bde8c27ec749e59d8d22ba6caca3780\": container with ID starting with 1ea6a10e96ba2d3597e999b2d473cb709bde8c27ec749e59d8d22ba6caca3780 not found: ID does not exist" containerID="1ea6a10e96ba2d3597e999b2d473cb709bde8c27ec749e59d8d22ba6caca3780" Oct 08 19:23:07 crc kubenswrapper[4859]: I1008 19:23:07.621336 4859 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"1ea6a10e96ba2d3597e999b2d473cb709bde8c27ec749e59d8d22ba6caca3780"} err="failed to get container status \"1ea6a10e96ba2d3597e999b2d473cb709bde8c27ec749e59d8d22ba6caca3780\": rpc error: code = NotFound desc = could not find container \"1ea6a10e96ba2d3597e999b2d473cb709bde8c27ec749e59d8d22ba6caca3780\": container with ID starting with 1ea6a10e96ba2d3597e999b2d473cb709bde8c27ec749e59d8d22ba6caca3780 not found: ID does not exist" Oct 08 19:23:07 crc kubenswrapper[4859]: I1008 19:23:07.621351 4859 scope.go:117] "RemoveContainer" containerID="50b23eb2303d4c5a18f4e6de83b53565b9e347da4a46dee75bc9bc512d3ad9a8" Oct 08 19:23:07 crc kubenswrapper[4859]: E1008 19:23:07.621596 4859 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"50b23eb2303d4c5a18f4e6de83b53565b9e347da4a46dee75bc9bc512d3ad9a8\": container with ID starting with 50b23eb2303d4c5a18f4e6de83b53565b9e347da4a46dee75bc9bc512d3ad9a8 not found: ID does not exist" containerID="50b23eb2303d4c5a18f4e6de83b53565b9e347da4a46dee75bc9bc512d3ad9a8" Oct 08 19:23:07 crc kubenswrapper[4859]: I1008 19:23:07.621629 4859 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"50b23eb2303d4c5a18f4e6de83b53565b9e347da4a46dee75bc9bc512d3ad9a8"} err="failed to get container status \"50b23eb2303d4c5a18f4e6de83b53565b9e347da4a46dee75bc9bc512d3ad9a8\": rpc error: code = NotFound desc = could not find container \"50b23eb2303d4c5a18f4e6de83b53565b9e347da4a46dee75bc9bc512d3ad9a8\": container with ID starting with 50b23eb2303d4c5a18f4e6de83b53565b9e347da4a46dee75bc9bc512d3ad9a8 not found: ID does not exist" Oct 08 19:23:07 crc kubenswrapper[4859]: I1008 19:23:07.671964 4859 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/03b56ca9-6072-4f44-b144-2e8654fc26af-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "03b56ca9-6072-4f44-b144-2e8654fc26af" (UID: "03b56ca9-6072-4f44-b144-2e8654fc26af"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 08 19:23:07 crc kubenswrapper[4859]: I1008 19:23:07.713753 4859 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-w2rq7\" (UniqueName: \"kubernetes.io/projected/23aa205e-275b-4feb-8b07-537fc2834e00-kube-api-access-w2rq7\") pod \"must-gather-rf7m5\" (UID: \"23aa205e-275b-4feb-8b07-537fc2834e00\") " pod="openshift-must-gather-4fzzd/must-gather-rf7m5" Oct 08 19:23:07 crc kubenswrapper[4859]: I1008 19:23:07.713902 4859 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"must-gather-output\" (UniqueName: \"kubernetes.io/empty-dir/23aa205e-275b-4feb-8b07-537fc2834e00-must-gather-output\") pod \"must-gather-rf7m5\" (UID: \"23aa205e-275b-4feb-8b07-537fc2834e00\") " pod="openshift-must-gather-4fzzd/must-gather-rf7m5" Oct 08 19:23:07 crc kubenswrapper[4859]: I1008 19:23:07.714046 4859 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/03b56ca9-6072-4f44-b144-2e8654fc26af-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 08 19:23:07 crc kubenswrapper[4859]: I1008 19:23:07.714068 4859 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/03b56ca9-6072-4f44-b144-2e8654fc26af-utilities\") on node \"crc\" DevicePath \"\"" Oct 08 19:23:07 crc kubenswrapper[4859]: I1008 19:23:07.714080 4859 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-pbzqj\" (UniqueName: \"kubernetes.io/projected/03b56ca9-6072-4f44-b144-2e8654fc26af-kube-api-access-pbzqj\") on node \"crc\" DevicePath \"\"" Oct 08 19:23:07 crc kubenswrapper[4859]: I1008 19:23:07.714379 4859 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"must-gather-output\" (UniqueName: \"kubernetes.io/empty-dir/23aa205e-275b-4feb-8b07-537fc2834e00-must-gather-output\") pod \"must-gather-rf7m5\" (UID: \"23aa205e-275b-4feb-8b07-537fc2834e00\") " pod="openshift-must-gather-4fzzd/must-gather-rf7m5" Oct 08 19:23:07 crc kubenswrapper[4859]: I1008 19:23:07.728911 4859 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-w2rq7\" (UniqueName: \"kubernetes.io/projected/23aa205e-275b-4feb-8b07-537fc2834e00-kube-api-access-w2rq7\") pod \"must-gather-rf7m5\" (UID: \"23aa205e-275b-4feb-8b07-537fc2834e00\") " pod="openshift-must-gather-4fzzd/must-gather-rf7m5" Oct 08 19:23:07 crc kubenswrapper[4859]: I1008 19:23:07.793961 4859 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-sjt8j"] Oct 08 19:23:07 crc kubenswrapper[4859]: I1008 19:23:07.801609 4859 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/community-operators-sjt8j"] Oct 08 19:23:07 crc kubenswrapper[4859]: I1008 19:23:07.834245 4859 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-4fzzd/must-gather-rf7m5" Oct 08 19:23:08 crc kubenswrapper[4859]: I1008 19:23:08.324364 4859 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-must-gather-4fzzd/must-gather-rf7m5"] Oct 08 19:23:08 crc kubenswrapper[4859]: I1008 19:23:08.461486 4859 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-4fzzd/must-gather-rf7m5" event={"ID":"23aa205e-275b-4feb-8b07-537fc2834e00","Type":"ContainerStarted","Data":"db009d983a9b2ebda63add58056fb9bbe53733e24a07fe196d5a327647835f46"} Oct 08 19:23:08 crc kubenswrapper[4859]: I1008 19:23:08.463389 4859 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-zltrt" Oct 08 19:23:08 crc kubenswrapper[4859]: I1008 19:23:08.481365 4859 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="f1c2e055-2fe8-40c5-a896-801d2b6d02bd" path="/var/lib/kubelet/pods/f1c2e055-2fe8-40c5-a896-801d2b6d02bd/volumes" Oct 08 19:23:08 crc kubenswrapper[4859]: I1008 19:23:08.513639 4859 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-zltrt"] Oct 08 19:23:08 crc kubenswrapper[4859]: I1008 19:23:08.521535 4859 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/certified-operators-zltrt"] Oct 08 19:23:09 crc kubenswrapper[4859]: I1008 19:23:09.472025 4859 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-4fzzd/must-gather-rf7m5" event={"ID":"23aa205e-275b-4feb-8b07-537fc2834e00","Type":"ContainerStarted","Data":"7e415e52ba985fe6878bc7957c47a8f0110b06082a1a375b930c4607c0d60f73"} Oct 08 19:23:09 crc kubenswrapper[4859]: I1008 19:23:09.472465 4859 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-4fzzd/must-gather-rf7m5" event={"ID":"23aa205e-275b-4feb-8b07-537fc2834e00","Type":"ContainerStarted","Data":"dfbc1b29c005089d554308d5526cc7452d53ddefdcf75e35f4611090b1cbc6db"} Oct 08 19:23:09 crc kubenswrapper[4859]: I1008 19:23:09.484570 4859 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-must-gather-4fzzd/must-gather-rf7m5" podStartSLOduration=2.484559102 podStartE2EDuration="2.484559102s" podCreationTimestamp="2025-10-08 19:23:07 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-08 19:23:09.484188071 +0000 UTC m=+3959.731027450" watchObservedRunningTime="2025-10-08 19:23:09.484559102 +0000 UTC m=+3959.731398481" Oct 08 19:23:10 crc kubenswrapper[4859]: I1008 19:23:10.482634 4859 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="03b56ca9-6072-4f44-b144-2e8654fc26af" path="/var/lib/kubelet/pods/03b56ca9-6072-4f44-b144-2e8654fc26af/volumes" Oct 08 19:23:12 crc kubenswrapper[4859]: I1008 19:23:12.091164 4859 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-must-gather-4fzzd/crc-debug-mjvd7"] Oct 08 19:23:12 crc kubenswrapper[4859]: I1008 19:23:12.092507 4859 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-4fzzd/crc-debug-mjvd7" Oct 08 19:23:12 crc kubenswrapper[4859]: I1008 19:23:12.195185 4859 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-kcdmk\" (UniqueName: \"kubernetes.io/projected/28e20e5f-7566-4eea-9814-befdc2fa7535-kube-api-access-kcdmk\") pod \"crc-debug-mjvd7\" (UID: \"28e20e5f-7566-4eea-9814-befdc2fa7535\") " pod="openshift-must-gather-4fzzd/crc-debug-mjvd7" Oct 08 19:23:12 crc kubenswrapper[4859]: I1008 19:23:12.195247 4859 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/28e20e5f-7566-4eea-9814-befdc2fa7535-host\") pod \"crc-debug-mjvd7\" (UID: \"28e20e5f-7566-4eea-9814-befdc2fa7535\") " pod="openshift-must-gather-4fzzd/crc-debug-mjvd7" Oct 08 19:23:12 crc kubenswrapper[4859]: I1008 19:23:12.296820 4859 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-kcdmk\" (UniqueName: \"kubernetes.io/projected/28e20e5f-7566-4eea-9814-befdc2fa7535-kube-api-access-kcdmk\") pod \"crc-debug-mjvd7\" (UID: \"28e20e5f-7566-4eea-9814-befdc2fa7535\") " pod="openshift-must-gather-4fzzd/crc-debug-mjvd7" Oct 08 19:23:12 crc kubenswrapper[4859]: I1008 19:23:12.296876 4859 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/28e20e5f-7566-4eea-9814-befdc2fa7535-host\") pod \"crc-debug-mjvd7\" (UID: \"28e20e5f-7566-4eea-9814-befdc2fa7535\") " pod="openshift-must-gather-4fzzd/crc-debug-mjvd7" Oct 08 19:23:12 crc kubenswrapper[4859]: I1008 19:23:12.297020 4859 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host\" (UniqueName: \"kubernetes.io/host-path/28e20e5f-7566-4eea-9814-befdc2fa7535-host\") pod \"crc-debug-mjvd7\" (UID: \"28e20e5f-7566-4eea-9814-befdc2fa7535\") " pod="openshift-must-gather-4fzzd/crc-debug-mjvd7" Oct 08 19:23:12 crc kubenswrapper[4859]: I1008 19:23:12.315884 4859 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-kcdmk\" (UniqueName: \"kubernetes.io/projected/28e20e5f-7566-4eea-9814-befdc2fa7535-kube-api-access-kcdmk\") pod \"crc-debug-mjvd7\" (UID: \"28e20e5f-7566-4eea-9814-befdc2fa7535\") " pod="openshift-must-gather-4fzzd/crc-debug-mjvd7" Oct 08 19:23:12 crc kubenswrapper[4859]: I1008 19:23:12.417328 4859 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-4fzzd/crc-debug-mjvd7" Oct 08 19:23:12 crc kubenswrapper[4859]: I1008 19:23:12.506605 4859 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-4fzzd/crc-debug-mjvd7" event={"ID":"28e20e5f-7566-4eea-9814-befdc2fa7535","Type":"ContainerStarted","Data":"b154433eb0a79c4cb0b9f88ae5b1318982fac6401f3c95f53d726e6a491cb2d4"} Oct 08 19:23:13 crc kubenswrapper[4859]: I1008 19:23:13.470655 4859 scope.go:117] "RemoveContainer" containerID="de71f0b717129dba48f3f0bdd90240a8e45404a17f0df79d6812d58c9209feff" Oct 08 19:23:13 crc kubenswrapper[4859]: E1008 19:23:13.471782 4859 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-82s52_openshift-machine-config-operator(b23a6a6c-9d92-4e7b-840e-55cfda873a2d)\"" pod="openshift-machine-config-operator/machine-config-daemon-82s52" podUID="b23a6a6c-9d92-4e7b-840e-55cfda873a2d" Oct 08 19:23:13 crc kubenswrapper[4859]: I1008 19:23:13.515864 4859 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-4fzzd/crc-debug-mjvd7" event={"ID":"28e20e5f-7566-4eea-9814-befdc2fa7535","Type":"ContainerStarted","Data":"3c1d50de8a22d6fd32dca7f0714e90e74ac2d271254e43f32073a96aeedf5a03"} Oct 08 19:23:13 crc kubenswrapper[4859]: I1008 19:23:13.536123 4859 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-must-gather-4fzzd/crc-debug-mjvd7" podStartSLOduration=1.536098792 podStartE2EDuration="1.536098792s" podCreationTimestamp="2025-10-08 19:23:12 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-08 19:23:13.52779897 +0000 UTC m=+3963.774638379" watchObservedRunningTime="2025-10-08 19:23:13.536098792 +0000 UTC m=+3963.782938191" Oct 08 19:23:28 crc kubenswrapper[4859]: I1008 19:23:28.470241 4859 scope.go:117] "RemoveContainer" containerID="de71f0b717129dba48f3f0bdd90240a8e45404a17f0df79d6812d58c9209feff" Oct 08 19:23:28 crc kubenswrapper[4859]: E1008 19:23:28.470963 4859 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-82s52_openshift-machine-config-operator(b23a6a6c-9d92-4e7b-840e-55cfda873a2d)\"" pod="openshift-machine-config-operator/machine-config-daemon-82s52" podUID="b23a6a6c-9d92-4e7b-840e-55cfda873a2d" Oct 08 19:23:42 crc kubenswrapper[4859]: I1008 19:23:42.470079 4859 scope.go:117] "RemoveContainer" containerID="de71f0b717129dba48f3f0bdd90240a8e45404a17f0df79d6812d58c9209feff" Oct 08 19:23:42 crc kubenswrapper[4859]: E1008 19:23:42.470905 4859 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-82s52_openshift-machine-config-operator(b23a6a6c-9d92-4e7b-840e-55cfda873a2d)\"" pod="openshift-machine-config-operator/machine-config-daemon-82s52" podUID="b23a6a6c-9d92-4e7b-840e-55cfda873a2d" Oct 08 19:23:56 crc kubenswrapper[4859]: I1008 19:23:56.470166 4859 scope.go:117] "RemoveContainer" containerID="de71f0b717129dba48f3f0bdd90240a8e45404a17f0df79d6812d58c9209feff" Oct 08 19:23:56 crc kubenswrapper[4859]: E1008 19:23:56.471087 4859 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-82s52_openshift-machine-config-operator(b23a6a6c-9d92-4e7b-840e-55cfda873a2d)\"" pod="openshift-machine-config-operator/machine-config-daemon-82s52" podUID="b23a6a6c-9d92-4e7b-840e-55cfda873a2d" Oct 08 19:24:10 crc kubenswrapper[4859]: I1008 19:24:10.482867 4859 scope.go:117] "RemoveContainer" containerID="de71f0b717129dba48f3f0bdd90240a8e45404a17f0df79d6812d58c9209feff" Oct 08 19:24:10 crc kubenswrapper[4859]: E1008 19:24:10.483975 4859 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-82s52_openshift-machine-config-operator(b23a6a6c-9d92-4e7b-840e-55cfda873a2d)\"" pod="openshift-machine-config-operator/machine-config-daemon-82s52" podUID="b23a6a6c-9d92-4e7b-840e-55cfda873a2d" Oct 08 19:24:13 crc kubenswrapper[4859]: I1008 19:24:13.481072 4859 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_barbican-api-588f98c7dd-c7l8h_3ecae548-eaa6-4c24-a538-b2a34fa0308f/barbican-api-log/0.log" Oct 08 19:24:13 crc kubenswrapper[4859]: I1008 19:24:13.503860 4859 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_barbican-api-588f98c7dd-c7l8h_3ecae548-eaa6-4c24-a538-b2a34fa0308f/barbican-api/0.log" Oct 08 19:24:13 crc kubenswrapper[4859]: I1008 19:24:13.703376 4859 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_barbican-keystone-listener-d798f6bcd-5zw8p_80660054-bd62-4742-a13d-5193b317b036/barbican-keystone-listener/0.log" Oct 08 19:24:13 crc kubenswrapper[4859]: I1008 19:24:13.722668 4859 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_barbican-keystone-listener-d798f6bcd-5zw8p_80660054-bd62-4742-a13d-5193b317b036/barbican-keystone-listener-log/0.log" Oct 08 19:24:13 crc kubenswrapper[4859]: I1008 19:24:13.915136 4859 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_barbican-worker-7849cf5875-6srbt_578aad66-818f-412f-9023-97ed6ae95de3/barbican-worker/0.log" Oct 08 19:24:13 crc kubenswrapper[4859]: I1008 19:24:13.924537 4859 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_barbican-worker-7849cf5875-6srbt_578aad66-818f-412f-9023-97ed6ae95de3/barbican-worker-log/0.log" Oct 08 19:24:14 crc kubenswrapper[4859]: I1008 19:24:14.161110 4859 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_bootstrap-edpm-deployment-openstack-edpm-ipam-nl8fp_e0bf2482-ef25-48bb-9491-282a3eabfec0/bootstrap-edpm-deployment-openstack-edpm-ipam/0.log" Oct 08 19:24:14 crc kubenswrapper[4859]: I1008 19:24:14.358363 4859 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ceilometer-0_53819555-42d9-40cb-9897-94c5496c8da9/ceilometer-central-agent/0.log" Oct 08 19:24:14 crc kubenswrapper[4859]: I1008 19:24:14.385384 4859 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ceilometer-0_53819555-42d9-40cb-9897-94c5496c8da9/ceilometer-notification-agent/0.log" Oct 08 19:24:14 crc kubenswrapper[4859]: I1008 19:24:14.434875 4859 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ceilometer-0_53819555-42d9-40cb-9897-94c5496c8da9/proxy-httpd/0.log" Oct 08 19:24:14 crc kubenswrapper[4859]: I1008 19:24:14.568322 4859 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ceilometer-0_53819555-42d9-40cb-9897-94c5496c8da9/sg-core/0.log" Oct 08 19:24:14 crc kubenswrapper[4859]: I1008 19:24:14.678233 4859 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_cinder-api-0_f8eeb421-2f6d-428f-93d5-5aca080785cc/cinder-api/0.log" Oct 08 19:24:14 crc kubenswrapper[4859]: I1008 19:24:14.791458 4859 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_cinder-api-0_f8eeb421-2f6d-428f-93d5-5aca080785cc/cinder-api-log/0.log" Oct 08 19:24:14 crc kubenswrapper[4859]: I1008 19:24:14.925127 4859 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_cinder-scheduler-0_f20f7045-6768-4824-9295-6e5ba74da387/cinder-scheduler/0.log" Oct 08 19:24:15 crc kubenswrapper[4859]: I1008 19:24:15.032249 4859 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_cinder-scheduler-0_f20f7045-6768-4824-9295-6e5ba74da387/probe/0.log" Oct 08 19:24:15 crc kubenswrapper[4859]: I1008 19:24:15.122246 4859 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_configure-network-edpm-deployment-openstack-edpm-ipam-6n7d4_70979534-d410-4768-95a9-5c9ff4ae1583/configure-network-edpm-deployment-openstack-edpm-ipam/0.log" Oct 08 19:24:15 crc kubenswrapper[4859]: I1008 19:24:15.228861 4859 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_configure-os-edpm-deployment-openstack-edpm-ipam-5dz4q_ff5a590b-fac3-4cf3-b472-db2f43030032/configure-os-edpm-deployment-openstack-edpm-ipam/0.log" Oct 08 19:24:15 crc kubenswrapper[4859]: I1008 19:24:15.452971 4859 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_configure-os-edpm-deployment-openstack-edpm-ipam-dnz8t_28feb067-adfe-4b48-aef3-321e18ac0e37/configure-os-edpm-deployment-openstack-edpm-ipam/0.log" Oct 08 19:24:15 crc kubenswrapper[4859]: I1008 19:24:15.550376 4859 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_dnsmasq-dns-6f586dc697-ncnrs_7f4d68a5-ccfe-48f2-a84f-e800d6718498/init/0.log" Oct 08 19:24:15 crc kubenswrapper[4859]: I1008 19:24:15.685358 4859 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_dnsmasq-dns-6f586dc697-ncnrs_7f4d68a5-ccfe-48f2-a84f-e800d6718498/init/0.log" Oct 08 19:24:15 crc kubenswrapper[4859]: I1008 19:24:15.739277 4859 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_dnsmasq-dns-6f586dc697-ncnrs_7f4d68a5-ccfe-48f2-a84f-e800d6718498/dnsmasq-dns/0.log" Oct 08 19:24:15 crc kubenswrapper[4859]: I1008 19:24:15.873971 4859 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_download-cache-edpm-deployment-openstack-edpm-ipam-hb4tw_2db15fd4-e39e-474c-b274-53262e92bccb/download-cache-edpm-deployment-openstack-edpm-ipam/0.log" Oct 08 19:24:15 crc kubenswrapper[4859]: I1008 19:24:15.953650 4859 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_glance-default-external-api-0_21a00512-7854-48c3-a040-795eb7d79744/glance-httpd/0.log" Oct 08 19:24:16 crc kubenswrapper[4859]: I1008 19:24:16.070899 4859 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_glance-default-external-api-0_21a00512-7854-48c3-a040-795eb7d79744/glance-log/0.log" Oct 08 19:24:16 crc kubenswrapper[4859]: I1008 19:24:16.185375 4859 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_glance-default-internal-api-0_d14936ce-b30c-44e7-95bc-9a4d516e71b1/glance-httpd/0.log" Oct 08 19:24:16 crc kubenswrapper[4859]: I1008 19:24:16.256710 4859 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_glance-default-internal-api-0_d14936ce-b30c-44e7-95bc-9a4d516e71b1/glance-log/0.log" Oct 08 19:24:16 crc kubenswrapper[4859]: I1008 19:24:16.489752 4859 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_horizon-6f569d5894-mx8v5_50b5de31-90f6-43cf-9e16-7b89b09f1e57/horizon/0.log" Oct 08 19:24:16 crc kubenswrapper[4859]: I1008 19:24:16.544712 4859 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_install-certs-edpm-deployment-openstack-edpm-ipam-w95ff_3b4b0004-0407-4671-a67e-0c14ef73e6ff/install-certs-edpm-deployment-openstack-edpm-ipam/0.log" Oct 08 19:24:16 crc kubenswrapper[4859]: I1008 19:24:16.746328 4859 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_install-os-edpm-deployment-openstack-edpm-ipam-twjcj_5d2f49c5-b079-4409-9ad7-e7c6b419e843/install-os-edpm-deployment-openstack-edpm-ipam/0.log" Oct 08 19:24:16 crc kubenswrapper[4859]: I1008 19:24:16.752291 4859 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_horizon-6f569d5894-mx8v5_50b5de31-90f6-43cf-9e16-7b89b09f1e57/horizon-log/0.log" Oct 08 19:24:16 crc kubenswrapper[4859]: I1008 19:24:16.949000 4859 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_keystone-cron-29332501-rxhgw_555ffac5-9b03-43eb-8610-fad69a164a42/keystone-cron/0.log" Oct 08 19:24:17 crc kubenswrapper[4859]: I1008 19:24:17.061636 4859 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_keystone-758ff8b98b-p6wm9_285bdd26-6109-4191-870a-2fde949a6a28/keystone-api/0.log" Oct 08 19:24:17 crc kubenswrapper[4859]: I1008 19:24:17.148155 4859 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_kube-state-metrics-0_3652de4f-2bd2-4fe2-b138-63dd20b42a61/kube-state-metrics/0.log" Oct 08 19:24:17 crc kubenswrapper[4859]: I1008 19:24:17.246610 4859 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_libvirt-edpm-deployment-openstack-edpm-ipam-rrrnw_a7cd7054-3a66-49ef-bcd5-405c815f8bfb/libvirt-edpm-deployment-openstack-edpm-ipam/0.log" Oct 08 19:24:17 crc kubenswrapper[4859]: I1008 19:24:17.593650 4859 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_neutron-5f9b9b6f45-wr5fx_f64840cf-331e-4d8d-b102-49225713df66/neutron-httpd/0.log" Oct 08 19:24:17 crc kubenswrapper[4859]: I1008 19:24:17.660114 4859 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_neutron-5f9b9b6f45-wr5fx_f64840cf-331e-4d8d-b102-49225713df66/neutron-api/0.log" Oct 08 19:24:18 crc kubenswrapper[4859]: I1008 19:24:18.388648 4859 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_neutron-metadata-edpm-deployment-openstack-edpm-ipam-fnxfs_3f1e49a1-2a8d-4d9e-9fde-529f8832d0dc/neutron-metadata-edpm-deployment-openstack-edpm-ipam/0.log" Oct 08 19:24:18 crc kubenswrapper[4859]: I1008 19:24:18.889858 4859 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-api-0_e56f2ed9-f7a3-4157-bb50-88fbd86d9c60/nova-api-log/0.log" Oct 08 19:24:19 crc kubenswrapper[4859]: I1008 19:24:19.160382 4859 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-cell0-conductor-0_8c9e1318-b507-4caa-832f-ea7c02f7c199/nova-cell0-conductor-conductor/0.log" Oct 08 19:24:19 crc kubenswrapper[4859]: I1008 19:24:19.207016 4859 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-api-0_e56f2ed9-f7a3-4157-bb50-88fbd86d9c60/nova-api-api/0.log" Oct 08 19:24:19 crc kubenswrapper[4859]: I1008 19:24:19.515551 4859 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-cell1-conductor-0_994d31f2-ad55-4817-a2e5-4e3743462cae/nova-cell1-conductor-conductor/0.log" Oct 08 19:24:19 crc kubenswrapper[4859]: I1008 19:24:19.576985 4859 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-cell1-novncproxy-0_f2fffadc-cd8a-4481-aafe-8f7e2dd338cf/nova-cell1-novncproxy-novncproxy/0.log" Oct 08 19:24:19 crc kubenswrapper[4859]: I1008 19:24:19.886616 4859 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-edpm-deployment-openstack-edpm-ipam-sg6sh_4035f4f1-5231-4112-89f6-d9d1c7c7eeb1/nova-edpm-deployment-openstack-edpm-ipam/0.log" Oct 08 19:24:20 crc kubenswrapper[4859]: I1008 19:24:20.172352 4859 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-metadata-0_6188ecfc-2b44-443d-a966-8b91ae535533/nova-metadata-log/0.log" Oct 08 19:24:20 crc kubenswrapper[4859]: I1008 19:24:20.597502 4859 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_openstack-cell1-galera-0_a2c834dc-2b23-4081-8c71-e7ae462ca063/mysql-bootstrap/0.log" Oct 08 19:24:20 crc kubenswrapper[4859]: I1008 19:24:20.650348 4859 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-scheduler-0_70906bc9-3b52-4436-8d68-cdb37f0d2478/nova-scheduler-scheduler/0.log" Oct 08 19:24:20 crc kubenswrapper[4859]: I1008 19:24:20.841714 4859 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_openstack-cell1-galera-0_a2c834dc-2b23-4081-8c71-e7ae462ca063/mysql-bootstrap/0.log" Oct 08 19:24:20 crc kubenswrapper[4859]: I1008 19:24:20.902454 4859 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_openstack-cell1-galera-0_a2c834dc-2b23-4081-8c71-e7ae462ca063/galera/0.log" Oct 08 19:24:21 crc kubenswrapper[4859]: I1008 19:24:21.151397 4859 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_openstack-galera-0_f1828698-e403-4643-a9ab-aedf302aae70/mysql-bootstrap/0.log" Oct 08 19:24:21 crc kubenswrapper[4859]: I1008 19:24:21.343957 4859 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_openstack-galera-0_f1828698-e403-4643-a9ab-aedf302aae70/mysql-bootstrap/0.log" Oct 08 19:24:21 crc kubenswrapper[4859]: I1008 19:24:21.360530 4859 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_openstack-galera-0_f1828698-e403-4643-a9ab-aedf302aae70/galera/0.log" Oct 08 19:24:21 crc kubenswrapper[4859]: I1008 19:24:21.504030 4859 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_openstackclient_1a631d33-4c0c-46c9-a3d5-4d9dc9f6de84/openstackclient/0.log" Oct 08 19:24:21 crc kubenswrapper[4859]: I1008 19:24:21.532621 4859 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-metadata-0_6188ecfc-2b44-443d-a966-8b91ae535533/nova-metadata-metadata/0.log" Oct 08 19:24:21 crc kubenswrapper[4859]: I1008 19:24:21.722522 4859 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-controller-b89lv_ec20820a-066c-4606-8f1b-a26b172b7720/ovn-controller/0.log" Oct 08 19:24:21 crc kubenswrapper[4859]: I1008 19:24:21.962120 4859 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-controller-metrics-bjb7d_f29e1ba4-3f25-45b7-ba3a-3592ce322840/openstack-network-exporter/0.log" Oct 08 19:24:22 crc kubenswrapper[4859]: I1008 19:24:22.114856 4859 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-controller-ovs-vthdk_eca01cc4-b7ef-429c-984d-1ccd1d95bcd2/ovsdb-server-init/0.log" Oct 08 19:24:22 crc kubenswrapper[4859]: I1008 19:24:22.318330 4859 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-controller-ovs-vthdk_eca01cc4-b7ef-429c-984d-1ccd1d95bcd2/ovsdb-server-init/0.log" Oct 08 19:24:22 crc kubenswrapper[4859]: I1008 19:24:22.318510 4859 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-controller-ovs-vthdk_eca01cc4-b7ef-429c-984d-1ccd1d95bcd2/ovs-vswitchd/0.log" Oct 08 19:24:22 crc kubenswrapper[4859]: I1008 19:24:22.327790 4859 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-controller-ovs-vthdk_eca01cc4-b7ef-429c-984d-1ccd1d95bcd2/ovsdb-server/0.log" Oct 08 19:24:22 crc kubenswrapper[4859]: I1008 19:24:22.572418 4859 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-edpm-deployment-openstack-edpm-ipam-gtbjg_4bbd12de-a0ea-4741-b6af-dc19464d2161/ovn-edpm-deployment-openstack-edpm-ipam/0.log" Oct 08 19:24:22 crc kubenswrapper[4859]: I1008 19:24:22.753623 4859 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-northd-0_3d31b043-3252-458e-8f23-afd68b624849/openstack-network-exporter/0.log" Oct 08 19:24:22 crc kubenswrapper[4859]: I1008 19:24:22.753800 4859 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-northd-0_3d31b043-3252-458e-8f23-afd68b624849/ovn-northd/0.log" Oct 08 19:24:22 crc kubenswrapper[4859]: I1008 19:24:22.969976 4859 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovsdbserver-nb-0_9fc7c8d2-a890-43c0-8081-6049a57c7474/openstack-network-exporter/0.log" Oct 08 19:24:22 crc kubenswrapper[4859]: I1008 19:24:22.986237 4859 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovsdbserver-nb-0_9fc7c8d2-a890-43c0-8081-6049a57c7474/ovsdbserver-nb/0.log" Oct 08 19:24:23 crc kubenswrapper[4859]: I1008 19:24:23.134652 4859 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovsdbserver-sb-0_fb46e63a-80c0-4141-8e6a-ad49c95bc924/openstack-network-exporter/0.log" Oct 08 19:24:23 crc kubenswrapper[4859]: I1008 19:24:23.187816 4859 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovsdbserver-sb-0_fb46e63a-80c0-4141-8e6a-ad49c95bc924/ovsdbserver-sb/0.log" Oct 08 19:24:23 crc kubenswrapper[4859]: I1008 19:24:23.410427 4859 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_placement-58c5f664f4-5lvrs_01262af8-82d8-4771-9fe8-472dc77e4b60/placement-api/0.log" Oct 08 19:24:23 crc kubenswrapper[4859]: I1008 19:24:23.538656 4859 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_placement-58c5f664f4-5lvrs_01262af8-82d8-4771-9fe8-472dc77e4b60/placement-log/0.log" Oct 08 19:24:23 crc kubenswrapper[4859]: I1008 19:24:23.601804 4859 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_rabbitmq-cell1-server-0_2c6db854-3378-4bc4-b622-480bc8c42bf5/setup-container/0.log" Oct 08 19:24:23 crc kubenswrapper[4859]: I1008 19:24:23.765354 4859 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_rabbitmq-cell1-server-0_2c6db854-3378-4bc4-b622-480bc8c42bf5/setup-container/0.log" Oct 08 19:24:23 crc kubenswrapper[4859]: I1008 19:24:23.793147 4859 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_rabbitmq-cell1-server-0_2c6db854-3378-4bc4-b622-480bc8c42bf5/rabbitmq/0.log" Oct 08 19:24:24 crc kubenswrapper[4859]: I1008 19:24:24.000536 4859 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_rabbitmq-server-0_94450380-649a-4476-b646-018ed24aa703/setup-container/0.log" Oct 08 19:24:24 crc kubenswrapper[4859]: I1008 19:24:24.173400 4859 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_rabbitmq-server-0_94450380-649a-4476-b646-018ed24aa703/setup-container/0.log" Oct 08 19:24:24 crc kubenswrapper[4859]: I1008 19:24:24.196232 4859 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_rabbitmq-server-0_94450380-649a-4476-b646-018ed24aa703/rabbitmq/0.log" Oct 08 19:24:24 crc kubenswrapper[4859]: I1008 19:24:24.377042 4859 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_reboot-os-edpm-deployment-openstack-edpm-ipam-cjdnl_5dc6ec8a-d555-43e6-b295-821ae62d0eb1/reboot-os-edpm-deployment-openstack-edpm-ipam/0.log" Oct 08 19:24:24 crc kubenswrapper[4859]: I1008 19:24:24.464369 4859 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_redhat-edpm-deployment-openstack-edpm-ipam-fz6lm_02a2d3c9-f778-4465-b0aa-62c051299f2b/redhat-edpm-deployment-openstack-edpm-ipam/0.log" Oct 08 19:24:24 crc kubenswrapper[4859]: I1008 19:24:24.469537 4859 scope.go:117] "RemoveContainer" containerID="de71f0b717129dba48f3f0bdd90240a8e45404a17f0df79d6812d58c9209feff" Oct 08 19:24:24 crc kubenswrapper[4859]: E1008 19:24:24.469848 4859 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-82s52_openshift-machine-config-operator(b23a6a6c-9d92-4e7b-840e-55cfda873a2d)\"" pod="openshift-machine-config-operator/machine-config-daemon-82s52" podUID="b23a6a6c-9d92-4e7b-840e-55cfda873a2d" Oct 08 19:24:24 crc kubenswrapper[4859]: I1008 19:24:24.682607 4859 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_repo-setup-edpm-deployment-openstack-edpm-ipam-n8rrq_53d801a5-a432-44f3-970a-23519033d1e3/repo-setup-edpm-deployment-openstack-edpm-ipam/0.log" Oct 08 19:24:24 crc kubenswrapper[4859]: I1008 19:24:24.870141 4859 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_run-os-edpm-deployment-openstack-edpm-ipam-746dn_e37b6e90-4b9a-41e8-9a83-db72281cc7f5/run-os-edpm-deployment-openstack-edpm-ipam/0.log" Oct 08 19:24:25 crc kubenswrapper[4859]: I1008 19:24:25.022338 4859 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ssh-known-hosts-edpm-deployment-f546c_252ee072-dfc8-4b4b-a8dc-c695f2be633a/ssh-known-hosts-edpm-deployment/0.log" Oct 08 19:24:25 crc kubenswrapper[4859]: I1008 19:24:25.279612 4859 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-proxy-776bbffb49-bdzz5_465dc520-34ae-41b8-869c-5d4ead3b16ab/proxy-server/0.log" Oct 08 19:24:25 crc kubenswrapper[4859]: I1008 19:24:25.378273 4859 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-proxy-776bbffb49-bdzz5_465dc520-34ae-41b8-869c-5d4ead3b16ab/proxy-httpd/0.log" Oct 08 19:24:25 crc kubenswrapper[4859]: I1008 19:24:25.462811 4859 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-ring-rebalance-qrpd6_431b4e6e-415d-4c14-8783-f94409691ae3/swift-ring-rebalance/0.log" Oct 08 19:24:25 crc kubenswrapper[4859]: I1008 19:24:25.598457 4859 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_c860866f-d309-4824-a31c-e88c5d820712/account-auditor/0.log" Oct 08 19:24:25 crc kubenswrapper[4859]: I1008 19:24:25.697044 4859 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_c860866f-d309-4824-a31c-e88c5d820712/account-reaper/0.log" Oct 08 19:24:25 crc kubenswrapper[4859]: I1008 19:24:25.778314 4859 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_c860866f-d309-4824-a31c-e88c5d820712/account-server/0.log" Oct 08 19:24:25 crc kubenswrapper[4859]: I1008 19:24:25.837546 4859 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_c860866f-d309-4824-a31c-e88c5d820712/account-replicator/0.log" Oct 08 19:24:25 crc kubenswrapper[4859]: I1008 19:24:25.901629 4859 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_c860866f-d309-4824-a31c-e88c5d820712/container-auditor/0.log" Oct 08 19:24:26 crc kubenswrapper[4859]: I1008 19:24:26.032498 4859 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_c860866f-d309-4824-a31c-e88c5d820712/container-replicator/0.log" Oct 08 19:24:26 crc kubenswrapper[4859]: I1008 19:24:26.047093 4859 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_c860866f-d309-4824-a31c-e88c5d820712/container-server/0.log" Oct 08 19:24:26 crc kubenswrapper[4859]: I1008 19:24:26.129322 4859 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_c860866f-d309-4824-a31c-e88c5d820712/container-updater/0.log" Oct 08 19:24:26 crc kubenswrapper[4859]: I1008 19:24:26.235218 4859 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_c860866f-d309-4824-a31c-e88c5d820712/object-auditor/0.log" Oct 08 19:24:26 crc kubenswrapper[4859]: I1008 19:24:26.251971 4859 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_c860866f-d309-4824-a31c-e88c5d820712/object-expirer/0.log" Oct 08 19:24:26 crc kubenswrapper[4859]: I1008 19:24:26.389093 4859 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_c860866f-d309-4824-a31c-e88c5d820712/object-replicator/0.log" Oct 08 19:24:26 crc kubenswrapper[4859]: I1008 19:24:26.429804 4859 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_c860866f-d309-4824-a31c-e88c5d820712/object-server/0.log" Oct 08 19:24:26 crc kubenswrapper[4859]: I1008 19:24:26.516395 4859 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_c860866f-d309-4824-a31c-e88c5d820712/object-updater/0.log" Oct 08 19:24:26 crc kubenswrapper[4859]: I1008 19:24:26.567543 4859 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_c860866f-d309-4824-a31c-e88c5d820712/rsync/0.log" Oct 08 19:24:26 crc kubenswrapper[4859]: I1008 19:24:26.634629 4859 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_c860866f-d309-4824-a31c-e88c5d820712/swift-recon-cron/0.log" Oct 08 19:24:27 crc kubenswrapper[4859]: I1008 19:24:27.321729 4859 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_tempest-tests-tempest_be909489-e6c5-4464-9292-aacd831885d7/tempest-tests-tempest-tests-runner/0.log" Oct 08 19:24:27 crc kubenswrapper[4859]: I1008 19:24:27.350033 4859 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_telemetry-edpm-deployment-openstack-edpm-ipam-twhvj_65244148-7971-47bf-b900-96932c6cd473/telemetry-edpm-deployment-openstack-edpm-ipam/0.log" Oct 08 19:24:27 crc kubenswrapper[4859]: I1008 19:24:27.501579 4859 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_test-operator-logs-pod-tempest-tempest-tests-tempest_f9597f73-13f2-4547-862b-fa23bba1c625/test-operator-logs-container/0.log" Oct 08 19:24:27 crc kubenswrapper[4859]: I1008 19:24:27.713165 4859 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_validate-network-edpm-deployment-openstack-edpm-ipam-2lvj9_67e6fffe-7caa-42a6-812c-efa6a8f2162b/validate-network-edpm-deployment-openstack-edpm-ipam/0.log" Oct 08 19:24:35 crc kubenswrapper[4859]: I1008 19:24:35.147033 4859 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_memcached-0_6dc3fa5d-109b-4140-93aa-61b855c46df2/memcached/0.log" Oct 08 19:24:37 crc kubenswrapper[4859]: I1008 19:24:37.469953 4859 scope.go:117] "RemoveContainer" containerID="de71f0b717129dba48f3f0bdd90240a8e45404a17f0df79d6812d58c9209feff" Oct 08 19:24:37 crc kubenswrapper[4859]: E1008 19:24:37.470872 4859 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-82s52_openshift-machine-config-operator(b23a6a6c-9d92-4e7b-840e-55cfda873a2d)\"" pod="openshift-machine-config-operator/machine-config-daemon-82s52" podUID="b23a6a6c-9d92-4e7b-840e-55cfda873a2d" Oct 08 19:24:52 crc kubenswrapper[4859]: I1008 19:24:52.470854 4859 scope.go:117] "RemoveContainer" containerID="de71f0b717129dba48f3f0bdd90240a8e45404a17f0df79d6812d58c9209feff" Oct 08 19:24:52 crc kubenswrapper[4859]: E1008 19:24:52.471825 4859 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-82s52_openshift-machine-config-operator(b23a6a6c-9d92-4e7b-840e-55cfda873a2d)\"" pod="openshift-machine-config-operator/machine-config-daemon-82s52" podUID="b23a6a6c-9d92-4e7b-840e-55cfda873a2d" Oct 08 19:24:58 crc kubenswrapper[4859]: I1008 19:24:58.524214 4859 generic.go:334] "Generic (PLEG): container finished" podID="28e20e5f-7566-4eea-9814-befdc2fa7535" containerID="3c1d50de8a22d6fd32dca7f0714e90e74ac2d271254e43f32073a96aeedf5a03" exitCode=0 Oct 08 19:24:58 crc kubenswrapper[4859]: I1008 19:24:58.524315 4859 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-4fzzd/crc-debug-mjvd7" event={"ID":"28e20e5f-7566-4eea-9814-befdc2fa7535","Type":"ContainerDied","Data":"3c1d50de8a22d6fd32dca7f0714e90e74ac2d271254e43f32073a96aeedf5a03"} Oct 08 19:24:59 crc kubenswrapper[4859]: I1008 19:24:59.974029 4859 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-4fzzd/crc-debug-mjvd7" Oct 08 19:25:00 crc kubenswrapper[4859]: I1008 19:25:00.013842 4859 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-must-gather-4fzzd/crc-debug-mjvd7"] Oct 08 19:25:00 crc kubenswrapper[4859]: I1008 19:25:00.023261 4859 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-must-gather-4fzzd/crc-debug-mjvd7"] Oct 08 19:25:00 crc kubenswrapper[4859]: I1008 19:25:00.114416 4859 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/28e20e5f-7566-4eea-9814-befdc2fa7535-host\") pod \"28e20e5f-7566-4eea-9814-befdc2fa7535\" (UID: \"28e20e5f-7566-4eea-9814-befdc2fa7535\") " Oct 08 19:25:00 crc kubenswrapper[4859]: I1008 19:25:00.114664 4859 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-kcdmk\" (UniqueName: \"kubernetes.io/projected/28e20e5f-7566-4eea-9814-befdc2fa7535-kube-api-access-kcdmk\") pod \"28e20e5f-7566-4eea-9814-befdc2fa7535\" (UID: \"28e20e5f-7566-4eea-9814-befdc2fa7535\") " Oct 08 19:25:00 crc kubenswrapper[4859]: I1008 19:25:00.114654 4859 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/28e20e5f-7566-4eea-9814-befdc2fa7535-host" (OuterVolumeSpecName: "host") pod "28e20e5f-7566-4eea-9814-befdc2fa7535" (UID: "28e20e5f-7566-4eea-9814-befdc2fa7535"). InnerVolumeSpecName "host". PluginName "kubernetes.io/host-path", VolumeGidValue "" Oct 08 19:25:00 crc kubenswrapper[4859]: I1008 19:25:00.115244 4859 reconciler_common.go:293] "Volume detached for volume \"host\" (UniqueName: \"kubernetes.io/host-path/28e20e5f-7566-4eea-9814-befdc2fa7535-host\") on node \"crc\" DevicePath \"\"" Oct 08 19:25:00 crc kubenswrapper[4859]: I1008 19:25:00.125582 4859 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/28e20e5f-7566-4eea-9814-befdc2fa7535-kube-api-access-kcdmk" (OuterVolumeSpecName: "kube-api-access-kcdmk") pod "28e20e5f-7566-4eea-9814-befdc2fa7535" (UID: "28e20e5f-7566-4eea-9814-befdc2fa7535"). InnerVolumeSpecName "kube-api-access-kcdmk". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 08 19:25:00 crc kubenswrapper[4859]: I1008 19:25:00.217144 4859 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-kcdmk\" (UniqueName: \"kubernetes.io/projected/28e20e5f-7566-4eea-9814-befdc2fa7535-kube-api-access-kcdmk\") on node \"crc\" DevicePath \"\"" Oct 08 19:25:00 crc kubenswrapper[4859]: I1008 19:25:00.483740 4859 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="28e20e5f-7566-4eea-9814-befdc2fa7535" path="/var/lib/kubelet/pods/28e20e5f-7566-4eea-9814-befdc2fa7535/volumes" Oct 08 19:25:00 crc kubenswrapper[4859]: I1008 19:25:00.554370 4859 scope.go:117] "RemoveContainer" containerID="3c1d50de8a22d6fd32dca7f0714e90e74ac2d271254e43f32073a96aeedf5a03" Oct 08 19:25:00 crc kubenswrapper[4859]: I1008 19:25:00.554521 4859 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-4fzzd/crc-debug-mjvd7" Oct 08 19:25:01 crc kubenswrapper[4859]: I1008 19:25:01.206146 4859 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-must-gather-4fzzd/crc-debug-gq474"] Oct 08 19:25:01 crc kubenswrapper[4859]: E1008 19:25:01.207547 4859 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="28e20e5f-7566-4eea-9814-befdc2fa7535" containerName="container-00" Oct 08 19:25:01 crc kubenswrapper[4859]: I1008 19:25:01.207640 4859 state_mem.go:107] "Deleted CPUSet assignment" podUID="28e20e5f-7566-4eea-9814-befdc2fa7535" containerName="container-00" Oct 08 19:25:01 crc kubenswrapper[4859]: I1008 19:25:01.208002 4859 memory_manager.go:354] "RemoveStaleState removing state" podUID="28e20e5f-7566-4eea-9814-befdc2fa7535" containerName="container-00" Oct 08 19:25:01 crc kubenswrapper[4859]: I1008 19:25:01.208868 4859 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-4fzzd/crc-debug-gq474" Oct 08 19:25:01 crc kubenswrapper[4859]: I1008 19:25:01.341473 4859 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/f1e2726d-3cf5-4427-af70-b68390aa35f0-host\") pod \"crc-debug-gq474\" (UID: \"f1e2726d-3cf5-4427-af70-b68390aa35f0\") " pod="openshift-must-gather-4fzzd/crc-debug-gq474" Oct 08 19:25:01 crc kubenswrapper[4859]: I1008 19:25:01.341618 4859 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-4xchb\" (UniqueName: \"kubernetes.io/projected/f1e2726d-3cf5-4427-af70-b68390aa35f0-kube-api-access-4xchb\") pod \"crc-debug-gq474\" (UID: \"f1e2726d-3cf5-4427-af70-b68390aa35f0\") " pod="openshift-must-gather-4fzzd/crc-debug-gq474" Oct 08 19:25:01 crc kubenswrapper[4859]: I1008 19:25:01.443340 4859 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/f1e2726d-3cf5-4427-af70-b68390aa35f0-host\") pod \"crc-debug-gq474\" (UID: \"f1e2726d-3cf5-4427-af70-b68390aa35f0\") " pod="openshift-must-gather-4fzzd/crc-debug-gq474" Oct 08 19:25:01 crc kubenswrapper[4859]: I1008 19:25:01.443420 4859 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-4xchb\" (UniqueName: \"kubernetes.io/projected/f1e2726d-3cf5-4427-af70-b68390aa35f0-kube-api-access-4xchb\") pod \"crc-debug-gq474\" (UID: \"f1e2726d-3cf5-4427-af70-b68390aa35f0\") " pod="openshift-must-gather-4fzzd/crc-debug-gq474" Oct 08 19:25:01 crc kubenswrapper[4859]: I1008 19:25:01.443473 4859 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host\" (UniqueName: \"kubernetes.io/host-path/f1e2726d-3cf5-4427-af70-b68390aa35f0-host\") pod \"crc-debug-gq474\" (UID: \"f1e2726d-3cf5-4427-af70-b68390aa35f0\") " pod="openshift-must-gather-4fzzd/crc-debug-gq474" Oct 08 19:25:01 crc kubenswrapper[4859]: I1008 19:25:01.468138 4859 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-4xchb\" (UniqueName: \"kubernetes.io/projected/f1e2726d-3cf5-4427-af70-b68390aa35f0-kube-api-access-4xchb\") pod \"crc-debug-gq474\" (UID: \"f1e2726d-3cf5-4427-af70-b68390aa35f0\") " pod="openshift-must-gather-4fzzd/crc-debug-gq474" Oct 08 19:25:01 crc kubenswrapper[4859]: I1008 19:25:01.524580 4859 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-4fzzd/crc-debug-gq474" Oct 08 19:25:01 crc kubenswrapper[4859]: W1008 19:25:01.564949 4859 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podf1e2726d_3cf5_4427_af70_b68390aa35f0.slice/crio-62083cbab235606e43ace181814fc9730ba60131682a0e13ceb72f6c07fd1b25 WatchSource:0}: Error finding container 62083cbab235606e43ace181814fc9730ba60131682a0e13ceb72f6c07fd1b25: Status 404 returned error can't find the container with id 62083cbab235606e43ace181814fc9730ba60131682a0e13ceb72f6c07fd1b25 Oct 08 19:25:02 crc kubenswrapper[4859]: I1008 19:25:02.577990 4859 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-4fzzd/crc-debug-gq474" event={"ID":"f1e2726d-3cf5-4427-af70-b68390aa35f0","Type":"ContainerStarted","Data":"30d34ee96e0138d3b4b0e028a347ca7c37500ad726355f082bcf94763c5e952b"} Oct 08 19:25:02 crc kubenswrapper[4859]: I1008 19:25:02.578403 4859 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-4fzzd/crc-debug-gq474" event={"ID":"f1e2726d-3cf5-4427-af70-b68390aa35f0","Type":"ContainerStarted","Data":"62083cbab235606e43ace181814fc9730ba60131682a0e13ceb72f6c07fd1b25"} Oct 08 19:25:02 crc kubenswrapper[4859]: I1008 19:25:02.599374 4859 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-must-gather-4fzzd/crc-debug-gq474" podStartSLOduration=1.599350807 podStartE2EDuration="1.599350807s" podCreationTimestamp="2025-10-08 19:25:01 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-08 19:25:02.59501279 +0000 UTC m=+4072.841852219" watchObservedRunningTime="2025-10-08 19:25:02.599350807 +0000 UTC m=+4072.846190176" Oct 08 19:25:02 crc kubenswrapper[4859]: E1008 19:25:02.639859 4859 log.go:32] "ReopenContainerLog from runtime service failed" err="rpc error: code = Unknown desc = container is not running" containerID="30d34ee96e0138d3b4b0e028a347ca7c37500ad726355f082bcf94763c5e952b" Oct 08 19:25:02 crc kubenswrapper[4859]: E1008 19:25:02.639954 4859 container_log_manager.go:307] "Failed to rotate log for container" err="failed to rotate log \"/var/log/pods/openshift-must-gather-4fzzd_crc-debug-gq474_f1e2726d-3cf5-4427-af70-b68390aa35f0/container-00/0.log\": failed to reopen container log \"30d34ee96e0138d3b4b0e028a347ca7c37500ad726355f082bcf94763c5e952b\": rpc error: code = Unknown desc = container is not running" worker=1 containerID="30d34ee96e0138d3b4b0e028a347ca7c37500ad726355f082bcf94763c5e952b" path="/var/log/pods/openshift-must-gather-4fzzd_crc-debug-gq474_f1e2726d-3cf5-4427-af70-b68390aa35f0/container-00/0.log" currentSize=53123022 maxSize=52428800 Oct 08 19:25:03 crc kubenswrapper[4859]: I1008 19:25:03.469770 4859 scope.go:117] "RemoveContainer" containerID="de71f0b717129dba48f3f0bdd90240a8e45404a17f0df79d6812d58c9209feff" Oct 08 19:25:03 crc kubenswrapper[4859]: E1008 19:25:03.470470 4859 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-82s52_openshift-machine-config-operator(b23a6a6c-9d92-4e7b-840e-55cfda873a2d)\"" pod="openshift-machine-config-operator/machine-config-daemon-82s52" podUID="b23a6a6c-9d92-4e7b-840e-55cfda873a2d" Oct 08 19:25:03 crc kubenswrapper[4859]: I1008 19:25:03.586934 4859 generic.go:334] "Generic (PLEG): container finished" podID="f1e2726d-3cf5-4427-af70-b68390aa35f0" containerID="30d34ee96e0138d3b4b0e028a347ca7c37500ad726355f082bcf94763c5e952b" exitCode=0 Oct 08 19:25:03 crc kubenswrapper[4859]: I1008 19:25:03.586983 4859 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-4fzzd/crc-debug-gq474" event={"ID":"f1e2726d-3cf5-4427-af70-b68390aa35f0","Type":"ContainerDied","Data":"30d34ee96e0138d3b4b0e028a347ca7c37500ad726355f082bcf94763c5e952b"} Oct 08 19:25:04 crc kubenswrapper[4859]: I1008 19:25:04.705246 4859 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-4fzzd/crc-debug-gq474" Oct 08 19:25:04 crc kubenswrapper[4859]: I1008 19:25:04.795193 4859 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/f1e2726d-3cf5-4427-af70-b68390aa35f0-host\") pod \"f1e2726d-3cf5-4427-af70-b68390aa35f0\" (UID: \"f1e2726d-3cf5-4427-af70-b68390aa35f0\") " Oct 08 19:25:04 crc kubenswrapper[4859]: I1008 19:25:04.795841 4859 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/f1e2726d-3cf5-4427-af70-b68390aa35f0-host" (OuterVolumeSpecName: "host") pod "f1e2726d-3cf5-4427-af70-b68390aa35f0" (UID: "f1e2726d-3cf5-4427-af70-b68390aa35f0"). InnerVolumeSpecName "host". PluginName "kubernetes.io/host-path", VolumeGidValue "" Oct 08 19:25:04 crc kubenswrapper[4859]: I1008 19:25:04.796503 4859 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-4xchb\" (UniqueName: \"kubernetes.io/projected/f1e2726d-3cf5-4427-af70-b68390aa35f0-kube-api-access-4xchb\") pod \"f1e2726d-3cf5-4427-af70-b68390aa35f0\" (UID: \"f1e2726d-3cf5-4427-af70-b68390aa35f0\") " Oct 08 19:25:04 crc kubenswrapper[4859]: I1008 19:25:04.797222 4859 reconciler_common.go:293] "Volume detached for volume \"host\" (UniqueName: \"kubernetes.io/host-path/f1e2726d-3cf5-4427-af70-b68390aa35f0-host\") on node \"crc\" DevicePath \"\"" Oct 08 19:25:04 crc kubenswrapper[4859]: I1008 19:25:04.802564 4859 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/f1e2726d-3cf5-4427-af70-b68390aa35f0-kube-api-access-4xchb" (OuterVolumeSpecName: "kube-api-access-4xchb") pod "f1e2726d-3cf5-4427-af70-b68390aa35f0" (UID: "f1e2726d-3cf5-4427-af70-b68390aa35f0"). InnerVolumeSpecName "kube-api-access-4xchb". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 08 19:25:04 crc kubenswrapper[4859]: I1008 19:25:04.898502 4859 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-4xchb\" (UniqueName: \"kubernetes.io/projected/f1e2726d-3cf5-4427-af70-b68390aa35f0-kube-api-access-4xchb\") on node \"crc\" DevicePath \"\"" Oct 08 19:25:05 crc kubenswrapper[4859]: I1008 19:25:05.605395 4859 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-4fzzd/crc-debug-gq474" event={"ID":"f1e2726d-3cf5-4427-af70-b68390aa35f0","Type":"ContainerDied","Data":"62083cbab235606e43ace181814fc9730ba60131682a0e13ceb72f6c07fd1b25"} Oct 08 19:25:05 crc kubenswrapper[4859]: I1008 19:25:05.605675 4859 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-4fzzd/crc-debug-gq474" Oct 08 19:25:05 crc kubenswrapper[4859]: I1008 19:25:05.605716 4859 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="62083cbab235606e43ace181814fc9730ba60131682a0e13ceb72f6c07fd1b25" Oct 08 19:25:09 crc kubenswrapper[4859]: I1008 19:25:09.449066 4859 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-must-gather-4fzzd/crc-debug-gq474"] Oct 08 19:25:09 crc kubenswrapper[4859]: I1008 19:25:09.456385 4859 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-must-gather-4fzzd/crc-debug-gq474"] Oct 08 19:25:10 crc kubenswrapper[4859]: I1008 19:25:10.481518 4859 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="f1e2726d-3cf5-4427-af70-b68390aa35f0" path="/var/lib/kubelet/pods/f1e2726d-3cf5-4427-af70-b68390aa35f0/volumes" Oct 08 19:25:10 crc kubenswrapper[4859]: I1008 19:25:10.645419 4859 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-must-gather-4fzzd/crc-debug-cs8l7"] Oct 08 19:25:10 crc kubenswrapper[4859]: E1008 19:25:10.645991 4859 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f1e2726d-3cf5-4427-af70-b68390aa35f0" containerName="container-00" Oct 08 19:25:10 crc kubenswrapper[4859]: I1008 19:25:10.646018 4859 state_mem.go:107] "Deleted CPUSet assignment" podUID="f1e2726d-3cf5-4427-af70-b68390aa35f0" containerName="container-00" Oct 08 19:25:10 crc kubenswrapper[4859]: I1008 19:25:10.646303 4859 memory_manager.go:354] "RemoveStaleState removing state" podUID="f1e2726d-3cf5-4427-af70-b68390aa35f0" containerName="container-00" Oct 08 19:25:10 crc kubenswrapper[4859]: I1008 19:25:10.647059 4859 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-4fzzd/crc-debug-cs8l7" Oct 08 19:25:10 crc kubenswrapper[4859]: I1008 19:25:10.690730 4859 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/61c91bad-599a-4a6d-b03c-75a84f05c57b-host\") pod \"crc-debug-cs8l7\" (UID: \"61c91bad-599a-4a6d-b03c-75a84f05c57b\") " pod="openshift-must-gather-4fzzd/crc-debug-cs8l7" Oct 08 19:25:10 crc kubenswrapper[4859]: I1008 19:25:10.690875 4859 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-8rnjw\" (UniqueName: \"kubernetes.io/projected/61c91bad-599a-4a6d-b03c-75a84f05c57b-kube-api-access-8rnjw\") pod \"crc-debug-cs8l7\" (UID: \"61c91bad-599a-4a6d-b03c-75a84f05c57b\") " pod="openshift-must-gather-4fzzd/crc-debug-cs8l7" Oct 08 19:25:10 crc kubenswrapper[4859]: I1008 19:25:10.792961 4859 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/61c91bad-599a-4a6d-b03c-75a84f05c57b-host\") pod \"crc-debug-cs8l7\" (UID: \"61c91bad-599a-4a6d-b03c-75a84f05c57b\") " pod="openshift-must-gather-4fzzd/crc-debug-cs8l7" Oct 08 19:25:10 crc kubenswrapper[4859]: I1008 19:25:10.793073 4859 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-8rnjw\" (UniqueName: \"kubernetes.io/projected/61c91bad-599a-4a6d-b03c-75a84f05c57b-kube-api-access-8rnjw\") pod \"crc-debug-cs8l7\" (UID: \"61c91bad-599a-4a6d-b03c-75a84f05c57b\") " pod="openshift-must-gather-4fzzd/crc-debug-cs8l7" Oct 08 19:25:10 crc kubenswrapper[4859]: I1008 19:25:10.793151 4859 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host\" (UniqueName: \"kubernetes.io/host-path/61c91bad-599a-4a6d-b03c-75a84f05c57b-host\") pod \"crc-debug-cs8l7\" (UID: \"61c91bad-599a-4a6d-b03c-75a84f05c57b\") " pod="openshift-must-gather-4fzzd/crc-debug-cs8l7" Oct 08 19:25:10 crc kubenswrapper[4859]: I1008 19:25:10.820418 4859 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-8rnjw\" (UniqueName: \"kubernetes.io/projected/61c91bad-599a-4a6d-b03c-75a84f05c57b-kube-api-access-8rnjw\") pod \"crc-debug-cs8l7\" (UID: \"61c91bad-599a-4a6d-b03c-75a84f05c57b\") " pod="openshift-must-gather-4fzzd/crc-debug-cs8l7" Oct 08 19:25:10 crc kubenswrapper[4859]: I1008 19:25:10.969349 4859 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-4fzzd/crc-debug-cs8l7" Oct 08 19:25:11 crc kubenswrapper[4859]: I1008 19:25:11.672531 4859 generic.go:334] "Generic (PLEG): container finished" podID="61c91bad-599a-4a6d-b03c-75a84f05c57b" containerID="aaf963761b8db41768acdc9d71ed0ffa8fb05bf52873eb3419d847f552dab74a" exitCode=0 Oct 08 19:25:11 crc kubenswrapper[4859]: I1008 19:25:11.672993 4859 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-4fzzd/crc-debug-cs8l7" event={"ID":"61c91bad-599a-4a6d-b03c-75a84f05c57b","Type":"ContainerDied","Data":"aaf963761b8db41768acdc9d71ed0ffa8fb05bf52873eb3419d847f552dab74a"} Oct 08 19:25:11 crc kubenswrapper[4859]: I1008 19:25:11.673039 4859 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-4fzzd/crc-debug-cs8l7" event={"ID":"61c91bad-599a-4a6d-b03c-75a84f05c57b","Type":"ContainerStarted","Data":"fb0c5c5ca1a3ac32cba9d51ccf87f278e12eb22667b06b7620192daff952dfc3"} Oct 08 19:25:11 crc kubenswrapper[4859]: I1008 19:25:11.719431 4859 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-must-gather-4fzzd/crc-debug-cs8l7"] Oct 08 19:25:11 crc kubenswrapper[4859]: I1008 19:25:11.729412 4859 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-must-gather-4fzzd/crc-debug-cs8l7"] Oct 08 19:25:12 crc kubenswrapper[4859]: I1008 19:25:12.983122 4859 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-4fzzd/crc-debug-cs8l7" Oct 08 19:25:13 crc kubenswrapper[4859]: I1008 19:25:13.037032 4859 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/61c91bad-599a-4a6d-b03c-75a84f05c57b-host\") pod \"61c91bad-599a-4a6d-b03c-75a84f05c57b\" (UID: \"61c91bad-599a-4a6d-b03c-75a84f05c57b\") " Oct 08 19:25:13 crc kubenswrapper[4859]: I1008 19:25:13.037116 4859 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-8rnjw\" (UniqueName: \"kubernetes.io/projected/61c91bad-599a-4a6d-b03c-75a84f05c57b-kube-api-access-8rnjw\") pod \"61c91bad-599a-4a6d-b03c-75a84f05c57b\" (UID: \"61c91bad-599a-4a6d-b03c-75a84f05c57b\") " Oct 08 19:25:13 crc kubenswrapper[4859]: I1008 19:25:13.037166 4859 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/61c91bad-599a-4a6d-b03c-75a84f05c57b-host" (OuterVolumeSpecName: "host") pod "61c91bad-599a-4a6d-b03c-75a84f05c57b" (UID: "61c91bad-599a-4a6d-b03c-75a84f05c57b"). InnerVolumeSpecName "host". PluginName "kubernetes.io/host-path", VolumeGidValue "" Oct 08 19:25:13 crc kubenswrapper[4859]: I1008 19:25:13.037589 4859 reconciler_common.go:293] "Volume detached for volume \"host\" (UniqueName: \"kubernetes.io/host-path/61c91bad-599a-4a6d-b03c-75a84f05c57b-host\") on node \"crc\" DevicePath \"\"" Oct 08 19:25:13 crc kubenswrapper[4859]: I1008 19:25:13.044065 4859 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/61c91bad-599a-4a6d-b03c-75a84f05c57b-kube-api-access-8rnjw" (OuterVolumeSpecName: "kube-api-access-8rnjw") pod "61c91bad-599a-4a6d-b03c-75a84f05c57b" (UID: "61c91bad-599a-4a6d-b03c-75a84f05c57b"). InnerVolumeSpecName "kube-api-access-8rnjw". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 08 19:25:13 crc kubenswrapper[4859]: I1008 19:25:13.139508 4859 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-8rnjw\" (UniqueName: \"kubernetes.io/projected/61c91bad-599a-4a6d-b03c-75a84f05c57b-kube-api-access-8rnjw\") on node \"crc\" DevicePath \"\"" Oct 08 19:25:13 crc kubenswrapper[4859]: I1008 19:25:13.692500 4859 scope.go:117] "RemoveContainer" containerID="aaf963761b8db41768acdc9d71ed0ffa8fb05bf52873eb3419d847f552dab74a" Oct 08 19:25:13 crc kubenswrapper[4859]: I1008 19:25:13.692512 4859 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-4fzzd/crc-debug-cs8l7" Oct 08 19:25:13 crc kubenswrapper[4859]: I1008 19:25:13.776062 4859 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_1bfea55cf6f540853d0c0d919aed51b5fd6e2d105cc44a68cb64399f8bxp4tb_70914ba1-396a-4131-835f-817debc89338/util/0.log" Oct 08 19:25:13 crc kubenswrapper[4859]: I1008 19:25:13.984975 4859 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_1bfea55cf6f540853d0c0d919aed51b5fd6e2d105cc44a68cb64399f8bxp4tb_70914ba1-396a-4131-835f-817debc89338/pull/0.log" Oct 08 19:25:13 crc kubenswrapper[4859]: I1008 19:25:13.997602 4859 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_1bfea55cf6f540853d0c0d919aed51b5fd6e2d105cc44a68cb64399f8bxp4tb_70914ba1-396a-4131-835f-817debc89338/util/0.log" Oct 08 19:25:14 crc kubenswrapper[4859]: I1008 19:25:14.042227 4859 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_1bfea55cf6f540853d0c0d919aed51b5fd6e2d105cc44a68cb64399f8bxp4tb_70914ba1-396a-4131-835f-817debc89338/pull/0.log" Oct 08 19:25:14 crc kubenswrapper[4859]: I1008 19:25:14.189739 4859 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_1bfea55cf6f540853d0c0d919aed51b5fd6e2d105cc44a68cb64399f8bxp4tb_70914ba1-396a-4131-835f-817debc89338/pull/0.log" Oct 08 19:25:14 crc kubenswrapper[4859]: I1008 19:25:14.190971 4859 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_1bfea55cf6f540853d0c0d919aed51b5fd6e2d105cc44a68cb64399f8bxp4tb_70914ba1-396a-4131-835f-817debc89338/util/0.log" Oct 08 19:25:14 crc kubenswrapper[4859]: I1008 19:25:14.225747 4859 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_1bfea55cf6f540853d0c0d919aed51b5fd6e2d105cc44a68cb64399f8bxp4tb_70914ba1-396a-4131-835f-817debc89338/extract/0.log" Oct 08 19:25:14 crc kubenswrapper[4859]: I1008 19:25:14.372298 4859 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_barbican-operator-controller-manager-658bdf4b74-8bfst_d097cd11-b23b-4d38-80d7-5d85c257cef4/kube-rbac-proxy/0.log" Oct 08 19:25:14 crc kubenswrapper[4859]: I1008 19:25:14.461614 4859 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_barbican-operator-controller-manager-658bdf4b74-8bfst_d097cd11-b23b-4d38-80d7-5d85c257cef4/manager/0.log" Oct 08 19:25:14 crc kubenswrapper[4859]: I1008 19:25:14.462198 4859 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_cinder-operator-controller-manager-7b7fb68549-lbw64_e2b104df-4109-45ab-8a5a-3569da53caa0/kube-rbac-proxy/0.log" Oct 08 19:25:14 crc kubenswrapper[4859]: I1008 19:25:14.480551 4859 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="61c91bad-599a-4a6d-b03c-75a84f05c57b" path="/var/lib/kubelet/pods/61c91bad-599a-4a6d-b03c-75a84f05c57b/volumes" Oct 08 19:25:14 crc kubenswrapper[4859]: I1008 19:25:14.606607 4859 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_cinder-operator-controller-manager-7b7fb68549-lbw64_e2b104df-4109-45ab-8a5a-3569da53caa0/manager/0.log" Oct 08 19:25:14 crc kubenswrapper[4859]: I1008 19:25:14.686321 4859 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_designate-operator-controller-manager-85d5d9dd78-nhqkh_d9a50bbe-f28f-46d8-9266-fd54e44388a3/kube-rbac-proxy/0.log" Oct 08 19:25:14 crc kubenswrapper[4859]: I1008 19:25:14.689524 4859 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_designate-operator-controller-manager-85d5d9dd78-nhqkh_d9a50bbe-f28f-46d8-9266-fd54e44388a3/manager/0.log" Oct 08 19:25:14 crc kubenswrapper[4859]: I1008 19:25:14.795039 4859 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_glance-operator-controller-manager-84b9b84486-d9gmv_0b068d16-0072-41f8-ad01-4194fd428014/kube-rbac-proxy/0.log" Oct 08 19:25:14 crc kubenswrapper[4859]: I1008 19:25:14.938916 4859 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_glance-operator-controller-manager-84b9b84486-d9gmv_0b068d16-0072-41f8-ad01-4194fd428014/manager/0.log" Oct 08 19:25:14 crc kubenswrapper[4859]: I1008 19:25:14.967155 4859 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_heat-operator-controller-manager-858f76bbdd-b5bfz_19fa1d1d-0731-4c21-9faa-b1665ec4af36/kube-rbac-proxy/0.log" Oct 08 19:25:15 crc kubenswrapper[4859]: I1008 19:25:15.027868 4859 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_heat-operator-controller-manager-858f76bbdd-b5bfz_19fa1d1d-0731-4c21-9faa-b1665ec4af36/manager/0.log" Oct 08 19:25:15 crc kubenswrapper[4859]: I1008 19:25:15.105618 4859 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_horizon-operator-controller-manager-7ffbcb7588-86z4n_ca71782e-3569-4ac2-a178-239f0ae200d0/kube-rbac-proxy/0.log" Oct 08 19:25:15 crc kubenswrapper[4859]: I1008 19:25:15.147301 4859 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_horizon-operator-controller-manager-7ffbcb7588-86z4n_ca71782e-3569-4ac2-a178-239f0ae200d0/manager/0.log" Oct 08 19:25:15 crc kubenswrapper[4859]: I1008 19:25:15.245169 4859 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_infra-operator-controller-manager-656bcbd775-ctjz9_1a1cfd66-310a-4077-8020-df7e6e069c53/kube-rbac-proxy/0.log" Oct 08 19:25:15 crc kubenswrapper[4859]: I1008 19:25:15.455792 4859 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_ironic-operator-controller-manager-9c5c78d49-z9952_2b67febf-c4c8-4e48-9350-29dd496439b4/manager/0.log" Oct 08 19:25:15 crc kubenswrapper[4859]: I1008 19:25:15.492001 4859 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_ironic-operator-controller-manager-9c5c78d49-z9952_2b67febf-c4c8-4e48-9350-29dd496439b4/kube-rbac-proxy/0.log" Oct 08 19:25:15 crc kubenswrapper[4859]: I1008 19:25:15.517910 4859 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_infra-operator-controller-manager-656bcbd775-ctjz9_1a1cfd66-310a-4077-8020-df7e6e069c53/manager/0.log" Oct 08 19:25:15 crc kubenswrapper[4859]: I1008 19:25:15.674954 4859 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_keystone-operator-controller-manager-55b6b7c7b8-rsbgr_0228033e-c835-4618-9603-2c67cf9ce57d/kube-rbac-proxy/0.log" Oct 08 19:25:15 crc kubenswrapper[4859]: I1008 19:25:15.747667 4859 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_keystone-operator-controller-manager-55b6b7c7b8-rsbgr_0228033e-c835-4618-9603-2c67cf9ce57d/manager/0.log" Oct 08 19:25:15 crc kubenswrapper[4859]: I1008 19:25:15.776575 4859 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_manila-operator-controller-manager-5f67fbc655-xg74g_fbce123a-545f-4c4c-86ca-2d68f81aa3f3/kube-rbac-proxy/0.log" Oct 08 19:25:15 crc kubenswrapper[4859]: I1008 19:25:15.894805 4859 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_manila-operator-controller-manager-5f67fbc655-xg74g_fbce123a-545f-4c4c-86ca-2d68f81aa3f3/manager/0.log" Oct 08 19:25:15 crc kubenswrapper[4859]: I1008 19:25:15.921529 4859 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_mariadb-operator-controller-manager-f9fb45f8f-ptpqs_b6a2cffa-8421-43c4-bad8-5d4d341a40cc/kube-rbac-proxy/0.log" Oct 08 19:25:15 crc kubenswrapper[4859]: I1008 19:25:15.997816 4859 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_mariadb-operator-controller-manager-f9fb45f8f-ptpqs_b6a2cffa-8421-43c4-bad8-5d4d341a40cc/manager/0.log" Oct 08 19:25:16 crc kubenswrapper[4859]: I1008 19:25:16.125762 4859 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_neutron-operator-controller-manager-79d585cb66-zn65q_e69529d1-1a2a-433c-a16b-3c6de62a0587/kube-rbac-proxy/0.log" Oct 08 19:25:16 crc kubenswrapper[4859]: I1008 19:25:16.167205 4859 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_neutron-operator-controller-manager-79d585cb66-zn65q_e69529d1-1a2a-433c-a16b-3c6de62a0587/manager/0.log" Oct 08 19:25:16 crc kubenswrapper[4859]: I1008 19:25:16.303460 4859 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_nova-operator-controller-manager-5df598886f-p99nw_e2889a4a-89ec-47de-99df-7e749c3e5514/kube-rbac-proxy/0.log" Oct 08 19:25:16 crc kubenswrapper[4859]: I1008 19:25:16.381525 4859 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_nova-operator-controller-manager-5df598886f-p99nw_e2889a4a-89ec-47de-99df-7e749c3e5514/manager/0.log" Oct 08 19:25:16 crc kubenswrapper[4859]: I1008 19:25:16.442915 4859 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_octavia-operator-controller-manager-69fdcfc5f5-lsl8l_550bb7bb-6f9c-4b20-9d92-2c85773e05a2/kube-rbac-proxy/0.log" Oct 08 19:25:16 crc kubenswrapper[4859]: I1008 19:25:16.505657 4859 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_octavia-operator-controller-manager-69fdcfc5f5-lsl8l_550bb7bb-6f9c-4b20-9d92-2c85773e05a2/manager/0.log" Oct 08 19:25:16 crc kubenswrapper[4859]: I1008 19:25:16.548324 4859 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_openstack-baremetal-operator-controller-manager-677c5f5bff5pg7p_dd924000-5690-490e-a4db-e29ed7f821ec/kube-rbac-proxy/0.log" Oct 08 19:25:16 crc kubenswrapper[4859]: I1008 19:25:16.636122 4859 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_openstack-baremetal-operator-controller-manager-677c5f5bff5pg7p_dd924000-5690-490e-a4db-e29ed7f821ec/manager/0.log" Oct 08 19:25:16 crc kubenswrapper[4859]: I1008 19:25:16.705758 4859 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_openstack-operator-controller-manager-598c4c5b5-6l55k_182ef4ca-109f-46bd-a935-091d05275271/kube-rbac-proxy/0.log" Oct 08 19:25:16 crc kubenswrapper[4859]: I1008 19:25:16.925123 4859 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_openstack-operator-controller-operator-bd6bc67fb-jkvgb_f60d2d10-2f0f-464f-9d1e-e63a5ef48972/kube-rbac-proxy/0.log" Oct 08 19:25:17 crc kubenswrapper[4859]: I1008 19:25:17.123798 4859 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_openstack-operator-index-z8f5z_6c1463a8-c973-42db-8671-f274a513230a/registry-server/0.log" Oct 08 19:25:17 crc kubenswrapper[4859]: I1008 19:25:17.140539 4859 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_openstack-operator-controller-operator-bd6bc67fb-jkvgb_f60d2d10-2f0f-464f-9d1e-e63a5ef48972/operator/0.log" Oct 08 19:25:17 crc kubenswrapper[4859]: I1008 19:25:17.197644 4859 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_ovn-operator-controller-manager-79db49b9fb-hhgrl_e2675d43-51ec-4a8c-aa2d-5aedc14f649a/kube-rbac-proxy/0.log" Oct 08 19:25:17 crc kubenswrapper[4859]: I1008 19:25:17.360424 4859 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_placement-operator-controller-manager-68b6c87b68-xgjpk_f56e7a0c-7b3a-4dc8-9367-cfeef5c5cef9/kube-rbac-proxy/0.log" Oct 08 19:25:17 crc kubenswrapper[4859]: I1008 19:25:17.412293 4859 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_ovn-operator-controller-manager-79db49b9fb-hhgrl_e2675d43-51ec-4a8c-aa2d-5aedc14f649a/manager/0.log" Oct 08 19:25:17 crc kubenswrapper[4859]: I1008 19:25:17.430987 4859 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_placement-operator-controller-manager-68b6c87b68-xgjpk_f56e7a0c-7b3a-4dc8-9367-cfeef5c5cef9/manager/0.log" Oct 08 19:25:17 crc kubenswrapper[4859]: I1008 19:25:17.692764 4859 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_rabbitmq-cluster-operator-manager-5f97d8c699-dxrl8_e1dbe8d7-f993-42b1-9a53-7d0c342780c1/operator/0.log" Oct 08 19:25:17 crc kubenswrapper[4859]: I1008 19:25:17.702400 4859 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_swift-operator-controller-manager-db6d7f97b-zjkwm_44980d79-924e-436f-950d-8fcdb4c42af1/kube-rbac-proxy/0.log" Oct 08 19:25:17 crc kubenswrapper[4859]: I1008 19:25:17.849024 4859 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_swift-operator-controller-manager-db6d7f97b-zjkwm_44980d79-924e-436f-950d-8fcdb4c42af1/manager/0.log" Oct 08 19:25:17 crc kubenswrapper[4859]: I1008 19:25:17.873364 4859 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_openstack-operator-controller-manager-598c4c5b5-6l55k_182ef4ca-109f-46bd-a935-091d05275271/manager/0.log" Oct 08 19:25:17 crc kubenswrapper[4859]: I1008 19:25:17.917173 4859 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_telemetry-operator-controller-manager-76796d4c6b-kjbfp_ffa2914a-717c-4763-8b1c-4a1ff18858f5/kube-rbac-proxy/0.log" Oct 08 19:25:17 crc kubenswrapper[4859]: I1008 19:25:17.963241 4859 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_telemetry-operator-controller-manager-76796d4c6b-kjbfp_ffa2914a-717c-4763-8b1c-4a1ff18858f5/manager/0.log" Oct 08 19:25:18 crc kubenswrapper[4859]: I1008 19:25:18.042964 4859 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_test-operator-controller-manager-56c698c775-d2g8b_2d6cf85d-c5cc-4466-a471-bb17aa8fca9d/kube-rbac-proxy/0.log" Oct 08 19:25:18 crc kubenswrapper[4859]: I1008 19:25:18.060576 4859 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_test-operator-controller-manager-56c698c775-d2g8b_2d6cf85d-c5cc-4466-a471-bb17aa8fca9d/manager/0.log" Oct 08 19:25:18 crc kubenswrapper[4859]: I1008 19:25:18.200902 4859 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_watcher-operator-controller-manager-7794bc6bd-2v65r_cf5b6741-336e-435c-8209-82b8e19f4896/manager/0.log" Oct 08 19:25:18 crc kubenswrapper[4859]: I1008 19:25:18.214833 4859 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_watcher-operator-controller-manager-7794bc6bd-2v65r_cf5b6741-336e-435c-8209-82b8e19f4896/kube-rbac-proxy/0.log" Oct 08 19:25:18 crc kubenswrapper[4859]: I1008 19:25:18.470306 4859 scope.go:117] "RemoveContainer" containerID="de71f0b717129dba48f3f0bdd90240a8e45404a17f0df79d6812d58c9209feff" Oct 08 19:25:18 crc kubenswrapper[4859]: E1008 19:25:18.470541 4859 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-82s52_openshift-machine-config-operator(b23a6a6c-9d92-4e7b-840e-55cfda873a2d)\"" pod="openshift-machine-config-operator/machine-config-daemon-82s52" podUID="b23a6a6c-9d92-4e7b-840e-55cfda873a2d" Oct 08 19:25:25 crc kubenswrapper[4859]: I1008 19:25:25.105225 4859 scope.go:117] "RemoveContainer" containerID="aaf7a34a7ba622ed67c2d3b385e9a8219112acf5608aeea4d037e11fd5595b1f" Oct 08 19:25:31 crc kubenswrapper[4859]: I1008 19:25:31.469712 4859 scope.go:117] "RemoveContainer" containerID="de71f0b717129dba48f3f0bdd90240a8e45404a17f0df79d6812d58c9209feff" Oct 08 19:25:31 crc kubenswrapper[4859]: E1008 19:25:31.470461 4859 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-82s52_openshift-machine-config-operator(b23a6a6c-9d92-4e7b-840e-55cfda873a2d)\"" pod="openshift-machine-config-operator/machine-config-daemon-82s52" podUID="b23a6a6c-9d92-4e7b-840e-55cfda873a2d" Oct 08 19:25:34 crc kubenswrapper[4859]: I1008 19:25:34.858248 4859 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-machine-api_control-plane-machine-set-operator-78cbb6b69f-q9nqj_492ef569-6d45-4dd0-89e5-506172689768/control-plane-machine-set-operator/0.log" Oct 08 19:25:35 crc kubenswrapper[4859]: I1008 19:25:35.047798 4859 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-machine-api_machine-api-operator-5694c8668f-gmbk9_a52b9536-1013-482b-9261-e024377687d2/kube-rbac-proxy/0.log" Oct 08 19:25:35 crc kubenswrapper[4859]: I1008 19:25:35.082814 4859 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-machine-api_machine-api-operator-5694c8668f-gmbk9_a52b9536-1013-482b-9261-e024377687d2/machine-api-operator/0.log" Oct 08 19:25:44 crc kubenswrapper[4859]: I1008 19:25:44.471035 4859 scope.go:117] "RemoveContainer" containerID="de71f0b717129dba48f3f0bdd90240a8e45404a17f0df79d6812d58c9209feff" Oct 08 19:25:44 crc kubenswrapper[4859]: E1008 19:25:44.471669 4859 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-82s52_openshift-machine-config-operator(b23a6a6c-9d92-4e7b-840e-55cfda873a2d)\"" pod="openshift-machine-config-operator/machine-config-daemon-82s52" podUID="b23a6a6c-9d92-4e7b-840e-55cfda873a2d" Oct 08 19:25:48 crc kubenswrapper[4859]: I1008 19:25:48.667339 4859 log.go:25] "Finished parsing log file" path="/var/log/pods/cert-manager_cert-manager-5b446d88c5-w95bk_2fa6fe28-fd3a-4611-ad50-2a4fc67ca563/cert-manager-controller/0.log" Oct 08 19:25:48 crc kubenswrapper[4859]: I1008 19:25:48.813402 4859 log.go:25] "Finished parsing log file" path="/var/log/pods/cert-manager_cert-manager-cainjector-7f985d654d-lqldn_ccaaab95-a4e1-43b8-a962-0b833beec038/cert-manager-cainjector/0.log" Oct 08 19:25:48 crc kubenswrapper[4859]: I1008 19:25:48.822616 4859 log.go:25] "Finished parsing log file" path="/var/log/pods/cert-manager_cert-manager-webhook-5655c58dd6-jkfxx_24b1f415-e420-420a-b6c1-7fde05b080b6/cert-manager-webhook/0.log" Oct 08 19:25:58 crc kubenswrapper[4859]: I1008 19:25:58.470856 4859 scope.go:117] "RemoveContainer" containerID="de71f0b717129dba48f3f0bdd90240a8e45404a17f0df79d6812d58c9209feff" Oct 08 19:25:58 crc kubenswrapper[4859]: E1008 19:25:58.471826 4859 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-82s52_openshift-machine-config-operator(b23a6a6c-9d92-4e7b-840e-55cfda873a2d)\"" pod="openshift-machine-config-operator/machine-config-daemon-82s52" podUID="b23a6a6c-9d92-4e7b-840e-55cfda873a2d" Oct 08 19:26:00 crc kubenswrapper[4859]: I1008 19:26:00.574831 4859 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-2cstr"] Oct 08 19:26:00 crc kubenswrapper[4859]: E1008 19:26:00.575722 4859 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="61c91bad-599a-4a6d-b03c-75a84f05c57b" containerName="container-00" Oct 08 19:26:00 crc kubenswrapper[4859]: I1008 19:26:00.575741 4859 state_mem.go:107] "Deleted CPUSet assignment" podUID="61c91bad-599a-4a6d-b03c-75a84f05c57b" containerName="container-00" Oct 08 19:26:00 crc kubenswrapper[4859]: I1008 19:26:00.575958 4859 memory_manager.go:354] "RemoveStaleState removing state" podUID="61c91bad-599a-4a6d-b03c-75a84f05c57b" containerName="container-00" Oct 08 19:26:00 crc kubenswrapper[4859]: I1008 19:26:00.577641 4859 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-2cstr" Oct 08 19:26:00 crc kubenswrapper[4859]: I1008 19:26:00.598412 4859 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-2cstr"] Oct 08 19:26:00 crc kubenswrapper[4859]: I1008 19:26:00.686263 4859 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/8be6447d-3aa6-4771-8dab-3fe8b4520187-catalog-content\") pod \"redhat-operators-2cstr\" (UID: \"8be6447d-3aa6-4771-8dab-3fe8b4520187\") " pod="openshift-marketplace/redhat-operators-2cstr" Oct 08 19:26:00 crc kubenswrapper[4859]: I1008 19:26:00.686331 4859 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-wrbmh\" (UniqueName: \"kubernetes.io/projected/8be6447d-3aa6-4771-8dab-3fe8b4520187-kube-api-access-wrbmh\") pod \"redhat-operators-2cstr\" (UID: \"8be6447d-3aa6-4771-8dab-3fe8b4520187\") " pod="openshift-marketplace/redhat-operators-2cstr" Oct 08 19:26:00 crc kubenswrapper[4859]: I1008 19:26:00.686582 4859 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/8be6447d-3aa6-4771-8dab-3fe8b4520187-utilities\") pod \"redhat-operators-2cstr\" (UID: \"8be6447d-3aa6-4771-8dab-3fe8b4520187\") " pod="openshift-marketplace/redhat-operators-2cstr" Oct 08 19:26:00 crc kubenswrapper[4859]: I1008 19:26:00.788286 4859 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/8be6447d-3aa6-4771-8dab-3fe8b4520187-utilities\") pod \"redhat-operators-2cstr\" (UID: \"8be6447d-3aa6-4771-8dab-3fe8b4520187\") " pod="openshift-marketplace/redhat-operators-2cstr" Oct 08 19:26:00 crc kubenswrapper[4859]: I1008 19:26:00.788382 4859 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/8be6447d-3aa6-4771-8dab-3fe8b4520187-catalog-content\") pod \"redhat-operators-2cstr\" (UID: \"8be6447d-3aa6-4771-8dab-3fe8b4520187\") " pod="openshift-marketplace/redhat-operators-2cstr" Oct 08 19:26:00 crc kubenswrapper[4859]: I1008 19:26:00.788422 4859 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-wrbmh\" (UniqueName: \"kubernetes.io/projected/8be6447d-3aa6-4771-8dab-3fe8b4520187-kube-api-access-wrbmh\") pod \"redhat-operators-2cstr\" (UID: \"8be6447d-3aa6-4771-8dab-3fe8b4520187\") " pod="openshift-marketplace/redhat-operators-2cstr" Oct 08 19:26:00 crc kubenswrapper[4859]: I1008 19:26:00.788796 4859 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/8be6447d-3aa6-4771-8dab-3fe8b4520187-utilities\") pod \"redhat-operators-2cstr\" (UID: \"8be6447d-3aa6-4771-8dab-3fe8b4520187\") " pod="openshift-marketplace/redhat-operators-2cstr" Oct 08 19:26:00 crc kubenswrapper[4859]: I1008 19:26:00.788909 4859 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/8be6447d-3aa6-4771-8dab-3fe8b4520187-catalog-content\") pod \"redhat-operators-2cstr\" (UID: \"8be6447d-3aa6-4771-8dab-3fe8b4520187\") " pod="openshift-marketplace/redhat-operators-2cstr" Oct 08 19:26:00 crc kubenswrapper[4859]: I1008 19:26:00.966722 4859 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-wrbmh\" (UniqueName: \"kubernetes.io/projected/8be6447d-3aa6-4771-8dab-3fe8b4520187-kube-api-access-wrbmh\") pod \"redhat-operators-2cstr\" (UID: \"8be6447d-3aa6-4771-8dab-3fe8b4520187\") " pod="openshift-marketplace/redhat-operators-2cstr" Oct 08 19:26:01 crc kubenswrapper[4859]: I1008 19:26:01.210370 4859 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-2cstr" Oct 08 19:26:01 crc kubenswrapper[4859]: I1008 19:26:01.708785 4859 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-2cstr"] Oct 08 19:26:02 crc kubenswrapper[4859]: I1008 19:26:02.124459 4859 generic.go:334] "Generic (PLEG): container finished" podID="8be6447d-3aa6-4771-8dab-3fe8b4520187" containerID="a41d8a13c3de59fc80a361d016e966bf03dc6254637e4defe5a7a4473c2f9567" exitCode=0 Oct 08 19:26:02 crc kubenswrapper[4859]: I1008 19:26:02.124576 4859 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-2cstr" event={"ID":"8be6447d-3aa6-4771-8dab-3fe8b4520187","Type":"ContainerDied","Data":"a41d8a13c3de59fc80a361d016e966bf03dc6254637e4defe5a7a4473c2f9567"} Oct 08 19:26:02 crc kubenswrapper[4859]: I1008 19:26:02.124807 4859 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-2cstr" event={"ID":"8be6447d-3aa6-4771-8dab-3fe8b4520187","Type":"ContainerStarted","Data":"a79fc33b1b2ce3da6aeec51530a27dff429ec09344df9d70b636fe83b11c4d40"} Oct 08 19:26:02 crc kubenswrapper[4859]: I1008 19:26:02.760361 4859 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-nmstate_nmstate-console-plugin-6b874cbd85-r5l52_a8be1e40-aedc-40bd-b47d-e381b5746080/nmstate-console-plugin/0.log" Oct 08 19:26:02 crc kubenswrapper[4859]: I1008 19:26:02.947408 4859 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-nmstate_nmstate-handler-htrgb_692be5e4-be66-4010-9547-610a2d3d85fd/nmstate-handler/0.log" Oct 08 19:26:03 crc kubenswrapper[4859]: I1008 19:26:03.090517 4859 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-nmstate_nmstate-metrics-fdff9cb8d-8vdx7_0beb8934-6616-4771-b93d-5b833f87600d/kube-rbac-proxy/0.log" Oct 08 19:26:03 crc kubenswrapper[4859]: I1008 19:26:03.143262 4859 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-nmstate_nmstate-metrics-fdff9cb8d-8vdx7_0beb8934-6616-4771-b93d-5b833f87600d/nmstate-metrics/0.log" Oct 08 19:26:03 crc kubenswrapper[4859]: I1008 19:26:03.263573 4859 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-nmstate_nmstate-operator-858ddd8f98-vbswp_3a114f4c-72fa-4640-86e4-cb4326069481/nmstate-operator/0.log" Oct 08 19:26:03 crc kubenswrapper[4859]: I1008 19:26:03.338097 4859 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-nmstate_nmstate-webhook-6cdbc54649-728jz_5af71d73-77d0-409b-8b48-cca9164c529e/nmstate-webhook/0.log" Oct 08 19:26:04 crc kubenswrapper[4859]: I1008 19:26:04.146891 4859 generic.go:334] "Generic (PLEG): container finished" podID="8be6447d-3aa6-4771-8dab-3fe8b4520187" containerID="aad7b7c3901df7d98e285866679c9249291ae54527a4bd9375a18be58db9bba1" exitCode=0 Oct 08 19:26:04 crc kubenswrapper[4859]: I1008 19:26:04.147756 4859 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-2cstr" event={"ID":"8be6447d-3aa6-4771-8dab-3fe8b4520187","Type":"ContainerDied","Data":"aad7b7c3901df7d98e285866679c9249291ae54527a4bd9375a18be58db9bba1"} Oct 08 19:26:05 crc kubenswrapper[4859]: I1008 19:26:05.157604 4859 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-2cstr" event={"ID":"8be6447d-3aa6-4771-8dab-3fe8b4520187","Type":"ContainerStarted","Data":"7ab713cc6b587f9b328d04fd73ab5ab644dc11e0e61cb5dc2fc0cd12feff3e02"} Oct 08 19:26:05 crc kubenswrapper[4859]: I1008 19:26:05.175325 4859 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-2cstr" podStartSLOduration=2.6833645329999998 podStartE2EDuration="5.175309267s" podCreationTimestamp="2025-10-08 19:26:00 +0000 UTC" firstStartedPulling="2025-10-08 19:26:02.127078626 +0000 UTC m=+4132.373918005" lastFinishedPulling="2025-10-08 19:26:04.61902336 +0000 UTC m=+4134.865862739" observedRunningTime="2025-10-08 19:26:05.173576887 +0000 UTC m=+4135.420416266" watchObservedRunningTime="2025-10-08 19:26:05.175309267 +0000 UTC m=+4135.422148646" Oct 08 19:26:11 crc kubenswrapper[4859]: I1008 19:26:11.210837 4859 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-2cstr" Oct 08 19:26:11 crc kubenswrapper[4859]: I1008 19:26:11.211858 4859 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-2cstr" Oct 08 19:26:11 crc kubenswrapper[4859]: I1008 19:26:11.276853 4859 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-2cstr" Oct 08 19:26:11 crc kubenswrapper[4859]: I1008 19:26:11.355905 4859 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-2cstr" Oct 08 19:26:11 crc kubenswrapper[4859]: I1008 19:26:11.525573 4859 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-2cstr"] Oct 08 19:26:13 crc kubenswrapper[4859]: I1008 19:26:13.239039 4859 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-operators-2cstr" podUID="8be6447d-3aa6-4771-8dab-3fe8b4520187" containerName="registry-server" containerID="cri-o://7ab713cc6b587f9b328d04fd73ab5ab644dc11e0e61cb5dc2fc0cd12feff3e02" gracePeriod=2 Oct 08 19:26:13 crc kubenswrapper[4859]: I1008 19:26:13.469781 4859 scope.go:117] "RemoveContainer" containerID="de71f0b717129dba48f3f0bdd90240a8e45404a17f0df79d6812d58c9209feff" Oct 08 19:26:13 crc kubenswrapper[4859]: E1008 19:26:13.470064 4859 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-82s52_openshift-machine-config-operator(b23a6a6c-9d92-4e7b-840e-55cfda873a2d)\"" pod="openshift-machine-config-operator/machine-config-daemon-82s52" podUID="b23a6a6c-9d92-4e7b-840e-55cfda873a2d" Oct 08 19:26:13 crc kubenswrapper[4859]: I1008 19:26:13.705266 4859 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-2cstr" Oct 08 19:26:13 crc kubenswrapper[4859]: I1008 19:26:13.848936 4859 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-wrbmh\" (UniqueName: \"kubernetes.io/projected/8be6447d-3aa6-4771-8dab-3fe8b4520187-kube-api-access-wrbmh\") pod \"8be6447d-3aa6-4771-8dab-3fe8b4520187\" (UID: \"8be6447d-3aa6-4771-8dab-3fe8b4520187\") " Oct 08 19:26:13 crc kubenswrapper[4859]: I1008 19:26:13.849152 4859 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/8be6447d-3aa6-4771-8dab-3fe8b4520187-catalog-content\") pod \"8be6447d-3aa6-4771-8dab-3fe8b4520187\" (UID: \"8be6447d-3aa6-4771-8dab-3fe8b4520187\") " Oct 08 19:26:13 crc kubenswrapper[4859]: I1008 19:26:13.849325 4859 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/8be6447d-3aa6-4771-8dab-3fe8b4520187-utilities\") pod \"8be6447d-3aa6-4771-8dab-3fe8b4520187\" (UID: \"8be6447d-3aa6-4771-8dab-3fe8b4520187\") " Oct 08 19:26:13 crc kubenswrapper[4859]: I1008 19:26:13.850542 4859 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/8be6447d-3aa6-4771-8dab-3fe8b4520187-utilities" (OuterVolumeSpecName: "utilities") pod "8be6447d-3aa6-4771-8dab-3fe8b4520187" (UID: "8be6447d-3aa6-4771-8dab-3fe8b4520187"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 08 19:26:13 crc kubenswrapper[4859]: I1008 19:26:13.856323 4859 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8be6447d-3aa6-4771-8dab-3fe8b4520187-kube-api-access-wrbmh" (OuterVolumeSpecName: "kube-api-access-wrbmh") pod "8be6447d-3aa6-4771-8dab-3fe8b4520187" (UID: "8be6447d-3aa6-4771-8dab-3fe8b4520187"). InnerVolumeSpecName "kube-api-access-wrbmh". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 08 19:26:13 crc kubenswrapper[4859]: I1008 19:26:13.951732 4859 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/8be6447d-3aa6-4771-8dab-3fe8b4520187-utilities\") on node \"crc\" DevicePath \"\"" Oct 08 19:26:13 crc kubenswrapper[4859]: I1008 19:26:13.951792 4859 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-wrbmh\" (UniqueName: \"kubernetes.io/projected/8be6447d-3aa6-4771-8dab-3fe8b4520187-kube-api-access-wrbmh\") on node \"crc\" DevicePath \"\"" Oct 08 19:26:14 crc kubenswrapper[4859]: I1008 19:26:14.072141 4859 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/8be6447d-3aa6-4771-8dab-3fe8b4520187-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "8be6447d-3aa6-4771-8dab-3fe8b4520187" (UID: "8be6447d-3aa6-4771-8dab-3fe8b4520187"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 08 19:26:14 crc kubenswrapper[4859]: I1008 19:26:14.155822 4859 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/8be6447d-3aa6-4771-8dab-3fe8b4520187-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 08 19:26:14 crc kubenswrapper[4859]: I1008 19:26:14.265077 4859 generic.go:334] "Generic (PLEG): container finished" podID="8be6447d-3aa6-4771-8dab-3fe8b4520187" containerID="7ab713cc6b587f9b328d04fd73ab5ab644dc11e0e61cb5dc2fc0cd12feff3e02" exitCode=0 Oct 08 19:26:14 crc kubenswrapper[4859]: I1008 19:26:14.265135 4859 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-2cstr" Oct 08 19:26:14 crc kubenswrapper[4859]: I1008 19:26:14.265139 4859 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-2cstr" event={"ID":"8be6447d-3aa6-4771-8dab-3fe8b4520187","Type":"ContainerDied","Data":"7ab713cc6b587f9b328d04fd73ab5ab644dc11e0e61cb5dc2fc0cd12feff3e02"} Oct 08 19:26:14 crc kubenswrapper[4859]: I1008 19:26:14.265256 4859 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-2cstr" event={"ID":"8be6447d-3aa6-4771-8dab-3fe8b4520187","Type":"ContainerDied","Data":"a79fc33b1b2ce3da6aeec51530a27dff429ec09344df9d70b636fe83b11c4d40"} Oct 08 19:26:14 crc kubenswrapper[4859]: I1008 19:26:14.265305 4859 scope.go:117] "RemoveContainer" containerID="7ab713cc6b587f9b328d04fd73ab5ab644dc11e0e61cb5dc2fc0cd12feff3e02" Oct 08 19:26:14 crc kubenswrapper[4859]: I1008 19:26:14.327821 4859 scope.go:117] "RemoveContainer" containerID="aad7b7c3901df7d98e285866679c9249291ae54527a4bd9375a18be58db9bba1" Oct 08 19:26:14 crc kubenswrapper[4859]: I1008 19:26:14.334010 4859 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-2cstr"] Oct 08 19:26:14 crc kubenswrapper[4859]: I1008 19:26:14.348794 4859 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-operators-2cstr"] Oct 08 19:26:14 crc kubenswrapper[4859]: I1008 19:26:14.395737 4859 scope.go:117] "RemoveContainer" containerID="a41d8a13c3de59fc80a361d016e966bf03dc6254637e4defe5a7a4473c2f9567" Oct 08 19:26:14 crc kubenswrapper[4859]: I1008 19:26:14.444427 4859 scope.go:117] "RemoveContainer" containerID="7ab713cc6b587f9b328d04fd73ab5ab644dc11e0e61cb5dc2fc0cd12feff3e02" Oct 08 19:26:14 crc kubenswrapper[4859]: E1008 19:26:14.444911 4859 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"7ab713cc6b587f9b328d04fd73ab5ab644dc11e0e61cb5dc2fc0cd12feff3e02\": container with ID starting with 7ab713cc6b587f9b328d04fd73ab5ab644dc11e0e61cb5dc2fc0cd12feff3e02 not found: ID does not exist" containerID="7ab713cc6b587f9b328d04fd73ab5ab644dc11e0e61cb5dc2fc0cd12feff3e02" Oct 08 19:26:14 crc kubenswrapper[4859]: I1008 19:26:14.445750 4859 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"7ab713cc6b587f9b328d04fd73ab5ab644dc11e0e61cb5dc2fc0cd12feff3e02"} err="failed to get container status \"7ab713cc6b587f9b328d04fd73ab5ab644dc11e0e61cb5dc2fc0cd12feff3e02\": rpc error: code = NotFound desc = could not find container \"7ab713cc6b587f9b328d04fd73ab5ab644dc11e0e61cb5dc2fc0cd12feff3e02\": container with ID starting with 7ab713cc6b587f9b328d04fd73ab5ab644dc11e0e61cb5dc2fc0cd12feff3e02 not found: ID does not exist" Oct 08 19:26:14 crc kubenswrapper[4859]: I1008 19:26:14.445815 4859 scope.go:117] "RemoveContainer" containerID="aad7b7c3901df7d98e285866679c9249291ae54527a4bd9375a18be58db9bba1" Oct 08 19:26:14 crc kubenswrapper[4859]: E1008 19:26:14.446197 4859 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"aad7b7c3901df7d98e285866679c9249291ae54527a4bd9375a18be58db9bba1\": container with ID starting with aad7b7c3901df7d98e285866679c9249291ae54527a4bd9375a18be58db9bba1 not found: ID does not exist" containerID="aad7b7c3901df7d98e285866679c9249291ae54527a4bd9375a18be58db9bba1" Oct 08 19:26:14 crc kubenswrapper[4859]: I1008 19:26:14.446230 4859 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"aad7b7c3901df7d98e285866679c9249291ae54527a4bd9375a18be58db9bba1"} err="failed to get container status \"aad7b7c3901df7d98e285866679c9249291ae54527a4bd9375a18be58db9bba1\": rpc error: code = NotFound desc = could not find container \"aad7b7c3901df7d98e285866679c9249291ae54527a4bd9375a18be58db9bba1\": container with ID starting with aad7b7c3901df7d98e285866679c9249291ae54527a4bd9375a18be58db9bba1 not found: ID does not exist" Oct 08 19:26:14 crc kubenswrapper[4859]: I1008 19:26:14.446252 4859 scope.go:117] "RemoveContainer" containerID="a41d8a13c3de59fc80a361d016e966bf03dc6254637e4defe5a7a4473c2f9567" Oct 08 19:26:14 crc kubenswrapper[4859]: E1008 19:26:14.447142 4859 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"a41d8a13c3de59fc80a361d016e966bf03dc6254637e4defe5a7a4473c2f9567\": container with ID starting with a41d8a13c3de59fc80a361d016e966bf03dc6254637e4defe5a7a4473c2f9567 not found: ID does not exist" containerID="a41d8a13c3de59fc80a361d016e966bf03dc6254637e4defe5a7a4473c2f9567" Oct 08 19:26:14 crc kubenswrapper[4859]: I1008 19:26:14.447165 4859 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"a41d8a13c3de59fc80a361d016e966bf03dc6254637e4defe5a7a4473c2f9567"} err="failed to get container status \"a41d8a13c3de59fc80a361d016e966bf03dc6254637e4defe5a7a4473c2f9567\": rpc error: code = NotFound desc = could not find container \"a41d8a13c3de59fc80a361d016e966bf03dc6254637e4defe5a7a4473c2f9567\": container with ID starting with a41d8a13c3de59fc80a361d016e966bf03dc6254637e4defe5a7a4473c2f9567 not found: ID does not exist" Oct 08 19:26:14 crc kubenswrapper[4859]: I1008 19:26:14.479679 4859 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="8be6447d-3aa6-4771-8dab-3fe8b4520187" path="/var/lib/kubelet/pods/8be6447d-3aa6-4771-8dab-3fe8b4520187/volumes" Oct 08 19:26:19 crc kubenswrapper[4859]: I1008 19:26:19.431838 4859 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_controller-68d546b9d8-cr5dk_7f1dbe0e-c03a-4ec2-ba3c-7179d2c67e41/kube-rbac-proxy/0.log" Oct 08 19:26:19 crc kubenswrapper[4859]: I1008 19:26:19.542176 4859 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_controller-68d546b9d8-cr5dk_7f1dbe0e-c03a-4ec2-ba3c-7179d2c67e41/controller/0.log" Oct 08 19:26:19 crc kubenswrapper[4859]: I1008 19:26:19.642756 4859 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-mfcd8_69cc95cc-1a58-40c2-9c84-4808c217c2e5/cp-frr-files/0.log" Oct 08 19:26:19 crc kubenswrapper[4859]: I1008 19:26:19.833437 4859 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-mfcd8_69cc95cc-1a58-40c2-9c84-4808c217c2e5/cp-reloader/0.log" Oct 08 19:26:19 crc kubenswrapper[4859]: I1008 19:26:19.840905 4859 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-mfcd8_69cc95cc-1a58-40c2-9c84-4808c217c2e5/cp-frr-files/0.log" Oct 08 19:26:19 crc kubenswrapper[4859]: I1008 19:26:19.909040 4859 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-mfcd8_69cc95cc-1a58-40c2-9c84-4808c217c2e5/cp-metrics/0.log" Oct 08 19:26:19 crc kubenswrapper[4859]: I1008 19:26:19.933354 4859 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-mfcd8_69cc95cc-1a58-40c2-9c84-4808c217c2e5/cp-reloader/0.log" Oct 08 19:26:20 crc kubenswrapper[4859]: I1008 19:26:20.059135 4859 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-mfcd8_69cc95cc-1a58-40c2-9c84-4808c217c2e5/cp-frr-files/0.log" Oct 08 19:26:20 crc kubenswrapper[4859]: I1008 19:26:20.072399 4859 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-mfcd8_69cc95cc-1a58-40c2-9c84-4808c217c2e5/cp-metrics/0.log" Oct 08 19:26:20 crc kubenswrapper[4859]: I1008 19:26:20.105538 4859 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-mfcd8_69cc95cc-1a58-40c2-9c84-4808c217c2e5/cp-reloader/0.log" Oct 08 19:26:20 crc kubenswrapper[4859]: I1008 19:26:20.166572 4859 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-mfcd8_69cc95cc-1a58-40c2-9c84-4808c217c2e5/cp-metrics/0.log" Oct 08 19:26:20 crc kubenswrapper[4859]: I1008 19:26:20.636262 4859 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-mfcd8_69cc95cc-1a58-40c2-9c84-4808c217c2e5/cp-reloader/0.log" Oct 08 19:26:20 crc kubenswrapper[4859]: I1008 19:26:20.652655 4859 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-mfcd8_69cc95cc-1a58-40c2-9c84-4808c217c2e5/cp-frr-files/0.log" Oct 08 19:26:20 crc kubenswrapper[4859]: I1008 19:26:20.691680 4859 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-mfcd8_69cc95cc-1a58-40c2-9c84-4808c217c2e5/controller/0.log" Oct 08 19:26:20 crc kubenswrapper[4859]: I1008 19:26:20.698726 4859 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-mfcd8_69cc95cc-1a58-40c2-9c84-4808c217c2e5/cp-metrics/0.log" Oct 08 19:26:20 crc kubenswrapper[4859]: I1008 19:26:20.823430 4859 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-mfcd8_69cc95cc-1a58-40c2-9c84-4808c217c2e5/frr-metrics/0.log" Oct 08 19:26:20 crc kubenswrapper[4859]: I1008 19:26:20.872106 4859 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-mfcd8_69cc95cc-1a58-40c2-9c84-4808c217c2e5/kube-rbac-proxy-frr/0.log" Oct 08 19:26:20 crc kubenswrapper[4859]: I1008 19:26:20.901122 4859 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-mfcd8_69cc95cc-1a58-40c2-9c84-4808c217c2e5/kube-rbac-proxy/0.log" Oct 08 19:26:21 crc kubenswrapper[4859]: I1008 19:26:21.058613 4859 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-mfcd8_69cc95cc-1a58-40c2-9c84-4808c217c2e5/reloader/0.log" Oct 08 19:26:21 crc kubenswrapper[4859]: I1008 19:26:21.090592 4859 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-webhook-server-64bf5d555-77xwq_4d682ed0-e5d8-48ad-8e52-cd29b2f65bd2/frr-k8s-webhook-server/0.log" Oct 08 19:26:21 crc kubenswrapper[4859]: I1008 19:26:21.343916 4859 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_metallb-operator-controller-manager-6789c5bc7c-k7fl4_47c9b7b8-561d-489b-a5d2-34ef1f6f4be3/manager/0.log" Oct 08 19:26:21 crc kubenswrapper[4859]: I1008 19:26:21.584296 4859 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_metallb-operator-webhook-server-c4d794b9c-6lnpl_c5192e5b-dc81-4df1-8ba2-f103a6b869eb/webhook-server/0.log" Oct 08 19:26:21 crc kubenswrapper[4859]: I1008 19:26:21.647384 4859 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_speaker-47524_7557fe67-9517-441e-944c-9d5286f5735c/kube-rbac-proxy/0.log" Oct 08 19:26:22 crc kubenswrapper[4859]: I1008 19:26:22.201758 4859 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_speaker-47524_7557fe67-9517-441e-944c-9d5286f5735c/speaker/0.log" Oct 08 19:26:22 crc kubenswrapper[4859]: I1008 19:26:22.328663 4859 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-mfcd8_69cc95cc-1a58-40c2-9c84-4808c217c2e5/frr/0.log" Oct 08 19:26:27 crc kubenswrapper[4859]: I1008 19:26:27.471069 4859 scope.go:117] "RemoveContainer" containerID="de71f0b717129dba48f3f0bdd90240a8e45404a17f0df79d6812d58c9209feff" Oct 08 19:26:28 crc kubenswrapper[4859]: I1008 19:26:28.437936 4859 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-82s52" event={"ID":"b23a6a6c-9d92-4e7b-840e-55cfda873a2d","Type":"ContainerStarted","Data":"9cb675f315140c9335fbc82b8aae4ac5d7ef2786da474fb4012875812abb974d"} Oct 08 19:26:33 crc kubenswrapper[4859]: I1008 19:26:33.979282 4859 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_8f2f4ee801e5826a37d84a7b1fc4ccbf6b79de668302737d0f1152d8d2lfzpq_a200c80e-a0f7-40ee-b697-acf6d429d14f/util/0.log" Oct 08 19:26:34 crc kubenswrapper[4859]: I1008 19:26:34.087067 4859 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_8f2f4ee801e5826a37d84a7b1fc4ccbf6b79de668302737d0f1152d8d2lfzpq_a200c80e-a0f7-40ee-b697-acf6d429d14f/util/0.log" Oct 08 19:26:34 crc kubenswrapper[4859]: I1008 19:26:34.087959 4859 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_8f2f4ee801e5826a37d84a7b1fc4ccbf6b79de668302737d0f1152d8d2lfzpq_a200c80e-a0f7-40ee-b697-acf6d429d14f/pull/0.log" Oct 08 19:26:34 crc kubenswrapper[4859]: I1008 19:26:34.133715 4859 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_8f2f4ee801e5826a37d84a7b1fc4ccbf6b79de668302737d0f1152d8d2lfzpq_a200c80e-a0f7-40ee-b697-acf6d429d14f/pull/0.log" Oct 08 19:26:34 crc kubenswrapper[4859]: I1008 19:26:34.311391 4859 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_8f2f4ee801e5826a37d84a7b1fc4ccbf6b79de668302737d0f1152d8d2lfzpq_a200c80e-a0f7-40ee-b697-acf6d429d14f/util/0.log" Oct 08 19:26:34 crc kubenswrapper[4859]: I1008 19:26:34.326361 4859 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_8f2f4ee801e5826a37d84a7b1fc4ccbf6b79de668302737d0f1152d8d2lfzpq_a200c80e-a0f7-40ee-b697-acf6d429d14f/pull/0.log" Oct 08 19:26:34 crc kubenswrapper[4859]: I1008 19:26:34.328267 4859 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_8f2f4ee801e5826a37d84a7b1fc4ccbf6b79de668302737d0f1152d8d2lfzpq_a200c80e-a0f7-40ee-b697-acf6d429d14f/extract/0.log" Oct 08 19:26:34 crc kubenswrapper[4859]: I1008 19:26:34.609328 4859 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-d8bkm_6e7e5636-231d-4788-9277-83549fc6b404/extract-utilities/0.log" Oct 08 19:26:34 crc kubenswrapper[4859]: I1008 19:26:34.772206 4859 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-d8bkm_6e7e5636-231d-4788-9277-83549fc6b404/extract-content/0.log" Oct 08 19:26:34 crc kubenswrapper[4859]: I1008 19:26:34.782575 4859 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-d8bkm_6e7e5636-231d-4788-9277-83549fc6b404/extract-utilities/0.log" Oct 08 19:26:34 crc kubenswrapper[4859]: I1008 19:26:34.821860 4859 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-d8bkm_6e7e5636-231d-4788-9277-83549fc6b404/extract-content/0.log" Oct 08 19:26:34 crc kubenswrapper[4859]: I1008 19:26:34.940230 4859 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-d8bkm_6e7e5636-231d-4788-9277-83549fc6b404/extract-utilities/0.log" Oct 08 19:26:34 crc kubenswrapper[4859]: I1008 19:26:34.953124 4859 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-d8bkm_6e7e5636-231d-4788-9277-83549fc6b404/extract-content/0.log" Oct 08 19:26:35 crc kubenswrapper[4859]: I1008 19:26:35.176363 4859 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-p7bgm_ad354068-21f9-40aa-a2ae-739f84e32469/extract-utilities/0.log" Oct 08 19:26:35 crc kubenswrapper[4859]: I1008 19:26:35.377530 4859 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-p7bgm_ad354068-21f9-40aa-a2ae-739f84e32469/extract-content/0.log" Oct 08 19:26:35 crc kubenswrapper[4859]: I1008 19:26:35.433196 4859 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-p7bgm_ad354068-21f9-40aa-a2ae-739f84e32469/extract-content/0.log" Oct 08 19:26:35 crc kubenswrapper[4859]: I1008 19:26:35.433496 4859 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-p7bgm_ad354068-21f9-40aa-a2ae-739f84e32469/extract-utilities/0.log" Oct 08 19:26:35 crc kubenswrapper[4859]: I1008 19:26:35.451903 4859 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-d8bkm_6e7e5636-231d-4788-9277-83549fc6b404/registry-server/0.log" Oct 08 19:26:35 crc kubenswrapper[4859]: I1008 19:26:35.653204 4859 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-p7bgm_ad354068-21f9-40aa-a2ae-739f84e32469/extract-utilities/0.log" Oct 08 19:26:35 crc kubenswrapper[4859]: I1008 19:26:35.690080 4859 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-p7bgm_ad354068-21f9-40aa-a2ae-739f84e32469/extract-content/0.log" Oct 08 19:26:35 crc kubenswrapper[4859]: I1008 19:26:35.901848 4859 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_fa9831ede5d93c33d525b70ce6ddf94e500d80992af75a3305fe98835cdxmbx_cc21f1af-ca05-4cf4-b66f-f7fcde5026ee/util/0.log" Oct 08 19:26:36 crc kubenswrapper[4859]: I1008 19:26:36.063769 4859 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_fa9831ede5d93c33d525b70ce6ddf94e500d80992af75a3305fe98835cdxmbx_cc21f1af-ca05-4cf4-b66f-f7fcde5026ee/pull/0.log" Oct 08 19:26:36 crc kubenswrapper[4859]: I1008 19:26:36.141005 4859 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_fa9831ede5d93c33d525b70ce6ddf94e500d80992af75a3305fe98835cdxmbx_cc21f1af-ca05-4cf4-b66f-f7fcde5026ee/util/0.log" Oct 08 19:26:36 crc kubenswrapper[4859]: I1008 19:26:36.181175 4859 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_fa9831ede5d93c33d525b70ce6ddf94e500d80992af75a3305fe98835cdxmbx_cc21f1af-ca05-4cf4-b66f-f7fcde5026ee/pull/0.log" Oct 08 19:26:36 crc kubenswrapper[4859]: I1008 19:26:36.380094 4859 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_fa9831ede5d93c33d525b70ce6ddf94e500d80992af75a3305fe98835cdxmbx_cc21f1af-ca05-4cf4-b66f-f7fcde5026ee/extract/0.log" Oct 08 19:26:36 crc kubenswrapper[4859]: I1008 19:26:36.380503 4859 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_fa9831ede5d93c33d525b70ce6ddf94e500d80992af75a3305fe98835cdxmbx_cc21f1af-ca05-4cf4-b66f-f7fcde5026ee/util/0.log" Oct 08 19:26:36 crc kubenswrapper[4859]: I1008 19:26:36.381654 4859 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_fa9831ede5d93c33d525b70ce6ddf94e500d80992af75a3305fe98835cdxmbx_cc21f1af-ca05-4cf4-b66f-f7fcde5026ee/pull/0.log" Oct 08 19:26:36 crc kubenswrapper[4859]: I1008 19:26:36.464890 4859 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-p7bgm_ad354068-21f9-40aa-a2ae-739f84e32469/registry-server/0.log" Oct 08 19:26:36 crc kubenswrapper[4859]: I1008 19:26:36.603990 4859 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_marketplace-operator-79b997595-h4tnd_8fdf9d2f-74e1-46e7-89ed-1947c3104386/marketplace-operator/0.log" Oct 08 19:26:36 crc kubenswrapper[4859]: I1008 19:26:36.708428 4859 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-wlwhh_997d41aa-a006-46d1-b488-b6d29186874d/extract-utilities/0.log" Oct 08 19:26:36 crc kubenswrapper[4859]: I1008 19:26:36.931181 4859 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-wlwhh_997d41aa-a006-46d1-b488-b6d29186874d/extract-content/0.log" Oct 08 19:26:36 crc kubenswrapper[4859]: I1008 19:26:36.934471 4859 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-wlwhh_997d41aa-a006-46d1-b488-b6d29186874d/extract-utilities/0.log" Oct 08 19:26:36 crc kubenswrapper[4859]: I1008 19:26:36.967951 4859 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-wlwhh_997d41aa-a006-46d1-b488-b6d29186874d/extract-content/0.log" Oct 08 19:26:37 crc kubenswrapper[4859]: I1008 19:26:37.099051 4859 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-wlwhh_997d41aa-a006-46d1-b488-b6d29186874d/extract-utilities/0.log" Oct 08 19:26:37 crc kubenswrapper[4859]: I1008 19:26:37.174055 4859 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-wlwhh_997d41aa-a006-46d1-b488-b6d29186874d/extract-content/0.log" Oct 08 19:26:37 crc kubenswrapper[4859]: I1008 19:26:37.268166 4859 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-hlqjs_748a4b84-e115-444a-b0f8-24029e8d1896/extract-utilities/0.log" Oct 08 19:26:37 crc kubenswrapper[4859]: I1008 19:26:37.407824 4859 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-wlwhh_997d41aa-a006-46d1-b488-b6d29186874d/registry-server/0.log" Oct 08 19:26:37 crc kubenswrapper[4859]: I1008 19:26:37.506079 4859 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-hlqjs_748a4b84-e115-444a-b0f8-24029e8d1896/extract-utilities/0.log" Oct 08 19:26:37 crc kubenswrapper[4859]: I1008 19:26:37.535515 4859 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-hlqjs_748a4b84-e115-444a-b0f8-24029e8d1896/extract-content/0.log" Oct 08 19:26:37 crc kubenswrapper[4859]: I1008 19:26:37.557744 4859 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-hlqjs_748a4b84-e115-444a-b0f8-24029e8d1896/extract-content/0.log" Oct 08 19:26:37 crc kubenswrapper[4859]: I1008 19:26:37.694013 4859 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-hlqjs_748a4b84-e115-444a-b0f8-24029e8d1896/extract-content/0.log" Oct 08 19:26:37 crc kubenswrapper[4859]: I1008 19:26:37.711168 4859 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-hlqjs_748a4b84-e115-444a-b0f8-24029e8d1896/extract-utilities/0.log" Oct 08 19:26:38 crc kubenswrapper[4859]: I1008 19:26:38.201816 4859 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-hlqjs_748a4b84-e115-444a-b0f8-24029e8d1896/registry-server/0.log" Oct 08 19:27:11 crc kubenswrapper[4859]: I1008 19:27:11.477912 4859 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-str2r"] Oct 08 19:27:11 crc kubenswrapper[4859]: E1008 19:27:11.478928 4859 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8be6447d-3aa6-4771-8dab-3fe8b4520187" containerName="registry-server" Oct 08 19:27:11 crc kubenswrapper[4859]: I1008 19:27:11.478942 4859 state_mem.go:107] "Deleted CPUSet assignment" podUID="8be6447d-3aa6-4771-8dab-3fe8b4520187" containerName="registry-server" Oct 08 19:27:11 crc kubenswrapper[4859]: E1008 19:27:11.478963 4859 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8be6447d-3aa6-4771-8dab-3fe8b4520187" containerName="extract-content" Oct 08 19:27:11 crc kubenswrapper[4859]: I1008 19:27:11.478968 4859 state_mem.go:107] "Deleted CPUSet assignment" podUID="8be6447d-3aa6-4771-8dab-3fe8b4520187" containerName="extract-content" Oct 08 19:27:11 crc kubenswrapper[4859]: E1008 19:27:11.478982 4859 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8be6447d-3aa6-4771-8dab-3fe8b4520187" containerName="extract-utilities" Oct 08 19:27:11 crc kubenswrapper[4859]: I1008 19:27:11.478989 4859 state_mem.go:107] "Deleted CPUSet assignment" podUID="8be6447d-3aa6-4771-8dab-3fe8b4520187" containerName="extract-utilities" Oct 08 19:27:11 crc kubenswrapper[4859]: I1008 19:27:11.479181 4859 memory_manager.go:354] "RemoveStaleState removing state" podUID="8be6447d-3aa6-4771-8dab-3fe8b4520187" containerName="registry-server" Oct 08 19:27:11 crc kubenswrapper[4859]: I1008 19:27:11.480631 4859 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-str2r" Oct 08 19:27:11 crc kubenswrapper[4859]: I1008 19:27:11.489552 4859 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-str2r"] Oct 08 19:27:11 crc kubenswrapper[4859]: I1008 19:27:11.579406 4859 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/0c5d8464-4994-4d09-8474-d484958f1d97-catalog-content\") pod \"redhat-marketplace-str2r\" (UID: \"0c5d8464-4994-4d09-8474-d484958f1d97\") " pod="openshift-marketplace/redhat-marketplace-str2r" Oct 08 19:27:11 crc kubenswrapper[4859]: I1008 19:27:11.579526 4859 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/0c5d8464-4994-4d09-8474-d484958f1d97-utilities\") pod \"redhat-marketplace-str2r\" (UID: \"0c5d8464-4994-4d09-8474-d484958f1d97\") " pod="openshift-marketplace/redhat-marketplace-str2r" Oct 08 19:27:11 crc kubenswrapper[4859]: I1008 19:27:11.579680 4859 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-6n255\" (UniqueName: \"kubernetes.io/projected/0c5d8464-4994-4d09-8474-d484958f1d97-kube-api-access-6n255\") pod \"redhat-marketplace-str2r\" (UID: \"0c5d8464-4994-4d09-8474-d484958f1d97\") " pod="openshift-marketplace/redhat-marketplace-str2r" Oct 08 19:27:11 crc kubenswrapper[4859]: I1008 19:27:11.681090 4859 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-6n255\" (UniqueName: \"kubernetes.io/projected/0c5d8464-4994-4d09-8474-d484958f1d97-kube-api-access-6n255\") pod \"redhat-marketplace-str2r\" (UID: \"0c5d8464-4994-4d09-8474-d484958f1d97\") " pod="openshift-marketplace/redhat-marketplace-str2r" Oct 08 19:27:11 crc kubenswrapper[4859]: I1008 19:27:11.681189 4859 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/0c5d8464-4994-4d09-8474-d484958f1d97-catalog-content\") pod \"redhat-marketplace-str2r\" (UID: \"0c5d8464-4994-4d09-8474-d484958f1d97\") " pod="openshift-marketplace/redhat-marketplace-str2r" Oct 08 19:27:11 crc kubenswrapper[4859]: I1008 19:27:11.681262 4859 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/0c5d8464-4994-4d09-8474-d484958f1d97-utilities\") pod \"redhat-marketplace-str2r\" (UID: \"0c5d8464-4994-4d09-8474-d484958f1d97\") " pod="openshift-marketplace/redhat-marketplace-str2r" Oct 08 19:27:11 crc kubenswrapper[4859]: I1008 19:27:11.681785 4859 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/0c5d8464-4994-4d09-8474-d484958f1d97-utilities\") pod \"redhat-marketplace-str2r\" (UID: \"0c5d8464-4994-4d09-8474-d484958f1d97\") " pod="openshift-marketplace/redhat-marketplace-str2r" Oct 08 19:27:11 crc kubenswrapper[4859]: I1008 19:27:11.682264 4859 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/0c5d8464-4994-4d09-8474-d484958f1d97-catalog-content\") pod \"redhat-marketplace-str2r\" (UID: \"0c5d8464-4994-4d09-8474-d484958f1d97\") " pod="openshift-marketplace/redhat-marketplace-str2r" Oct 08 19:27:12 crc kubenswrapper[4859]: I1008 19:27:12.052101 4859 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-6n255\" (UniqueName: \"kubernetes.io/projected/0c5d8464-4994-4d09-8474-d484958f1d97-kube-api-access-6n255\") pod \"redhat-marketplace-str2r\" (UID: \"0c5d8464-4994-4d09-8474-d484958f1d97\") " pod="openshift-marketplace/redhat-marketplace-str2r" Oct 08 19:27:12 crc kubenswrapper[4859]: I1008 19:27:12.104095 4859 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-str2r" Oct 08 19:27:12 crc kubenswrapper[4859]: I1008 19:27:12.609474 4859 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-str2r"] Oct 08 19:27:12 crc kubenswrapper[4859]: I1008 19:27:12.870483 4859 generic.go:334] "Generic (PLEG): container finished" podID="0c5d8464-4994-4d09-8474-d484958f1d97" containerID="5acafc12088cb41323804fa5014521817625406075b5e5bf10c7050572ac4287" exitCode=0 Oct 08 19:27:12 crc kubenswrapper[4859]: I1008 19:27:12.870587 4859 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-str2r" event={"ID":"0c5d8464-4994-4d09-8474-d484958f1d97","Type":"ContainerDied","Data":"5acafc12088cb41323804fa5014521817625406075b5e5bf10c7050572ac4287"} Oct 08 19:27:12 crc kubenswrapper[4859]: I1008 19:27:12.871283 4859 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-str2r" event={"ID":"0c5d8464-4994-4d09-8474-d484958f1d97","Type":"ContainerStarted","Data":"3279ff5480639646687db37a393eb7702f1b9c6e061eed9cbaac664d70c3c0e6"} Oct 08 19:27:14 crc kubenswrapper[4859]: I1008 19:27:14.891966 4859 generic.go:334] "Generic (PLEG): container finished" podID="0c5d8464-4994-4d09-8474-d484958f1d97" containerID="26fd50082ecc8246c5b8803a5ee155e6c662008433ffc305a7b7a6e8d1ffe306" exitCode=0 Oct 08 19:27:14 crc kubenswrapper[4859]: I1008 19:27:14.892070 4859 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-str2r" event={"ID":"0c5d8464-4994-4d09-8474-d484958f1d97","Type":"ContainerDied","Data":"26fd50082ecc8246c5b8803a5ee155e6c662008433ffc305a7b7a6e8d1ffe306"} Oct 08 19:27:15 crc kubenswrapper[4859]: I1008 19:27:15.907253 4859 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-str2r" event={"ID":"0c5d8464-4994-4d09-8474-d484958f1d97","Type":"ContainerStarted","Data":"fe8d3532d4bb6e595eb55adc0cb6f42b4d3257ba3b2f369e5c3b49fe4259b7b8"} Oct 08 19:27:15 crc kubenswrapper[4859]: I1008 19:27:15.939593 4859 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-str2r" podStartSLOduration=2.359832026 podStartE2EDuration="4.93956679s" podCreationTimestamp="2025-10-08 19:27:11 +0000 UTC" firstStartedPulling="2025-10-08 19:27:12.872033275 +0000 UTC m=+4203.118872654" lastFinishedPulling="2025-10-08 19:27:15.451768009 +0000 UTC m=+4205.698607418" observedRunningTime="2025-10-08 19:27:15.926106087 +0000 UTC m=+4206.172945476" watchObservedRunningTime="2025-10-08 19:27:15.93956679 +0000 UTC m=+4206.186406209" Oct 08 19:27:22 crc kubenswrapper[4859]: I1008 19:27:22.104676 4859 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-str2r" Oct 08 19:27:22 crc kubenswrapper[4859]: I1008 19:27:22.105464 4859 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-str2r" Oct 08 19:27:22 crc kubenswrapper[4859]: I1008 19:27:22.154626 4859 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-str2r" Oct 08 19:27:23 crc kubenswrapper[4859]: I1008 19:27:23.048994 4859 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-str2r" Oct 08 19:27:23 crc kubenswrapper[4859]: I1008 19:27:23.100968 4859 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-str2r"] Oct 08 19:27:25 crc kubenswrapper[4859]: I1008 19:27:25.010292 4859 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-marketplace-str2r" podUID="0c5d8464-4994-4d09-8474-d484958f1d97" containerName="registry-server" containerID="cri-o://fe8d3532d4bb6e595eb55adc0cb6f42b4d3257ba3b2f369e5c3b49fe4259b7b8" gracePeriod=2 Oct 08 19:27:25 crc kubenswrapper[4859]: I1008 19:27:25.487722 4859 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-str2r" Oct 08 19:27:25 crc kubenswrapper[4859]: I1008 19:27:25.654329 4859 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/0c5d8464-4994-4d09-8474-d484958f1d97-catalog-content\") pod \"0c5d8464-4994-4d09-8474-d484958f1d97\" (UID: \"0c5d8464-4994-4d09-8474-d484958f1d97\") " Oct 08 19:27:25 crc kubenswrapper[4859]: I1008 19:27:25.654485 4859 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-6n255\" (UniqueName: \"kubernetes.io/projected/0c5d8464-4994-4d09-8474-d484958f1d97-kube-api-access-6n255\") pod \"0c5d8464-4994-4d09-8474-d484958f1d97\" (UID: \"0c5d8464-4994-4d09-8474-d484958f1d97\") " Oct 08 19:27:25 crc kubenswrapper[4859]: I1008 19:27:25.654539 4859 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/0c5d8464-4994-4d09-8474-d484958f1d97-utilities\") pod \"0c5d8464-4994-4d09-8474-d484958f1d97\" (UID: \"0c5d8464-4994-4d09-8474-d484958f1d97\") " Oct 08 19:27:25 crc kubenswrapper[4859]: I1008 19:27:25.655397 4859 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/0c5d8464-4994-4d09-8474-d484958f1d97-utilities" (OuterVolumeSpecName: "utilities") pod "0c5d8464-4994-4d09-8474-d484958f1d97" (UID: "0c5d8464-4994-4d09-8474-d484958f1d97"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 08 19:27:25 crc kubenswrapper[4859]: I1008 19:27:25.656915 4859 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/0c5d8464-4994-4d09-8474-d484958f1d97-utilities\") on node \"crc\" DevicePath \"\"" Oct 08 19:27:25 crc kubenswrapper[4859]: I1008 19:27:25.661844 4859 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/0c5d8464-4994-4d09-8474-d484958f1d97-kube-api-access-6n255" (OuterVolumeSpecName: "kube-api-access-6n255") pod "0c5d8464-4994-4d09-8474-d484958f1d97" (UID: "0c5d8464-4994-4d09-8474-d484958f1d97"). InnerVolumeSpecName "kube-api-access-6n255". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 08 19:27:25 crc kubenswrapper[4859]: I1008 19:27:25.678631 4859 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/0c5d8464-4994-4d09-8474-d484958f1d97-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "0c5d8464-4994-4d09-8474-d484958f1d97" (UID: "0c5d8464-4994-4d09-8474-d484958f1d97"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 08 19:27:25 crc kubenswrapper[4859]: I1008 19:27:25.758165 4859 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/0c5d8464-4994-4d09-8474-d484958f1d97-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 08 19:27:25 crc kubenswrapper[4859]: I1008 19:27:25.758201 4859 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-6n255\" (UniqueName: \"kubernetes.io/projected/0c5d8464-4994-4d09-8474-d484958f1d97-kube-api-access-6n255\") on node \"crc\" DevicePath \"\"" Oct 08 19:27:26 crc kubenswrapper[4859]: I1008 19:27:26.041284 4859 generic.go:334] "Generic (PLEG): container finished" podID="0c5d8464-4994-4d09-8474-d484958f1d97" containerID="fe8d3532d4bb6e595eb55adc0cb6f42b4d3257ba3b2f369e5c3b49fe4259b7b8" exitCode=0 Oct 08 19:27:26 crc kubenswrapper[4859]: I1008 19:27:26.041380 4859 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-str2r" event={"ID":"0c5d8464-4994-4d09-8474-d484958f1d97","Type":"ContainerDied","Data":"fe8d3532d4bb6e595eb55adc0cb6f42b4d3257ba3b2f369e5c3b49fe4259b7b8"} Oct 08 19:27:26 crc kubenswrapper[4859]: I1008 19:27:26.041459 4859 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-str2r" Oct 08 19:27:26 crc kubenswrapper[4859]: I1008 19:27:26.041665 4859 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-str2r" event={"ID":"0c5d8464-4994-4d09-8474-d484958f1d97","Type":"ContainerDied","Data":"3279ff5480639646687db37a393eb7702f1b9c6e061eed9cbaac664d70c3c0e6"} Oct 08 19:27:26 crc kubenswrapper[4859]: I1008 19:27:26.041731 4859 scope.go:117] "RemoveContainer" containerID="fe8d3532d4bb6e595eb55adc0cb6f42b4d3257ba3b2f369e5c3b49fe4259b7b8" Oct 08 19:27:26 crc kubenswrapper[4859]: I1008 19:27:26.080583 4859 scope.go:117] "RemoveContainer" containerID="26fd50082ecc8246c5b8803a5ee155e6c662008433ffc305a7b7a6e8d1ffe306" Oct 08 19:27:26 crc kubenswrapper[4859]: I1008 19:27:26.096371 4859 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-str2r"] Oct 08 19:27:26 crc kubenswrapper[4859]: I1008 19:27:26.108101 4859 scope.go:117] "RemoveContainer" containerID="5acafc12088cb41323804fa5014521817625406075b5e5bf10c7050572ac4287" Oct 08 19:27:26 crc kubenswrapper[4859]: I1008 19:27:26.109471 4859 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-marketplace-str2r"] Oct 08 19:27:26 crc kubenswrapper[4859]: I1008 19:27:26.152974 4859 scope.go:117] "RemoveContainer" containerID="fe8d3532d4bb6e595eb55adc0cb6f42b4d3257ba3b2f369e5c3b49fe4259b7b8" Oct 08 19:27:26 crc kubenswrapper[4859]: E1008 19:27:26.153385 4859 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"fe8d3532d4bb6e595eb55adc0cb6f42b4d3257ba3b2f369e5c3b49fe4259b7b8\": container with ID starting with fe8d3532d4bb6e595eb55adc0cb6f42b4d3257ba3b2f369e5c3b49fe4259b7b8 not found: ID does not exist" containerID="fe8d3532d4bb6e595eb55adc0cb6f42b4d3257ba3b2f369e5c3b49fe4259b7b8" Oct 08 19:27:26 crc kubenswrapper[4859]: I1008 19:27:26.153427 4859 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"fe8d3532d4bb6e595eb55adc0cb6f42b4d3257ba3b2f369e5c3b49fe4259b7b8"} err="failed to get container status \"fe8d3532d4bb6e595eb55adc0cb6f42b4d3257ba3b2f369e5c3b49fe4259b7b8\": rpc error: code = NotFound desc = could not find container \"fe8d3532d4bb6e595eb55adc0cb6f42b4d3257ba3b2f369e5c3b49fe4259b7b8\": container with ID starting with fe8d3532d4bb6e595eb55adc0cb6f42b4d3257ba3b2f369e5c3b49fe4259b7b8 not found: ID does not exist" Oct 08 19:27:26 crc kubenswrapper[4859]: I1008 19:27:26.153453 4859 scope.go:117] "RemoveContainer" containerID="26fd50082ecc8246c5b8803a5ee155e6c662008433ffc305a7b7a6e8d1ffe306" Oct 08 19:27:26 crc kubenswrapper[4859]: E1008 19:27:26.153836 4859 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"26fd50082ecc8246c5b8803a5ee155e6c662008433ffc305a7b7a6e8d1ffe306\": container with ID starting with 26fd50082ecc8246c5b8803a5ee155e6c662008433ffc305a7b7a6e8d1ffe306 not found: ID does not exist" containerID="26fd50082ecc8246c5b8803a5ee155e6c662008433ffc305a7b7a6e8d1ffe306" Oct 08 19:27:26 crc kubenswrapper[4859]: I1008 19:27:26.153897 4859 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"26fd50082ecc8246c5b8803a5ee155e6c662008433ffc305a7b7a6e8d1ffe306"} err="failed to get container status \"26fd50082ecc8246c5b8803a5ee155e6c662008433ffc305a7b7a6e8d1ffe306\": rpc error: code = NotFound desc = could not find container \"26fd50082ecc8246c5b8803a5ee155e6c662008433ffc305a7b7a6e8d1ffe306\": container with ID starting with 26fd50082ecc8246c5b8803a5ee155e6c662008433ffc305a7b7a6e8d1ffe306 not found: ID does not exist" Oct 08 19:27:26 crc kubenswrapper[4859]: I1008 19:27:26.153925 4859 scope.go:117] "RemoveContainer" containerID="5acafc12088cb41323804fa5014521817625406075b5e5bf10c7050572ac4287" Oct 08 19:27:26 crc kubenswrapper[4859]: E1008 19:27:26.154198 4859 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"5acafc12088cb41323804fa5014521817625406075b5e5bf10c7050572ac4287\": container with ID starting with 5acafc12088cb41323804fa5014521817625406075b5e5bf10c7050572ac4287 not found: ID does not exist" containerID="5acafc12088cb41323804fa5014521817625406075b5e5bf10c7050572ac4287" Oct 08 19:27:26 crc kubenswrapper[4859]: I1008 19:27:26.154223 4859 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"5acafc12088cb41323804fa5014521817625406075b5e5bf10c7050572ac4287"} err="failed to get container status \"5acafc12088cb41323804fa5014521817625406075b5e5bf10c7050572ac4287\": rpc error: code = NotFound desc = could not find container \"5acafc12088cb41323804fa5014521817625406075b5e5bf10c7050572ac4287\": container with ID starting with 5acafc12088cb41323804fa5014521817625406075b5e5bf10c7050572ac4287 not found: ID does not exist" Oct 08 19:27:26 crc kubenswrapper[4859]: I1008 19:27:26.480294 4859 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="0c5d8464-4994-4d09-8474-d484958f1d97" path="/var/lib/kubelet/pods/0c5d8464-4994-4d09-8474-d484958f1d97/volumes" Oct 08 19:28:36 crc kubenswrapper[4859]: I1008 19:28:36.811855 4859 generic.go:334] "Generic (PLEG): container finished" podID="23aa205e-275b-4feb-8b07-537fc2834e00" containerID="dfbc1b29c005089d554308d5526cc7452d53ddefdcf75e35f4611090b1cbc6db" exitCode=0 Oct 08 19:28:36 crc kubenswrapper[4859]: I1008 19:28:36.811959 4859 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-4fzzd/must-gather-rf7m5" event={"ID":"23aa205e-275b-4feb-8b07-537fc2834e00","Type":"ContainerDied","Data":"dfbc1b29c005089d554308d5526cc7452d53ddefdcf75e35f4611090b1cbc6db"} Oct 08 19:28:36 crc kubenswrapper[4859]: I1008 19:28:36.813595 4859 scope.go:117] "RemoveContainer" containerID="dfbc1b29c005089d554308d5526cc7452d53ddefdcf75e35f4611090b1cbc6db" Oct 08 19:28:36 crc kubenswrapper[4859]: I1008 19:28:36.940566 4859 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-must-gather-4fzzd_must-gather-rf7m5_23aa205e-275b-4feb-8b07-537fc2834e00/gather/0.log" Oct 08 19:28:47 crc kubenswrapper[4859]: I1008 19:28:47.924538 4859 patch_prober.go:28] interesting pod/machine-config-daemon-82s52 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 08 19:28:47 crc kubenswrapper[4859]: I1008 19:28:47.925204 4859 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-82s52" podUID="b23a6a6c-9d92-4e7b-840e-55cfda873a2d" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 08 19:28:48 crc kubenswrapper[4859]: I1008 19:28:48.766319 4859 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-must-gather-4fzzd/must-gather-rf7m5"] Oct 08 19:28:48 crc kubenswrapper[4859]: I1008 19:28:48.766609 4859 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-must-gather-4fzzd/must-gather-rf7m5" podUID="23aa205e-275b-4feb-8b07-537fc2834e00" containerName="copy" containerID="cri-o://7e415e52ba985fe6878bc7957c47a8f0110b06082a1a375b930c4607c0d60f73" gracePeriod=2 Oct 08 19:28:48 crc kubenswrapper[4859]: I1008 19:28:48.778358 4859 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-must-gather-4fzzd/must-gather-rf7m5"] Oct 08 19:28:48 crc kubenswrapper[4859]: I1008 19:28:48.938266 4859 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-must-gather-4fzzd_must-gather-rf7m5_23aa205e-275b-4feb-8b07-537fc2834e00/copy/0.log" Oct 08 19:28:48 crc kubenswrapper[4859]: I1008 19:28:48.938641 4859 generic.go:334] "Generic (PLEG): container finished" podID="23aa205e-275b-4feb-8b07-537fc2834e00" containerID="7e415e52ba985fe6878bc7957c47a8f0110b06082a1a375b930c4607c0d60f73" exitCode=143 Oct 08 19:28:49 crc kubenswrapper[4859]: I1008 19:28:49.181360 4859 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-must-gather-4fzzd_must-gather-rf7m5_23aa205e-275b-4feb-8b07-537fc2834e00/copy/0.log" Oct 08 19:28:49 crc kubenswrapper[4859]: I1008 19:28:49.181945 4859 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-4fzzd/must-gather-rf7m5" Oct 08 19:28:49 crc kubenswrapper[4859]: I1008 19:28:49.336061 4859 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"must-gather-output\" (UniqueName: \"kubernetes.io/empty-dir/23aa205e-275b-4feb-8b07-537fc2834e00-must-gather-output\") pod \"23aa205e-275b-4feb-8b07-537fc2834e00\" (UID: \"23aa205e-275b-4feb-8b07-537fc2834e00\") " Oct 08 19:28:49 crc kubenswrapper[4859]: I1008 19:28:49.336217 4859 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-w2rq7\" (UniqueName: \"kubernetes.io/projected/23aa205e-275b-4feb-8b07-537fc2834e00-kube-api-access-w2rq7\") pod \"23aa205e-275b-4feb-8b07-537fc2834e00\" (UID: \"23aa205e-275b-4feb-8b07-537fc2834e00\") " Oct 08 19:28:49 crc kubenswrapper[4859]: I1008 19:28:49.342107 4859 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/23aa205e-275b-4feb-8b07-537fc2834e00-kube-api-access-w2rq7" (OuterVolumeSpecName: "kube-api-access-w2rq7") pod "23aa205e-275b-4feb-8b07-537fc2834e00" (UID: "23aa205e-275b-4feb-8b07-537fc2834e00"). InnerVolumeSpecName "kube-api-access-w2rq7". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 08 19:28:49 crc kubenswrapper[4859]: I1008 19:28:49.437972 4859 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-w2rq7\" (UniqueName: \"kubernetes.io/projected/23aa205e-275b-4feb-8b07-537fc2834e00-kube-api-access-w2rq7\") on node \"crc\" DevicePath \"\"" Oct 08 19:28:49 crc kubenswrapper[4859]: I1008 19:28:49.514251 4859 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/23aa205e-275b-4feb-8b07-537fc2834e00-must-gather-output" (OuterVolumeSpecName: "must-gather-output") pod "23aa205e-275b-4feb-8b07-537fc2834e00" (UID: "23aa205e-275b-4feb-8b07-537fc2834e00"). InnerVolumeSpecName "must-gather-output". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 08 19:28:49 crc kubenswrapper[4859]: I1008 19:28:49.539831 4859 reconciler_common.go:293] "Volume detached for volume \"must-gather-output\" (UniqueName: \"kubernetes.io/empty-dir/23aa205e-275b-4feb-8b07-537fc2834e00-must-gather-output\") on node \"crc\" DevicePath \"\"" Oct 08 19:28:49 crc kubenswrapper[4859]: I1008 19:28:49.952629 4859 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-must-gather-4fzzd_must-gather-rf7m5_23aa205e-275b-4feb-8b07-537fc2834e00/copy/0.log" Oct 08 19:28:49 crc kubenswrapper[4859]: I1008 19:28:49.954102 4859 scope.go:117] "RemoveContainer" containerID="7e415e52ba985fe6878bc7957c47a8f0110b06082a1a375b930c4607c0d60f73" Oct 08 19:28:49 crc kubenswrapper[4859]: I1008 19:28:49.954153 4859 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-4fzzd/must-gather-rf7m5" Oct 08 19:28:49 crc kubenswrapper[4859]: I1008 19:28:49.996381 4859 scope.go:117] "RemoveContainer" containerID="dfbc1b29c005089d554308d5526cc7452d53ddefdcf75e35f4611090b1cbc6db" Oct 08 19:28:50 crc kubenswrapper[4859]: I1008 19:28:50.483042 4859 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="23aa205e-275b-4feb-8b07-537fc2834e00" path="/var/lib/kubelet/pods/23aa205e-275b-4feb-8b07-537fc2834e00/volumes" Oct 08 19:29:17 crc kubenswrapper[4859]: I1008 19:29:17.924818 4859 patch_prober.go:28] interesting pod/machine-config-daemon-82s52 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 08 19:29:17 crc kubenswrapper[4859]: I1008 19:29:17.925580 4859 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-82s52" podUID="b23a6a6c-9d92-4e7b-840e-55cfda873a2d" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 08 19:29:25 crc kubenswrapper[4859]: I1008 19:29:25.334075 4859 scope.go:117] "RemoveContainer" containerID="bb99e93a03ec37774df06deef51aa5ae1dff8f4a7d8c8a6b1e6a269df29aab88" Oct 08 19:29:25 crc kubenswrapper[4859]: I1008 19:29:25.679824 4859 scope.go:117] "RemoveContainer" containerID="3bafd7fad7bba7601dad2292419fd75fb4073cc5ae13ec8f6b640f2f87db5e72" Oct 08 19:29:25 crc kubenswrapper[4859]: I1008 19:29:25.760866 4859 scope.go:117] "RemoveContainer" containerID="6a807051dcf3a501a57a6451e555c30692c2d0ea55d98f63864f27651ddb4f56" Oct 08 19:29:47 crc kubenswrapper[4859]: I1008 19:29:47.924791 4859 patch_prober.go:28] interesting pod/machine-config-daemon-82s52 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 08 19:29:47 crc kubenswrapper[4859]: I1008 19:29:47.925283 4859 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-82s52" podUID="b23a6a6c-9d92-4e7b-840e-55cfda873a2d" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 08 19:29:47 crc kubenswrapper[4859]: I1008 19:29:47.925334 4859 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-82s52" Oct 08 19:29:47 crc kubenswrapper[4859]: I1008 19:29:47.926085 4859 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"9cb675f315140c9335fbc82b8aae4ac5d7ef2786da474fb4012875812abb974d"} pod="openshift-machine-config-operator/machine-config-daemon-82s52" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Oct 08 19:29:47 crc kubenswrapper[4859]: I1008 19:29:47.926212 4859 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-82s52" podUID="b23a6a6c-9d92-4e7b-840e-55cfda873a2d" containerName="machine-config-daemon" containerID="cri-o://9cb675f315140c9335fbc82b8aae4ac5d7ef2786da474fb4012875812abb974d" gracePeriod=600 Oct 08 19:29:48 crc kubenswrapper[4859]: I1008 19:29:48.631009 4859 generic.go:334] "Generic (PLEG): container finished" podID="b23a6a6c-9d92-4e7b-840e-55cfda873a2d" containerID="9cb675f315140c9335fbc82b8aae4ac5d7ef2786da474fb4012875812abb974d" exitCode=0 Oct 08 19:29:48 crc kubenswrapper[4859]: I1008 19:29:48.631277 4859 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-82s52" event={"ID":"b23a6a6c-9d92-4e7b-840e-55cfda873a2d","Type":"ContainerDied","Data":"9cb675f315140c9335fbc82b8aae4ac5d7ef2786da474fb4012875812abb974d"} Oct 08 19:29:48 crc kubenswrapper[4859]: I1008 19:29:48.631760 4859 scope.go:117] "RemoveContainer" containerID="de71f0b717129dba48f3f0bdd90240a8e45404a17f0df79d6812d58c9209feff" Oct 08 19:29:49 crc kubenswrapper[4859]: I1008 19:29:49.641294 4859 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-82s52" event={"ID":"b23a6a6c-9d92-4e7b-840e-55cfda873a2d","Type":"ContainerStarted","Data":"546fde72862cb19b5ebe8d53ca513a224a6a40575daefef96ff950dbb05ba77d"} Oct 08 19:30:00 crc kubenswrapper[4859]: I1008 19:30:00.170214 4859 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29332530-mq4hw"] Oct 08 19:30:00 crc kubenswrapper[4859]: E1008 19:30:00.171347 4859 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0c5d8464-4994-4d09-8474-d484958f1d97" containerName="extract-utilities" Oct 08 19:30:00 crc kubenswrapper[4859]: I1008 19:30:00.171362 4859 state_mem.go:107] "Deleted CPUSet assignment" podUID="0c5d8464-4994-4d09-8474-d484958f1d97" containerName="extract-utilities" Oct 08 19:30:00 crc kubenswrapper[4859]: E1008 19:30:00.171389 4859 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="23aa205e-275b-4feb-8b07-537fc2834e00" containerName="copy" Oct 08 19:30:00 crc kubenswrapper[4859]: I1008 19:30:00.171395 4859 state_mem.go:107] "Deleted CPUSet assignment" podUID="23aa205e-275b-4feb-8b07-537fc2834e00" containerName="copy" Oct 08 19:30:00 crc kubenswrapper[4859]: E1008 19:30:00.171414 4859 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0c5d8464-4994-4d09-8474-d484958f1d97" containerName="extract-content" Oct 08 19:30:00 crc kubenswrapper[4859]: I1008 19:30:00.171420 4859 state_mem.go:107] "Deleted CPUSet assignment" podUID="0c5d8464-4994-4d09-8474-d484958f1d97" containerName="extract-content" Oct 08 19:30:00 crc kubenswrapper[4859]: E1008 19:30:00.171435 4859 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="23aa205e-275b-4feb-8b07-537fc2834e00" containerName="gather" Oct 08 19:30:00 crc kubenswrapper[4859]: I1008 19:30:00.171440 4859 state_mem.go:107] "Deleted CPUSet assignment" podUID="23aa205e-275b-4feb-8b07-537fc2834e00" containerName="gather" Oct 08 19:30:00 crc kubenswrapper[4859]: E1008 19:30:00.171454 4859 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0c5d8464-4994-4d09-8474-d484958f1d97" containerName="registry-server" Oct 08 19:30:00 crc kubenswrapper[4859]: I1008 19:30:00.171460 4859 state_mem.go:107] "Deleted CPUSet assignment" podUID="0c5d8464-4994-4d09-8474-d484958f1d97" containerName="registry-server" Oct 08 19:30:00 crc kubenswrapper[4859]: I1008 19:30:00.171653 4859 memory_manager.go:354] "RemoveStaleState removing state" podUID="0c5d8464-4994-4d09-8474-d484958f1d97" containerName="registry-server" Oct 08 19:30:00 crc kubenswrapper[4859]: I1008 19:30:00.171666 4859 memory_manager.go:354] "RemoveStaleState removing state" podUID="23aa205e-275b-4feb-8b07-537fc2834e00" containerName="gather" Oct 08 19:30:00 crc kubenswrapper[4859]: I1008 19:30:00.171678 4859 memory_manager.go:354] "RemoveStaleState removing state" podUID="23aa205e-275b-4feb-8b07-537fc2834e00" containerName="copy" Oct 08 19:30:00 crc kubenswrapper[4859]: I1008 19:30:00.172405 4859 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29332530-mq4hw" Oct 08 19:30:00 crc kubenswrapper[4859]: I1008 19:30:00.178089 4859 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"collect-profiles-config" Oct 08 19:30:00 crc kubenswrapper[4859]: I1008 19:30:00.178538 4859 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"collect-profiles-dockercfg-kzf4t" Oct 08 19:30:00 crc kubenswrapper[4859]: I1008 19:30:00.198044 4859 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29332530-mq4hw"] Oct 08 19:30:00 crc kubenswrapper[4859]: I1008 19:30:00.229989 4859 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/b00a4f1d-9aed-4fd5-83fe-bbe55a689582-secret-volume\") pod \"collect-profiles-29332530-mq4hw\" (UID: \"b00a4f1d-9aed-4fd5-83fe-bbe55a689582\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29332530-mq4hw" Oct 08 19:30:00 crc kubenswrapper[4859]: I1008 19:30:00.230398 4859 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-9jpsq\" (UniqueName: \"kubernetes.io/projected/b00a4f1d-9aed-4fd5-83fe-bbe55a689582-kube-api-access-9jpsq\") pod \"collect-profiles-29332530-mq4hw\" (UID: \"b00a4f1d-9aed-4fd5-83fe-bbe55a689582\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29332530-mq4hw" Oct 08 19:30:00 crc kubenswrapper[4859]: I1008 19:30:00.230516 4859 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/b00a4f1d-9aed-4fd5-83fe-bbe55a689582-config-volume\") pod \"collect-profiles-29332530-mq4hw\" (UID: \"b00a4f1d-9aed-4fd5-83fe-bbe55a689582\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29332530-mq4hw" Oct 08 19:30:00 crc kubenswrapper[4859]: I1008 19:30:00.332714 4859 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/b00a4f1d-9aed-4fd5-83fe-bbe55a689582-config-volume\") pod \"collect-profiles-29332530-mq4hw\" (UID: \"b00a4f1d-9aed-4fd5-83fe-bbe55a689582\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29332530-mq4hw" Oct 08 19:30:00 crc kubenswrapper[4859]: I1008 19:30:00.332834 4859 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/b00a4f1d-9aed-4fd5-83fe-bbe55a689582-secret-volume\") pod \"collect-profiles-29332530-mq4hw\" (UID: \"b00a4f1d-9aed-4fd5-83fe-bbe55a689582\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29332530-mq4hw" Oct 08 19:30:00 crc kubenswrapper[4859]: I1008 19:30:00.332943 4859 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-9jpsq\" (UniqueName: \"kubernetes.io/projected/b00a4f1d-9aed-4fd5-83fe-bbe55a689582-kube-api-access-9jpsq\") pod \"collect-profiles-29332530-mq4hw\" (UID: \"b00a4f1d-9aed-4fd5-83fe-bbe55a689582\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29332530-mq4hw" Oct 08 19:30:00 crc kubenswrapper[4859]: I1008 19:30:00.333734 4859 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/b00a4f1d-9aed-4fd5-83fe-bbe55a689582-config-volume\") pod \"collect-profiles-29332530-mq4hw\" (UID: \"b00a4f1d-9aed-4fd5-83fe-bbe55a689582\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29332530-mq4hw" Oct 08 19:30:00 crc kubenswrapper[4859]: I1008 19:30:00.341596 4859 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/b00a4f1d-9aed-4fd5-83fe-bbe55a689582-secret-volume\") pod \"collect-profiles-29332530-mq4hw\" (UID: \"b00a4f1d-9aed-4fd5-83fe-bbe55a689582\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29332530-mq4hw" Oct 08 19:30:00 crc kubenswrapper[4859]: I1008 19:30:00.349118 4859 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-9jpsq\" (UniqueName: \"kubernetes.io/projected/b00a4f1d-9aed-4fd5-83fe-bbe55a689582-kube-api-access-9jpsq\") pod \"collect-profiles-29332530-mq4hw\" (UID: \"b00a4f1d-9aed-4fd5-83fe-bbe55a689582\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29332530-mq4hw" Oct 08 19:30:00 crc kubenswrapper[4859]: I1008 19:30:00.494706 4859 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29332530-mq4hw" Oct 08 19:30:00 crc kubenswrapper[4859]: I1008 19:30:00.911142 4859 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29332530-mq4hw"] Oct 08 19:30:01 crc kubenswrapper[4859]: I1008 19:30:01.752849 4859 generic.go:334] "Generic (PLEG): container finished" podID="b00a4f1d-9aed-4fd5-83fe-bbe55a689582" containerID="700cd62c20ec227161dd6ac52021689cd94349e2384737d660a5cd8675481d57" exitCode=0 Oct 08 19:30:01 crc kubenswrapper[4859]: I1008 19:30:01.753194 4859 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29332530-mq4hw" event={"ID":"b00a4f1d-9aed-4fd5-83fe-bbe55a689582","Type":"ContainerDied","Data":"700cd62c20ec227161dd6ac52021689cd94349e2384737d660a5cd8675481d57"} Oct 08 19:30:01 crc kubenswrapper[4859]: I1008 19:30:01.753243 4859 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29332530-mq4hw" event={"ID":"b00a4f1d-9aed-4fd5-83fe-bbe55a689582","Type":"ContainerStarted","Data":"a73a19133f3e8510aed2cf51fe6886b686fceda1b951b626a789f56853c5fb43"} Oct 08 19:30:03 crc kubenswrapper[4859]: I1008 19:30:03.221676 4859 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29332530-mq4hw" Oct 08 19:30:03 crc kubenswrapper[4859]: I1008 19:30:03.287359 4859 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/b00a4f1d-9aed-4fd5-83fe-bbe55a689582-config-volume\") pod \"b00a4f1d-9aed-4fd5-83fe-bbe55a689582\" (UID: \"b00a4f1d-9aed-4fd5-83fe-bbe55a689582\") " Oct 08 19:30:03 crc kubenswrapper[4859]: I1008 19:30:03.287428 4859 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-9jpsq\" (UniqueName: \"kubernetes.io/projected/b00a4f1d-9aed-4fd5-83fe-bbe55a689582-kube-api-access-9jpsq\") pod \"b00a4f1d-9aed-4fd5-83fe-bbe55a689582\" (UID: \"b00a4f1d-9aed-4fd5-83fe-bbe55a689582\") " Oct 08 19:30:03 crc kubenswrapper[4859]: I1008 19:30:03.287568 4859 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/b00a4f1d-9aed-4fd5-83fe-bbe55a689582-secret-volume\") pod \"b00a4f1d-9aed-4fd5-83fe-bbe55a689582\" (UID: \"b00a4f1d-9aed-4fd5-83fe-bbe55a689582\") " Oct 08 19:30:03 crc kubenswrapper[4859]: I1008 19:30:03.288287 4859 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/b00a4f1d-9aed-4fd5-83fe-bbe55a689582-config-volume" (OuterVolumeSpecName: "config-volume") pod "b00a4f1d-9aed-4fd5-83fe-bbe55a689582" (UID: "b00a4f1d-9aed-4fd5-83fe-bbe55a689582"). InnerVolumeSpecName "config-volume". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 08 19:30:03 crc kubenswrapper[4859]: I1008 19:30:03.289220 4859 reconciler_common.go:293] "Volume detached for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/b00a4f1d-9aed-4fd5-83fe-bbe55a689582-config-volume\") on node \"crc\" DevicePath \"\"" Oct 08 19:30:03 crc kubenswrapper[4859]: I1008 19:30:03.295836 4859 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b00a4f1d-9aed-4fd5-83fe-bbe55a689582-kube-api-access-9jpsq" (OuterVolumeSpecName: "kube-api-access-9jpsq") pod "b00a4f1d-9aed-4fd5-83fe-bbe55a689582" (UID: "b00a4f1d-9aed-4fd5-83fe-bbe55a689582"). InnerVolumeSpecName "kube-api-access-9jpsq". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 08 19:30:03 crc kubenswrapper[4859]: I1008 19:30:03.296563 4859 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b00a4f1d-9aed-4fd5-83fe-bbe55a689582-secret-volume" (OuterVolumeSpecName: "secret-volume") pod "b00a4f1d-9aed-4fd5-83fe-bbe55a689582" (UID: "b00a4f1d-9aed-4fd5-83fe-bbe55a689582"). InnerVolumeSpecName "secret-volume". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 08 19:30:03 crc kubenswrapper[4859]: I1008 19:30:03.390751 4859 reconciler_common.go:293] "Volume detached for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/b00a4f1d-9aed-4fd5-83fe-bbe55a689582-secret-volume\") on node \"crc\" DevicePath \"\"" Oct 08 19:30:03 crc kubenswrapper[4859]: I1008 19:30:03.390791 4859 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-9jpsq\" (UniqueName: \"kubernetes.io/projected/b00a4f1d-9aed-4fd5-83fe-bbe55a689582-kube-api-access-9jpsq\") on node \"crc\" DevicePath \"\"" Oct 08 19:30:03 crc kubenswrapper[4859]: I1008 19:30:03.772836 4859 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29332530-mq4hw" event={"ID":"b00a4f1d-9aed-4fd5-83fe-bbe55a689582","Type":"ContainerDied","Data":"a73a19133f3e8510aed2cf51fe6886b686fceda1b951b626a789f56853c5fb43"} Oct 08 19:30:03 crc kubenswrapper[4859]: I1008 19:30:03.773117 4859 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="a73a19133f3e8510aed2cf51fe6886b686fceda1b951b626a789f56853c5fb43" Oct 08 19:30:03 crc kubenswrapper[4859]: I1008 19:30:03.772888 4859 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29332530-mq4hw" Oct 08 19:30:04 crc kubenswrapper[4859]: I1008 19:30:04.297387 4859 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29332485-lsgdn"] Oct 08 19:30:04 crc kubenswrapper[4859]: I1008 19:30:04.306410 4859 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29332485-lsgdn"] Oct 08 19:30:04 crc kubenswrapper[4859]: I1008 19:30:04.478383 4859 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="46b24a07-d369-4899-ad78-b6a8df8a77fe" path="/var/lib/kubelet/pods/46b24a07-d369-4899-ad78-b6a8df8a77fe/volumes" var/home/core/zuul-output/logs/crc-cloud-workdir-crc-all-logs.tar.gz0000644000175000000000000000005515071535706024456 0ustar coreroot‹íÁ  ÷Om7 €7šÞ'(var/home/core/zuul-output/logs/crc-cloud/0000755000175000000000000000000015071535707017374 5ustar corerootvar/home/core/zuul-output/artifacts/0000755000175000017500000000000015071524742016514 5ustar corecorevar/home/core/zuul-output/docs/0000755000175000017500000000000015071524743015465 5ustar corecore